-
Notifications
You must be signed in to change notification settings - Fork 96
Cannot complete the test run #32
Comments
@dilipbiswal can you give this a look when you have a minute ^ |
Hello @stevemart @dilipbiswal, Also, I have questions about tpcdsenv.sh variables. For the error above, I used default except point the root directory to my TPC-DS installation directory. Here is the tpcdsenv.sh: harry.li@perf84:/usr/local/harry/tpcds/spark-tpc-ds-performance-test$ cat bin/tpcdsenv.sh ####################################################################### ####################################################################### export TPCDS_ROOT_DIR=/usr/local/harry/tpcds/spark-tpc-ds-performance-test My questions are:
export TPCDS_ROOT_DIR=/usr/local/harry/tpcds/spark-tpc-ds-performance-test Please advise and thanks in advance. |
@HarryLiUS Can you run step 4 (cleanup) to clean all data and start from scratch? I think you may have run dsdgen to generate data at a different scale factor. |
@HarryLiUS Could you solve the problem? I am facing the same problem. |
has anyone resolved this? |
I could not make it work with Spark 3.0.0. But after switching to Spark 2.4.5 the problem went away. |
Works without any modifications with Spark 2.4.5 and Spark 2.4.7. However, requires some modifications to run with Spark version 3.0.1. Actually, the solution does not even relate to Spark. There is a check which compares row counts from generated data and expected results. The check fails because it compares the contents of files. The newer version of Spark has some new warnings that get added to the beginning of the generated file and thus fails the comparison with the expected result.
|
This error occurs when the file
And the
This cause the error in check_createtable. So here's my solution: |
For spark 3.3.0, it need more filter. i make it work with adding the follow filter:
|
hi, @HarryLiUS Have you solved this problem? I checked my rowcounts.rrn and it is also all 0. |
This is what fixed it for me. |
Hello,
I following the instruction to do a local test run. First 3 steps completed successfully. At step 4, the table creation completed in about 10+ minutes. It is longer than I expected, but it is completed. Here is the output:
==============================================
TPC-DS On Spark Menu
SETUP
(1) Create spark tables
RUN
(2) Run a subset of TPC-DS queries
(3) Run All (99) TPC-DS Queries
CLEANUP
(4) Cleanup
(Q) Quit
Please enter your choice followed by [ENTER]: 1
INFO: Creating tables. Will take a few minutes ...
INFO: Progress : [########################################] 100%
INFO: Spark tables created successfully..
Press any key to continue
After succeeded with table creation, I tried to run query 1 and here is what I got:
==============================================
TPC-DS On Spark Menu
SETUP
(1) Create spark tables
RUN
(2) Run a subset of TPC-DS queries
(3) Run All (99) TPC-DS Queries
CLEANUP
(4) Cleanup
(Q) Quit
Please enter your choice followed by [ENTER]: 2
Enter a comma separated list of queries to run (ex: 1, 2), followed by [ENTER]:
1
INFO: Checking pre-reqs for running TPC-DS queries. May take a few seconds..
ERROR: The rowcounts for TPC-DS tables are not correct. Please make sure option 1
is run before continuing with currently selected option
Press any key to continue
I repeated this again and no help.
Checking rowcounts.rrn, it is all 0.
And, here is the output from spark-shell from step 3.
scala> spark.conf
res0: org.apache.spark.sql.RuntimeConfig = org.apache.spark.sql.RuntimeConfig@505bc480
scala> spark.conf.get("spark.sql.catalogImplementation")
res1: String = hive
Thank you for the help,
Harry
The text was updated successfully, but these errors were encountered: