-
Notifications
You must be signed in to change notification settings - Fork 232
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] Failed to cast value false
to BooleanType
for partition column k1
#6026
Comments
When trying to reproduce this issue, I found another #6029 that Spark RAPIDS(without alluxio) can not read a Hive partition table if the partition key is boolean if |
Now I can kind of reproduce the issue with a different error message which is the same as #6029:
Below is the minimum repro:
Make sure to use the newer guava-27.0-jre.jar from hive
Here I am using Hive(apache-hive-3.1.2-bin), so make sure below jars are there:
To access S3 using access key, create $SPARK_HOME/conf/hdfs-site.xml and $HIVE_HOME/conf/hdfs-site.xml as:
To access S3 using Spark, make sure add below settings in spark-defaults.conf:
Error:
|
I can reproduce the original error by just creating a Hive delta partition table, then below error will show up: |
so this has the same root cause as #6029. I was experimenting with ways to update the paths by updating the PartitionSpecs but that isn't accessible to all FileIndex types. And some CSPs have custom versions of that so we can't just use the built in Spark classes like PartitioningAwareFileIndex. |
Env:
Databricks 10.4ML LTS
Spark RAPIDS 22.08 snapshot jar with the alluxio-auto-mount feature
When using Alluxio auto-mount feature --
spark.rapids.alluxio.automount.enabled=true
, a simple query failed:Full stacktrace:
This table is partitioned delta partition table(hive external partition table with underline storage on s3), and column "k1" is the leading partition key column.
Eg:
If we disable auto-mount feature, then the query runs fine on GPU:
Still trying to figure out a minimum repro as of now.
The text was updated successfully, but these errors were encountered: