diff --git a/shims/spark300db/src/main/scala/com/nvidia/spark/rapids/shims/spark300db/Spark300dbShims.scala b/shims/spark300db/src/main/scala/com/nvidia/spark/rapids/shims/spark300db/Spark300dbShims.scala index fcf42aff1de..5895739e37a 100644 --- a/shims/spark300db/src/main/scala/com/nvidia/spark/rapids/shims/spark300db/Spark300dbShims.scala +++ b/shims/spark300db/src/main/scala/com/nvidia/spark/rapids/shims/spark300db/Spark300dbShims.scala @@ -70,7 +70,7 @@ class Spark300dbShims extends Spark300Shims { override def getExecs: Map[Class[_ <: SparkPlan], ExecRule[_ <: SparkPlan]] = { Seq( - GpuOverrides.exec[FileSourceScanExec]( + GpuOverrides.exec[FileSourceScanExec]( "Reading data from files, often from Hive tables", (fsse, conf, p, r) => new SparkPlanMeta[FileSourceScanExec](fsse, conf, p, r) { // partition filters and data filters are not run on the GPU @@ -104,7 +104,7 @@ class Spark300dbShims extends Spark300Shims { (join, conf, p, r) => new GpuBroadcastHashJoinMeta(join, conf, p, r)), GpuOverrides.exec[ShuffledHashJoinExec]( "Implementation of join using hashed shuffled data", - (join, conf, p, r) => new GpuShuffledHashJoinMeta(join, conf, p, r)), + (join, conf, p, r) => new GpuShuffledHashJoinMeta(join, conf, p, r)) ).map(r => (r.getClassFor.asSubclass(classOf[SparkPlan]), r)).toMap }