File tree Expand file tree Collapse file tree 2 files changed +3
-2
lines changed
external/docker-integration-tests/src/test/scala/org/apache/spark/sql/jdbc
sql/core/src/test/scala/org/apache/spark/sql/test Expand file tree Collapse file tree 2 files changed +3
-2
lines changed Original file line number Diff line number Diff line change @@ -260,11 +260,11 @@ class OracleIntegrationSuite extends DockerJDBCIntegrationSuite with SharedSQLCo
260
260
assert(parentPlan.isInstanceOf [WholeStageCodegenExec ])
261
261
val node = parentPlan.asInstanceOf [WholeStageCodegenExec ]
262
262
val metadata = node.child.asInstanceOf [RowDataSourceScanExec ].metadata
263
- // The "PushedFilters" part should be exist in Dataframe's
263
+ // The "PushedFilters" part should exist in Dataframe's
264
264
// physical plan and the existence of right literals in
265
265
// "PushedFilters" is used to prove that the predicates
266
266
// pushing down have been effective.
267
- assert(metadata.get(" PushedFilters" ).ne( None ) )
267
+ assert(metadata.get(" PushedFilters" ).isDefined )
268
268
assert(metadata(" PushedFilters" ).contains(dt.toString))
269
269
assert(metadata(" PushedFilters" ).contains(ts.toString))
270
270
Original file line number Diff line number Diff line change @@ -39,6 +39,7 @@ import org.apache.spark.sql.catalyst.plans.PlanTest
39
39
import org .apache .spark .sql .catalyst .plans .logical .LogicalPlan
40
40
import org .apache .spark .sql .catalyst .util ._
41
41
import org .apache .spark .sql .execution .FilterExec
42
+ import org .apache .spark .sql .internal .SQLConf
42
43
import org .apache .spark .util .{UninterruptibleThread , Utils }
43
44
44
45
/**
You can’t perform that action at this time.
0 commit comments