File tree Expand file tree Collapse file tree 2 files changed +4
-6
lines changed
core/src/main/scala/org/apache/spark/sql
hive/src/main/scala/org/apache/spark/sql/hive Expand file tree Collapse file tree 2 files changed +4
-6
lines changed Original file line number Diff line number Diff line change 17
17
18
18
package org .apache .spark .sql
19
19
20
- import org . apache . spark . sql . catalyst . CatalystConf
20
+ import java . util . Properties
21
21
22
22
import scala .collection .immutable
23
23
import scala .collection .JavaConversions ._
24
24
25
- import java . util . Properties
25
+ import org . apache . spark . sql . catalyst . CatalystConf
26
26
27
27
private [spark] object SQLConf {
28
28
val COMPRESS_CACHED = " spark.sql.inMemoryColumnarStorage.compressed"
@@ -266,3 +266,4 @@ private[sql] class SQLConf extends Serializable with CatalystConf {
266
266
settings.clear()
267
267
}
268
268
}
269
+
Original file line number Diff line number Diff line change @@ -50,10 +50,6 @@ import org.apache.spark.sql.types._
50
50
class HiveContext (sc : SparkContext ) extends SQLContext (sc) {
51
51
self =>
52
52
53
- protected [sql] override lazy val conf : SQLConf = new SQLConf {
54
- override def dialect : String = getConf(SQLConf .DIALECT , " hiveql" )
55
- }
56
-
57
53
/**
58
54
* When true, enables an experimental feature where metastore tables that use the parquet SerDe
59
55
* are automatically converted to use the Spark SQL parquet table scan, instead of the Hive
@@ -266,6 +262,7 @@ class HiveContext(sc: SparkContext) extends SQLContext(sc) {
266
262
protected [hive] class SQLSession extends super .SQLSession {
267
263
protected [sql] override lazy val conf : SQLConf = new SQLConf {
268
264
override def dialect : String = getConf(SQLConf .DIALECT , " hiveql" )
265
+ setConf(CatalystConf .CASE_SENSITIVE , " false" )
269
266
}
270
267
271
268
protected [hive] lazy val hiveconf : HiveConf = {
You can’t perform that action at this time.
0 commit comments