Compare commits
4 Commits
master
...
branch-0.3
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
530cdbf241 | ||
|
|
bd434c3c8b | ||
|
|
b91e2df5cd | ||
|
|
01e13dfde1 |
@ -22,7 +22,7 @@
|
||||
<parent>
|
||||
<artifactId>kyuubi</artifactId>
|
||||
<groupId>yaooqinn</groupId>
|
||||
<version>0.3.0-SNAPSHOT</version>
|
||||
<version>0.3.2-SNAPSHOT</version>
|
||||
</parent>
|
||||
<modelVersion>4.0.0</modelVersion>
|
||||
|
||||
|
||||
@ -99,6 +99,7 @@ object KyuubiSparkUtil extends Logging {
|
||||
val HDFS_CLIENT_CACHE_DEFAULT = "true"
|
||||
val FILE_CLIENT_CACHE: String = SPARK_HADOOP_PREFIX + "fs.file.impl.disable.cache"
|
||||
val FILE_CLIENT_CACHE_DEFAULT = "true"
|
||||
val RDD_PAR_LISTING: String = SPARK_PREFIX + "rdd.parallelListingThreshold"
|
||||
|
||||
// Runtime Spark Version
|
||||
val SPARK_VERSION: String = org.apache.spark.SPARK_VERSION
|
||||
@ -275,6 +276,14 @@ object KyuubiSparkUtil extends Logging {
|
||||
if (UserGroupInformation.isSecurityEnabled) {
|
||||
conf.setIfMissing(HDFS_CLIENT_CACHE, HDFS_CLIENT_CACHE_DEFAULT)
|
||||
conf.setIfMissing(FILE_CLIENT_CACHE, FILE_CLIENT_CACHE_DEFAULT)
|
||||
// If you are using Kyuubi against kerberized HDFS, you will run into HDFS_DELEGATION_TOKEN
|
||||
// expiration in some particular sql queries. This exception is usually caught in
|
||||
// HadoopRDD.getPartitions, where the JobConf has no Credentials because it is generated by
|
||||
// Configuration, and the UGI.getCurrentUser contains only the oldest tokens which are
|
||||
// destined to expire. The reason seems to be parallel listing UnionRDD's sub RDDs using a
|
||||
// ForkJoinPool which points to another calling context. Turn off parallel listing seems
|
||||
// to be a solution to this issue.
|
||||
conf.setIfMissing(RDD_PAR_LISTING, Int.MaxValue.toString)
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user