diff --git a/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/KyuubiSparkUtil.scala b/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/KyuubiSparkUtil.scala index a7a7621b7..274cbc6cc 100644 --- a/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/KyuubiSparkUtil.scala +++ b/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/KyuubiSparkUtil.scala @@ -23,7 +23,8 @@ import org.apache.spark.sql.SparkSession object KyuubiSparkUtil { - def diagnostics(spark: SparkSession): String = { + lazy val diagnostics: String = { + val spark = SparkSession.active val sc = spark.sparkContext val webUrl = sc.getConf.getOption( "spark.org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter.param.PROXY_URI_BASES") diff --git a/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/SparkSQLEngine.scala b/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/SparkSQLEngine.scala index 51bf633c5..cf313ef2f 100644 --- a/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/SparkSQLEngine.scala +++ b/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/SparkSQLEngine.scala @@ -116,7 +116,7 @@ object SparkSQLEngine extends Logging { try { spark = createSpark() engine = startEngine(spark) - info(KyuubiSparkUtil.diagnostics(spark)) + info(KyuubiSparkUtil.diagnostics) // blocking main thread countDownLatch.await() } catch { diff --git a/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/operation/ExecuteStatement.scala b/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/operation/ExecuteStatement.scala index 11142d41b..c16ce9e9e 100644 --- a/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/operation/ExecuteStatement.scala +++ b/externals/kyuubi-spark-sql-engine/src/main/scala/org/apache/kyuubi/engine/spark/operation/ExecuteStatement.scala @@ -67,7 +67,7 @@ class ExecuteStatement( private def executeStatement(): Unit = { try { setState(OperationState.RUNNING) - info(KyuubiSparkUtil.diagnostics(spark)) + info(KyuubiSparkUtil.diagnostics) Thread.currentThread().setContextClassLoader(spark.sharedState.jarClassLoader) spark.sparkContext.setJobGroup(statementId, statement, forceCancel) result = spark.sql(statement)