From cab07c3001ad82a3a5e3d5a6755f77e7b2a79a68 Mon Sep 17 00:00:00 2001 From: Kent Yao Date: Mon, 21 Sep 2020 18:44:46 +0800 Subject: [PATCH] cleanup --- .travis.yml | 3 +- .../2ecb7654-7c83-4bd9-8ded-4551eee0911b | 0 .../kyuubi/engine/ProcessBuilderLike.scala | 6 +-- .../engine/spark/SparkProcessBuilder.scala | 42 +++---------------- 4 files changed, 7 insertions(+), 44 deletions(-) delete mode 100644 kyuubi-main/kentyao/kyuubi-9d0834d8-0bc4-471b-bcb6-a9b346b0e148/2ecb7654-7c83-4bd9-8ded-4551eee0911b diff --git a/.travis.yml b/.travis.yml index de3ed603d..0560beacf 100644 --- a/.travis.yml +++ b/.travis.yml @@ -32,9 +32,8 @@ install: jobs: include: - stage: spark3.0.0 - language: scala script: - - mvn --no-transfer-progress clean install -pl :kyuubi-common,:kyuubi-ha,:kyuubi-main,:kyuubi-spark-sql-engine,:kyuubi-codecov,:kyuubi-download,:kyuubi-assembly -Dmaven.javadoc.skip=true -B -V + - build/mvn --no-transfer-progress clean install -pl :kyuubi-common,:kyuubi-ha,:kyuubi-main,:kyuubi-spark-sql-engine,:kyuubi-codecov,:kyuubi-download,:kyuubi-assembly -Dmaven.javadoc.skip=true -B -V after_success: - bash <(curl -s https://codecov.io/bash) diff --git a/kyuubi-main/kentyao/kyuubi-9d0834d8-0bc4-471b-bcb6-a9b346b0e148/2ecb7654-7c83-4bd9-8ded-4551eee0911b b/kyuubi-main/kentyao/kyuubi-9d0834d8-0bc4-471b-bcb6-a9b346b0e148/2ecb7654-7c83-4bd9-8ded-4551eee0911b deleted file mode 100644 index e69de29bb..000000000 diff --git a/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/ProcessBuilderLike.scala b/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/ProcessBuilderLike.scala index 5bf345e82..d197cabd2 100644 --- a/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/ProcessBuilderLike.scala +++ b/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/ProcessBuilderLike.scala @@ -17,15 +17,11 @@ package org.apache.kyuubi.engine -import java.io.File -import java.lang.ProcessBuilder.Redirect -import java.nio.file.{Files, Path, Paths} +import java.nio.file.{Path, Paths} import java.util.UUID import scala.collection.JavaConverters._ -import org.apache.kyuubi.Utils - trait ProcessBuilderLike { protected def executable: String diff --git a/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/spark/SparkProcessBuilder.scala b/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/spark/SparkProcessBuilder.scala index a6681db57..61f082ff8 100644 --- a/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/spark/SparkProcessBuilder.scala +++ b/kyuubi-main/src/main/scala/org/apache/kyuubi/engine/spark/SparkProcessBuilder.scala @@ -17,9 +17,7 @@ package org.apache.kyuubi.engine.spark -import java.io.{BufferedReader, InputStreamReader} import java.nio.file.{Files, Path, Paths} -import java.util.concurrent.TimeUnit import scala.collection.mutable.ArrayBuffer @@ -40,12 +38,15 @@ class SparkProcessBuilder( val path = env.get("SPARK_HOME").map { sparkHome => Paths.get(sparkHome, "bin", "spark-submit").toAbsolutePath } getOrElse { + val sparkVer = SPARK_COMPILE_VERSION + val hadoopVer = HADOOP_COMPILE_VERSION.take(3) + val hiveVer = if (HIVE_COMPILE_VERSION.take(3).toDouble < 2.3) "-hive1.2" else "" Paths.get( "..", "externals", "kyuubi-download", "target", - s"spark-$SPARK_COMPILE_VERSION-bin-hadoop2.7", + s"spark-$sparkVer-bin-hadoop$hadoopVer$hiveVer", "bin", "spark-submit") } path.toAbsolutePath.toFile.getCanonicalPath @@ -76,7 +77,7 @@ class SparkProcessBuilder( env.get("KYUUBI_WORK_DIR_ROOT").map { root => Utils.createTempDir(root, proxyUser) }.getOrElse { - Utils.createTempDir(proxyUser) + Utils.createTempDir(namePrefix = proxyUser) } } @@ -101,41 +102,8 @@ class SparkProcessBuilder( } -/** - * May need download spark release packages first. - * - * (build/)mvn clean package -pl :kyuubi-download -DskipTests - */ object SparkProcessBuilder { - private final val CONF = "--conf" private final val CLASS = "--class" private final val PROXY_USER = "--proxy-user" - - def main(args: Array[String]): Unit = { - val conf = Map("spark.abc" -> "1", "spark.xyz" -> "2", "spark.master" -> "hello") - val sparkProcessBuilder = new SparkProcessBuilder("kent", conf) - print(sparkProcessBuilder.toString) - val start = sparkProcessBuilder.start - - // scalastyle:off - if (start.waitFor(1, TimeUnit.MINUTES)) { - val reader = new BufferedReader(new InputStreamReader(start.getInputStream)) - var line = reader.readLine() - while(line != null) { - println(line) - line = reader.readLine() - } - reader.close() - } else { - val reader = new BufferedReader(new InputStreamReader(start.getErrorStream)) - var line = reader.readLine() - while(line != null) { - println(line) - line = reader.readLine() - } - reader.close() - println("\nnot started") - } - } }