This is an automated email from the ASF dual-hosted git repository. zjffdu pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/zeppelin.git
The following commit(s) were added to refs/heads/master by this push: new 0b85cac Revert "[ZEPPELIN-4950]. Support for manually specifying the Java version of Spark Interpreter Scala REPL" 0b85cac is described below commit 0b85cac1abfda3c85defc185d33b2d846c370525 Author: Jeff Zhang <zjf...@apache.org> AuthorDate: Tue Jul 21 10:09:54 2020 +0800 Revert "[ZEPPELIN-4950]. Support for manually specifying the Java version of Spark Interpreter Scala REPL" This reverts commit 6045079b5f5cb89def74bef55518351d4966c2dd. --- docs/interpreter/spark.md | 4 ---- .../scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala | 3 --- .../scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala | 2 -- .../scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala | 3 --- 4 files changed, 12 deletions(-) diff --git a/docs/interpreter/spark.md b/docs/interpreter/spark.md index 21069b4..5fc9305 100644 --- a/docs/interpreter/spark.md +++ b/docs/interpreter/spark.md @@ -199,10 +199,6 @@ You can also set other Spark properties which are not listed in the table. For a <td>false</td> <td>whether use yarn proxy url as spark weburl, e.g. http://localhost:8088/proxy/application_1583396598068_0004</td> </tr> - <td>spark.repl.target</td> - <td>jvm-1.8</td> - <td>Manually specifying the Java version of Spark Interpreter Scala REPL,Available options:[jvm-1.5, jvm-1.6, jvm-1.7, jvm-1.8] </td> - </tr> </table> Without any configuration, Spark interpreter works out of box in local mode. But if you want to connect to your Spark cluster, you'll need to follow below two simple steps. diff --git a/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala b/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala index f59f137..0eac200 100644 --- a/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala +++ b/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala @@ -67,13 +67,10 @@ class SparkScala210Interpreter(override val conf: SparkConf, sparkHttpServer = server conf.set("spark.repl.class.uri", uri) } - val target = conf.get("spark.repl.target", "jvm-1.8") val settings = new Settings() settings.embeddedDefaults(sparkInterpreterClassLoader) settings.usejavacp.value = true - settings.target.value = target - this.userJars = getUserJars() LOGGER.info("UserJars: " + userJars.mkString(File.pathSeparator)) settings.classpath.value = userJars.mkString(File.pathSeparator) diff --git a/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala b/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala index d2fb971..cb5a016 100644 --- a/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala +++ b/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala @@ -66,14 +66,12 @@ class SparkScala211Interpreter(override val conf: SparkConf, sparkHttpServer = server conf.set("spark.repl.class.uri", uri) } - val target = conf.get("spark.repl.target", "jvm-1.8") val settings = new Settings() settings.processArguments(List("-Yrepl-class-based", "-Yrepl-outdir", s"${outputDir.getAbsolutePath}"), true) settings.embeddedDefaults(sparkInterpreterClassLoader) settings.usejavacp.value = true - settings.target.value = target this.userJars = getUserJars() LOGGER.info("UserJars: " + userJars.mkString(File.pathSeparator)) diff --git a/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala b/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala index 7f35125..2b04a1d 100644 --- a/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala +++ b/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala @@ -60,15 +60,12 @@ class SparkScala212Interpreter(override val conf: SparkConf, LOGGER.info("Scala shell repl output dir: " + outputDir.getAbsolutePath) outputDir.deleteOnExit() conf.set("spark.repl.class.outputDir", outputDir.getAbsolutePath) - val target = conf.get("spark.repl.target", "jvm-1.8") val settings = new Settings() settings.processArguments(List("-Yrepl-class-based", "-Yrepl-outdir", s"${outputDir.getAbsolutePath}"), true) settings.embeddedDefaults(sparkInterpreterClassLoader) settings.usejavacp.value = true - settings.target.value = target - this.userJars = getUserJars() LOGGER.info("UserJars: " + userJars.mkString(File.pathSeparator)) settings.classpath.value = userJars.mkString(File.pathSeparator)