Toree removed support for PySpark and R kernels. See: https://issues.apache.org/jira/browse/TOREE-487 https://issues.apache.org/jira/browse/TOREE-488
On Fri, Feb 18, 2022 at 11:21 AM Artemis User <[email protected]> wrote: > I was able to install version 0.5.0 and it works great with JupyterHub! > Thanks again for the help! However, when I tried to install the PySpark > and the SparkR kernel (using the command line option > --interpreters=Scala,PySpark,SparkR,SQL), I got the following error > messages: > > [ToreeInstall] ERROR | Unknown interpreter PySpark. Skipping > installation of PySpark interpreter > [ToreeInstall] ERROR | Unknown interpreter SparkR. Skipping installation > of SparkR interpreter > > Are there two kernels available in 0.5.0? > > > On 2/17/22 8:30 PM, Kevin Bates wrote: > > Hello, > > > > You should checkout v0.5.0-rc5: > https://github.com/apache/incubator-toree/releases/tag/v0.5.0-incubating-rc5 > which includes support for Spark 3.2. > > > > On 2022/02/17 22:27:30 Artemis User wrote: > >> After looking at the Toree's Readme on github, I realized that the toree > >> version 0.4.x only supports Spark version 2.x, whereas the master branch > >> supports Spark 3.2.x. Could someone confirm this? In addition, is a > >> distribution package of toree from the master branch available? > >> > >> Thanks! > >> > >> On 2/17/22 4:57 PM, Artemis User wrote: > >>> Hi Toree Dev Team, Could someone please help with resolving the > >>> following error when starting the Toree Scala Kernel inside Jupyter? > >>> My configuration settings: > >>> > >>> * JupyterHub version 3.2.9 > >>> * OpenJDK 11 > >>> * Spark 3.2.0 with Scala version 2.12 > >>> * Apache Toree 0.4.0-incubating > >>> > >>> Thanks a lot for your help! > >>> > >>> Exception in thread "main" java.lang.NoClassDefFoundError: > >>> scala/App$class > >>> at org.apache.toree.Main$.<init>(Main.scala:24) > >>> at org.apache.toree.Main$.<clinit>(Main.scala) > >>> at org.apache.toree.Main.main(Main.scala) > >>> at > >>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native > >>> Method) > >>> at > >>> > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > >>> at > >>> > java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > >>> at java.base/java.lang.reflect.Method.invoke(Method.java:566) > >>> at > >>> > org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) > >>> at > >>> org.apache.spark.deploy.SparkSubmit.org > $apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:955) > >>> at > >>> org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180) > >>> at > org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203) > >>> at > org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90) > >>> at > >>> > org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043) > >>> at > org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052) > >>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > >>> Caused by: java.lang.ClassNotFoundException: scala.App$class > >>> at > >>> java.base/java.net.URLClassLoader.findClass(URLClassLoader.java:476) > >>> at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:589) > >>> at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:522) > >>> ... 15 more > >>> > >> > > -- Luciano Resende http://twitter.com/lresende1975 http://lresende.blogspot.com/
