From 63470afc997fb9d6b6f8a911c25964743556c9cc Mon Sep 17 00:00:00 2001 From: Nezih Yigitbasi Date: Thu, 16 Jun 2016 18:19:29 -0700 Subject: [SPARK-15782][YARN] Fix spark.jars and spark.yarn.dist.jars handling When `--packages` is specified with spark-shell the classes from those packages cannot be found, which I think is due to some of the changes in SPARK-12343. Tested manually with both scala 2.10 and 2.11 repls. vanzin davies can you guys please review? Author: Marcelo Vanzin Author: Nezih Yigitbasi Closes #13709 from nezihyigitbasi/SPARK-15782. --- repl/scala-2.11/src/main/scala/org/apache/spark/repl/Main.scala | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) (limited to 'repl/scala-2.11') diff --git a/repl/scala-2.11/src/main/scala/org/apache/spark/repl/Main.scala b/repl/scala-2.11/src/main/scala/org/apache/spark/repl/Main.scala index 771670fa55..28fe84d6fe 100644 --- a/repl/scala-2.11/src/main/scala/org/apache/spark/repl/Main.scala +++ b/repl/scala-2.11/src/main/scala/org/apache/spark/repl/Main.scala @@ -54,9 +54,7 @@ object Main extends Logging { // Visible for testing private[repl] def doMain(args: Array[String], _interp: SparkILoop): Unit = { interp = _interp - val jars = conf.getOption("spark.jars") - .map(_.replace(",", File.pathSeparator)) - .getOrElse("") + val jars = Utils.getUserJars(conf).mkString(File.pathSeparator) val interpArguments = List( "-Yrepl-class-based", "-Yrepl-outdir", s"${outputDir.getAbsolutePath}", -- cgit v1.2.3