diff options
author | Prashant Sharma <prashant.s@imaginea.com> | 2014-01-03 15:12:09 +0530 |
---|---|---|
committer | Prashant Sharma <prashant.s@imaginea.com> | 2014-01-03 15:12:29 +0530 |
commit | 9ae382c363202eac8ef0b8e0fe1a7eab3de545b1 (patch) | |
tree | 4f8c448dbd71e746b217c10a509c86826f72a65d | |
parent | 74ba97fcf75ae61833295a896dcdf775036ae1d8 (diff) | |
download | spark-9ae382c363202eac8ef0b8e0fe1a7eab3de545b1.tar.gz spark-9ae382c363202eac8ef0b8e0fe1a7eab3de545b1.tar.bz2 spark-9ae382c363202eac8ef0b8e0fe1a7eab3de545b1.zip |
sbin/compute-classpath* bin/compute-classpath*
-rw-r--r-- | bin/compute-classpath.cmd (renamed from sbin/compute-classpath.cmd) | 0 | ||||
-rwxr-xr-x | bin/compute-classpath.sh (renamed from sbin/compute-classpath.sh) | 0 | ||||
-rwxr-xr-x | bin/run-example | 2 | ||||
-rwxr-xr-x | bin/spark-class | 2 | ||||
-rw-r--r-- | core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala | 2 |
5 files changed, 3 insertions, 3 deletions
diff --git a/sbin/compute-classpath.cmd b/bin/compute-classpath.cmd index 4f60bff19c..4f60bff19c 100644 --- a/sbin/compute-classpath.cmd +++ b/bin/compute-classpath.cmd diff --git a/sbin/compute-classpath.sh b/bin/compute-classpath.sh index 0c82310421..0c82310421 100755 --- a/sbin/compute-classpath.sh +++ b/bin/compute-classpath.sh diff --git a/bin/run-example b/bin/run-example index f2699c38a9..6c5d4a6a8f 100755 --- a/bin/run-example +++ b/bin/run-example @@ -61,7 +61,7 @@ fi # Since the examples JAR ideally shouldn't include spark-core (that dependency should be # "provided"), also add our standard Spark classpath, built using compute-classpath.sh. -CLASSPATH=`$FWDIR/sbin/compute-classpath.sh` +CLASSPATH=`$FWDIR/bin/compute-classpath.sh` CLASSPATH="$SPARK_EXAMPLES_JAR:$CLASSPATH" if $cygwin; then diff --git a/bin/spark-class b/bin/spark-class index 4e440d8729..c4225a392d 100755 --- a/bin/spark-class +++ b/bin/spark-class @@ -128,7 +128,7 @@ if [ -e "$TOOLS_DIR"/target/spark-tools*[0-9Tg].jar ]; then fi # Compute classpath using external script -CLASSPATH=`$FWDIR/sbin/compute-classpath.sh` +CLASSPATH=`$FWDIR/bin/compute-classpath.sh` if [ "$1" == "org.apache.spark.tools.JavaAPICompletenessChecker" ]; then CLASSPATH="$CLASSPATH:$SPARK_TOOLS_JAR" diff --git a/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala b/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala index d4084820f6..fff9cb60c7 100644 --- a/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala +++ b/core/src/main/scala/org/apache/spark/deploy/worker/ExecutorRunner.scala @@ -122,7 +122,7 @@ private[spark] class ExecutorRunner( // Figure out our classpath with the external compute-classpath script val ext = if (System.getProperty("os.name").startsWith("Windows")) ".cmd" else ".sh" val classPath = Utils.executeAndGetOutput( - Seq(sparkHome + "/sbin/compute-classpath" + ext), + Seq(sparkHome + "/bin/compute-classpath" + ext), extraEnvironment=appDesc.command.environment) Seq("-cp", classPath) ++ libraryOpts ++ workerLocalOpts ++ userOpts ++ memoryOpts |