diff options
author | Andre Schumacher <schumach@icsi.berkeley.edu> | 2013-08-12 18:00:35 -0700 |
---|---|---|
committer | Andre Schumacher <schumach@icsi.berkeley.edu> | 2013-08-12 20:22:52 -0700 |
commit | 8fd5c7bc00b1104e4282959ec95b699955ded976 (patch) | |
tree | e665d241bcb5a866f7bbb5c39ed99bf0aad096c1 | |
parent | a4611d66f0d2ebc4425f385988d541b8f930e505 (diff) | |
download | spark-8fd5c7bc00b1104e4282959ec95b699955ded976.tar.gz spark-8fd5c7bc00b1104e4282959ec95b699955ded976.tar.bz2 spark-8fd5c7bc00b1104e4282959ec95b699955ded976.zip |
Implementing SPARK-865: Add the equivalent of ADD_JARS to PySpark
Now ADD_FILES uses a comma as file name separator.
-rw-r--r-- | python/pyspark/shell.py | 7 |
1 files changed, 6 insertions, 1 deletions
diff --git a/python/pyspark/shell.py b/python/pyspark/shell.py index 54ff1bf8e7..c8297b662e 100644 --- a/python/pyspark/shell.py +++ b/python/pyspark/shell.py @@ -7,10 +7,15 @@ import os import pyspark from pyspark.context import SparkContext +# this is the equivalent of ADD_JARS +add_files = os.environ.get("ADD_FILES").split(',') if os.environ.get("ADD_FILES") != None else None -sc = SparkContext(os.environ.get("MASTER", "local"), "PySparkShell") +sc = SparkContext(os.environ.get("MASTER", "local"), "PySparkShell", pyFiles=add_files) print "Spark context avaiable as sc." +if add_files != None: + print "Adding files: [%s]" % ", ".join(add_files) + # The ./pyspark script stores the old PYTHONSTARTUP value in OLD_PYTHONSTARTUP, # which allows us to execute the user's PYTHONSTARTUP file: _pythonstartup = os.environ.get('OLD_PYTHONSTARTUP') |