aboutsummaryrefslogtreecommitdiff
path: root/repl
Commit message (Collapse)AuthorAgeFilesLines
* Merge pull request #557 from ScrapCodes/style. Closes #557.Patrick Wendell2014-02-099-6/+25
| | | | | | | | | | | | | | | | | | | | | SPARK-1058, Fix Style Errors and Add Scala Style to Spark Build. Author: Patrick Wendell <pwendell@gmail.com> Author: Prashant Sharma <scrapcodes@gmail.com> == Merge branch commits == commit 1a8bd1c059b842cb95cc246aaea74a79fec684f4 Author: Prashant Sharma <scrapcodes@gmail.com> Date: Sun Feb 9 17:39:07 2014 +0530 scala style fixes commit f91709887a8e0b608c5c2b282db19b8a44d53a43 Author: Patrick Wendell <pwendell@gmail.com> Date: Fri Jan 24 11:22:53 2014 -0800 Adding scalastyle snapshot
* Merge pull request #542 from markhamstra/versionBump. Closes #542.Mark Hamstra2014-02-082-2/+2
| | | | | | | | | | | | | | | | | | Version number to 1.0.0-SNAPSHOT Since 0.9.0-incubating is done and out the door, we shouldn't be building 0.9.0-incubating-SNAPSHOT anymore. @pwendell Author: Mark Hamstra <markhamstra@gmail.com> == Merge branch commits == commit 1b00a8a7c1a7f251b4bb3774b84b9e64758eaa71 Author: Mark Hamstra <markhamstra@gmail.com> Date: Wed Feb 5 09:30:32 2014 -0800 Version number to 1.0.0-SNAPSHOT
* Add missing header filesPatrick Wendell2014-01-141-0/+17
|
* Removing mentions in testsPatrick Wendell2014-01-121-2/+0
|
* Merge pull request #327 from lucarosellini/masterMatei Zaharia2014-01-083-3/+73
|\ | | | | | | | | | | | | | | | | Added ‘-i’ command line option to Spark REPL We had to create a new implementation of both scala.tools.nsc.CompilerCommand and scala.tools.nsc.Settings, because using scala.tools.nsc.GenericRunnerSettings would bring in other options (-howtorun, -save and -execute) which don’t make sense in Spark. Any new Spark specific command line option could now be added to org.apache.spark.repl.SparkRunnerSettings class. Since the behavior of loading a script from the command line should be the same as loading it using the “:load” command inside the shell, the script should be loaded when the SparkContext is available, that’s why we had to move the call to ‘loadfiles(settings)’ _after_ the call to postInitialization(). This still doesn’t work if ‘isAsync = true’.
| * Added license header and removed @author tagLuca Rosellini2014-01-072-4/+34
| |
| * Added ‘-i’ command line option to spark REPL.Luca Rosellini2014-01-033-3/+43
| | | | | | | | | | | | | | We had to create a new implementation of both scala.tools.nsc.CompilerCommand and scala.tools.nsc.Settings, because using scala.tools.nsc.GenericRunnerSettings would bring in other options (-howtorun, -save and -execute) which don’t make sense in Spark. Any new Spark specific command line option could now be added to org.apache.spark.repl.SparkRunnerSettings class. Since the behavior of loading a script from the command line should be the same as loading it using the “:load” command inside the shell, the script should be loaded when the SparkContext is available, that’s why we had to move the call to ‘loadfiles(settings)’ _after_ the call to postInitialization(). This still doesn’t work if ‘isAsync = true’.
* | Merge remote-tracking branch 'apache-github/master' into remove-binariesPatrick Wendell2014-01-031-1/+0
|\ \ | | | | | | | | | | | | | | | Conflicts: core/src/test/scala/org/apache/spark/DriverSuite.scala docs/python-programming-guide.md
| * \ Merge branch 'scripts-reorg' of github.com:shane-huang/incubator-spark into ↵Prashant Sharma2014-01-021-1/+0
| |\ \ | | |/ | |/| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | spark-915-segregate-scripts Conflicts: bin/spark-shell core/pom.xml core/src/main/scala/org/apache/spark/SparkContext.scala core/src/main/scala/org/apache/spark/scheduler/cluster/mesos/CoarseMesosSchedulerBackend.scala core/src/main/scala/org/apache/spark/ui/UIWorkloadGenerator.scala core/src/test/scala/org/apache/spark/DriverSuite.scala python/run-tests sbin/compute-classpath.sh sbin/spark-class sbin/stop-slaves.sh
| | * deprecate "spark" script and SPAKR_CLASSPATH environment variableAndrew xia2013-10-121-1/+0
| | |
* | | fixed review commentsPrashant Sharma2014-01-031-1/+3
|/ /
* | Miscellaneous fixes from code review.Matei Zaharia2014-01-011-1/+1
| | | | | | | | | | | | Also replaced SparkConf.getOrElse with just a "get" that takes a default value, and added getInt, getLong, etc to make code that uses this simpler later on.
* | Various fixes to configuration codeMatei Zaharia2013-12-282-7/+13
| | | | | | | | | | | | | | | | | | | | | | | | | | | | - Got rid of global SparkContext.globalConf - Pass SparkConf to serializers and compression codecs - Made SparkConf public instead of private[spark] - Improved API of SparkContext and SparkConf - Switched executor environment vars to be passed through SparkConf - Fixed some places that were still using system properties - Fixed some tests, though others are still failing This still fails several tests in core, repl and streaming, likely due to properties not being set or cleared correctly (some of the tests run fine in isolation).
* | spark-544, introducing SparkConf and related configuration overhaul.Prashant Sharma2013-12-252-8/+6
| |
* | Use scala.binary.version in POMsMark Hamstra2013-12-151-7/+7
| |
* | Review comments on the PR for scala 2.10 migration.Prashant Sharma2013-12-131-1/+0
| |
* | Style fixes and addressed review comments at #221Prashant Sharma2013-12-101-7/+7
| |
* | Incorporated Patrick's feedback comment on #211 and made maven ↵Prashant Sharma2013-12-071-1/+1
| | | | | | | | build/dep-resolution atleast a bit faster.
* | Fixed compile time warnings and formatting post merge.Prashant Sharma2013-11-261-65/+74
| |
* | Various merge correctionsAaron Davidson2013-11-142-12/+4
| | | | | | | | | | | | | | | | | | | | I've diff'd this patch against my own -- since they were both created independently, this means that two sets of eyes have gone over all the merge conflicts that were created, so I'm feeling significantly more confident in the resulting PR. @rxin has looked at the changes to the repl and is resoundingly confident that they are correct.
* | Merge branch 'master' into scala-2.10Raymond Liu2013-11-141-2/+34
|\ \
| * | Propagate the SparkContext local property from the thread that calls the ↵Reynold Xin2013-11-092-4/+42
| | | | | | | | | | | | spark-repl to the actual execution thread.
* | | Merge branch 'master' into scala-2.10Raymond Liu2013-11-131-7/+29
|\| |
| * | Makes Spark SIMR ready.Ali Ghodsi2013-10-241-0/+14
| | |
| * | Spark shell exits if it cannot create SparkContextAaron Davidson2013-10-171-1/+8
| | | | | | | | | | | | | | | | | | Mainly, this occurs if you provide a messed up MASTER url (one that doesn't match one of our regexes). Previously, we would default to Mesos, fail, and then start the shell anyway, except that any Spark command would fail.
* | | Merge branch 'master' into wip-merge-masterPrashant Sharma2013-10-081-8/+8
|\| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Conflicts: bagel/pom.xml core/pom.xml core/src/test/scala/org/apache/spark/ui/UISuite.scala examples/pom.xml mllib/pom.xml pom.xml project/SparkBuild.scala repl/pom.xml streaming/pom.xml tools/pom.xml In scala 2.10, a shorter representation is used for naming artifacts so changed to shorter scala version for artifacts and made it a property in pom.
| * | Merging build changes in from 0.8Patrick Wendell2013-10-051-10/+10
| |/
* | Merge branch 'master' into scala-2.10Prashant Sharma2013-10-012-2/+2
|\| | | | | | | | | | | | | | | Conflicts: core/src/main/scala/org/apache/spark/ui/jobs/JobProgressUI.scala docs/_config.yml project/SparkBuild.scala repl/src/main/scala/org/apache/spark/repl/SparkILoop.scala
| * Bug fix in master buildPatrick Wendell2013-09-261-1/+1
| |
| * Update build version in masterPatrick Wendell2013-09-241-1/+1
| |
* | fixed maven build for scala 2.10Prashant Sharma2013-09-261-6/+6
| |
* | ported repl improvements from masterPrashant Sharma2013-09-152-2/+11
| |
* | Merge branch 'master' of git://github.com/mesos/spark into scala-2.10Prashant Sharma2013-09-151-12/+0
|\| | | | | | | | | | | Conflicts: core/src/main/scala/org/apache/spark/SparkContext.scala project/SparkBuild.scala
| * Minor YARN build cleanupsJey Kottalam2013-09-061-12/+0
| |
* | Fixed repl suitePrashant Sharma2013-09-152-7/+7
| |
* | Few more fixes to tests broken during mergePrashant Sharma2013-09-101-50/+0
| |
* | Merged with masterPrashant Sharma2013-09-0618-372/+369
|\|
| * Updated LICENSE with third-party licensesMatei Zaharia2013-09-021-0/+17
| |
| * Move some classes to more appropriate packages:Matei Zaharia2013-09-011-1/+1
| | | | | | | | | | | | * RDD, *RDDFunctions -> org.apache.spark.rdd * Utils, ClosureCleaner, SizeEstimator -> org.apache.spark.util * JavaSerializer, KryoSerializer -> org.apache.spark.serializer
| * Fix some URLsMatei Zaharia2013-09-011-1/+1
| |
| * Initial work to rename package to org.apache.sparkMatei Zaharia2013-09-0112-30/+30
| |
| * Synced sbt and maven buildsMark Hamstra2013-08-211-0/+6
| |
| * Remove redundant dependencies from POMsJey Kottalam2013-08-181-15/+0
| |
| * Updates to repl and example POMs to match SBT buildJey Kottalam2013-08-161-6/+0
| |
| * Maven build now also works with YARNJey Kottalam2013-08-161-63/+1
| |
| * Maven build now works with CDH hadoop-2.0.0-mr1Jey Kottalam2013-08-161-54/+10
| |
| * Initial changes to make Maven build agnostic of hadoop versionJey Kottalam2013-08-161-50/+22
| |
| * Set SPARK_CLASSPATH for maven repl testsShivaram Venkataraman2013-08-131-0/+30
| |
| * Format cleanup.Benjamin Hindman2013-07-301-1/+3
| |
| * Added property 'spark.executor.uri' for launching on Mesos withoutBenjamin Hindman2013-07-291-0/+2
| | | | | | | | | | | | | | requiring Spark to be installed. Using 'make_distribution.sh' a user can put a Spark distribution at a URI supported by Mesos (e.g., 'hdfs://...') and then set that when launching their job. Also added SPARK_EXECUTOR_URI for the REPL.