aboutsummaryrefslogtreecommitdiff
Commit message (Collapse)AuthorAgeFilesLines
* Merge pull request #245 from gregakespret/task-maxfailures-fixReynold Xin2013-12-163-5/+5
|\ | | | | | | | | | | | | | | | | | | | | | | | | | | Fix for spark.task.maxFailures not enforced correctly. Docs at http://spark.incubator.apache.org/docs/latest/configuration.html say: ``` spark.task.maxFailures Number of individual task failures before giving up on the job. Should be greater than or equal to 1. Number of allowed retries = this value - 1. ``` Previous implementation worked incorrectly. When for example `spark.task.maxFailures` was set to 1, the job was aborted only after the second task failure, not after the first one.
| * Fix tests.Grega Kespret2013-12-102-2/+2
| |
| * Fix for spark.task.maxFailures not enforced correctly.Grega Kespret2013-12-091-3/+3
| |
* | Merge pull request #265 from markhamstra/scala.binary.versionPatrick Wendell2013-12-1511-70/+71
|\ \ | | | | | | | | | | | | | | | | | | | | | DRY out the POMs with scala.binary.version ...instead of hard-coding 2.10 repeatedly. As long as it's not a `<project>`-level `<artifactId>`, I think that we are okay parameterizing these.
| * | Use scala.binary.version in POMsMark Hamstra2013-12-1511-70/+71
| | |
* | | Merge pull request #256 from MLnick/masterJosh Rosen2013-12-151-2/+6
|\ \ \ | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Fix 'IPYTHON=1 ./pyspark' throwing ValueError This fixes an annoying issue where running ```IPYTHON=1 ./pyspark``` resulted in: ``` Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /__ / .__/\_,_/_/ /_/\_\ version 0.8.0 /_/ Using Python version 2.7.5 (default, Jun 20 2013 11:06:30) Spark context avaiable as sc. --------------------------------------------------------------------------- ValueError Traceback (most recent call last) /usr/local/lib/python2.7/site-packages/IPython/utils/py3compat.pyc in execfile(fname, *where) 202 else: 203 filename = fname --> 204 __builtin__.execfile(filename, *where) /Users/Nick/workspace/scala/spark-0.8.0-incubating-bin-hadoop1/python/pyspark/shell.py in <module>() 30 add_files = os.environ.get("ADD_FILES").split(',') if os.environ.get("ADD_FILES") != None else None 31 ---> 32 sc = SparkContext(os.environ.get("MASTER", "local"), "PySparkShell", pyFiles=add_files) 33 34 print """Welcome to /Users/Nick/workspace/scala/spark-0.8.0-incubating-bin-hadoop1/python/pyspark/context.pyc in __init__(self, master, jobName, sparkHome, pyFiles, environment, batchSize) 70 with SparkContext._lock: 71 if SparkContext._active_spark_context: ---> 72 raise ValueError("Cannot run multiple SparkContexts at once") 73 else: 74 SparkContext._active_spark_context = self ValueError: Cannot run multiple SparkContexts at once ``` The issue arises since previously IPython didn't seem to respect ```$PYTHONSTARTUP```, but since at least 1.0.0 it has. Technically this might break for older versions of IPython, but most users should be able to upgrade IPython to at least 1.0.0 (and should be encouraged to do so :). New behaviour: ``` Nicks-MacBook-Pro:incubator-spark-mlnick Nick$ IPYTHON=1 ./pyspark Python 2.7.5 (default, Jun 20 2013, 11:06:30) Type "copyright", "credits" or "license" for more information. IPython 1.1.0 -- An enhanced Interactive Python. ? -> Introduction and overview of IPython's features. %quickref -> Quick reference. help -> Python's own help system. object? -> Details about 'object', use 'object??' for extra details. SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/Users/Nick/workspace/scala/incubator-spark-mlnick/tools/target/scala-2.9.3/spark-tools-assembly-0.9.0-incubating-SNAPSHOT.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/Users/Nick/workspace/scala/incubator-spark-mlnick/assembly/target/scala-2.9.3/spark-assembly-0.9.0-incubating-SNAPSHOT-hadoop1.0.4.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] 13/12/12 13:08:15 WARN Utils: Your hostname, Nicks-MacBook-Pro.local resolves to a loopback address: 127.0.0.1; using 10.0.0.4 instead (on interface en0) 13/12/12 13:08:15 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address 13/12/12 13:08:15 INFO Slf4jEventHandler: Slf4jEventHandler started 13/12/12 13:08:15 INFO SparkEnv: Registering BlockManagerMaster 13/12/12 13:08:15 INFO DiskBlockManager: Created local directory at /var/folders/_l/06wxljt13wqgm7r08jlc44_r0000gn/T/spark-local-20131212130815-0e76 13/12/12 13:08:15 INFO MemoryStore: MemoryStore started with capacity 326.7 MB. 13/12/12 13:08:15 INFO ConnectionManager: Bound socket to port 53732 with id = ConnectionManagerId(10.0.0.4,53732) 13/12/12 13:08:15 INFO BlockManagerMaster: Trying to register BlockManager 13/12/12 13:08:15 INFO BlockManagerMasterActor$BlockManagerInfo: Registering block manager 10.0.0.4:53732 with 326.7 MB RAM 13/12/12 13:08:15 INFO BlockManagerMaster: Registered BlockManager 13/12/12 13:08:15 INFO HttpBroadcast: Broadcast server started at http://10.0.0.4:53733 13/12/12 13:08:15 INFO SparkEnv: Registering MapOutputTracker 13/12/12 13:08:15 INFO HttpFileServer: HTTP File server directory is /var/folders/_l/06wxljt13wqgm7r08jlc44_r0000gn/T/spark-8f40e897-8211-4628-a7a8-755562d5244c 13/12/12 13:08:16 INFO SparkUI: Started Spark Web UI at http://10.0.0.4:4040 2013-12-12 13:08:16.337 java[56801:4003] Unable to load realm info from SCDynamicStore Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /__ / .__/\_,_/_/ /_/\_\ version 0.9.0-SNAPSHOT /_/ Using Python version 2.7.5 (default, Jun 20 2013 11:06:30) Spark context avaiable as sc. ```
| * | | Making IPython PySpark compatible across versions <1.0.0. Also cleaned up ↵Nick Pentreath2013-12-151-1/+6
| | | | | | | | | | | | | | | | '-i' option and made IPYTHON_OPTS work
| * | | Merge remote-tracking branch 'upstream/master'Nick Pentreath2013-12-15197-2905/+3470
| |\| |
| * | | Fix 'IPYTHON=1 ./pyspark' throwing 'ValueError: Cannot run multiple ↵Nick Pentreath2013-12-121-2/+1
| | | | | | | | | | | | | | | | SparkContexts at once'
* | | | Merge pull request #257 from tgravescs/sparkYarnFixNameReynold Xin2013-12-152-0/+2
|\ \ \ \ | | | | | | | | | | | | | | | | | | | | | | | | | Fix the --name option for Spark on Yarn Looks like the --name option accidentally got broken in one of the merges. The Client hangs if the --name option is used right now.
| * | | | Fix the --name option for Spark on YarnThomas Graves2013-12-122-0/+2
| |/ / /
* | | | Merge pull request #264 from shivaram/spark-class-fixReynold Xin2013-12-151-1/+1
|\ \ \ \ | |_|/ / |/| | | | | | | Use CoarseGrainedExecutorBackend in spark-class
| * | | Use CoarseGrainedExecutorBackend in spark-classShivaram Venkataraman2013-12-151-1/+1
|/ / /
* | | Merge pull request #251 from pwendell/masterReynold Xin2013-12-141-5/+7
|\ \ \ | | | | | | | | | | | | | | | | | | | | Fix list rendering in YARN markdown docs. This is some minor clean-up which makes the list render correctly.
| * | | Fix list rendering in YARN markdown docs.Patrick Wendell2013-12-101-5/+7
| | | |
* | | | Merge pull request #249 from ngbinh/partitionInJavaSortByKeyJosh Rosen2013-12-141-0/+14
|\ \ \ \ | | | | | | | | | | | | | | | | | | | | | | | | | Expose numPartitions parameter in JavaPairRDD.sortByKey() This change makes Java and Scala API on sortByKey() the same.
| * | | | Hook directly to Scala APIBinh Nguyen2013-12-101-8/+6
| | | | |
| * | | | Leave default value of numPartitions to Scala code.Binh Nguyen2013-12-101-2/+8
| | | | |
| * | | | Use braces to shorten the line.Binh Nguyen2013-12-101-1/+3
| | | | |
| * | | | Expose numPartitions parameter in JavaPairRDD.sortByKey()Binh Nguyen2013-12-101-2/+10
| | | | | | | | | | | | | | | | | | | | This change make Java and Scala API on sortByKey() the same.
* | | | | Merge pull request #259 from pwendell/scala-2.10Patrick Wendell2013-12-14196-2900/+3449
|\ \ \ \ \ | |_|_|/ / |/| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Migration to Scala 2.10 == Below description was written by Prashant Sharma == This PR migrates spark to scala 2.10. Summary of changes apart from scala 2.10 migration: (has no implications for user.) 1. Migrated Akka to 2.2.3. Does not use remote death watch for it has a bug, where it tries to send message to dead node infinitely. Uses an indestructible actorsystem which tolerates errors only on executors. (Might be useful for user.) 4. New configuration settings introduced: System.getProperty("spark.akka.heartbeat.pauses", "600") System.getProperty("spark.akka.failure-detector.threshold", "300.0") System.getProperty("spark.akka.heartbeat.interval", "1000") Defaults for these are fairly large to only disable Failure detector that comes with akka. The reason for doing so is we have our own failure detector like mechanism in place and then this is just an overhead on top of that + it leads to a lot of false positives. But with these properties it is possible to enable them. A good use case for enabling it could be when someone wants spark to be sensitive (in a controllable manner ofc.) to GC pauses/Network lags and quickly evict executors that experienced it. More information is included in configuration.md Once we have the SPARK-544 merged, I had like to deprecate atleast these akka properties and may be others too. This PR is duplicate of #221(where all the discussion happened.) for that one pointed to master this one points to scala-2.10 branch.
| * | | | Merge pull request #262 from pwendell/mvn-fixPatrick Wendell2013-12-132-1/+5
| |\ \ \ \ | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Fix maven build issues in 2.10 branch Found some issues when locally testing maven.
| | * | | | Fix maven build issues in 2.10 branchPatrick Wendell2013-12-132-1/+5
| |/ / / /
| * | | | Merge pull request #261 from ScrapCodes/scala-2.10Reynold Xin2013-12-131-0/+7
| |\ \ \ \ | | | | | | | | | | | | | | | | | | Added a comment about ActorRef and ActorSelection difference.
| | * | | | Added a comment about ActorRef and ActorSelection difference.Prashant Sharma2013-12-141-0/+7
| |/ / / /
| * | | | Merge pull request #260 from ScrapCodes/scala-2.10Reynold Xin2013-12-1315-49/+29
| |\ \ \ \ | | | | | | | | | | | | | | | | | | Review comments on the PR for scala 2.10 migration.
| | * | | | Review comments on the PR for scala 2.10 migration.Prashant Sharma2013-12-1315-49/+29
| |/ / / /
| * | | | Merge pull request #255 from ScrapCodes/scala-2.10Patrick Wendell2013-12-122-37/+47
| |\ \ \ \ | | | | | | | | | | | | | | | | | | Disabled yarn 2.2 in sbt and mvn build and added a message in the sbt build.
| | * | | | Disabled yarn 2.2 and added a message in the sbt buildPrashant Sharma2013-12-122-37/+47
| |/ / / /
| * | | | Merge pull request #254 from ScrapCodes/scala-2.10Patrick Wendell2013-12-11337-6632/+18488
| |\ \ \ \ | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Scala 2.10 migration This PR migrates spark to scala 2.10. Summary of changes apart from scala 2.10 migration: (has no implications for user.) 1. Migrated Akka to 2.2.3. Does not use remote death watch for it has a bug, where it tries to send message to dead node infinitely. Uses an indestructible actorsystem which tolerates errors only on executors. (Might be useful for user.) 4. New configuration settings introduced: System.getProperty("spark.akka.heartbeat.pauses", "600") System.getProperty("spark.akka.failure-detector.threshold", "300.0") System.getProperty("spark.akka.heartbeat.interval", "1000") Defaults for these are fairly large to only disable Failure detector that comes with akka. The reason for doing so is we have our own failure detector like mechanism in place and then this is just an overhead on top of that + it leads to a lot of false positives. But with these properties it is possible to enable them. A good use case for enabling it could be when someone wants spark to be sensitive (in a controllable manner ofc.) to GC pauses/Network lags and quickly evict executors that experienced it. More information is included in configuration.md Once we have the SPARK-544 merged, I had like to deprecate atleast these akka properties and may be others too. This PR is duplicate of #221(where all the discussion happened.) for that one pointed to master this one points to scala-2.10 branch.
| | * | | | A few corrections to documentation.Prashant Sharma2013-12-121-7/+7
| | | | | |
| | * | | | Merge branch 'akka-bug-fix' of github.com:ScrapCodes/incubator-spark into ↵Prashant Sharma2013-12-111-1/+1
| | |\ \ \ \ | | | | | | | | | | | | | | | | | | | | | akka-bug-fix
| | | * | | | added eclipse repository for spark streaming.Prashant Sharma2013-12-111-1/+1
| | | | | | |
| | * | | | | Merge branch 'master' into akka-bug-fixPrashant Sharma2013-12-1164-645/+4656
| | |\ \ \ \ \ | |_|/ / / / / |/| | / / / / | | |/ / / / | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Conflicts: core/pom.xml core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala pom.xml project/SparkBuild.scala streaming/pom.xml yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocationHandler.scala
* | | | | | Merge pull request #250 from pwendell/masterPatrick Wendell2013-12-101-3/+0
|\ \ \ \ \ \ | |_|_|/ / / |/| | | / / | | |_|/ / | |/| | | | | | | | | | | | | | | | | | | | | | | README incorrectly suggests build sources spark-env.sh This is misleading because the build doesn't source that file. IMO it's better to force people to specify build environment variables on the command line always, like we do in every example, so I'm just removing this doc.
| * | | | README incorrectly suggests build sources spark-env.shPatrick Wendell2013-12-101-3/+0
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This is misleading because the build doesn't source that file. IMO it's better to force people to specify build environment variables on the command line always, like we do in every example.
* | | | | Merge pull request #246 from pwendell/masterPatrick Wendell2013-12-097-0/+119
|\| | | | | |_|_|/ |/| | | | | | | | | | | | | | | Add missing license headers I found this when doing further audits on the 0.8.1 release candidate.
| * | | License headersPatrick Wendell2013-12-097-0/+119
| | | |
* | | | Merge pull request #195 from dhardy92/fix_DebScriptPackagePatrick Wendell2013-12-082-2/+2
|\ \ \ \ | | | | | | | | | | | | | | | [Deb] fix package of Spark classes adding org.apache prefix in scripts embeded in .deb
| * | | | [Deb] fix package of Spark classes adding org.apache prefix in scripts ↵dhardy922013-11-212-2/+2
| | | | | | | | | | | | | | | | | | | | embeded in .deb
* | | | | Merge pull request #242 from pwendell/masterPatrick Wendell2013-12-076-10/+11
|\ \ \ \ \ | | |/ / / | |/| | | | | | | | | | | | | | | | | | Update broken links and add HDP 2.0 version string I ran a link checker on the UI and found several broken links.
| * | | | Small fixPatrick Wendell2013-12-071-1/+1
| | | | |
| * | | | Adding HDP 2.0 versionPatrick Wendell2013-12-071-1/+2
| | | | |
| * | | | Various broken links in documentationPatrick Wendell2013-12-076-10/+10
|/ / / /
* | | | Merge pull request #240 from pwendell/masterPatrick Wendell2013-12-071-4/+4
|\ \ \ \ | | | | | | | | | | | | | | | SPARK-917 Improve API links in nav bar
| * | | | SPARK-917 Improve API links in nav barPatrick Wendell2013-12-071-4/+4
| | | | |
* | | | | Merge pull request #239 from aarondav/nitPatrick Wendell2013-12-071-1/+1
|\ \ \ \ \ | |/ / / / |/| | | | | | | | | Correct spellling error in configuration.md
| * | | | Correct spellling error in configuration.mdAaron Davidson2013-12-071-1/+1
|/ / / /
* | | | Merge pull request #237 from pwendell/formatting-fixPatrick Wendell2013-12-061-1/+0
|\ \ \ \ | | | | | | | | | | | | | | | | | | | | | | | | | Formatting fix This is a single-line change. The diff appears larger here due to github being out of sync.
| * | | | Minor formatting fix in config filePatrick Wendell2013-12-061-1/+0
|/ / / /