aboutsummaryrefslogtreecommitdiff
path: root/bagel
Commit message (Collapse)AuthorAgeFilesLines
...
* Maven build now works with CDH hadoop-2.0.0-mr1Jey Kottalam2013-08-161-27/+0
|
* Initial changes to make Maven build agnostic of hadoop versionJey Kottalam2013-08-161-33/+10
|
* Add Apache license headers and LICENSE and NOTICE filesMatei Zaharia2013-07-167-1/+120
|
* Attempt to fix streaming test failures after yarn branch mergeMridul Muralidharan2013-04-281-0/+1
|
* Attempt at fixing merge conflictMridul Muralidharan2013-04-242-4/+4
|\
| * Fix passing of superstep in Bagel to avoid seeing new values of theMatei Zaharia2013-04-081-3/+3
| | | | | | | | | | superstep value upon recomputation, and set the default storage level in Bagel to MEMORY_AND_DISK
| * Bump development version to 0.8.0Jey Kottalam2013-03-281-1/+1
| |
* | Add support for building against hadoop2-yarn : adding new maven profile for itMridul Muralidharan2013-04-071-0/+37
|/
* Add a log4j compile dependency to fix build in IntelliJMikhail Bautin2013-03-151-1/+1
| | | | | Also rename parent project to spark-parent (otherwise it shows up as "parent" in IntelliJ, which is very confusing).
* Fix doc styleNick Pentreath2013-03-111-7/+13
|
* Adding test for non-default persistence levelNick Pentreath2013-03-091-0/+18
|
* Added choice of persitance level to Bagel. Also added documentation.Nick Pentreath2013-03-091-8/+83
|
* bump version to 0.7.1-SNAPSHOT in the subproject poms to keep the maven ↵Mark Hamstra2013-02-281-1/+1
| | | | build building.
* Renamed "splits" to "partitions"Matei Zaharia2013-02-173-14/+14
|
* Formatting fixesMatei Zaharia2013-02-111-13/+9
|
* Fixed an exponential recursion that could happen with doCheckpoint dueMatei Zaharia2013-02-111-8/+27
| | | | to lack of memoization
* Remove activation of profiles by defaultMikhail Bautin2013-01-311-11/+0
| | | | | See the discussion at https://github.com/mesos/spark/pull/355 for why default profile activation is a problem.
* Replace old 'master' term with 'driver'.Stephen Haberman2013-01-251-1/+1
|
* Merge branch 'master' into streamingTathagata Das2013-01-151-0/+11
|\ | | | | | | | | | | | | | | | | Conflicts: core/src/main/scala/spark/rdd/CoGroupedRDD.scala core/src/main/scala/spark/rdd/FilteredRDD.scala docs/_layouts/global.html docs/index.md run
| * Rename environment variable for hadoop profiles to hadoopVersionShivaram Venkataraman2013-01-121-2/+2
| |
| * Activate hadoop2 profile in pom.xml with -Dhadoop=2Shivaram Venkataraman2013-01-101-0/+6
| |
| * Activate hadoop1 if property hadoop is missing. hadoop2 can be activated nowShivaram Venkataraman2013-01-081-1/+3
| | | | | | | | by using -Dhadoop -Phadoop2.
| * Activate hadoop1 profile by default for maven buildsShivaram Venkataraman2013-01-071-0/+3
| |
* | Changed locations for unit test logs.Tathagata Das2013-01-071-2/+2
|/
* Mark hadoop dependencies provided in all library artifactsThomas Dudziak2012-12-101-0/+3
|
* Use the same output directories that SBT had in subprojectsMatei Zaharia2012-12-101-1/+3
| | | | This will make it easier to make the "run" script work with a Maven build
* Updated versions in the pom.xml files to match current masterThomas Dudziak2012-11-271-1/+1
|
* Addressed code review commentsThomas Dudziak2012-11-271-0/+1
|
* Added maven and debian build filesThomas Dudziak2012-11-201-0/+100
|
* Some doc fixes, including showing version number in nav bar againMatei Zaharia2012-10-131-0/+5
|
* More doc updates, and moved Serializer to a subpackage.Matei Zaharia2012-10-121-5/+6
|
* Removed the need to sleep in tests due to waiting for Akka to shut downMatei Zaharia2012-10-071-0/+2
|
* Write all unit test output to a fileMatei Zaharia2012-10-011-4/+6
|
* Changed the way tasks' dependency files are sent to workers so thatMatei Zaharia2012-09-281-1/+4
| | | | custom serializers or Kryo registrators can be loaded.
* Set log level in tests to WARNMatei Zaharia2012-08-231-0/+8
|
* Fix further issues with tests and broadcast.Matei Zaharia2012-08-231-1/+4
| | | | | The broadcast fix is to store values as MEMORY_ONLY_DESER instead of MEMORY_ONLY, which will save substantial time on serialization.
* Stlystic changesDenny2012-07-231-2/+2
| | | | | | Conflicts: core/src/test/scala/spark/MesosSchedulerSuite.scala
* Always destroy SparkContext in after block for the unit tests.Denny2012-07-231-6/+11
| | | | | | Conflicts: core/src/test/scala/spark/ShuffleSuite.scala
* Merge branch 'master' into devMatei Zaharia2012-06-152-3/+1
|\
| * Performance improvements to shuffle operations: in particular, preserveMatei Zaharia2012-06-092-3/+1
| | | | | | | | | | RDD partitioning in more cases where it's possible, and use iterators instead of materializing collections when doing joins.
* | Merge in engine improvements from the Spark Streaming project, developedMatei Zaharia2012-06-071-5/+6
|/ | | | | | jointly with Tathagata Das and Haoyuan Li. This commit imports the changes and ports them to Mesos 0.9, but does not yet pass unit tests due to various classes not supporting a graceful stop() yet.
* Added an option (spark.closure.serializer) to specify the serializer forReynold Xin2012-04-091-0/+4
| | | | closures. This enables using Kryo as the closure serializer.
* Update Bagel unit tests to reflect API changeAnkur Dave2011-11-081-23/+21
|
* Implement standalone WikipediaPageRank with custom serializerAnkur Dave2011-10-091-0/+198
|
* Update WikipediaPageRank to reflect Bagel API changesAnkur Dave2011-10-092-100/+129
|
* Remove ShortestPath for nowAnkur Dave2011-10-091-95/+0
|
* Simplify and genericize type parameters in BagelAnkur Dave2011-10-091-85/+129
|
* Fix issue #65: Change @serializable to extends Serializable in 2.9 branchIsmael Juma2011-08-024-23/+18
| | | | | | Note that we use scala.Serializable introduced in Scala 2.9 instead of java.io.Serializable. Also, case classes inherit from scala.Serializable by default.
* Cleaned up a few issues to do with default parallelism levels. AlsoMatei Zaharia2011-07-141-1/+1
| | | | | renamed HadoopFileWriter to HadoopWriter (since it's not only for files) and fixed a bug for lookup().
* Fixed unit tests by making them clean up the SparkContext after use andMatei Zaharia2011-05-131-0/+2
| | | | | | thus clean up the various singletons (RDDCache, MapOutputTracker, etc). This isn't perfect yet (ideally we shouldn't use singleton objects at all) but we can fix that later.