diff options
author | Tathagata Das <tathagata.das1565@gmail.com> | 2013-12-31 00:42:10 -0800 |
---|---|---|
committer | Tathagata Das <tathagata.das1565@gmail.com> | 2013-12-31 00:42:10 -0800 |
commit | 87b915f22105ced8b9cad2a1262a0fd26542ee4f (patch) | |
tree | 227d14e03c1ad28c43e6d332fa7d27420df6f3b7 | |
parent | 3ab297adaae153cd76c9893ab62e0a815ec21fa4 (diff) | |
download | spark-87b915f22105ced8b9cad2a1262a0fd26542ee4f.tar.gz spark-87b915f22105ced8b9cad2a1262a0fd26542ee4f.tar.bz2 spark-87b915f22105ced8b9cad2a1262a0fd26542ee4f.zip |
Removed extra empty lines.
3 files changed, 0 insertions, 3 deletions
diff --git a/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala b/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala index fd69328aba..a2cd49c573 100644 --- a/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala +++ b/external/kafka/src/main/scala/org/apache/spark/streaming/kafka/KafkaInputDStream.scala @@ -34,7 +34,6 @@ import org.apache.spark.storage.StorageLevel import org.apache.spark.streaming.StreamingContext import org.apache.spark.streaming.dstream._ - /** * Input stream that pulls messages from a Kafka Broker. * diff --git a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala index 97e48ebeca..5cc721d7f9 100644 --- a/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala +++ b/external/twitter/src/main/scala/org/apache/spark/streaming/twitter/TwitterInputDStream.scala @@ -98,4 +98,3 @@ class TwitterReceiver( logInfo("Twitter receiver stopped") } } - diff --git a/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala b/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala index 4a8e15db21..75f7244643 100644 --- a/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala +++ b/streaming/src/main/scala/org/apache/spark/streaming/scheduler/NetworkInputTracker.scala @@ -164,7 +164,6 @@ class NetworkInputTracker( ssc.sparkContext.makeRDD(1 to 50, 50).map(x => (x, 1)).reduceByKey(_ + _, 20).collect() } - // Distribute the receivers and start them ssc.sparkContext.runJob(tempRDD, startReceiver) } |