diff options
author | Denny <dennybritz@gmail.com> | 2012-11-11 11:15:31 -0800 |
---|---|---|
committer | Denny <dennybritz@gmail.com> | 2012-11-11 11:15:31 -0800 |
commit | 0fd4c93f1c349f052f633fea64f975d53976bd9c (patch) | |
tree | 1d5f10c76dd2ed67d408f0609e2c3683ca1ab216 | |
parent | deb2c4df72f65f2bd90cc97a9abcd59a12eecabc (diff) | |
download | spark-0fd4c93f1c349f052f633fea64f975d53976bd9c.tar.gz spark-0fd4c93f1c349f052f633fea64f975d53976bd9c.tar.bz2 spark-0fd4c93f1c349f052f633fea64f975d53976bd9c.zip |
Updated comment.
-rw-r--r-- | streaming/src/main/scala/spark/streaming/input/KafkaInputDStream.scala | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/streaming/src/main/scala/spark/streaming/input/KafkaInputDStream.scala b/streaming/src/main/scala/spark/streaming/input/KafkaInputDStream.scala index cc74855983..318537532c 100644 --- a/streaming/src/main/scala/spark/streaming/input/KafkaInputDStream.scala +++ b/streaming/src/main/scala/spark/streaming/input/KafkaInputDStream.scala @@ -104,7 +104,7 @@ class KafkaReceiver(streamId: Int, host: String, port: Int, groupId: String, // Handles pushing data into the BlockManager lazy protected val dataHandler = new KafkaDataHandler(this, storageLevel) - // Keeps track of the current offsets. Maps from (topic, partitionID) -> Offset + // Keeps track of the current offsets. Maps from (broker, topic, group, part) -> Offset lazy val offsets = HashMap[KafkaPartitionKey, Long]() // Connection to Kafka var consumerConnector : ZookeeperConsumerConnector = null |