diff options
Diffstat (limited to 'external')
2 files changed, 6 insertions, 9 deletions
diff --git a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaDirectKafkaStreamSuite.java b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaDirectKafkaStreamSuite.java index fbdfbf7e50..4891e4f4a1 100644 --- a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaDirectKafkaStreamSuite.java +++ b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaDirectKafkaStreamSuite.java @@ -35,6 +35,7 @@ import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaRDD; import org.apache.spark.api.java.JavaPairRDD; import org.apache.spark.api.java.function.Function; +import org.apache.spark.api.java.function.VoidFunction; import org.apache.spark.streaming.Durations; import org.apache.spark.streaming.api.java.JavaDStream; import org.apache.spark.streaming.api.java.JavaStreamingContext; @@ -130,17 +131,15 @@ public class JavaDirectKafkaStreamSuite implements Serializable { JavaDStream<String> unifiedStream = stream1.union(stream2); final Set<String> result = Collections.synchronizedSet(new HashSet<String>()); - unifiedStream.foreachRDD( - new Function<JavaRDD<String>, Void>() { + unifiedStream.foreachRDD(new VoidFunction<JavaRDD<String>>() { @Override - public Void call(JavaRDD<String> rdd) { + public void call(JavaRDD<String> rdd) { result.addAll(rdd.collect()); for (OffsetRange o : offsetRanges.get()) { System.out.println( o.topic() + " " + o.partition() + " " + o.fromOffset() + " " + o.untilOffset() ); } - return null; } } ); diff --git a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java index 1e69de46cd..617c92a008 100644 --- a/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java +++ b/external/kafka/src/test/java/org/apache/spark/streaming/kafka/JavaKafkaStreamSuite.java @@ -31,6 +31,7 @@ import org.junit.Test; import org.apache.spark.SparkConf; import org.apache.spark.api.java.JavaPairRDD; import org.apache.spark.api.java.function.Function; +import org.apache.spark.api.java.function.VoidFunction; import org.apache.spark.storage.StorageLevel; import org.apache.spark.streaming.Duration; import org.apache.spark.streaming.api.java.JavaDStream; @@ -103,10 +104,9 @@ public class JavaKafkaStreamSuite implements Serializable { } ); - words.countByValue().foreachRDD( - new Function<JavaPairRDD<String, Long>, Void>() { + words.countByValue().foreachRDD(new VoidFunction<JavaPairRDD<String, Long>>() { @Override - public Void call(JavaPairRDD<String, Long> rdd) { + public void call(JavaPairRDD<String, Long> rdd) { List<Tuple2<String, Long>> ret = rdd.collect(); for (Tuple2<String, Long> r : ret) { if (result.containsKey(r._1())) { @@ -115,8 +115,6 @@ public class JavaKafkaStreamSuite implements Serializable { result.put(r._1(), r._2()); } } - - return null; } } ); |