diff options
-rw-r--r-- | core/src/main/java/org/apache/spark/io/LZ4BlockInputStream.java (renamed from core/src/main/scala/org/apache/spark/io/LZ4BlockInputStream.java) | 14 | ||||
-rw-r--r-- | external/java8-tests/src/test/java/org/apache/spark/streaming/Java8APISuite.java | 6 | ||||
-rw-r--r-- | pom.xml | 2 | ||||
-rw-r--r-- | sql/core/src/main/java/org/apache/spark/sql/execution/BufferedRowIterator.java (renamed from sql/core/src/main/scala/org/apache/spark/sql/execution/BufferedRowIterator.java) | 2 | ||||
-rw-r--r-- | sql/core/src/main/java/org/apache/spark/sql/expressions/java/typed.java (renamed from sql/core/src/main/scala/org/apache/spark/sql/expressions/java/typed.java) | 11 |
5 files changed, 17 insertions, 18 deletions
diff --git a/core/src/main/scala/org/apache/spark/io/LZ4BlockInputStream.java b/core/src/main/java/org/apache/spark/io/LZ4BlockInputStream.java index 27b6f0d4a3..8783b5f56e 100644 --- a/core/src/main/scala/org/apache/spark/io/LZ4BlockInputStream.java +++ b/core/src/main/java/org/apache/spark/io/LZ4BlockInputStream.java @@ -20,20 +20,17 @@ import java.io.IOException; import java.io.InputStream; import java.util.zip.Checksum; -import net.jpountz.lz4.LZ4BlockOutputStream; import net.jpountz.lz4.LZ4Exception; import net.jpountz.lz4.LZ4Factory; import net.jpountz.lz4.LZ4FastDecompressor; import net.jpountz.util.SafeUtils; -import net.jpountz.xxhash.StreamingXXHash32; -import net.jpountz.xxhash.XXHash32; import net.jpountz.xxhash.XXHashFactory; /** * {@link InputStream} implementation to decode data written with - * {@link LZ4BlockOutputStream}. This class is not thread-safe and does not + * {@link net.jpountz.lz4.LZ4BlockOutputStream}. This class is not thread-safe and does not * support {@link #mark(int)}/{@link #reset()}. - * @see LZ4BlockOutputStream + * @see net.jpountz.lz4.LZ4BlockOutputStream * * This is based on net.jpountz.lz4.LZ4BlockInputStream * @@ -90,12 +87,13 @@ public final class LZ4BlockInputStream extends FilterInputStream { } /** - * Create a new instance using {@link XXHash32} for checksuming. + * Create a new instance using {@link net.jpountz.xxhash.XXHash32} for checksuming. * @see #LZ4BlockInputStream(InputStream, LZ4FastDecompressor, Checksum) - * @see StreamingXXHash32#asChecksum() + * @see net.jpountz.xxhash.StreamingXXHash32#asChecksum() */ public LZ4BlockInputStream(InputStream in, LZ4FastDecompressor decompressor) { - this(in, decompressor, XXHashFactory.fastestInstance().newStreamingHash32(DEFAULT_SEED).asChecksum()); + this(in, decompressor, + XXHashFactory.fastestInstance().newStreamingHash32(DEFAULT_SEED).asChecksum()); } /** diff --git a/external/java8-tests/src/test/java/org/apache/spark/streaming/Java8APISuite.java b/external/java8-tests/src/test/java/org/apache/spark/streaming/Java8APISuite.java index 67bc64a444..d0fed303e6 100644 --- a/external/java8-tests/src/test/java/org/apache/spark/streaming/Java8APISuite.java +++ b/external/java8-tests/src/test/java/org/apache/spark/streaming/Java8APISuite.java @@ -377,7 +377,9 @@ public class Java8APISuite extends LocalJavaStreamingContext implements Serializ }); // This is a test to make sure foreachRDD(VoidFunction2) can be called from Java - stream.foreachRDD((rdd, time) -> { return; }); + stream.foreachRDD((rdd, time) -> { + return; + }); JavaTestUtils.runStreams(ssc, 2, 2); @@ -873,7 +875,7 @@ public class Java8APISuite extends LocalJavaStreamingContext implements Serializ JavaMapWithStateDStream<String, Integer, Boolean, Double> stateDstream = wordsDstream.mapWithState( - StateSpec.<String, Integer, Boolean, Double> function((time, key, value, state) -> { + StateSpec.<String, Integer, Boolean, Double>function((time, key, value, state) -> { // Use all State's methods here state.exists(); state.get(); @@ -2253,7 +2253,7 @@ <failOnViolation>false</failOnViolation> <includeTestSourceDirectory>true</includeTestSourceDirectory> <failOnWarning>false</failOnWarning> - <sourceDirectory>${basedir}/src/main/java</sourceDirectory> + <sourceDirectories>${basedir}/src/main/java,${basedir}/src/main/scala</sourceDirectories> <testSourceDirectory>${basedir}/src/test/java</testSourceDirectory> <configLocation>dev/checkstyle.xml</configLocation> <outputFile>${basedir}/target/checkstyle-output.xml</outputFile> diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/BufferedRowIterator.java b/sql/core/src/main/java/org/apache/spark/sql/execution/BufferedRowIterator.java index c2633a9f8c..086547c793 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/BufferedRowIterator.java +++ b/sql/core/src/main/java/org/apache/spark/sql/execution/BufferedRowIterator.java @@ -60,7 +60,7 @@ public abstract class BufferedRowIterator { /** * Initializes from array of iterators of InternalRow. */ - public abstract void init(int index, Iterator<InternalRow> iters[]); + public abstract void init(int index, Iterator<InternalRow>[] iters); /** * Append a row to currentRows. diff --git a/sql/core/src/main/scala/org/apache/spark/sql/expressions/java/typed.java b/sql/core/src/main/java/org/apache/spark/sql/expressions/java/typed.java index 8ff7b6549b..c7c6e3868f 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/expressions/java/typed.java +++ b/sql/core/src/main/java/org/apache/spark/sql/expressions/java/typed.java @@ -19,7 +19,6 @@ package org.apache.spark.sql.expressions.java; import org.apache.spark.annotation.Experimental; import org.apache.spark.api.java.function.MapFunction; -import org.apache.spark.sql.Dataset; import org.apache.spark.sql.TypedColumn; import org.apache.spark.sql.execution.aggregate.TypedAverage; import org.apache.spark.sql.execution.aggregate.TypedCount; @@ -28,7 +27,7 @@ import org.apache.spark.sql.execution.aggregate.TypedSumLong; /** * :: Experimental :: - * Type-safe functions available for {@link Dataset} operations in Java. + * Type-safe functions available for {@link org.apache.spark.sql.Dataset} operations in Java. * * Scala users should use {@link org.apache.spark.sql.expressions.scala.typed}. * @@ -43,7 +42,7 @@ public class typed { * * @since 2.0.0 */ - public static<T> TypedColumn<T, Double> avg(MapFunction<T, Double> f) { + public static <T> TypedColumn<T, Double> avg(MapFunction<T, Double> f) { return new TypedAverage<T>(f).toColumnJava(); } @@ -52,7 +51,7 @@ public class typed { * * @since 2.0.0 */ - public static<T> TypedColumn<T, Long> count(MapFunction<T, Object> f) { + public static <T> TypedColumn<T, Long> count(MapFunction<T, Object> f) { return new TypedCount<T>(f).toColumnJava(); } @@ -61,7 +60,7 @@ public class typed { * * @since 2.0.0 */ - public static<T> TypedColumn<T, Double> sum(MapFunction<T, Double> f) { + public static <T> TypedColumn<T, Double> sum(MapFunction<T, Double> f) { return new TypedSumDouble<T>(f).toColumnJava(); } @@ -70,7 +69,7 @@ public class typed { * * @since 2.0.0 */ - public static<T> TypedColumn<T, Long> sumLong(MapFunction<T, Long> f) { + public static <T> TypedColumn<T, Long> sumLong(MapFunction<T, Long> f) { return new TypedSumLong<T>(f).toColumnJava(); } } |