aboutsummaryrefslogtreecommitdiff
path: root/examples/src/main/java
diff options
context:
space:
mode:
authorDongjoon Hyun <dongjoon@apache.org>2016-05-05 14:37:50 -0700
committerAndrew Or <andrew@databricks.com>2016-05-05 14:37:50 -0700
commit2c170dd3d731bd848d62265431795e1c141d75d7 (patch)
treed81ec5e4a6adfda683d7882680d50d2261b06818 /examples/src/main/java
parentbb9991dec5dd631b22a05e2e1b83b9082a845e8f (diff)
downloadspark-2c170dd3d731bd848d62265431795e1c141d75d7.tar.gz
spark-2c170dd3d731bd848d62265431795e1c141d75d7.tar.bz2
spark-2c170dd3d731bd848d62265431795e1c141d75d7.zip
[SPARK-15134][EXAMPLE] Indent SparkSession builder patterns and update binary_classification_metrics_example.py
## What changes were proposed in this pull request? This issue addresses the comments in SPARK-15031 and also fix java-linter errors. - Use multiline format in SparkSession builder patterns. - Update `binary_classification_metrics_example.py` to use `SparkSession`. - Fix Java Linter errors (in SPARK-13745, SPARK-15031, and so far) ## How was this patch tested? After passing the Jenkins tests and run `dev/lint-java` manually. Author: Dongjoon Hyun <dongjoon@apache.org> Closes #12911 from dongjoon-hyun/SPARK-15134.
Diffstat (limited to 'examples/src/main/java')
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java7
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java10
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java14
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java10
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java5
38 files changed, 165 insertions, 58 deletions
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java
index ecb7084e03..2c2aa6df47 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java
@@ -33,7 +33,10 @@ import org.apache.spark.sql.types.*;
public class JavaAFTSurvivalRegressionExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaAFTSurvivalRegressionExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaAFTSurvivalRegressionExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java
index 9a9a10489b..4b13ba6f9c 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java
@@ -81,7 +81,10 @@ public class JavaALSExample {
// $example off$
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaALSExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaALSExample")
+ .getOrCreate();
// $example on$
JavaRDD<Rating> ratingsRDD = spark
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java
index 88e4298a61..5f964aca92 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -26,7 +24,6 @@ import org.apache.spark.sql.SparkSession;
import java.util.Arrays;
import java.util.List;
-import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.ml.feature.Binarizer;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.RowFactory;
@@ -38,7 +35,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaBinarizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaBinarizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaBinarizerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java
index 51aa35084e..810ad905c5 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java
@@ -42,7 +42,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaBisectingKMeansExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaBisectingKMeansExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaBisectingKMeansExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java
index 0c24f52cf5..691df3887a 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java
@@ -35,7 +35,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaBucketizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaBucketizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaBucketizerExample")
+ .getOrCreate();
// $example on$
double[] splits = {Double.NEGATIVE_INFINITY, -0.5, 0.0, 0.5, Double.POSITIVE_INFINITY};
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java
index 684cf9a714..f8f2fb14be 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java
@@ -17,9 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaRDD;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -40,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaChiSqSelectorExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaChiSqSelectorExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaChiSqSelectorExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java
index 0631f9d6d5..0a6b136014 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java
@@ -32,7 +32,10 @@ import org.apache.spark.sql.types.*;
public class JavaCountVectorizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaCountVectorizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaCountVectorizerExample")
+ .getOrCreate();
// $example on$
// Input data: Each row is a bag of words from a sentence or document.
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java
index ec57a24451..eee92c77a8 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -26,7 +24,6 @@ import org.apache.spark.sql.SparkSession;
import java.util.Arrays;
import java.util.List;
-import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.ml.feature.DCT;
import org.apache.spark.mllib.linalg.VectorUDT;
import org.apache.spark.mllib.linalg.Vectors;
@@ -39,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaDCTExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaDCTExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaDCTExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
index 90023ac06b..49bad0afc0 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
@@ -49,7 +49,10 @@ import org.apache.spark.sql.SparkSession;
public class JavaDeveloperApiExample {
public static void main(String[] args) throws Exception {
- SparkSession spark = SparkSession.builder().appName("JavaDeveloperApiExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaDeveloperApiExample")
+ .getOrCreate();
// Prepare training data.
List<LabeledPoint> localTraining = Lists.newArrayList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java
index a062a6fcd0..9126242f9e 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -27,7 +25,6 @@ import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
-import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.ml.feature.ElementwiseProduct;
import org.apache.spark.mllib.linalg.Vector;
import org.apache.spark.mllib.linalg.VectorUDT;
@@ -42,7 +39,9 @@ import org.apache.spark.sql.types.StructType;
public class JavaElementwiseProductExample {
public static void main(String[] args) {
SparkSession spark = SparkSession
- .builder().appName("JavaElementwiseProductExample").getOrCreate();
+ .builder()
+ .appName("JavaElementwiseProductExample")
+ .getOrCreate();
// $example on$
// Create some vector data; also works for sparse vectors
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
index a7c89b9d19..baacd796a0 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
// $example on$
import org.apache.spark.ml.Pipeline;
import org.apache.spark.ml.PipelineModel;
@@ -35,11 +33,15 @@ import org.apache.spark.sql.SparkSession;
public class JavaGradientBoostedTreeClassifierExample {
public static void main(String[] args) {
SparkSession spark = SparkSession
- .builder().appName("JavaGradientBoostedTreeClassifierExample").getOrCreate();
+ .builder()
+ .appName("JavaGradientBoostedTreeClassifierExample")
+ .getOrCreate();
// $example on$
// Load and parse the data file, converting it to a DataFrame.
- Dataset<Row> data = spark.read().format("libsvm")
+ Dataset<Row> data = spark
+ .read()
+ .format("libsvm")
.load("data/mllib/sample_libsvm_data.txt");
// Index labels, adding metadata to the label column.
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java
index ccd74f2920..0064beb8c8 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaIndexToStringExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaIndexToStringExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaIndexToStringExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java
index e6d82a0513..65e29ade29 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java
@@ -70,7 +70,10 @@ public class JavaKMeansExample {
int k = Integer.parseInt(args[1]);
// Parses the arguments
- SparkSession spark = SparkSession.builder().appName("JavaKMeansExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaKMeansExample")
+ .getOrCreate();
// $example on$
// Loads data
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java
index b8baca5920..1c52f37867 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java
@@ -65,7 +65,10 @@ public class JavaLDAExample {
String inputFile = "data/mllib/sample_lda_data.txt";
// Parses the arguments
- SparkSession spark = SparkSession.builder().appName("JavaLDAExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaLDAExample")
+ .getOrCreate();
// Loads data
JavaRDD<Row> points = spark.read().text(inputFile).javaRDD().map(new ParseVector());
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java
index 80cdd364b9..9a27b0e9e2 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java
@@ -28,13 +28,19 @@ import org.apache.spark.sql.SparkSession;
public class JavaMaxAbsScalerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaMaxAbsScalerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaMaxAbsScalerExample")
+ .getOrCreate();
// $example on$
- Dataset<Row> dataFrame = spark.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
+ Dataset<Row> dataFrame = spark
+ .read()
+ .format("libsvm")
+ .load("data/mllib/sample_libsvm_data.txt");
MaxAbsScaler scaler = new MaxAbsScaler()
- .setInputCol("features")
- .setOutputCol("scaledFeatures");
+ .setInputCol("features")
+ .setOutputCol("scaledFeatures");
// Compute summary statistics and generate MaxAbsScalerModel
MaxAbsScalerModel scalerModel = scaler.fit(dataFrame);
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java
index 022940fd1e..37fa1c5434 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java
@@ -28,10 +28,16 @@ import org.apache.spark.sql.Row;
public class JavaMinMaxScalerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaMinMaxScalerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaMinMaxScalerExample")
+ .getOrCreate();
// $example on$
- Dataset<Row> dataFrame = spark.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
+ Dataset<Row> dataFrame = spark
+ .read()
+ .format("libsvm")
+ .load("data/mllib/sample_libsvm_data.txt");
MinMaxScaler scaler = new MinMaxScaler()
.setInputCol("features")
.setOutputCol("scaledFeatures");
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java
index 325b7b5874..899815f57c 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java
@@ -35,7 +35,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaNGramExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaNGramExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaNGramExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java
index 1f24a23609..50a46a5774 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java
@@ -32,7 +32,10 @@ import org.apache.spark.sql.SparkSession;
public class JavaNaiveBayesExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaNaiveBayesExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaNaiveBayesExample")
+ .getOrCreate();
// $example on$
// Load training data
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java
index 4b3a718ea9..abc38f85ea 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java
@@ -27,7 +27,10 @@ import org.apache.spark.sql.Row;
public class JavaNormalizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaNormalizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaNormalizerExample")
+ .getOrCreate();
// $example on$
Dataset<Row> dataFrame =
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java
index d6e4d21ead..5d29e54549 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaOneHotEncoderExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaOneHotEncoderExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaOneHotEncoderExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java
index 9cc983bd11..e0cb752224 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java
@@ -58,7 +58,10 @@ public class JavaOneVsRestExample {
public static void main(String[] args) {
// parse the arguments
Params params = parse(args);
- SparkSession spark = SparkSession.builder().appName("JavaOneVsRestExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaOneVsRestExample")
+ .getOrCreate();
// $example on$
// configure the base classifier
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java
index 6b1dcb68ba..ffa979ee01 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaPCAExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaPCAExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaPCAExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java
index 556a457326..9a43189c91 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.SparkSession;
*/
public class JavaPipelineExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaPipelineExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaPipelineExample")
+ .getOrCreate();
// $example on$
// Prepare training documents, which are labeled.
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java
index e328454c70..7afcd0e50c 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaPolynomialExpansionExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaPolynomialExpansionExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaPolynomialExpansionExample")
+ .getOrCreate();
// $example on$
PolynomialExpansion polyExpansion = new PolynomialExpansion()
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java
index 8282ce01d3..428067e0f7 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java
@@ -35,7 +35,10 @@ import static org.apache.spark.sql.types.DataTypes.*;
public class JavaRFormulaExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaRFormulaExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaRFormulaExample")
+ .getOrCreate();
// $example on$
StructType schema = createStructType(new StructField[]{
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java
index 492718bbdb..2a3d62de41 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java
@@ -31,7 +31,10 @@ import org.apache.spark.sql.types.*;
public class JavaSQLTransformerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaSQLTransformerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaSQLTransformerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java
index f906843640..0787079ba4 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java
@@ -40,7 +40,10 @@ import org.apache.spark.sql.SparkSession;
public class JavaSimpleParamsExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaSimpleParamsExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaSimpleParamsExample")
+ .getOrCreate();
// Prepare training data.
// We use LabeledPoint, which is a JavaBean. Spark SQL can convert RDDs of JavaBeans
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java
index 10f82f2233..08ea285a0d 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java
@@ -28,7 +28,10 @@ import org.apache.spark.sql.Row;
public class JavaStandardScalerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaStandardScalerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaStandardScalerExample")
+ .getOrCreate();
// $example on$
Dataset<Row> dataFrame =
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java
index 23ed071c9f..def5994429 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaStopWordsRemoverExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaStopWordsRemoverExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaStopWordsRemoverExample")
+ .getOrCreate();
// $example on$
StopWordsRemover remover = new StopWordsRemover()
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java
index d4c2cf96a7..7533c1835e 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java
@@ -35,7 +35,10 @@ import static org.apache.spark.sql.types.DataTypes.*;
public class JavaStringIndexerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaStringIndexerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaStringIndexerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java
index a816991777..6e0753959e 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java
@@ -38,7 +38,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaTfIdfExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaTfIdfExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaTfIdfExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java
index a65735a5e5..1cc16bb60d 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaTokenizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaTokenizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaTokenizerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java
index 9569bc2412..41f1d8750a 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java
@@ -35,7 +35,10 @@ import static org.apache.spark.sql.types.DataTypes.*;
public class JavaVectorAssemblerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaVectorAssemblerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaVectorAssemblerExample")
+ .getOrCreate();
// $example on$
StructType schema = createStructType(new StructField[]{
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java
index 217d5a06d1..dd9d757dd6 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java
@@ -30,7 +30,10 @@ import org.apache.spark.sql.Row;
public class JavaVectorIndexerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaVectorIndexerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaVectorIndexerExample")
+ .getOrCreate();
// $example on$
Dataset<Row> data = spark.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java
index 4f1ea824a3..24959c0e10 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.*;
public class JavaVectorSlicerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaVectorSlicerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaVectorSlicerExample")
+ .getOrCreate();
// $example on$
Attribute[] attrs = new Attribute[]{
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java
index d9b1a79b52..9be6e6353a 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java
@@ -32,7 +32,10 @@ import org.apache.spark.sql.types.*;
public class JavaWord2VecExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaWord2VecExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaWord2VecExample")
+ .getOrCreate();
// $example on$
// Input data: Each row is a bag of words from a sentence or document.
diff --git a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
index ec2142e756..755b4f5381 100644
--- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
+++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
@@ -51,7 +51,10 @@ public class JavaSparkSQL {
}
public static void main(String[] args) throws Exception {
- SparkSession spark = SparkSession.builder().appName("JavaSparkSQL").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaSparkSQL")
+ .getOrCreate();
System.out.println("=== Data source: RDD ===");
// Load a text file and convert each line to a Java Bean.
@@ -147,7 +150,8 @@ public class JavaSparkSQL {
// a RDD[String] storing one JSON object per string.
List<String> jsonData = Arrays.asList(
"{\"name\":\"Yin\",\"address\":{\"city\":\"Columbus\",\"state\":\"Ohio\"}}");
- JavaRDD<String> anotherPeopleRDD = spark.createDataFrame(jsonData, String.class).toJSON().javaRDD();
+ JavaRDD<String> anotherPeopleRDD = spark
+ .createDataFrame(jsonData, String.class).toJSON().javaRDD();
Dataset<Row> peopleFromJsonRDD = spark.read().json(anotherPeopleRDD);
// Take a look at the schema of this new DataFrame.
diff --git a/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java b/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java
index 44f1e800fe..57953ef74f 100644
--- a/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java
+++ b/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java
@@ -115,7 +115,10 @@ class JavaSparkSessionSingleton {
private static transient SparkSession instance = null;
public static SparkSession getInstance(SparkConf sparkConf) {
if (instance == null) {
- instance = SparkSession.builder().config(sparkConf).getOrCreate();
+ instance = SparkSession
+ .builder()
+ .config(sparkConf)
+ .getOrCreate();
}
return instance;
}