aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java7
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java7
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java10
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java14
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java10
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java5
-rw-r--r--examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java8
-rw-r--r--examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java5
-rw-r--r--examples/src/main/python/ml/als_example.py5
-rw-r--r--examples/src/main/python/ml/binarizer_example.py5
-rw-r--r--examples/src/main/python/ml/bisecting_k_means_example.py5
-rw-r--r--examples/src/main/python/ml/bucketizer_example.py5
-rw-r--r--examples/src/main/python/ml/chisq_selector_example.py5
-rw-r--r--examples/src/main/python/ml/count_vectorizer_example.py5
-rw-r--r--examples/src/main/python/ml/cross_validator.py5
-rw-r--r--examples/src/main/python/ml/dataframe_example.py5
-rw-r--r--examples/src/main/python/ml/dct_example.py5
-rw-r--r--examples/src/main/python/ml/decision_tree_classification_example.py5
-rw-r--r--examples/src/main/python/ml/decision_tree_regression_example.py5
-rw-r--r--examples/src/main/python/ml/elementwise_product_example.py5
-rw-r--r--examples/src/main/python/ml/estimator_transformer_param_example.py5
-rw-r--r--examples/src/main/python/ml/gradient_boosted_tree_classifier_example.py5
-rw-r--r--examples/src/main/python/ml/gradient_boosted_tree_regressor_example.py5
-rw-r--r--examples/src/main/python/ml/index_to_string_example.py5
-rw-r--r--examples/src/main/python/ml/kmeans_example.py5
-rw-r--r--examples/src/main/python/ml/linear_regression_with_elastic_net.py5
-rw-r--r--examples/src/main/python/ml/logistic_regression_with_elastic_net.py5
-rw-r--r--examples/src/main/python/ml/max_abs_scaler_example.py5
-rw-r--r--examples/src/main/python/ml/min_max_scaler_example.py5
-rw-r--r--examples/src/main/python/ml/n_gram_example.py5
-rw-r--r--examples/src/main/python/ml/naive_bayes_example.py5
-rw-r--r--examples/src/main/python/ml/normalizer_example.py5
-rw-r--r--examples/src/main/python/ml/onehot_encoder_example.py5
-rw-r--r--examples/src/main/python/ml/pca_example.py5
-rw-r--r--examples/src/main/python/ml/pipeline_example.py5
-rw-r--r--examples/src/main/python/ml/polynomial_expansion_example.py5
-rw-r--r--examples/src/main/python/ml/random_forest_classifier_example.py5
-rw-r--r--examples/src/main/python/ml/random_forest_regressor_example.py5
-rw-r--r--examples/src/main/python/ml/rformula_example.py5
-rw-r--r--examples/src/main/python/ml/simple_text_classification_pipeline.py5
-rw-r--r--examples/src/main/python/ml/sql_transformer.py5
-rw-r--r--examples/src/main/python/ml/standard_scaler_example.py5
-rw-r--r--examples/src/main/python/ml/stopwords_remover_example.py5
-rw-r--r--examples/src/main/python/ml/string_indexer_example.py5
-rw-r--r--examples/src/main/python/ml/tf_idf_example.py5
-rw-r--r--examples/src/main/python/ml/tokenizer_example.py5
-rw-r--r--examples/src/main/python/ml/train_validation_split.py5
-rw-r--r--examples/src/main/python/ml/vector_assembler_example.py5
-rw-r--r--examples/src/main/python/ml/vector_indexer_example.py5
-rw-r--r--examples/src/main/python/ml/vector_slicer_example.py5
-rw-r--r--examples/src/main/python/ml/word2vec_example.py5
-rw-r--r--examples/src/main/python/mllib/binary_classification_metrics_example.py15
-rw-r--r--examples/src/main/python/sql.py5
-rw-r--r--examples/src/main/python/streaming/sql_network_wordcount.py6
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/ALSExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala4
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/KMeansExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/MaxAbsScalerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/PipelineExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/SimpleParamsExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/SimpleTextClassificationPipeline.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala4
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala5
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala5
-rw-r--r--sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/VectorizedPlainValuesReader.java5
-rw-r--r--sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OffHeapColumnVector.java15
-rw-r--r--sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OnHeapColumnVector.java7
-rw-r--r--sql/hive-thriftserver/src/main/java/org/apache/hive/service/cli/session/SessionManager.java2
142 files changed, 585 insertions, 178 deletions
diff --git a/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java b/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java
index fb1226c09e..22fd592a32 100644
--- a/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java
+++ b/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockHandler.java
@@ -87,8 +87,11 @@ public class ExternalShuffleBlockHandler extends RpcHandler {
blocks.add(blockManager.getBlockData(msg.appId, msg.execId, blockId));
}
long streamId = streamManager.registerStream(client.getClientId(), blocks.iterator());
- logger.trace("Registered streamId {} with {} buffers for client {} from host {}", streamId,
- msg.blockIds.length, client.getClientId(), NettyUtils.getRemoteAddress(client.getChannel()));
+ logger.trace("Registered streamId {} with {} buffers for client {} from host {}",
+ streamId,
+ msg.blockIds.length,
+ client.getClientId(),
+ NettyUtils.getRemoteAddress(client.getChannel()));
callback.onSuccess(new StreamHandle(streamId, msg.blockIds.length).toByteBuffer());
} else if (msgObj instanceof RegisterExecutor) {
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java
index ecb7084e03..2c2aa6df47 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaAFTSurvivalRegressionExample.java
@@ -33,7 +33,10 @@ import org.apache.spark.sql.types.*;
public class JavaAFTSurvivalRegressionExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaAFTSurvivalRegressionExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaAFTSurvivalRegressionExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java
index 9a9a10489b..4b13ba6f9c 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaALSExample.java
@@ -81,7 +81,10 @@ public class JavaALSExample {
// $example off$
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaALSExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaALSExample")
+ .getOrCreate();
// $example on$
JavaRDD<Rating> ratingsRDD = spark
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java
index 88e4298a61..5f964aca92 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaBinarizerExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -26,7 +24,6 @@ import org.apache.spark.sql.SparkSession;
import java.util.Arrays;
import java.util.List;
-import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.ml.feature.Binarizer;
import org.apache.spark.sql.Row;
import org.apache.spark.sql.RowFactory;
@@ -38,7 +35,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaBinarizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaBinarizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaBinarizerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java
index 51aa35084e..810ad905c5 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaBisectingKMeansExample.java
@@ -42,7 +42,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaBisectingKMeansExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaBisectingKMeansExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaBisectingKMeansExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java
index 0c24f52cf5..691df3887a 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaBucketizerExample.java
@@ -35,7 +35,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaBucketizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaBucketizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaBucketizerExample")
+ .getOrCreate();
// $example on$
double[] splits = {Double.NEGATIVE_INFINITY, -0.5, 0.0, 0.5, Double.POSITIVE_INFINITY};
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java
index 684cf9a714..f8f2fb14be 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaChiSqSelectorExample.java
@@ -17,9 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaRDD;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -40,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaChiSqSelectorExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaChiSqSelectorExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaChiSqSelectorExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java
index 0631f9d6d5..0a6b136014 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaCountVectorizerExample.java
@@ -32,7 +32,10 @@ import org.apache.spark.sql.types.*;
public class JavaCountVectorizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaCountVectorizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaCountVectorizerExample")
+ .getOrCreate();
// $example on$
// Input data: Each row is a bag of words from a sentence or document.
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java
index ec57a24451..eee92c77a8 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaDCTExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -26,7 +24,6 @@ import org.apache.spark.sql.SparkSession;
import java.util.Arrays;
import java.util.List;
-import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.ml.feature.DCT;
import org.apache.spark.mllib.linalg.VectorUDT;
import org.apache.spark.mllib.linalg.Vectors;
@@ -39,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaDCTExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaDCTExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaDCTExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
index 90023ac06b..49bad0afc0 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
@@ -49,7 +49,10 @@ import org.apache.spark.sql.SparkSession;
public class JavaDeveloperApiExample {
public static void main(String[] args) throws Exception {
- SparkSession spark = SparkSession.builder().appName("JavaDeveloperApiExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaDeveloperApiExample")
+ .getOrCreate();
// Prepare training data.
List<LabeledPoint> localTraining = Lists.newArrayList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java
index a062a6fcd0..9126242f9e 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaElementwiseProductExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.SparkSession;
@@ -27,7 +25,6 @@ import java.util.ArrayList;
import java.util.Arrays;
import java.util.List;
-import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.ml.feature.ElementwiseProduct;
import org.apache.spark.mllib.linalg.Vector;
import org.apache.spark.mllib.linalg.VectorUDT;
@@ -42,7 +39,9 @@ import org.apache.spark.sql.types.StructType;
public class JavaElementwiseProductExample {
public static void main(String[] args) {
SparkSession spark = SparkSession
- .builder().appName("JavaElementwiseProductExample").getOrCreate();
+ .builder()
+ .appName("JavaElementwiseProductExample")
+ .getOrCreate();
// $example on$
// Create some vector data; also works for sparse vectors
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
index a7c89b9d19..baacd796a0 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
@@ -17,8 +17,6 @@
package org.apache.spark.examples.ml;
-import org.apache.spark.SparkConf;
-import org.apache.spark.api.java.JavaSparkContext;
// $example on$
import org.apache.spark.ml.Pipeline;
import org.apache.spark.ml.PipelineModel;
@@ -35,11 +33,15 @@ import org.apache.spark.sql.SparkSession;
public class JavaGradientBoostedTreeClassifierExample {
public static void main(String[] args) {
SparkSession spark = SparkSession
- .builder().appName("JavaGradientBoostedTreeClassifierExample").getOrCreate();
+ .builder()
+ .appName("JavaGradientBoostedTreeClassifierExample")
+ .getOrCreate();
// $example on$
// Load and parse the data file, converting it to a DataFrame.
- Dataset<Row> data = spark.read().format("libsvm")
+ Dataset<Row> data = spark
+ .read()
+ .format("libsvm")
.load("data/mllib/sample_libsvm_data.txt");
// Index labels, adding metadata to the label column.
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java
index ccd74f2920..0064beb8c8 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaIndexToStringExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaIndexToStringExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaIndexToStringExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaIndexToStringExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java
index e6d82a0513..65e29ade29 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaKMeansExample.java
@@ -70,7 +70,10 @@ public class JavaKMeansExample {
int k = Integer.parseInt(args[1]);
// Parses the arguments
- SparkSession spark = SparkSession.builder().appName("JavaKMeansExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaKMeansExample")
+ .getOrCreate();
// $example on$
// Loads data
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java
index b8baca5920..1c52f37867 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaLDAExample.java
@@ -65,7 +65,10 @@ public class JavaLDAExample {
String inputFile = "data/mllib/sample_lda_data.txt";
// Parses the arguments
- SparkSession spark = SparkSession.builder().appName("JavaLDAExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaLDAExample")
+ .getOrCreate();
// Loads data
JavaRDD<Row> points = spark.read().text(inputFile).javaRDD().map(new ParseVector());
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java
index 80cdd364b9..9a27b0e9e2 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaMaxAbsScalerExample.java
@@ -28,13 +28,19 @@ import org.apache.spark.sql.SparkSession;
public class JavaMaxAbsScalerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaMaxAbsScalerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaMaxAbsScalerExample")
+ .getOrCreate();
// $example on$
- Dataset<Row> dataFrame = spark.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
+ Dataset<Row> dataFrame = spark
+ .read()
+ .format("libsvm")
+ .load("data/mllib/sample_libsvm_data.txt");
MaxAbsScaler scaler = new MaxAbsScaler()
- .setInputCol("features")
- .setOutputCol("scaledFeatures");
+ .setInputCol("features")
+ .setOutputCol("scaledFeatures");
// Compute summary statistics and generate MaxAbsScalerModel
MaxAbsScalerModel scalerModel = scaler.fit(dataFrame);
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java
index 022940fd1e..37fa1c5434 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaMinMaxScalerExample.java
@@ -28,10 +28,16 @@ import org.apache.spark.sql.Row;
public class JavaMinMaxScalerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaMinMaxScalerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaMinMaxScalerExample")
+ .getOrCreate();
// $example on$
- Dataset<Row> dataFrame = spark.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
+ Dataset<Row> dataFrame = spark
+ .read()
+ .format("libsvm")
+ .load("data/mllib/sample_libsvm_data.txt");
MinMaxScaler scaler = new MinMaxScaler()
.setInputCol("features")
.setOutputCol("scaledFeatures");
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java
index 325b7b5874..899815f57c 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaNGramExample.java
@@ -35,7 +35,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaNGramExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaNGramExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaNGramExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java
index 1f24a23609..50a46a5774 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaNaiveBayesExample.java
@@ -32,7 +32,10 @@ import org.apache.spark.sql.SparkSession;
public class JavaNaiveBayesExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaNaiveBayesExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaNaiveBayesExample")
+ .getOrCreate();
// $example on$
// Load training data
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java
index 4b3a718ea9..abc38f85ea 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaNormalizerExample.java
@@ -27,7 +27,10 @@ import org.apache.spark.sql.Row;
public class JavaNormalizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaNormalizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaNormalizerExample")
+ .getOrCreate();
// $example on$
Dataset<Row> dataFrame =
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java
index d6e4d21ead..5d29e54549 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneHotEncoderExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaOneHotEncoderExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaOneHotEncoderExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaOneHotEncoderExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java
index 9cc983bd11..e0cb752224 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaOneVsRestExample.java
@@ -58,7 +58,10 @@ public class JavaOneVsRestExample {
public static void main(String[] args) {
// parse the arguments
Params params = parse(args);
- SparkSession spark = SparkSession.builder().appName("JavaOneVsRestExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaOneVsRestExample")
+ .getOrCreate();
// $example on$
// configure the base classifier
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java
index 6b1dcb68ba..ffa979ee01 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaPCAExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaPCAExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaPCAExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaPCAExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java
index 556a457326..9a43189c91 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaPipelineExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.SparkSession;
*/
public class JavaPipelineExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaPipelineExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaPipelineExample")
+ .getOrCreate();
// $example on$
// Prepare training documents, which are labeled.
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java
index e328454c70..7afcd0e50c 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaPolynomialExpansionExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaPolynomialExpansionExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaPolynomialExpansionExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaPolynomialExpansionExample")
+ .getOrCreate();
// $example on$
PolynomialExpansion polyExpansion = new PolynomialExpansion()
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java
index 8282ce01d3..428067e0f7 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaRFormulaExample.java
@@ -35,7 +35,10 @@ import static org.apache.spark.sql.types.DataTypes.*;
public class JavaRFormulaExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaRFormulaExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaRFormulaExample")
+ .getOrCreate();
// $example on$
StructType schema = createStructType(new StructField[]{
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java
index 492718bbdb..2a3d62de41 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaSQLTransformerExample.java
@@ -31,7 +31,10 @@ import org.apache.spark.sql.types.*;
public class JavaSQLTransformerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaSQLTransformerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaSQLTransformerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java
index f906843640..0787079ba4 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaSimpleParamsExample.java
@@ -40,7 +40,10 @@ import org.apache.spark.sql.SparkSession;
public class JavaSimpleParamsExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaSimpleParamsExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaSimpleParamsExample")
+ .getOrCreate();
// Prepare training data.
// We use LabeledPoint, which is a JavaBean. Spark SQL can convert RDDs of JavaBeans
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java
index 10f82f2233..08ea285a0d 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaStandardScalerExample.java
@@ -28,7 +28,10 @@ import org.apache.spark.sql.Row;
public class JavaStandardScalerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaStandardScalerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaStandardScalerExample")
+ .getOrCreate();
// $example on$
Dataset<Row> dataFrame =
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java
index 23ed071c9f..def5994429 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaStopWordsRemoverExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaStopWordsRemoverExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaStopWordsRemoverExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaStopWordsRemoverExample")
+ .getOrCreate();
// $example on$
StopWordsRemover remover = new StopWordsRemover()
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java
index d4c2cf96a7..7533c1835e 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaStringIndexerExample.java
@@ -35,7 +35,10 @@ import static org.apache.spark.sql.types.DataTypes.*;
public class JavaStringIndexerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaStringIndexerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaStringIndexerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java
index a816991777..6e0753959e 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaTfIdfExample.java
@@ -38,7 +38,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaTfIdfExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaTfIdfExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaTfIdfExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java
index a65735a5e5..1cc16bb60d 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaTokenizerExample.java
@@ -36,7 +36,10 @@ import org.apache.spark.sql.types.StructType;
public class JavaTokenizerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaTokenizerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaTokenizerExample")
+ .getOrCreate();
// $example on$
List<Row> data = Arrays.asList(
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java
index 9569bc2412..41f1d8750a 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorAssemblerExample.java
@@ -35,7 +35,10 @@ import static org.apache.spark.sql.types.DataTypes.*;
public class JavaVectorAssemblerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaVectorAssemblerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaVectorAssemblerExample")
+ .getOrCreate();
// $example on$
StructType schema = createStructType(new StructField[]{
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java
index 217d5a06d1..dd9d757dd6 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorIndexerExample.java
@@ -30,7 +30,10 @@ import org.apache.spark.sql.Row;
public class JavaVectorIndexerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaVectorIndexerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaVectorIndexerExample")
+ .getOrCreate();
// $example on$
Dataset<Row> data = spark.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java
index 4f1ea824a3..24959c0e10 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaVectorSlicerExample.java
@@ -37,7 +37,10 @@ import org.apache.spark.sql.types.*;
public class JavaVectorSlicerExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaVectorSlicerExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaVectorSlicerExample")
+ .getOrCreate();
// $example on$
Attribute[] attrs = new Attribute[]{
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java
index d9b1a79b52..9be6e6353a 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaWord2VecExample.java
@@ -32,7 +32,10 @@ import org.apache.spark.sql.types.*;
public class JavaWord2VecExample {
public static void main(String[] args) {
- SparkSession spark = SparkSession.builder().appName("JavaWord2VecExample").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaWord2VecExample")
+ .getOrCreate();
// $example on$
// Input data: Each row is a bag of words from a sentence or document.
diff --git a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
index ec2142e756..755b4f5381 100644
--- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
+++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
@@ -51,7 +51,10 @@ public class JavaSparkSQL {
}
public static void main(String[] args) throws Exception {
- SparkSession spark = SparkSession.builder().appName("JavaSparkSQL").getOrCreate();
+ SparkSession spark = SparkSession
+ .builder()
+ .appName("JavaSparkSQL")
+ .getOrCreate();
System.out.println("=== Data source: RDD ===");
// Load a text file and convert each line to a Java Bean.
@@ -147,7 +150,8 @@ public class JavaSparkSQL {
// a RDD[String] storing one JSON object per string.
List<String> jsonData = Arrays.asList(
"{\"name\":\"Yin\",\"address\":{\"city\":\"Columbus\",\"state\":\"Ohio\"}}");
- JavaRDD<String> anotherPeopleRDD = spark.createDataFrame(jsonData, String.class).toJSON().javaRDD();
+ JavaRDD<String> anotherPeopleRDD = spark
+ .createDataFrame(jsonData, String.class).toJSON().javaRDD();
Dataset<Row> peopleFromJsonRDD = spark.read().json(anotherPeopleRDD);
// Take a look at the schema of this new DataFrame.
diff --git a/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java b/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java
index 44f1e800fe..57953ef74f 100644
--- a/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java
+++ b/examples/src/main/java/org/apache/spark/examples/streaming/JavaSqlNetworkWordCount.java
@@ -115,7 +115,10 @@ class JavaSparkSessionSingleton {
private static transient SparkSession instance = null;
public static SparkSession getInstance(SparkConf sparkConf) {
if (instance == null) {
- instance = SparkSession.builder().config(sparkConf).getOrCreate();
+ instance = SparkSession
+ .builder()
+ .config(sparkConf)
+ .getOrCreate();
}
return instance;
}
diff --git a/examples/src/main/python/ml/als_example.py b/examples/src/main/python/ml/als_example.py
index e36444f185..ff0829b0dd 100644
--- a/examples/src/main/python/ml/als_example.py
+++ b/examples/src/main/python/ml/als_example.py
@@ -30,7 +30,10 @@ from pyspark.sql import Row
# $example off$
if __name__ == "__main__":
- spark = SparkSession.builder.appName("ALSExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("ALSExample")\
+ .getOrCreate()
# $example on$
lines = spark.read.text("data/mllib/als/sample_movielens_ratings.txt").rdd
diff --git a/examples/src/main/python/ml/binarizer_example.py b/examples/src/main/python/ml/binarizer_example.py
index 072187e645..4224a27dbe 100644
--- a/examples/src/main/python/ml/binarizer_example.py
+++ b/examples/src/main/python/ml/binarizer_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import Binarizer
# $example off$
if __name__ == "__main__":
- spark = SparkSession.builder.appName("BinarizerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("BinarizerExample")\
+ .getOrCreate()
# $example on$
continuousDataFrame = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/bisecting_k_means_example.py b/examples/src/main/python/ml/bisecting_k_means_example.py
index 836a89cde0..540a4bc3e4 100644
--- a/examples/src/main/python/ml/bisecting_k_means_example.py
+++ b/examples/src/main/python/ml/bisecting_k_means_example.py
@@ -30,7 +30,10 @@ A simple example demonstrating a bisecting k-means clustering.
"""
if __name__ == "__main__":
- spark = SparkSession.builder.appName("PythonBisectingKMeansExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("PythonBisectingKMeansExample")\
+ .getOrCreate()
# $example on$
data = spark.read.text("data/mllib/kmeans_data.txt").rdd
diff --git a/examples/src/main/python/ml/bucketizer_example.py b/examples/src/main/python/ml/bucketizer_example.py
index 288ec62bdf..8177e560dd 100644
--- a/examples/src/main/python/ml/bucketizer_example.py
+++ b/examples/src/main/python/ml/bucketizer_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import Bucketizer
# $example off$
if __name__ == "__main__":
- spark = SparkSession.builder.appName("BucketizerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("BucketizerExample")\
+ .getOrCreate()
# $example on$
splits = [-float("inf"), -0.5, 0.0, 0.5, float("inf")]
diff --git a/examples/src/main/python/ml/chisq_selector_example.py b/examples/src/main/python/ml/chisq_selector_example.py
index 8f58fc28de..8bafb942e0 100644
--- a/examples/src/main/python/ml/chisq_selector_example.py
+++ b/examples/src/main/python/ml/chisq_selector_example.py
@@ -24,7 +24,10 @@ from pyspark.mllib.linalg import Vectors
# $example off$
if __name__ == "__main__":
- spark = SparkSession.builder.appName("ChiSqSelectorExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("ChiSqSelectorExample")\
+ .getOrCreate()
# $example on$
df = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/count_vectorizer_example.py b/examples/src/main/python/ml/count_vectorizer_example.py
index 9dbf9959d1..38cfac82fb 100644
--- a/examples/src/main/python/ml/count_vectorizer_example.py
+++ b/examples/src/main/python/ml/count_vectorizer_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import CountVectorizer
# $example off$
if __name__ == "__main__":
- spark = SparkSession.builder.appName("CountVectorizerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("CountVectorizerExample")\
+ .getOrCreate()
# $example on$
# Input data: Each row is a bag of words with a ID.
diff --git a/examples/src/main/python/ml/cross_validator.py b/examples/src/main/python/ml/cross_validator.py
index a61d0f63d2..a41df6cf94 100644
--- a/examples/src/main/python/ml/cross_validator.py
+++ b/examples/src/main/python/ml/cross_validator.py
@@ -35,7 +35,10 @@ Run with:
"""
if __name__ == "__main__":
- spark = SparkSession.builder.appName("CrossValidatorExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("CrossValidatorExample")\
+ .getOrCreate()
# $example on$
# Prepare training documents, which are labeled.
training = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/dataframe_example.py b/examples/src/main/python/ml/dataframe_example.py
index b3e671038e..a7d8b9056d 100644
--- a/examples/src/main/python/ml/dataframe_example.py
+++ b/examples/src/main/python/ml/dataframe_example.py
@@ -33,7 +33,10 @@ if __name__ == "__main__":
if len(sys.argv) > 2:
print("Usage: dataframe_example.py <libsvm file>", file=sys.stderr)
exit(-1)
- spark = SparkSession.builder.appName("DataFrameExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("DataFrameExample")\
+ .getOrCreate()
if len(sys.argv) == 2:
input = sys.argv[1]
else:
diff --git a/examples/src/main/python/ml/dct_example.py b/examples/src/main/python/ml/dct_example.py
index 1bf8fc6d14..e36fcdeaee 100644
--- a/examples/src/main/python/ml/dct_example.py
+++ b/examples/src/main/python/ml/dct_example.py
@@ -24,7 +24,10 @@ from pyspark.mllib.linalg import Vectors
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("DCTExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("DCTExample")\
+ .getOrCreate()
# $example on$
df = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/decision_tree_classification_example.py b/examples/src/main/python/ml/decision_tree_classification_example.py
index d2318e2436..9b40b701ec 100644
--- a/examples/src/main/python/ml/decision_tree_classification_example.py
+++ b/examples/src/main/python/ml/decision_tree_classification_example.py
@@ -29,7 +29,10 @@ from pyspark.ml.evaluation import MulticlassClassificationEvaluator
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("decision_tree_classification_example").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("decision_tree_classification_example")\
+ .getOrCreate()
# $example on$
# Load the data stored in LIBSVM format as a DataFrame.
diff --git a/examples/src/main/python/ml/decision_tree_regression_example.py b/examples/src/main/python/ml/decision_tree_regression_example.py
index 9e8cb382a9..b734d4974a 100644
--- a/examples/src/main/python/ml/decision_tree_regression_example.py
+++ b/examples/src/main/python/ml/decision_tree_regression_example.py
@@ -29,7 +29,10 @@ from pyspark.ml.evaluation import RegressionEvaluator
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("decision_tree_classification_example").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("decision_tree_classification_example")\
+ .getOrCreate()
# $example on$
# Load the data stored in LIBSVM format as a DataFrame.
diff --git a/examples/src/main/python/ml/elementwise_product_example.py b/examples/src/main/python/ml/elementwise_product_example.py
index 6fa641b772..41727edcdb 100644
--- a/examples/src/main/python/ml/elementwise_product_example.py
+++ b/examples/src/main/python/ml/elementwise_product_example.py
@@ -24,7 +24,10 @@ from pyspark.mllib.linalg import Vectors
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("ElementwiseProductExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("ElementwiseProductExample")\
+ .getOrCreate()
# $example on$
data = [(Vectors.dense([1.0, 2.0, 3.0]),), (Vectors.dense([4.0, 5.0, 6.0]),)]
diff --git a/examples/src/main/python/ml/estimator_transformer_param_example.py b/examples/src/main/python/ml/estimator_transformer_param_example.py
index 4993b5a984..0fcae0e3fc 100644
--- a/examples/src/main/python/ml/estimator_transformer_param_example.py
+++ b/examples/src/main/python/ml/estimator_transformer_param_example.py
@@ -26,7 +26,10 @@ from pyspark.ml.classification import LogisticRegression
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("EstimatorTransformerParamExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("EstimatorTransformerParamExample")\
+ .getOrCreate()
# $example on$
# Prepare training data from a list of (label, features) tuples.
diff --git a/examples/src/main/python/ml/gradient_boosted_tree_classifier_example.py b/examples/src/main/python/ml/gradient_boosted_tree_classifier_example.py
index b09ad41da3..50026d7b7e 100644
--- a/examples/src/main/python/ml/gradient_boosted_tree_classifier_example.py
+++ b/examples/src/main/python/ml/gradient_boosted_tree_classifier_example.py
@@ -29,7 +29,10 @@ from pyspark.ml.evaluation import MulticlassClassificationEvaluator
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("gradient_boosted_tree_classifier_example").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("gradient_boosted_tree_classifier_example")\
+ .getOrCreate()
# $example on$
# Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/python/ml/gradient_boosted_tree_regressor_example.py b/examples/src/main/python/ml/gradient_boosted_tree_regressor_example.py
index caa7cfc4e1..5dd2272748 100644
--- a/examples/src/main/python/ml/gradient_boosted_tree_regressor_example.py
+++ b/examples/src/main/python/ml/gradient_boosted_tree_regressor_example.py
@@ -29,7 +29,10 @@ from pyspark.ml.evaluation import RegressionEvaluator
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("gradient_boosted_tree_regressor_example").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("gradient_boosted_tree_regressor_example")\
+ .getOrCreate()
# $example on$
# Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/python/ml/index_to_string_example.py b/examples/src/main/python/ml/index_to_string_example.py
index dd04b2c4b0..523caac00c 100644
--- a/examples/src/main/python/ml/index_to_string_example.py
+++ b/examples/src/main/python/ml/index_to_string_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import IndexToString, StringIndexer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("IndexToStringExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("IndexToStringExample")\
+ .getOrCreate()
# $example on$
df = spark.createDataFrame(
diff --git a/examples/src/main/python/ml/kmeans_example.py b/examples/src/main/python/ml/kmeans_example.py
index 7d9d80e645..7382396955 100644
--- a/examples/src/main/python/ml/kmeans_example.py
+++ b/examples/src/main/python/ml/kmeans_example.py
@@ -49,7 +49,10 @@ if __name__ == "__main__":
path = sys.argv[1]
k = sys.argv[2]
- spark = SparkSession.builder.appName("PythonKMeansExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("PythonKMeansExample")\
+ .getOrCreate()
lines = spark.read.text(path).rdd
data = lines.map(parseVector)
diff --git a/examples/src/main/python/ml/linear_regression_with_elastic_net.py b/examples/src/main/python/ml/linear_regression_with_elastic_net.py
index 99b7f7fe99..620ab5b87e 100644
--- a/examples/src/main/python/ml/linear_regression_with_elastic_net.py
+++ b/examples/src/main/python/ml/linear_regression_with_elastic_net.py
@@ -23,7 +23,10 @@ from pyspark.ml.regression import LinearRegression
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("LinearRegressionWithElasticNet").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("LinearRegressionWithElasticNet")\
+ .getOrCreate()
# $example on$
# Load training data
diff --git a/examples/src/main/python/ml/logistic_regression_with_elastic_net.py b/examples/src/main/python/ml/logistic_regression_with_elastic_net.py
index 0d7112e723..33d0689f75 100644
--- a/examples/src/main/python/ml/logistic_regression_with_elastic_net.py
+++ b/examples/src/main/python/ml/logistic_regression_with_elastic_net.py
@@ -23,7 +23,10 @@ from pyspark.ml.classification import LogisticRegression
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("LogisticRegressionWithElasticNet").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("LogisticRegressionWithElasticNet")\
+ .getOrCreate()
# $example on$
# Load training data
diff --git a/examples/src/main/python/ml/max_abs_scaler_example.py b/examples/src/main/python/ml/max_abs_scaler_example.py
index 1cb95a98f0..ab91198b08 100644
--- a/examples/src/main/python/ml/max_abs_scaler_example.py
+++ b/examples/src/main/python/ml/max_abs_scaler_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import MaxAbsScaler
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("MaxAbsScalerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("MaxAbsScalerExample")\
+ .getOrCreate()
# $example on$
dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/python/ml/min_max_scaler_example.py b/examples/src/main/python/ml/min_max_scaler_example.py
index 8d91a59e2b..e3e7bc205b 100644
--- a/examples/src/main/python/ml/min_max_scaler_example.py
+++ b/examples/src/main/python/ml/min_max_scaler_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import MinMaxScaler
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("MinMaxScalerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("MinMaxScalerExample")\
+ .getOrCreate()
# $example on$
dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/python/ml/n_gram_example.py b/examples/src/main/python/ml/n_gram_example.py
index b7fecf0d68..9ac07f2c8e 100644
--- a/examples/src/main/python/ml/n_gram_example.py
+++ b/examples/src/main/python/ml/n_gram_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import NGram
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("NGramExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("NGramExample")\
+ .getOrCreate()
# $example on$
wordDataFrame = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/naive_bayes_example.py b/examples/src/main/python/ml/naive_bayes_example.py
index e37035542c..89255a2bae 100644
--- a/examples/src/main/python/ml/naive_bayes_example.py
+++ b/examples/src/main/python/ml/naive_bayes_example.py
@@ -24,7 +24,10 @@ from pyspark.ml.evaluation import MulticlassClassificationEvaluator
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("naive_bayes_example").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("naive_bayes_example")\
+ .getOrCreate()
# $example on$
# Load training data
diff --git a/examples/src/main/python/ml/normalizer_example.py b/examples/src/main/python/ml/normalizer_example.py
index ae25537619..19012f51f4 100644
--- a/examples/src/main/python/ml/normalizer_example.py
+++ b/examples/src/main/python/ml/normalizer_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import Normalizer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("NormalizerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("NormalizerExample")\
+ .getOrCreate()
# $example on$
dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/python/ml/onehot_encoder_example.py b/examples/src/main/python/ml/onehot_encoder_example.py
index 9acc363dc9..b9fceef68e 100644
--- a/examples/src/main/python/ml/onehot_encoder_example.py
+++ b/examples/src/main/python/ml/onehot_encoder_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import OneHotEncoder, StringIndexer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("OneHotEncoderExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("OneHotEncoderExample")\
+ .getOrCreate()
# $example on$
df = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/pca_example.py b/examples/src/main/python/ml/pca_example.py
index adab151734..f1b3cdec7b 100644
--- a/examples/src/main/python/ml/pca_example.py
+++ b/examples/src/main/python/ml/pca_example.py
@@ -24,7 +24,10 @@ from pyspark.mllib.linalg import Vectors
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("PCAExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("PCAExample")\
+ .getOrCreate()
# $example on$
data = [(Vectors.sparse(5, [(1, 1.0), (3, 7.0)]),),
diff --git a/examples/src/main/python/ml/pipeline_example.py b/examples/src/main/python/ml/pipeline_example.py
index ed9765d961..bd10cfd7a2 100644
--- a/examples/src/main/python/ml/pipeline_example.py
+++ b/examples/src/main/python/ml/pipeline_example.py
@@ -27,7 +27,10 @@ from pyspark.ml.feature import HashingTF, Tokenizer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("PipelineExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("PipelineExample")\
+ .getOrCreate()
# $example on$
# Prepare training documents from a list of (id, text, label) tuples.
diff --git a/examples/src/main/python/ml/polynomial_expansion_example.py b/examples/src/main/python/ml/polynomial_expansion_example.py
index 328b559320..08882bcb25 100644
--- a/examples/src/main/python/ml/polynomial_expansion_example.py
+++ b/examples/src/main/python/ml/polynomial_expansion_example.py
@@ -24,7 +24,10 @@ from pyspark.mllib.linalg import Vectors
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("PolynomialExpansionExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("PolynomialExpansionExample")\
+ .getOrCreate()
# $example on$
df = spark\
diff --git a/examples/src/main/python/ml/random_forest_classifier_example.py b/examples/src/main/python/ml/random_forest_classifier_example.py
index b0a93e050c..c618eaf60c 100644
--- a/examples/src/main/python/ml/random_forest_classifier_example.py
+++ b/examples/src/main/python/ml/random_forest_classifier_example.py
@@ -29,7 +29,10 @@ from pyspark.ml.evaluation import MulticlassClassificationEvaluator
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("random_forest_classifier_example").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("random_forest_classifier_example")\
+ .getOrCreate()
# $example on$
# Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/python/ml/random_forest_regressor_example.py b/examples/src/main/python/ml/random_forest_regressor_example.py
index 4bb84f0de8..3a793737db 100644
--- a/examples/src/main/python/ml/random_forest_regressor_example.py
+++ b/examples/src/main/python/ml/random_forest_regressor_example.py
@@ -29,7 +29,10 @@ from pyspark.ml.evaluation import RegressionEvaluator
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("random_forest_regressor_example").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("random_forest_regressor_example")\
+ .getOrCreate()
# $example on$
# Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/python/ml/rformula_example.py b/examples/src/main/python/ml/rformula_example.py
index 45cc116ac2..d5df3ce4f5 100644
--- a/examples/src/main/python/ml/rformula_example.py
+++ b/examples/src/main/python/ml/rformula_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import RFormula
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("RFormulaExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("RFormulaExample")\
+ .getOrCreate()
# $example on$
dataset = spark.createDataFrame(
diff --git a/examples/src/main/python/ml/simple_text_classification_pipeline.py b/examples/src/main/python/ml/simple_text_classification_pipeline.py
index 3600c12211..886f43c0b0 100644
--- a/examples/src/main/python/ml/simple_text_classification_pipeline.py
+++ b/examples/src/main/python/ml/simple_text_classification_pipeline.py
@@ -33,7 +33,10 @@ pipeline in Python. Run with:
if __name__ == "__main__":
- spark = SparkSession.builder.appName("SimpleTextClassificationPipeline").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("SimpleTextClassificationPipeline")\
+ .getOrCreate()
# Prepare training documents, which are labeled.
training = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/sql_transformer.py b/examples/src/main/python/ml/sql_transformer.py
index 26045db4be..0bf8f35720 100644
--- a/examples/src/main/python/ml/sql_transformer.py
+++ b/examples/src/main/python/ml/sql_transformer.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import SQLTransformer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("SQLTransformerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("SQLTransformerExample")\
+ .getOrCreate()
# $example on$
df = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/standard_scaler_example.py b/examples/src/main/python/ml/standard_scaler_example.py
index c50804f6bf..c0027480e6 100644
--- a/examples/src/main/python/ml/standard_scaler_example.py
+++ b/examples/src/main/python/ml/standard_scaler_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import StandardScaler
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("StandardScalerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("StandardScalerExample")\
+ .getOrCreate()
# $example on$
dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/python/ml/stopwords_remover_example.py b/examples/src/main/python/ml/stopwords_remover_example.py
index 57362673df..395fdeffc5 100644
--- a/examples/src/main/python/ml/stopwords_remover_example.py
+++ b/examples/src/main/python/ml/stopwords_remover_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import StopWordsRemover
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("StopWordsRemoverExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("StopWordsRemoverExample")\
+ .getOrCreate()
# $example on$
sentenceData = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/string_indexer_example.py b/examples/src/main/python/ml/string_indexer_example.py
index aacd4f999b..a328e040f5 100644
--- a/examples/src/main/python/ml/string_indexer_example.py
+++ b/examples/src/main/python/ml/string_indexer_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import StringIndexer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("StringIndexerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("StringIndexerExample")\
+ .getOrCreate()
# $example on$
df = spark.createDataFrame(
diff --git a/examples/src/main/python/ml/tf_idf_example.py b/examples/src/main/python/ml/tf_idf_example.py
index 25df8166ef..fb4ad992fb 100644
--- a/examples/src/main/python/ml/tf_idf_example.py
+++ b/examples/src/main/python/ml/tf_idf_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import HashingTF, IDF, Tokenizer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("TfIdfExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("TfIdfExample")\
+ .getOrCreate()
# $example on$
sentenceData = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/tokenizer_example.py b/examples/src/main/python/ml/tokenizer_example.py
index 5be4b4cfe3..e61ec920d2 100644
--- a/examples/src/main/python/ml/tokenizer_example.py
+++ b/examples/src/main/python/ml/tokenizer_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import Tokenizer, RegexTokenizer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("TokenizerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("TokenizerExample")\
+ .getOrCreate()
# $example on$
sentenceDataFrame = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/train_validation_split.py b/examples/src/main/python/ml/train_validation_split.py
index 2e43a0f8ae..5f5c52aca8 100644
--- a/examples/src/main/python/ml/train_validation_split.py
+++ b/examples/src/main/python/ml/train_validation_split.py
@@ -31,7 +31,10 @@ Run with:
"""
if __name__ == "__main__":
- spark = SparkSession.builder.appName("TrainValidationSplit").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("TrainValidationSplit")\
+ .getOrCreate()
# $example on$
# Prepare training and test data.
data = spark.read.format("libsvm")\
diff --git a/examples/src/main/python/ml/vector_assembler_example.py b/examples/src/main/python/ml/vector_assembler_example.py
index 019a9ea6f7..b955ff00a8 100644
--- a/examples/src/main/python/ml/vector_assembler_example.py
+++ b/examples/src/main/python/ml/vector_assembler_example.py
@@ -24,7 +24,10 @@ from pyspark.ml.feature import VectorAssembler
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("VectorAssemblerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("VectorAssemblerExample")\
+ .getOrCreate()
# $example on$
dataset = spark.createDataFrame(
diff --git a/examples/src/main/python/ml/vector_indexer_example.py b/examples/src/main/python/ml/vector_indexer_example.py
index 3cf5b8ebf1..9b00e0f841 100644
--- a/examples/src/main/python/ml/vector_indexer_example.py
+++ b/examples/src/main/python/ml/vector_indexer_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import VectorIndexer
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("VectorIndexerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("VectorIndexerExample")\
+ .getOrCreate()
# $example on$
data = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/python/ml/vector_slicer_example.py b/examples/src/main/python/ml/vector_slicer_example.py
index 0531bcdb06..b833a894eb 100644
--- a/examples/src/main/python/ml/vector_slicer_example.py
+++ b/examples/src/main/python/ml/vector_slicer_example.py
@@ -25,7 +25,10 @@ from pyspark.sql.types import Row
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("VectorSlicerExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("VectorSlicerExample")\
+ .getOrCreate()
# $example on$
df = spark.createDataFrame([
diff --git a/examples/src/main/python/ml/word2vec_example.py b/examples/src/main/python/ml/word2vec_example.py
index 6766a7b6aa..66500bee15 100644
--- a/examples/src/main/python/ml/word2vec_example.py
+++ b/examples/src/main/python/ml/word2vec_example.py
@@ -23,7 +23,10 @@ from pyspark.ml.feature import Word2Vec
from pyspark.sql import SparkSession
if __name__ == "__main__":
- spark = SparkSession.builder.appName("Word2VecExample").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("Word2VecExample")\
+ .getOrCreate()
# $example on$
# Input data: Each row is a bag of words from a sentence or document.
diff --git a/examples/src/main/python/mllib/binary_classification_metrics_example.py b/examples/src/main/python/mllib/binary_classification_metrics_example.py
index 8f0fc9d45d..daf000e38d 100644
--- a/examples/src/main/python/mllib/binary_classification_metrics_example.py
+++ b/examples/src/main/python/mllib/binary_classification_metrics_example.py
@@ -18,20 +18,25 @@
Binary Classification Metrics Example.
"""
from __future__ import print_function
-from pyspark import SparkContext
+from pyspark.sql import SparkSession
# $example on$
from pyspark.mllib.classification import LogisticRegressionWithLBFGS
from pyspark.mllib.evaluation import BinaryClassificationMetrics
-from pyspark.mllib.util import MLUtils
+from pyspark.mllib.regression import LabeledPoint
# $example off$
if __name__ == "__main__":
- sc = SparkContext(appName="BinaryClassificationMetricsExample")
+ spark = SparkSession\
+ .builder\
+ .appName("BinaryClassificationMetricsExample")\
+ .getOrCreate()
# $example on$
# Several of the methods available in scala are currently missing from pyspark
# Load training data in LIBSVM format
- data = MLUtils.loadLibSVMFile(sc, "data/mllib/sample_binary_classification_data.txt")
+ data = spark\
+ .read.format("libsvm").load("data/mllib/sample_binary_classification_data.txt")\
+ .rdd.map(lambda row: LabeledPoint(row[0], row[1]))
# Split data into training (60%) and test (40%)
training, test = data.randomSplit([0.6, 0.4], seed=11L)
@@ -53,4 +58,4 @@ if __name__ == "__main__":
print("Area under ROC = %s" % metrics.areaUnderROC)
# $example off$
- sc.stop()
+ spark.stop()
diff --git a/examples/src/main/python/sql.py b/examples/src/main/python/sql.py
index 59a46cb283..5594223465 100644
--- a/examples/src/main/python/sql.py
+++ b/examples/src/main/python/sql.py
@@ -25,7 +25,10 @@ from pyspark.sql.types import Row, StructField, StructType, StringType, IntegerT
if __name__ == "__main__":
- spark = SparkSession.builder.appName("PythonSQL").getOrCreate()
+ spark = SparkSession\
+ .builder\
+ .appName("PythonSQL")\
+ .getOrCreate()
# A list of Rows. Infer schema from the first row, create a DataFrame and print the schema
rows = [Row(name="John", age=19), Row(name="Smith", age=23), Row(name="Sarah", age=18)]
diff --git a/examples/src/main/python/streaming/sql_network_wordcount.py b/examples/src/main/python/streaming/sql_network_wordcount.py
index 588cbfee14..f8801d4ea6 100644
--- a/examples/src/main/python/streaming/sql_network_wordcount.py
+++ b/examples/src/main/python/streaming/sql_network_wordcount.py
@@ -38,8 +38,10 @@ from pyspark.sql import Row, SparkSession
def getSparkSessionInstance(sparkConf):
if ('sparkSessionSingletonInstance' not in globals()):
- globals()['sparkSessionSingletonInstance'] =\
- SparkSession.builder.config(conf=sparkConf).getOrCreate()
+ globals()['sparkSessionSingletonInstance'] = SparkSession\
+ .builder\
+ .config(conf=sparkConf)\
+ .getOrCreate()
return globals()['sparkSessionSingletonInstance']
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala
index 3795af8309..2b224d50a0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala
@@ -30,7 +30,10 @@ import org.apache.spark.sql.SparkSession
object AFTSurvivalRegressionExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("AFTSurvivalRegressionExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("AFTSurvivalRegressionExample")
+ .getOrCreate()
// $example on$
val training = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/ALSExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/ALSExample.scala
index 41750ca779..7c1cfe2937 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/ALSExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/ALSExample.scala
@@ -42,7 +42,10 @@ object ALSExample {
// $example off$
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("ALSExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("ALSExample")
+ .getOrCreate()
import spark.implicits._
// $example on$
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala
index 93c153f923..82bc14789b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.{DataFrame, SparkSession}
object BinarizerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("BinarizerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("BinarizerExample")
+ .getOrCreate()
// $example on$
val data = Array((0, 0.1), (1, 0.8), (2, 0.2))
val dataFrame: DataFrame = spark.createDataFrame(data).toDF("label", "feature")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala
index 779ad33dbd..38cce34bb5 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object BucketizerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("BucketizerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("BucketizerExample")
+ .getOrCreate()
// $example on$
val splits = Array(Double.NegativeInfinity, -0.5, 0.0, 0.5, Double.PositiveInfinity)
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala
index 84ca1f0b56..80f50cd355 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object ChiSqSelectorExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("ChiSqSelectorExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("ChiSqSelectorExample")
+ .getOrCreate()
import spark.implicits._
// $example on$
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala
index 9ab43a48bf..51aa5179fa 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object CountVectorizerExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("CounterVectorizerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("CounterVectorizerExample")
+ .getOrCreate()
// $example on$
val df = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala
index b415333c71..5a888b15eb 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object DCTExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("DCTExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("DCTExample")
+ .getOrCreate()
// $example on$
val data = Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala
index 2f892f8d72..6cb81cde6f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala
@@ -61,7 +61,10 @@ object DataFrameExample {
}
def run(params: Params) {
- val spark = SparkSession.builder.appName(s"DataFrameExample with $params").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName(s"DataFrameExample with $params")
+ .getOrCreate()
// Load input data
println(s"Loading LIBSVM file with UDT from ${params.input}.")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
index a0a2e1fb33..7f6c8de967 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
@@ -29,7 +29,10 @@ import org.apache.spark.sql.SparkSession
object DecisionTreeClassificationExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("DecisionTreeClassificationExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("DecisionTreeClassificationExample")
+ .getOrCreate()
// $example on$
// Load the data stored in LIBSVM format as a DataFrame.
val data = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala
index cea1d801aa..eadb02ab0d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala
@@ -167,7 +167,9 @@ object DecisionTreeExample {
testInput: String,
algo: String,
fracTest: Double): (DataFrame, DataFrame) = {
- val spark = SparkSession.builder.getOrCreate()
+ val spark = SparkSession
+ .builder
+ .getOrCreate()
// Load training data
val origExamples: DataFrame = loadData(spark, input, dataFormat)
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
index 26b52d0489..799070ef47 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
@@ -29,7 +29,10 @@ import org.apache.spark.sql.SparkSession
object DecisionTreeRegressionExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("DecisionTreeRegressionExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("DecisionTreeRegressionExample")
+ .getOrCreate()
// $example on$
// Load the data stored in LIBSVM format as a DataFrame.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala
index 2aa1ab1ec8..a522d2127e 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala
@@ -37,7 +37,10 @@ import org.apache.spark.sql.{Dataset, Row, SparkSession}
object DeveloperApiExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("DeveloperApiExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("DeveloperApiExample")
+ .getOrCreate()
import spark.implicits._
// Prepare training data.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala
index f289c28df9..b99b76e58c 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object ElementwiseProductExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("ElementwiseProductExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("ElementwiseProductExample")
+ .getOrCreate()
// $example on$
// Create some vector data; also works for sparse vectors
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala
index 91076ccbc1..972241e769 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala
@@ -29,7 +29,10 @@ import org.apache.spark.sql.SparkSession
object EstimatorTransformerParamExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("EstimatorTransformerParamExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("EstimatorTransformerParamExample")
+ .getOrCreate()
// $example on$
// Prepare training data from a list of (label, features) tuples.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
index 412c54db7d..b6a8baba2d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
@@ -28,7 +28,10 @@ import org.apache.spark.sql.SparkSession
object GradientBoostedTreeClassifierExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("GradientBoostedTreeClassifierExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("GradientBoostedTreeClassifierExample")
+ .getOrCreate()
// $example on$
// Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
index fd43553cc6..62285b83cb 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
@@ -28,7 +28,10 @@ import org.apache.spark.sql.SparkSession
object GradientBoostedTreeRegressorExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("GradientBoostedTreeRegressorExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("GradientBoostedTreeRegressorExample")
+ .getOrCreate()
// $example on$
// Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala
index d873618726..950733831c 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object IndexToStringExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("IndexToStringExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("IndexToStringExample")
+ .getOrCreate()
// $example on$
val df = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/KMeansExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/KMeansExample.scala
index d2573fad35..2abd588c6f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/KMeansExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/KMeansExample.scala
@@ -36,7 +36,10 @@ object KMeansExample {
def main(args: Array[String]): Unit = {
// Creates a Spark context and a SQL context
- val spark = SparkSession.builder.appName(s"${this.getClass.getSimpleName}").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName(s"${this.getClass.getSimpleName}")
+ .getOrCreate()
// $example on$
// Crates a DataFrame
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala
index c23adee1a3..c2920f6a5d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala
@@ -40,7 +40,10 @@ object LDAExample {
val input = "data/mllib/sample_lda_data.txt"
// Creates a Spark context and a SQL context
- val spark = SparkSession.builder.appName(s"${this.getClass.getSimpleName}").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName(s"${this.getClass.getSimpleName}")
+ .getOrCreate()
// $example on$
// Loads data
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala
index cb6e2492f5..94cf286623 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object LinearRegressionWithElasticNetExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("LinearRegressionWithElasticNetExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("LinearRegressionWithElasticNetExample")
+ .getOrCreate()
// $example on$
// Load training data
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala
index 50670d7b38..cd8775c942 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala
@@ -27,7 +27,10 @@ import org.apache.spark.sql.functions.max
object LogisticRegressionSummaryExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("LogisticRegressionSummaryExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("LogisticRegressionSummaryExample")
+ .getOrCreate()
import spark.implicits._
// Load training data
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/MaxAbsScalerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/MaxAbsScalerExample.scala
index 896d8fadbe..572adce657 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/MaxAbsScalerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/MaxAbsScalerExample.scala
@@ -24,7 +24,10 @@ import org.apache.spark.sql.SparkSession
object MaxAbsScalerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("MaxAbsScalerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("MaxAbsScalerExample")
+ .getOrCreate()
// $example on$
val dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala
index bcdca0fa04..d728019a62 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object MinMaxScalerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("MinMaxScalerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("MinMaxScalerExample")
+ .getOrCreate()
// $example on$
val dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
index a11fe1b4b2..0e780fb7d3 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
@@ -30,7 +30,10 @@ import org.apache.spark.sql.SparkSession
object MultilayerPerceptronClassifierExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("MultilayerPerceptronClassifierExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("MultilayerPerceptronClassifierExample")
+ .getOrCreate()
// $example on$
// Load the data stored in LIBSVM format as a DataFrame.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala
index 1b71a39890..e0b52e7a36 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object NGramExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("NGramExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("NGramExample")
+ .getOrCreate()
// $example on$
val wordDataFrame = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala
index 8d54555cd3..90cdebfcb0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object NaiveBayesExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("NaiveBayesExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("NaiveBayesExample")
+ .getOrCreate()
// $example on$
// Load the data stored in LIBSVM format as a DataFrame.
val data = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala
index 4622d69ef9..75ba33a7e7 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object NormalizerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("NormalizerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("NormalizerExample")
+ .getOrCreate()
// $example on$
val dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala
index 338436100c..4aa649b133 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object OneHotEncoderExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("OneHotEncoderExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("OneHotEncoderExample")
+ .getOrCreate()
// $example on$
val df = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala
index e2351c682d..fc73ae07ff 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala
@@ -109,7 +109,10 @@ object OneVsRestExample {
}
private def run(params: Params) {
- val spark = SparkSession.builder.appName(s"OneVsRestExample with $params").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName(s"OneVsRestExample with $params")
+ .getOrCreate()
// $example on$
val inputData = spark.read.format("libsvm").load(params.input)
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala
index 14394d5624..7927323b42 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object PCAExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("PCAExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("PCAExample")
+ .getOrCreate()
// $example on$
val data = Array(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/PipelineExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/PipelineExample.scala
index 61b34aebd9..e5e916ac16 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/PipelineExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/PipelineExample.scala
@@ -30,7 +30,10 @@ import org.apache.spark.sql.SparkSession
object PipelineExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("PipelineExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("PipelineExample")
+ .getOrCreate()
// $example on$
// Prepare training documents from a list of (id, text, label) tuples.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala
index 4d8c672a55..94b17a3cd7 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object PolynomialExpansionExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("PolynomialExpansionExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("PolynomialExpansionExample")
+ .getOrCreate()
// $example on$
val data = Array(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
index 0839c609f1..1a16515594 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
@@ -24,7 +24,10 @@ import org.apache.spark.sql.SparkSession
object QuantileDiscretizerExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("QuantileDiscretizerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("QuantileDiscretizerExample")
+ .getOrCreate()
import spark.implicits._
// $example on$
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala
index 699b621db9..9ea4920146 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object RFormulaExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("RFormulaExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("RFormulaExample")
+ .getOrCreate()
// $example on$
val dataset = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
index 4192a9c737..ae0bd945d8 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
@@ -28,7 +28,10 @@ import org.apache.spark.sql.SparkSession
object RandomForestClassifierExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("RandomForestClassifierExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("RandomForestClassifierExample")
+ .getOrCreate()
// $example on$
// Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
index 5632f0419a..96dc2f05be 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
@@ -28,7 +28,10 @@ import org.apache.spark.sql.SparkSession
object RandomForestRegressorExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("RandomForestRegressorExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("RandomForestRegressorExample")
+ .getOrCreate()
// $example on$
// Load and parse the data file, converting it to a DataFrame.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala
index f03b29ba32..bb4587b82c 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object SQLTransformerExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("SQLTransformerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("SQLTransformerExample")
+ .getOrCreate()
// $example on$
val df = spark.createDataFrame(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/SimpleParamsExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/SimpleParamsExample.scala
index dff7719507..3547dd95bd 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/SimpleParamsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/SimpleParamsExample.scala
@@ -34,7 +34,10 @@ import org.apache.spark.sql.{Row, SparkSession}
object SimpleParamsExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("SimpleParamsExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("SimpleParamsExample")
+ .getOrCreate()
import spark.implicits._
// Prepare training data.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/SimpleTextClassificationPipeline.scala b/examples/src/main/scala/org/apache/spark/examples/ml/SimpleTextClassificationPipeline.scala
index 05199007f0..c78ff2378b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/SimpleTextClassificationPipeline.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/SimpleTextClassificationPipeline.scala
@@ -42,7 +42,10 @@ case class Document(id: Long, text: String)
object SimpleTextClassificationPipeline {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("SimpleTextClassificationPipeline").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("SimpleTextClassificationPipeline")
+ .getOrCreate()
import spark.implicits._
// Prepare training documents, which are labeled.
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala
index 55f777c6e2..4d668e8ab9 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object StandardScalerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("StandardScalerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("StandardScalerExample")
+ .getOrCreate()
// $example on$
val dataFrame = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala
index 85e79c8cb3..fb1a43e962 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object StopWordsRemoverExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("StopWordsRemoverExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("StopWordsRemoverExample")
+ .getOrCreate()
// $example on$
val remover = new StopWordsRemover()
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala
index e01a768da9..63f273e87a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object StringIndexerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("StringIndexerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("StringIndexerExample")
+ .getOrCreate()
// $example on$
val df = spark.createDataFrame(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala
index 910ef62a26..33b5daec59 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object TfIdfExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("TfIdfExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("TfIdfExample")
+ .getOrCreate()
// $example on$
val sentenceData = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala
index 4f0c47b3c8..1c70dc700b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object TokenizerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("TokenizerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("TokenizerExample")
+ .getOrCreate()
// $example on$
val sentenceDataFrame = spark.createDataFrame(Seq(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala
index 56b7263b19..8e382ccc48 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object VectorAssemblerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("VectorAssemblerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("VectorAssemblerExample")
+ .getOrCreate()
// $example on$
val dataset = spark.createDataFrame(
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
index 214ad91634..afa761aee0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object VectorIndexerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("VectorIndexerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("VectorIndexerExample")
+ .getOrCreate()
// $example on$
val data = spark.read.format("libsvm").load("data/mllib/sample_libsvm_data.txt")
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala
index 716bf023a8..b1a3997f48 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala
@@ -31,7 +31,10 @@ import org.apache.spark.sql.SparkSession
object VectorSlicerExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("VectorSlicerExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("VectorSlicerExample")
+ .getOrCreate()
// $example on$
val data = Arrays.asList(Row(Vectors.dense(-2.0, 2.3, 0.0)))
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala
index 292b6d9f77..9ac5623607 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala
@@ -25,7 +25,10 @@ import org.apache.spark.sql.SparkSession
object Word2VecExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("Word2Vec example").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("Word2Vec example")
+ .getOrCreate()
// $example on$
// Input data: Each row is a bag of words from a sentence or document.
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala
index c2bf1548b5..7651aade49 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala
@@ -189,7 +189,9 @@ object LDAExample {
vocabSize: Int,
stopwordFile: String): (RDD[(Long, Vector)], Array[String], Long) = {
- val spark = SparkSession.builder.getOrCreate()
+ val spark = SparkSession
+ .builder
+ .getOrCreate()
import spark.implicits._
// Get dataset of document texts
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala
index cd4f0bb0de..781a934df6 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala
@@ -26,7 +26,10 @@ import org.apache.spark.sql.SparkSession
object RankingMetricsExample {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("RankingMetricsExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("RankingMetricsExample")
+ .getOrCreate()
import spark.implicits._
// $example on$
// Read in the ratings data
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala
index 22c47a694d..abeaaa00b5 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala
@@ -27,7 +27,10 @@ import org.apache.spark.sql.SparkSession
object RegressionMetricsExample {
def main(args: Array[String]): Unit = {
- val spark = SparkSession.builder.appName("RegressionMetricsExample").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("RegressionMetricsExample")
+ .getOrCreate()
// $example on$
// Load the data
val data = spark
diff --git a/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala b/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala
index b4118b16e2..94c378ae4b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala
@@ -26,7 +26,10 @@ case class Record(key: Int, value: String)
object RDDRelation {
def main(args: Array[String]) {
- val spark = SparkSession.builder.appName("RDDRelation").getOrCreate()
+ val spark = SparkSession
+ .builder
+ .appName("RDDRelation")
+ .getOrCreate()
// Importing the SparkSession gives access to all the SQL functions and implicit conversions.
import spark.implicits._
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala
index 2f0fe704f7..9aba4a05a8 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala
@@ -93,7 +93,10 @@ object SparkSessionSingleton {
def getInstance(sparkConf: SparkConf): SparkSession = {
if (instance == null) {
- instance = SparkSession.builder.config(sparkConf).getOrCreate()
+ instance = SparkSession
+ .builder
+ .config(sparkConf)
+ .getOrCreate()
}
instance
}
diff --git a/sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/VectorizedPlainValuesReader.java b/sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/VectorizedPlainValuesReader.java
index 9475c853a0..9def4559d2 100644
--- a/sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/VectorizedPlainValuesReader.java
+++ b/sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/VectorizedPlainValuesReader.java
@@ -34,8 +34,9 @@ public class VectorizedPlainValuesReader extends ValuesReader implements Vectori
private int offset;
private int bitOffset; // Only used for booleans.
private ByteBuffer byteBuffer; // used to wrap the byte array buffer
-
- private final static boolean bigEndianPlatform = ByteOrder.nativeOrder().equals(ByteOrder.BIG_ENDIAN);
+
+ private static final boolean bigEndianPlatform =
+ ByteOrder.nativeOrder().equals(ByteOrder.BIG_ENDIAN);
public VectorizedPlainValuesReader() {
}
diff --git a/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OffHeapColumnVector.java b/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OffHeapColumnVector.java
index b8dd16227e..70b4a68331 100644
--- a/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OffHeapColumnVector.java
+++ b/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OffHeapColumnVector.java
@@ -19,8 +19,6 @@ package org.apache.spark.sql.execution.vectorized;
import java.nio.ByteBuffer;
import java.nio.ByteOrder;
-import org.apache.commons.lang.NotImplementedException;
-
import org.apache.spark.memory.MemoryMode;
import org.apache.spark.sql.types.*;
import org.apache.spark.unsafe.Platform;
@@ -29,9 +27,10 @@ import org.apache.spark.unsafe.Platform;
* Column data backed using offheap memory.
*/
public final class OffHeapColumnVector extends ColumnVector {
-
- private final static boolean bigEndianPlatform = ByteOrder.nativeOrder().equals(ByteOrder.BIG_ENDIAN);
-
+
+ private static final boolean bigEndianPlatform =
+ ByteOrder.nativeOrder().equals(ByteOrder.BIG_ENDIAN);
+
// The data stored in these two allocations need to maintain binary compatible. We can
// directly pass this buffer to external components.
private long nulls;
@@ -230,7 +229,8 @@ public final class OffHeapColumnVector extends ColumnVector {
int srcOffset = srcIndex + Platform.BYTE_ARRAY_OFFSET;
long offset = data + 4 * rowId;
for (int i = 0; i < count; ++i, offset += 4, srcOffset += 4) {
- Platform.putInt(null, offset, java.lang.Integer.reverseBytes(Platform.getInt(src, srcOffset)));
+ Platform.putInt(null, offset,
+ java.lang.Integer.reverseBytes(Platform.getInt(src, srcOffset)));
}
}
}
@@ -276,7 +276,8 @@ public final class OffHeapColumnVector extends ColumnVector {
int srcOffset = srcIndex + Platform.BYTE_ARRAY_OFFSET;
long offset = data + 8 * rowId;
for (int i = 0; i < count; ++i, offset += 8, srcOffset += 8) {
- Platform.putLong(null, offset, java.lang.Long.reverseBytes(Platform.getLong(src, srcOffset)));
+ Platform.putLong(null, offset,
+ java.lang.Long.reverseBytes(Platform.getLong(src, srcOffset)));
}
}
}
diff --git a/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OnHeapColumnVector.java b/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OnHeapColumnVector.java
index b1ffe4c210..7fb7617050 100644
--- a/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OnHeapColumnVector.java
+++ b/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/OnHeapColumnVector.java
@@ -29,9 +29,10 @@ import org.apache.spark.unsafe.Platform;
* and a java array for the values.
*/
public final class OnHeapColumnVector extends ColumnVector {
-
- private final static boolean bigEndianPlatform = ByteOrder.nativeOrder().equals(ByteOrder.BIG_ENDIAN);
-
+
+ private static final boolean bigEndianPlatform =
+ ByteOrder.nativeOrder().equals(ByteOrder.BIG_ENDIAN);
+
// The data stored in these arrays need to maintain binary compatible. We can
// directly pass this buffer to external components.
diff --git a/sql/hive-thriftserver/src/main/java/org/apache/hive/service/cli/session/SessionManager.java b/sql/hive-thriftserver/src/main/java/org/apache/hive/service/cli/session/SessionManager.java
index 0457b3781e..de066dd406 100644
--- a/sql/hive-thriftserver/src/main/java/org/apache/hive/service/cli/session/SessionManager.java
+++ b/sql/hive-thriftserver/src/main/java/org/apache/hive/service/cli/session/SessionManager.java
@@ -22,7 +22,6 @@ import java.io.File;
import java.io.IOException;
import java.util.ArrayList;
import java.util.Date;
-import java.util.List;
import java.util.Map;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.Future;
@@ -35,7 +34,6 @@ import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.conf.HiveConf.ConfVars;
-import org.apache.hadoop.hive.ql.hooks.HookUtils;
import org.apache.hive.service.CompositeService;
import org.apache.hive.service.cli.HiveSQLException;
import org.apache.hive.service.cli.SessionHandle;