From b208229ba1768ee72eae7b6b0da0ac9ccff3be62 Mon Sep 17 00:00:00 2001 From: Jacek Laskowski Date: Tue, 26 Apr 2016 11:51:12 +0100 Subject: [MINOR][DOCS] Minor typo fixes ## What changes were proposed in this pull request? Minor typo fixes (too minor to deserve separate a JIRA) ## How was this patch tested? local build Author: Jacek Laskowski Closes #12469 from jaceklaskowski/minor-typo-fixes. --- sql/catalyst/src/main/scala/org/apache/spark/sql/Row.scala | 2 +- sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala | 4 ++-- .../scala/org/apache/spark/sql/execution/datasources/DataSource.scala | 2 +- .../org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala | 3 +-- .../main/scala/org/apache/spark/sql/execution/streaming/Source.scala | 4 ++-- .../org/apache/spark/sql/execution/streaming/StreamExecution.scala | 2 +- .../org/apache/spark/sql/execution/streaming/StreamingRelation.scala | 2 +- .../scala/org/apache/spark/sql/streaming/FileStreamSourceSuite.scala | 4 ++-- 8 files changed, 11 insertions(+), 12 deletions(-) (limited to 'sql') diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/Row.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/Row.scala index 1219d4d453..726291b96c 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/Row.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/Row.scala @@ -338,7 +338,7 @@ trait Row extends Serializable { * Returns the index of a given field name. * * @throws UnsupportedOperationException when schema is not defined. - * @throws IllegalArgumentException when fieldName do not exist. + * @throws IllegalArgumentException when a field `name` does not exist. */ def fieldIndex(name: String): Int = { throw new UnsupportedOperationException("fieldIndex on a Row without schema is undefined.") diff --git a/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala b/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala index b3064fd531..4c2bf12ac9 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala @@ -458,8 +458,8 @@ class Dataset[T] protected[sql]( * Returns true if this [[Dataset]] contains one or more sources that continuously * return data as it arrives. A [[Dataset]] that reads data from a streaming source * must be executed as a [[ContinuousQuery]] using the `startStream()` method in - * [[DataFrameWriter]]. Methods that return a single answer, (e.g., `count()` or - * `collect()`) will throw an [[AnalysisException]] when there is a streaming + * [[DataFrameWriter]]. Methods that return a single answer, e.g. `count()` or + * `collect()`, will throw an [[AnalysisException]] when there is a streaming * source present. * * @group basic diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala index ef626ef5fc..8e72e06b1f 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala @@ -143,7 +143,7 @@ case class DataSource( caseInsensitiveOptions, fileCatalog.allFiles()) }.getOrElse { - throw new AnalysisException("Unable to infer schema. It must be specified manually.") + throw new AnalysisException("Unable to infer schema. It must be specified manually.") } } diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala index ddba3ccb1b..99bf20c746 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/HDFSMetadataLog.scala @@ -46,8 +46,7 @@ import org.apache.spark.sql.SparkSession * files in a directory always shows the latest files. */ class HDFSMetadataLog[T: ClassTag](sparkSession: SparkSession, path: String) - extends MetadataLog[T] - with Logging { + extends MetadataLog[T] with Logging { import HDFSMetadataLog._ diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Source.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Source.scala index 6457f928ed..1d2f7a8753 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Source.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/Source.scala @@ -34,8 +34,8 @@ trait Source { def getOffset: Option[Offset] /** - * Returns the data that is is between the offsets (`start`, `end`]. When `start` is `None` then - * the batch should begin with the first available record. This method must always return the + * Returns the data that is between the offsets (`start`, `end`]. When `start` is `None` then + * the batch should begin with the first available record. This method must always return the * same data for a particular `start` and `end` pair. */ def getBatch(start: Option[Offset], end: Offset): DataFrame diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamExecution.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamExecution.scala index ea3c73d984..fc18e5f065 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamExecution.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamExecution.scala @@ -75,7 +75,7 @@ class StreamExecution( /** The current batchId or -1 if execution has not yet been initialized. */ private var currentBatchId: Long = -1 - /** All stream sources present the query plan. */ + /** All stream sources present in the query plan. */ private val sources = logicalPlan.collect { case s: StreamingExecutionRelation => s.source } diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamingRelation.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamingRelation.scala index 3341580fc4..4d65d2f4f5 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamingRelation.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/StreamingRelation.scala @@ -33,7 +33,7 @@ object StreamingRelation { * [[org.apache.spark.sql.catalyst.plans.logical.LogicalPlan]]. This is only used for creating * a streaming [[org.apache.spark.sql.DataFrame]] from [[org.apache.spark.sql.DataFrameReader]]. * It should be used to create [[Source]] and converted to [[StreamingExecutionRelation]] when - * passing to [StreamExecution]] to run a query. + * passing to [[StreamExecution]] to run a query. */ case class StreamingRelation(dataSource: DataSource, sourceName: String, output: Seq[Attribute]) extends LeafNode { diff --git a/sql/core/src/test/scala/org/apache/spark/sql/streaming/FileStreamSourceSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/streaming/FileStreamSourceSuite.scala index 8ff82a3c74..6b1ecd08c1 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/streaming/FileStreamSourceSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/streaming/FileStreamSourceSuite.scala @@ -204,7 +204,7 @@ class FileStreamSourceSuite extends FileStreamSourceTest with SharedSQLContext { createFileStreamSourceAndGetSchema( format = Some("parquet"), path = Some(new File(src, "1").getCanonicalPath), schema = None) } - assert("Unable to infer schema. It must be specified manually.;" === e.getMessage) + assert("Unable to infer schema. It must be specified manually.;" === e.getMessage) } } @@ -235,7 +235,7 @@ class FileStreamSourceSuite extends FileStreamSourceTest with SharedSQLContext { createFileStreamSourceAndGetSchema( format = Some("json"), path = Some(src.getCanonicalPath), schema = None) } - assert("Unable to infer schema. It must be specified manually.;" === e.getMessage) + assert("Unable to infer schema. It must be specified manually.;" === e.getMessage) } } -- cgit v1.2.3