From b162cc0c2810c1a9fa2eee8e664ffae84f9eea11 Mon Sep 17 00:00:00 2001 From: Jacek Laskowski Date: Fri, 9 Dec 2016 18:45:57 +0800 Subject: [MINOR][CORE][SQL][DOCS] Typo fixes ## What changes were proposed in this pull request? Typo fixes ## How was this patch tested? Local build. Awaiting the official build. Author: Jacek Laskowski Closes #16144 from jaceklaskowski/typo-fixes. --- .../src/main/java/org/apache/spark/sql/streaming/OutputMode.java | 2 +- .../src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala | 2 +- .../org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala | 2 +- .../org/apache/spark/sql/catalyst/expressions/Expression.scala | 6 +++--- .../org/apache/spark/sql/catalyst/expressions/objects/objects.scala | 2 +- 5 files changed, 7 insertions(+), 7 deletions(-) (limited to 'sql') diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/streaming/OutputMode.java b/sql/catalyst/src/main/java/org/apache/spark/sql/streaming/OutputMode.java index 49a18df2c7..a515c1a109 100644 --- a/sql/catalyst/src/main/java/org/apache/spark/sql/streaming/OutputMode.java +++ b/sql/catalyst/src/main/java/org/apache/spark/sql/streaming/OutputMode.java @@ -46,7 +46,7 @@ public class OutputMode { /** * OutputMode in which all the rows in the streaming DataFrame/Dataset will be written - * to the sink every time these is some updates. This output mode can only be used in queries + * to the sink every time there are some updates. This output mode can only be used in queries * that contain aggregations. * * @since 2.0.0 diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala index f498e071b5..256f64e320 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/InternalRow.scala @@ -21,7 +21,7 @@ import org.apache.spark.sql.catalyst.expressions._ import org.apache.spark.sql.types.{DataType, Decimal, StructType} /** - * An abstract class for row used internal in Spark SQL, which only contain the columns as + * An abstract class for row used internally in Spark SQL, which only contains the columns as * internal types. */ abstract class InternalRow extends SpecializedGetters with Serializable { diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala index 4b8cac8f32..78897daec8 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala @@ -22,7 +22,7 @@ import org.apache.spark.sql.catalyst.expressions.Expression /** - * Interface for the system catalog (of columns, partitions, tables, and databases). + * Interface for the system catalog (of functions, partitions, tables, and databases). * * This is only used for non-temporary items, and implementations must be thread-safe as they * can be accessed in multiple threads. This is an external catalog because it is expected to diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Expression.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Expression.scala index 221f830aa8..b93a5d0b7a 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Expression.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/Expression.scala @@ -70,9 +70,9 @@ abstract class Expression extends TreeNode[Expression] { * children. * * Note that this means that an expression should be considered as non-deterministic if: - * - if it relies on some mutable internal state, or - * - if it relies on some implicit input that is not part of the children expression list. - * - if it has non-deterministic child or children. + * - it relies on some mutable internal state, or + * - it relies on some implicit input that is not part of the children expression list. + * - it has non-deterministic child or children. * * An example would be `SparkPartitionID` that relies on the partition id returned by TaskContext. * By default leaf expressions are deterministic as Nil.forall(_.deterministic) returns true. diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/objects/objects.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/objects/objects.scala index a8aa1e7255..fc323693a2 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/objects/objects.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/objects/objects.scala @@ -930,7 +930,7 @@ case class InitializeJavaBean(beanInstance: Expression, setters: Map[String, Exp /** * Asserts that input values of a non-nullable child expression are not null. * - * Note that there are cases where `child.nullable == true`, while we still needs to add this + * Note that there are cases where `child.nullable == true`, while we still need to add this * assertion. Consider a nullable column `s` whose data type is a struct containing a non-nullable * `Int` field named `i`. Expression `s.i` is nullable because `s` can be null. However, for all * non-null `s`, `s.i` can't be null. -- cgit v1.2.3