diff options
author | Dongjoon Hyun <dongjoon@apache.org> | 2016-02-22 09:52:07 +0000 |
---|---|---|
committer | Sean Owen <sowen@cloudera.com> | 2016-02-22 09:52:07 +0000 |
commit | 024482bf51e8158eed08a7dc0758f585baf86e1f (patch) | |
tree | e51f2c53b027178bb4e485d2781e266d96ff6e3d /sql/catalyst | |
parent | 1b144455b620861d8cc790d3fc69902717f14524 (diff) | |
download | spark-024482bf51e8158eed08a7dc0758f585baf86e1f.tar.gz spark-024482bf51e8158eed08a7dc0758f585baf86e1f.tar.bz2 spark-024482bf51e8158eed08a7dc0758f585baf86e1f.zip |
[MINOR][DOCS] Fix all typos in markdown files of `doc` and similar patterns in other comments
## What changes were proposed in this pull request?
This PR tries to fix all typos in all markdown files under `docs` module,
and fixes similar typos in other comments, too.
## How was the this patch tested?
manual tests.
Author: Dongjoon Hyun <dongjoon@apache.org>
Closes #11300 from dongjoon-hyun/minor_fix_typos.
Diffstat (limited to 'sql/catalyst')
4 files changed, 5 insertions, 5 deletions
diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/DistinctAggregationRewriter.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/DistinctAggregationRewriter.scala index b49885d469..7518946a94 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/DistinctAggregationRewriter.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/DistinctAggregationRewriter.scala @@ -88,7 +88,7 @@ import org.apache.spark.sql.types.IntegerType * this aggregate consists of the original group by clause, all the requested distinct columns * and the group id. Both de-duplication of distinct column and the aggregation of the * non-distinct group take advantage of the fact that we group by the group id (gid) and that we - * have nulled out all non-relevant columns for the the given group. + * have nulled out all non-relevant columns the given group. * 3. Aggregating the distinct groups and combining this with the results of the non-distinct * aggregation. In this step we use the group id to filter the inputs for the aggregate * functions. The result of the non-distinct group are 'aggregated' by using the first operator, diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HyperLogLogPlusPlus.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HyperLogLogPlusPlus.scala index ec833d6789..a474017221 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HyperLogLogPlusPlus.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/aggregate/HyperLogLogPlusPlus.scala @@ -238,7 +238,7 @@ case class HyperLogLogPlusPlus( diff * diff } - // Keep moving bounds as long as the the (exclusive) high bound is closer to the estimate than + // Keep moving bounds as long as the (exclusive) high bound is closer to the estimate than // the lower (inclusive) bound. var low = math.max(nearestEstimateIndex - K + 1, 0) var high = math.min(low + K, numEstimates) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GenerateMutableProjection.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GenerateMutableProjection.scala index 5b4dc8df86..9abe92b1e7 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GenerateMutableProjection.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/codegen/GenerateMutableProjection.scala @@ -83,7 +83,7 @@ object GenerateMutableProjection extends CodeGenerator[Seq[Expression], () => Mu } } - // Evaluate all the the subexpressions. + // Evaluate all the subexpressions. val evalSubexpr = ctx.subexprFunctions.mkString("\n") val updates = validExpr.zip(index).map { diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/RandomDataGenerator.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/RandomDataGenerator.scala index 7c173cbcee..8207d64798 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/RandomDataGenerator.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/RandomDataGenerator.scala @@ -148,7 +148,7 @@ object RandomDataGenerator { // for "0001-01-01 00:00:00.000000". We need to find a // number that is greater or equals to this number as a valid timestamp value. while (milliseconds < -62135740800000L) { - // 253402329599999L is the the number of milliseconds since + // 253402329599999L is the number of milliseconds since // January 1, 1970, 00:00:00 GMT for "9999-12-31 23:59:59.999999". milliseconds = rand.nextLong() % 253402329599999L } @@ -163,7 +163,7 @@ object RandomDataGenerator { // for "0001-01-01 00:00:00.000000". We need to find a // number that is greater or equals to this number as a valid timestamp value. while (milliseconds < -62135740800000L) { - // 253402329599999L is the the number of milliseconds since + // 253402329599999L is the number of milliseconds since // January 1, 1970, 00:00:00 GMT for "9999-12-31 23:59:59.999999". milliseconds = rand.nextLong() % 253402329599999L } |