aboutsummaryrefslogtreecommitdiff
path: root/mllib
diff options
context:
space:
mode:
authorJoseph K. Bradley <joseph@databricks.com>2015-06-02 22:56:56 -0700
committerXiangrui Meng <meng@databricks.com>2015-06-02 22:56:56 -0700
commit07c16cb5ba9cb0bfe34e8c0efbf06540a22d4e4e (patch)
treea8d45d54e5ed58299bfa0cd1336ae3dbf85fe8c9 /mllib
parentcafd5056e12a15f0ebf8015d52dfab999c4443b8 (diff)
downloadspark-07c16cb5ba9cb0bfe34e8c0efbf06540a22d4e4e.tar.gz
spark-07c16cb5ba9cb0bfe34e8c0efbf06540a22d4e4e.tar.bz2
spark-07c16cb5ba9cb0bfe34e8c0efbf06540a22d4e4e.zip
[SPARK-8053] [MLLIB] renamed scalingVector to scalingVec
I searched the Spark codebase for all occurrences of "scalingVector" CC: mengxr Author: Joseph K. Bradley <joseph@databricks.com> Closes #6596 from jkbradley/scalingVec-rename and squashes the following commits: d3812f8 [Joseph K. Bradley] renamed scalingVector to scalingVec
Diffstat (limited to 'mllib')
-rw-r--r--mllib/src/main/scala/org/apache/spark/ml/feature/ElementwiseProduct.scala2
-rw-r--r--mllib/src/main/scala/org/apache/spark/mllib/feature/ElementwiseProduct.scala14
2 files changed, 8 insertions, 8 deletions
diff --git a/mllib/src/main/scala/org/apache/spark/ml/feature/ElementwiseProduct.scala b/mllib/src/main/scala/org/apache/spark/ml/feature/ElementwiseProduct.scala
index 3ae1833390..1e758cb775 100644
--- a/mllib/src/main/scala/org/apache/spark/ml/feature/ElementwiseProduct.scala
+++ b/mllib/src/main/scala/org/apache/spark/ml/feature/ElementwiseProduct.scala
@@ -41,7 +41,7 @@ class ElementwiseProduct(override val uid: String)
* the vector to multiply with input vectors
* @group param
*/
- val scalingVec: Param[Vector] = new Param(this, "scalingVector", "vector for hadamard product")
+ val scalingVec: Param[Vector] = new Param(this, "scalingVec", "vector for hadamard product")
/** @group setParam */
def setScalingVec(value: Vector): this.type = set(scalingVec, value)
diff --git a/mllib/src/main/scala/org/apache/spark/mllib/feature/ElementwiseProduct.scala b/mllib/src/main/scala/org/apache/spark/mllib/feature/ElementwiseProduct.scala
index b0985baf9b..d67fe6c3ee 100644
--- a/mllib/src/main/scala/org/apache/spark/mllib/feature/ElementwiseProduct.scala
+++ b/mllib/src/main/scala/org/apache/spark/mllib/feature/ElementwiseProduct.scala
@@ -25,10 +25,10 @@ import org.apache.spark.mllib.linalg._
* Outputs the Hadamard product (i.e., the element-wise product) of each input vector with a
* provided "weight" vector. In other words, it scales each column of the dataset by a scalar
* multiplier.
- * @param scalingVector The values used to scale the reference vector's individual components.
+ * @param scalingVec The values used to scale the reference vector's individual components.
*/
@Experimental
-class ElementwiseProduct(val scalingVector: Vector) extends VectorTransformer {
+class ElementwiseProduct(val scalingVec: Vector) extends VectorTransformer {
/**
* Does the hadamard product transformation.
@@ -37,15 +37,15 @@ class ElementwiseProduct(val scalingVector: Vector) extends VectorTransformer {
* @return transformed vector.
*/
override def transform(vector: Vector): Vector = {
- require(vector.size == scalingVector.size,
- s"vector sizes do not match: Expected ${scalingVector.size} but found ${vector.size}")
+ require(vector.size == scalingVec.size,
+ s"vector sizes do not match: Expected ${scalingVec.size} but found ${vector.size}")
vector match {
case dv: DenseVector =>
val values: Array[Double] = dv.values.clone()
- val dim = scalingVector.size
+ val dim = scalingVec.size
var i = 0
while (i < dim) {
- values(i) *= scalingVector(i)
+ values(i) *= scalingVec(i)
i += 1
}
Vectors.dense(values)
@@ -54,7 +54,7 @@ class ElementwiseProduct(val scalingVector: Vector) extends VectorTransformer {
val dim = values.length
var i = 0
while (i < dim) {
- values(i) *= scalingVector(indices(i))
+ values(i) *= scalingVec(indices(i))
i += 1
}
Vectors.sparse(size, indices, values)