From 8896b0ee3931285942a28ff9f3b54bf289a0944d Mon Sep 17 00:00:00 2001 From: jrabary Date: Sun, 19 Apr 2015 21:00:50 +0200 Subject: [PATCH] Comments fix --- .../scala/org/apache/spark/ml/feature/StandardScaler.scala | 7 +++---- 1 file changed, 3 insertions(+), 4 deletions(-) diff --git a/mllib/src/main/scala/org/apache/spark/ml/feature/StandardScaler.scala b/mllib/src/main/scala/org/apache/spark/ml/feature/StandardScaler.scala index 978f10269ebda..e540ddcb22104 100644 --- a/mllib/src/main/scala/org/apache/spark/ml/feature/StandardScaler.scala +++ b/mllib/src/main/scala/org/apache/spark/ml/feature/StandardScaler.scala @@ -32,20 +32,19 @@ import org.apache.spark.sql.types.{StructField, StructType} private[feature] trait StandardScalerParams extends Params with HasInputCol with HasOutputCol { /** - * Whether to center the data before scaling + * False by default. Centers the data with mean before scaling. + * It will build a dense output, so this does not work on sparse input and will raise an exception. * @group param */ val withMean: BooleanParam = new BooleanParam(this, "withMean", "Center data with mean") /** - * Whether to scale the data to have unit standard deviation + * True by default. Scales the data to unit standard deviation. * @group param */ val withStd: BooleanParam = new BooleanParam(this, "withStd", "Scale to unit standard deviation") } - - /** * :: AlphaComponent :: * Standardizes features by removing the mean and scaling to unit variance using column summary