Serializable
, org.apache.spark.internal.Logging
, Params
, HasHandleInvalid
, HasInputCol
, HasInputCols
, HasOutputCol
, HasOutputCols
, DefaultParamsWritable
, Identifiable
, MLWritable
Bucketizer
maps a column of continuous features to a column of feature buckets.
Since 2.3.0, Bucketizer
can map multiple columns at once by setting the inputCols
parameter. Note that when both the inputCol
and inputCols
parameters are set, an Exception will be thrown. The splits
parameter is only used for single column usage, and splitsArray
is for multiple columns.
org.apache.spark.internal.Logging.LogStringContext, org.apache.spark.internal.Logging.SparkShellLoggingFilter
Constructors
Creates a copy of this instance with the same UID and some extra params.
double[]
double[][]
Param for how to handle invalid entries containing NaN values.
Param for input column name.
Param for input column names.
Param for output column name.
Param for output column names.
Parameter for mapping continuous features into buckets.
Parameter for specifying multiple splits parameters.
Transforms the input dataset.
Check transform validity and derive the output schema from the input schema.
An immutable unique ID for the object and its derivatives.
Methods inherited from interface org.apache.spark.internal.LogginginitializeForcefully, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, isTraceEnabled, log, logDebug, logDebug, logDebug, logDebug, logError, logError, logError, logError, logInfo, logInfo, logInfo, logInfo, logName, LogStringContext, logTrace, logTrace, logTrace, logTrace, logWarning, logWarning, logWarning, logWarning, org$apache$spark$internal$Logging$$log_, org$apache$spark$internal$Logging$$log__$eq, withLogContext
Methods inherited from interface org.apache.spark.ml.util.MLWritablesave
Methods inherited from interface org.apache.spark.ml.param.Paramsclear, copyValues, defaultCopy, defaultParamMap, explainParam, explainParams, extractParamMap, extractParamMap, get, getDefault, getOrDefault, getParam, hasDefault, hasParam, isDefined, isSet, onParamChange, paramMap, params, set, set, set, setDefault, setDefault, shouldOwn
public Bucketizer()
Param for output column names.
outputCols
in interface HasOutputCols
Param for input column names.
inputCols
in interface HasInputCols
Param for output column name.
outputCol
in interface HasOutputCol
Param for input column name.
inputCol
in interface HasInputCol
An immutable unique ID for the object and its derivatives.
uid
in interface Identifiable
Parameter for mapping continuous features into buckets. With n+1 splits, there are n buckets. A bucket defined by splits x,y holds values in the range [x,y) except the last bucket, which also includes y. Splits should be of length greater than or equal to 3 and strictly increasing. Values at -inf, inf must be explicitly provided to cover all Double values; otherwise, values outside the splits specified will be treated as errors.
See also handleInvalid()
, which can optionally create an additional bucket for NaN values.
public double[] getSplits()
Param for how to handle invalid entries containing NaN values. Values outside the splits will always be treated as errors. Options are 'skip' (filter out rows with invalid values), 'error' (throw an error), or 'keep' (keep invalid values in a special additional bucket). Note that in the multiple column case, the invalid handling is applied to all columns. That said for 'error' it will throw an error if any invalids are found in any column, for 'skip' it will skip rows with any invalids in any columns, etc. Default: "error"
handleInvalid
in interface HasHandleInvalid
public double[][] getSplitsArray()
Transforms the input dataset.
transform
in class Transformer
dataset
- (undocumented)
Check transform validity and derive the output schema from the input schema.
We check validity for interactions between parameters during transformSchema
and raise an exception if any parameter value is invalid. Parameter value checks which do not depend on other parameters are handled by Param.validate()
.
Typical implementation should first conduct verification on schema change and parameter validity, including complex parameter interaction checks.
transformSchema
in class PipelineStage
schema
- (undocumented)
Params
Creates a copy of this instance with the same UID and some extra params. Subclasses should implement this method and set the return type properly. See defaultCopy()
.
copy
in interface Params
copy
in class Model<Bucketizer>
extra
- (undocumented)
toString
in interface Identifiable
toString
in class Object
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4