Package org.apache.spark.ml
Class Transformer
Object
org.apache.spark.ml.PipelineStage
org.apache.spark.ml.Transformer
- All Implemented Interfaces:
- Serializable,- org.apache.spark.internal.Logging,- Params,- Identifiable
- Direct Known Subclasses:
- Binarizer,- ColumnPruner,- FeatureHasher,- HashingTF,- IndexToString,- Interaction,- Model,- SQLTransformer,- StopWordsRemover,- UnaryTransformer,- VectorAssembler,- VectorAttributeRewriter,- VectorSizeHint,- VectorSlicer
Abstract class for transformers that transform one dataset into another.
- See Also:
- 
Nested Class SummaryNested classes/interfaces inherited from interface org.apache.spark.internal.Loggingorg.apache.spark.internal.Logging.LogStringContext, org.apache.spark.internal.Logging.SparkShellLoggingFilter
- 
Constructor SummaryConstructors
- 
Method SummaryModifier and TypeMethodDescriptionabstract TransformerCreates a copy of this instance with the same UID and some extra params.Transforms the input dataset.Transforms the dataset with provided parameter map as additional parameters.Transforms the dataset with optional parameterstransform(Dataset<?> dataset, ParamPair<?> firstParamPair, scala.collection.immutable.Seq<ParamPair<?>> otherParamPairs) Transforms the dataset with optional parametersMethods inherited from class org.apache.spark.ml.PipelineStageparams, transformSchemaMethods inherited from class java.lang.Objectequals, getClass, hashCode, notify, notifyAll, toString, wait, wait, waitMethods inherited from interface org.apache.spark.ml.util.IdentifiabletoString, uidMethods inherited from interface org.apache.spark.internal.LogginginitializeForcefully, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, isTraceEnabled, log, logDebug, logDebug, logDebug, logDebug, logError, logError, logError, logError, logInfo, logInfo, logInfo, logInfo, logName, LogStringContext, logTrace, logTrace, logTrace, logTrace, logWarning, logWarning, logWarning, logWarning, org$apache$spark$internal$Logging$$log_, org$apache$spark$internal$Logging$$log__$eq, withLogContextMethods inherited from interface org.apache.spark.ml.param.Paramsclear, copyValues, defaultCopy, explainParam, explainParams, extractParamMap, extractParamMap, get, getDefault, getOrDefault, getParam, hasDefault, hasParam, isDefined, isSet, onParamChange, set, set, set, setDefault, setDefault, shouldOwn
- 
Constructor Details- 
Transformerpublic Transformer()
 
- 
- 
Method Details- 
copyDescription copied from interface:ParamsCreates a copy of this instance with the same UID and some extra params. Subclasses should implement this method and set the return type properly. SeedefaultCopy().- Specified by:
- copyin interface- Params
- Specified by:
- copyin class- PipelineStage
- Parameters:
- extra- (undocumented)
- Returns:
- (undocumented)
 
- 
transformpublic Dataset<Row> transform(Dataset<?> dataset, ParamPair<?> firstParamPair, ParamPair<?>... otherParamPairs) Transforms the dataset with optional parameters- Parameters:
- dataset- input dataset
- firstParamPair- the first param pair, overwrite embedded params
- otherParamPairs- other param pairs, overwrite embedded params
- Returns:
- transformed dataset
 
- 
transformpublic Dataset<Row> transform(Dataset<?> dataset, ParamPair<?> firstParamPair, scala.collection.immutable.Seq<ParamPair<?>> otherParamPairs) Transforms the dataset with optional parameters- Parameters:
- dataset- input dataset
- firstParamPair- the first param pair, overwrite embedded params
- otherParamPairs- other param pairs, overwrite embedded params
- Returns:
- transformed dataset
 
- 
transformTransforms the dataset with provided parameter map as additional parameters.- Parameters:
- dataset- input dataset
- paramMap- additional parameters, overwrite embedded params
- Returns:
- transformed dataset
 
- 
transformTransforms the input dataset.- Parameters:
- dataset- (undocumented)
- Returns:
- (undocumented)
 
 
-