If your number of features is high, it may be useful to reduce it with an unsupervised step prior to supervised steps. Many of the Unsupervised learning methods implement a transform
method that can be used to reduce the dimensionality. Below we discuss two specific example of this pattern that are heavily used.
Pipelining
The unsupervised data reduction and the supervised estimator can be chained in one step. See Pipeline: chaining estimators.
decomposition.PCA
looks for a combination of features that capture well the variance of the original features. See Decomposing signals in components (matrix factorization problems).
The module: random_projection
provides several tools for data reduction by random projections. See the relevant section of the documentation: Random Projection.
cluster.FeatureAgglomeration
applies Hierarchical clustering to group together features that behave similarly.
Feature scaling
Note that if features have very different scaling or statistical properties, cluster.FeatureAgglomeration
may not be able to capture the links between related features. Using a preprocessing.StandardScaler
can be useful in these settings.
© 2007–2016 The scikit-learn developers
Licensed under the 3-clause BSD License.
http://scikit-learn.org/stable/modules/unsupervised_reduction.html