|
3 | 3 | Hashing feature transformation using Random Forests |
4 | 4 | =================================================== |
5 | 5 |
|
6 | | -RandomForestEmbedding provide a way to map data to a |
| 6 | +RandomForestEmbedding provides a way to map data to a |
7 | 7 | very high-dimensional, sparse representation, which might |
8 | 8 | be beneficial for classification. |
9 | 9 | The mapping is completely unsupervised and very efficient. |
10 | 10 |
|
11 | | -This example visualizes the partitionings given by several |
| 11 | +This example visualizes the partitions given by several |
12 | 12 | trees and shows how the transformation can also be used for |
13 | | -non-linear dimensionality reduction or manifold learning. |
| 13 | +non-linear dimensionality reduction or non-linear classification. |
14 | 14 |
|
15 | | -Points that are neighboring often share the same leaf of a tree will share |
16 | | -large parts of their hashed representation. This allows to |
17 | | -separate two circles simply based on the principal components of the |
| 15 | +Points that are neighboring often share the same leaf of a tree and therefore |
| 16 | +share large parts of their hashed representation. This allows to |
| 17 | +separate two concentric circles simply based on the principal components of the |
18 | 18 | transformed data. |
19 | 19 |
|
20 | | -In the high-dimensional space, a simple classifier if often |
21 | | -enough for a good fit. For sparse binary data, BernoulliNB |
| 20 | +In high-dimensional spaces, linear classifiers often achieve |
| 21 | +excellent accuracy. For sparse binary data, BernoulliNB |
22 | 22 | is particularly well-suited. The bottom row compares the |
23 | 23 | decision boundary obtained by BernoulliNB in the transformed |
24 | 24 | space with an ExtraTreesClassifier forests learned on the |
|
0 commit comments