Orange: Manifold Learning: Difference between revisions

From OnnoCenterWiki
Jump to navigationJump to search
Onnowpurbo (talk | contribs)
No edit summary
Onnowpurbo (talk | contribs)
No edit summary
Line 5: Line 5:
Nonlinear dimensionality reduction.
Nonlinear dimensionality reduction.


Inputs
==Input==


    Data: input dataset
Data: input dataset


Outputs
==Output==


    Transformed Data: dataset with reduced coordinates
Transformed Data: dataset with reduced coordinates


Manifold Learning is a technique which finds a non-linear manifold within the higher-dimensional space. The widget then outputs new coordinates which correspond to a two-dimensional space. Such data can be later visualized with Scatter Plot or other visualization widgets.
Manifold Learning is a technique which finds a non-linear manifold within the higher-dimensional space. The widget then outputs new coordinates which correspond to a two-dimensional space. Such data can be later visualized with Scatter Plot or other visualization widgets.
Line 18: Line 18:




    Method for manifold learning:
* Method for manifold learning:
        t-SNE
** t-SNE
        MDS, see also MDS widget
** MDS, see also MDS widget
        Isomap
** Isomap
        Locally Linear Embedding
** Locally Linear Embedding
        Spectral Embedding
** Spectral Embedding
    Set parameters for the method:
* Set parameters for the method:
        t-SNE (distance measures):
** t-SNE (distance measures):
            Euclidean distance
*** Euclidean distance
            Manhattan
*** Manhattan
            Chebyshev
*** Chebyshev
            Jaccard
*** Jaccard
            Mahalanobis
*** Mahalanobis
            Cosine
*** Cosine
        MDS (iterations and initialization):
** MDS (iterations and initialization):
            max iterations: maximum number of optimization interactions
*** max iterations: maximum number of optimization interactions
            initialization: method for initialization of the algorithm (PCA or random)
*** initialization: method for initialization of the algorithm (PCA or random)
        Isomap:
** Isomap:
            number of neighbors
*** number of neighbors
        Locally Linear Embedding:
** Locally Linear Embedding:
            method:
*** method:
                standard
**** standard
                modified
**** modified
                hessian eigenmap
**** hessian eigenmap
                local
**** local
            number of neighbors
*** number of neighbors
            max iterations
*** max iterations
        Spectral Embedding:
** Spectral Embedding:
            affinity:
*** affinity:
                nearest neighbors
**** nearest neighbors
                RFB kernel
**** RFB kernel
    Output: the number of reduced features (components).
* Output: the number of reduced features (components).
    If Apply automatically is ticked, changes will be propagated automatically. Alternatively, click Apply.
* If Apply automatically is ticked, changes will be propagated automatically. Alternatively, click Apply.
    Produce a report.
* Produce a report.


Manifold Learning widget produces different embeddings for high-dimensional data.
Manifold Learning widget produces different embeddings for high-dimensional data.

Revision as of 01:54, 29 January 2020

Sumber: https://docs.biolab.si//3/visual-programming/widgets/unsupervised/manifoldlearning.html


Nonlinear dimensionality reduction.

Input

Data: input dataset

Output

Transformed Data: dataset with reduced coordinates

Manifold Learning is a technique which finds a non-linear manifold within the higher-dimensional space. The widget then outputs new coordinates which correspond to a two-dimensional space. Such data can be later visualized with Scatter Plot or other visualization widgets.


  • Method for manifold learning:
    • t-SNE
    • MDS, see also MDS widget
    • Isomap
    • Locally Linear Embedding
    • Spectral Embedding
  • Set parameters for the method:
    • t-SNE (distance measures):
      • Euclidean distance
      • Manhattan
      • Chebyshev
      • Jaccard
      • Mahalanobis
      • Cosine
    • MDS (iterations and initialization):
      • max iterations: maximum number of optimization interactions
      • initialization: method for initialization of the algorithm (PCA or random)
    • Isomap:
      • number of neighbors
    • Locally Linear Embedding:
      • method:
        • standard
        • modified
        • hessian eigenmap
        • local
      • number of neighbors
      • max iterations
    • Spectral Embedding:
      • affinity:
        • nearest neighbors
        • RFB kernel
  • Output: the number of reduced features (components).
  • If Apply automatically is ticked, changes will be propagated automatically. Alternatively, click Apply.
  • Produce a report.

Manifold Learning widget produces different embeddings for high-dimensional data.

From left to right, top to bottom: t-SNE, MDS, Isomap, Locally Linear Embedding and Spectral Embedding.

Contoh

Manifold Learning widget transforms high-dimensional data into a lower dimensional approximation. This makes it great for visualizing datasets with many features. We used voting.tab to map 16-dimensional data onto a 2D graph. Then we used Scatter Plot to plot the embeddings.


Referensi

Pranala Menarik