Site icon i2tutorials

What is Dimensionality Reduction in Machine Learning?

Dimensionality Reduction in Machine Learning (i2tutorials)

Dimensionality reduction is the process of reducing the number of random variables under consideration, by obtaining a set of principal variables. It can be divided into feature selection and feature extraction. The higher the number of features, the harder it gets to visualize the training set and then work on it. Sometimes, most of these features are correlated, and hence redundant. This is where dimensionality reduction algorithms come into play.

There are two components of dimensionality reduction:

Feature selection:

In this, we try to find a subset of the original set of variables, or features, to get a smaller subset which can be used to model the problem. It usually involves three ways:

Feature extraction:

This reduces the data in a high dimensional space to a lower dimension space, i.e. a space with lesser no. of dimensions.

 

The various methods used for dimensionality reduction include:

  1. Principal Component Analysis (PCA)
  2. Linear Discriminant Analysis (LDA)
  3. Generalized Discriminant Analysis (GDA)

Advantages of Dimensionality Reduction

Disadvantages of Dimensionality Reduction

Exit mobile version