Abstract

From the analysis of big data, dimensionality reduction techniques play a significant role in various fields where the data is huge with multiple columns or classes. Data with high dimensions contains thousands of features where many of these features contain useful information. Along with this there contains a lot of redundant or irrelevant features which reduce the quality, performance of data and decrease the efficiency in computation. Procedures which are done mathematically for reducing dimensions are known as dimensionality reduction techniques. The main aim of the Dimensionality Reduction algorithms such as Principal Component Analysis (PCA), Random Projection (RP) and Non Negative Matrix Factorization (NMF) is used to decrease the inappropriate information from the data and moreover the features and attributes taken from these algorithms were not able to characterize data as different divisions. This paper gives a review about the traditional methods used in Machine algorithm for reducing the dimension and proposes a view, how deep learning can be used for dimensionality reduction.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.