Abstract

Modern methods for nonlinear dimensionality reduction have been used extensively in the machine learning community for discovering the intrinsic dimension of several datasets. In this paper we apply one of the most successful ones Maximum Variance Unfolding on a big sample of the well known speech benchmark TIMIT. Although MVU is not generally scalable, we managed to apply to 1 million 39-dimensional points and successfully reduced the dimension down to 15. In this paper we apply some of the state-of-the-art techniques for handling big datasets. The biggest bottleneck is the local neighborhood computation. For 300K points it took 9 hours while for 1M points it took 3.5 days. We also demonstrate the weakness of MFCC representation under the k-nearest neighborhood classification since the error rate is more than 50%.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call