Abstract

With the fast development of computer and information technology, multimedia data has become the most important form of information media. Auditory information plays an important role in information location, this comes from the fact that it can be difficult to find useful information. Thus audio classification becomes more important in audio analysis as it prepares for content-based audio retrieval. There is quite a bit of research on the topic of audio classification methods, audio feature analysis, and extraction based on audio classification. Many works of literature extract features of audio signals based on time or Fourier transform frequency domain. The emergence of the wavelet theory provides a time-frequency analysis tool for signal analysis. Wavelet transformation is a local transformation of the signal in time and frequency which can effectively extract information from the signal, and perform multi-scale refinement analysis on functions or signals through operations such as stretching and translation instead of the traditional Fourier transformation. In the time-frequency analysis of the signal, the wavelet analysis captures the local time and frequency characters of the signal which can improve the ability of signal analysis. It can also change certain locals of the signal without affecting other aspects of it. In this paper, the frequency domain features are combined with the wavelet domain features. At the same time that the MFCC features are extracted, the discrete wavelet transform is used to extract the features of the wavelet domain. Then the statistical features are extracted for each audio example, and the SVM model is used to realize the different forms of audio classification identification.

Highlights

  • People can experience a colorful world from perceiving sounds around them, including the process of continuous recognition

  • Audio data is an indispensable part of multimedia information, and research on content-based audio retrieval technology is booming and has achieved considerable results

  • Based on the work of predecessors, this paper attempts to solve the problems of audio feature analysis, classifier design and speech information retrieval

Read more

Summary

INTRODUCTION

People can experience a colorful world from perceiving sounds around them, including the process of continuous recognition. The model training techniques applied in the field of audio recognition mainly include dynamic time warping (DTW) technology, vector quantization (VQ) technology, artificial neural network (ANN) technology, and support vector machine (SVM). Incremental Learning SVM algorithm is suitable for mining large numbers of data sets, and it can get efficient training results when new training is generated over time. This paper considers the mechanism of improving the incremental learning algorithm to process audio based on the research from the incremental learning support vector machine. The unit of audio processing is in a frame, so before the feature extraction, the original audio data needs to be pre-processed: into a unified format, pre-emphasis, segmentation, and windowing framing.

AUDIO SIGNAL FEATURE EXTRACTION
RELATIONSHIP BETWEEN KKT CONDITION
PRINCIPAL COMPONENT ANALYSIS
RESULTS AND DISCUSSION
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call