Abstract

The speech denoising model based on adversarial generative network has achieved better results than the traditional machine learning model. In this paper, for the short cut connection in the generator, we discuss its influence on the information transfer between encoder and decoder, and propose SDGAN at target. SDGAN sets linear and convolution filters in the short cut connection which adaptively learn the optimal information processing. The information filter still enables the generator to solve the gradient vanishing problem, and it can also avoid information redundancy and improve expression ability. In addition, SDGAN replaces the L1 regularization term in loss function with the L2 regularization term, which not only makes the output speech of the generator closer to the clean speech, but also avoids sparsity. In the experiments, SDGAN significantly performs better than other traditional GAN in five performance metrics (such as PESQ), and the effect of convolution filter is better than that of linear filter.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call