Abstract

In this paper, a neural network is proposed based on a convolutional recurrent network (CRN) architecture for speech enhancement in the time-frequency (T-F) domain. The proposed model is composed of a convolutional encoder-decoder (CED) with gated convolutional units, a skip connection module named attention gate units (AGUs) and a two-stage modeling module, which is called a grouped F-T-LSTM using multiply long short-term memory (LSTM) layers. The encoder transforms the input noisy speech into high dimensional feature maps and the gated convolutional units control the information flows through the network. Instead of directly concatenating the output of each encoder layer to the input of the corresponding decoder layer, attention gate units that better integrate both information are developed as the skip connections. The grouped F-T-LSTM module is developed to model the relations between frequency bands and temporal dependencies. Finally, the decoder uses the output of AGUs and the grouped F-T-LSTM to produce a mask. The proposed model is trained using a loss function which is a weighted time domain and time-frequency loss. Experimental results show that the proposed model achieves a superior performance when compared with other existing speech enhancement methods.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.