Abstract

This paper deals with the unit-norm tight frames (UNTFs) that better represent and solve sparse representation problems. Under this sparse representation framework, a novel model termed tight dictionary learning (TDL) is proposed. Unlike those dictionaries that only focus on the signals of interest, TDL intends to learn a more comprehensive dictionary that better represents the signals with the UNTF properties to facilitate sparse recovery. To better describe the physical meanings, the normalization constraints on the dictionary columns are imposed. A gradient-based approach is developed to obtain the tight dictionary where the normalization constraints are embedded into the cost function by a parametrization method. The learned tight dictionary is applied to the speech inpainting problem to restore speech signals corrupted by data missing. In order to compensate the degradation effect of the data missing, a UNTF-based preconditioning technique is developed to reform the frame properties of the degraded system. Parametrization and gradient-based approaches similar to those applied in the TDL are adopted to design a tight preconditioner. Extensive experiments on speech denoising and speech segment restoration are conducted to demonstrate the superiority of the proposed schemes.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.