Abstract

AbstractIn this paper, we present several methods to reduce the computational and memory cost to embed HMM-based TTS system. We firstly decrease the number of HMMs by applying decision tree based context clustering technique. Secondly propose address-based model compression technique to compress the model size without degradation in synthesis speech quality. Thirdly reduce the feature vector size to decrease computational and memory resources. Finally, fixed-point implementation is taken to fit the TTS system requirements to embedded devices’ resource. Experimental results show that the system size can be compressed to 3.61MB from 293MB, memory and computational cost are low enough for real-time embedded application. Subjective evaluation shows that the synthesis speech quality is fairly good.KeywordsHMM-based speech synthesislarger speech databasemodel compressionembedded application

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.