Abstract
Discriminative training is difficult to implement but essential to attaining state‐of‐the‐art performance in automatic speech recognition systems today. Most of the discriminative training results for large scale recognition tasks (with vocabularies well over 10<th>000 words) so far use the maximum mutual information (MMI) framework, but recent results for the minimum classification error (MCE) framework suggest that MCE too yields significant improvements in recognition accuracy and system compactness on large‐scale tasks. MCE embodies rather well the general intuition that recognition system design should attempt to improve performance (i.e., recognition accuracy) directly, by optimizing a criterion function that is closely related to performance, rather than indirectly, by optimizing a criterion such as overall log likelihood that does not reflect performance. This presentation provides an overview of the MCE framework, and describes recent MCE speech recognition results obtained with both the MIT Galaxy system and the NTT Communication Science Labs SOLON system. The tasks examined include a 33<th>000‐word vocabulary telephone‐based spontaneous speech weather information task, a 22<th>000‐word telephone‐based name recognition task, and a 100<th>000‐word Japanese lecture speech transcription task.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.