Abstract
Until recently, state-of-the-art, large-vocabulary, continuous speech recognition has employed hidden Markov modeling (HMM) to model speech sounds. The authors previously (ICASSP-92 p.625-8) presented the concept of a segmental neural network (SNN) for phonetic modeling in continuous speech recognition and demonstrated that a feedforward neural network, used within a hybrid SNN/HMM system, is able to reduce by 20% the word error rate over the baseline HMM system. They describe two developments over the initial system. First, a novel way to generate fixed length segment representations based on the discrete cosine transform (DCT) is presented. Second, it is demonstrated that an elliptical basis function (EBF) network can be used in the same hybrid framework. >
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.