Abstract

Traditional English classroom teaching is difficult to meet the oral learning needs of most learners. Thanks to the continuous advancement of speech processing technology, computer-assisted language learning systems are becoming more intelligent, not only pointing out learners' pronunciation errors but also assessing their overall pronunciation level. Method. This paper uses the method of tree kernel function to measure the similarity of two ontology trees. According to the features of nodes in ontology tree, methods to calculate the external features and internal features of nodes are proposed, respectively. External features are mainly obtained by calculating the hierarchical centrality, node density, and node coverage of nodes in the ontology tree; internal features are mainly obtained by measuring the richness of internal information. According to the similarity of ontology tree and the external features and internal features of nodes, the calculation formula of structural comprehensive similarity is improved, and the features of ontology itself can be fully considered in the calculation. According to the difference of the structure, the weights of the corresponding features during the calculation are adjusted autonomously, so that the calculation results are closer to reality. In spectral image preprocessing, endpoint detection utilizes the harmonic characteristics presented by narrowband spectrograms with high frequency resolution and eliminates useless nonspeech segments by detecting the presence of voiced segments. When building the neural network model, four convolutional layers, two fully connected layers, and one softmax output layer were conceived, and dropout was used to randomly suspend the work of some neurons to avoid overfitting. Results/Discussion. Through the data analysis of mean and variance and verified by one-way analysis of variance, it proves that the sentiment evaluation method in this paper is effective. The traditional multiple linear regression method is not suitable for the corpus and application scenarios of this paper. This paper proposes a decision tree structure, which is similar to the overall scoring process of raters, and uses the Interactive Dicremiser version 3 (ID3) algorithm to build a comprehensive evaluation decision tree for pitch, rhythm, intonation, speech rate, and emotion indicators. It is proved by experiments that the accurate consistency rate of the human-machine evaluation in this paper is 93%, the adjacent consistency rate is 96%, and the Pearson correlation coefficient value of the human-machine evaluation results is 0.89. The data results prove that the evaluation method in this paper is credible.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call