This paper designs a hotel human-computer interaction (HCI) system which is based on deep learning. The whole system mainly includes face detection, speech synthesis, and speech recognition. Real time face detection is realized by transplanting the Opencv library into the Android system and combining that with the AdaBoost algorithm. Furthermore, the local speech recognition is realized through the special chip for speech recognition hbr740, while the local speech synthesis is realized through the special chip for speech synthesis syn6288. Subsequently, the massive speech resources are obtained through the network connection into the iFLYTEK’s open platform, which can realize online speech recognition, semantic understanding, and speech synthesis. In the data flow phase, we transmit data to the lower computer through the serial port of tiny4412 in order to realize the motion control of the lower computer. This is achieved through the SQLite database for system voice interaction and motion control, which is built through the jar package of the Litepal and POI. Finally, a hybrid voice interaction system is designed combining the local voice interaction with the online voice interaction. Through numerical simulation, the suggested system is tested to verify the feasibility of the hybrid voice interaction scheme. We observed that when the network is in good condition, the speech recognition rate of the whole system reaches as high as 94.67%; while without the network, the speech recognition rate can still reaches up to 84.67%. The attained outcomes demonstrate the superiority of the suggested hybrid system.
Read full abstract