Abstract
Natural language processing (NLP) has three basic tasks divided into two levels, lexical, which includes Tokenization task and syntactic level which includes Part Of Speech tasks (POS) and Name Entity Recognition (NER) tasks. Recent research has demonstrated the effectiveness of deep learning in many NLP tasks including NER, POS, sentiment analysis, language modeling, and other tasks. This study focused on the utilizing of Long Short-Term Memory (LSTM), Bidirectional Long Short-Term Memory (BLSTM), Bidirectional Long Short-Term Memory with Conditional Random Field (BLSTM-CRF), and Long Short-Term Memory with Conditional Random Field (LSTM-CRF) deep learning techniques for tasks at the syntactic level; and to compare their performance on noisy data. The models were trained and tested by using KALIMAT corpus with simulated noise on testing dataset. For evaluation purpose, F1-score was used, where the results of our experiments showed that a BLSTM-CRF model surpassed the rest of other models in the NER task at a low noise level, while the LSTM-CRF model obtained a higher F1-score at a higher noise level. With respect to the POS task, the BLSTM-CRF model gained the highest F1-score in all noise levels equated to the other competitive models.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.