Abstract

To unlock information present in clinical description, automatic medical text classification is highly useful in the arena of natural language processing (NLP). For medical text classification tasks, machine learning techniques seem to be quite effective; however, it requires extensive effort from human side, so that the labeled training data can be created. For clinical and translational research, a huge quantity of detailed patient information, such as disease status, lab tests, medication history, side effects, and treatment outcomes, has been collected in an electronic format, and it serves as a valuable data source for further analysis. Therefore, a huge quantity of detailed patient information is present in the medical text, and it is quite a huge challenge to process it efficiently. In this work, a medical text classification paradigm, using two novel deep learning architectures, is proposed to mitigate the human efforts. The first approach is that a quad channel hybrid long short-term memory (QC-LSTM) deep learning model is implemented utilizing four channels, and the second approach is that a hybrid bidirectional gated recurrent unit (BiGRU) deep learning model with multihead attention is developed and implemented successfully. The proposed methodology is validated on two medical text datasets, and a comprehensive analysis is conducted. The best results in terms of classification accuracy of 96.72% is obtained with the proposed QC-LSTM deep learning model, and a classification accuracy of 95.76% is obtained with the proposed hybrid BiGRU deep learning model.

Highlights

  • To unlock information present in clinical description, automatic medical text classification is highly useful in the arena of natural language processing (NLP)

  • A medical text classification paradigm, using two novel deep learning architectures, is proposed to mitigate the human efforts. e first approach is that a quad channel hybrid long short-term memory (QC-LSTM) deep learning model is implemented utilizing four channels, and the second approach is that a hybrid bidirectional gated recurrent unit (BiGRU) deep learning model with multihead attention is developed and implemented successfully. e proposed methodology is validated on two medical text datasets, and a comprehensive analysis is conducted. e best results in terms of classification accuracy of 96.72% is obtained with the proposed QC-LSTM deep learning model, and a classification accuracy of 95.76% is obtained with the proposed hybrid BiGRU deep learning model

  • In the text classification field, a special emphasis is always given to the medical text classification as a lot of medical records along with medical literature are contained in the medical text [4]. e medical records include the doctor’s examination, diagnosis procedures, treatment protocols, and notification of improvement of the disease in the patient. e entire medical history along with the prescription effect of the medicine on the patient is stored in the medical record. e medical literature includes the oldest and recent documents of the medical techniques used for diagnosis and treatment of a particular disease [5]

Read more

Summary

Proposed Method 1

Erefore, to train the word embedding, Word2vec model dependent on skip-gram is utilized in this paper. A sigmoid function homogeneous to input gate, which generates a value ot between 0 and 1, shows the amount of cell state information determined to project it as output. When the multiplication of the cell state information happens with ot, it is activated by means of utilizing tanh layer, and so, the output details of the LSTM representation ht are modeled. A four-channel mechanism is introduced in the CNN-LSTM model by means of giving multiple labels of embeddings as input simultaneously at a given instant of time, so that multiple aspects of features are acquired. E word-level semantic features can be well extracted by this model, and so, the input data along with the output size can be reduced by means of mitigating the overfitting aspect.

Proposed Method 2
Results and Discussion
Analysis with Proposed Model 1
Analysis with Proposed Model 2
Conclusion and Future Work
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call