Abstract

Biomedical named entity recognition (BioNER) is the task of categorizing biomedical entities. Due to the specific characteristics of the names of biomedical entities, such as ambiguity among different concepts or different ways of referring to the same entity, the BioNER task is usually considered more challenging compared to standard named entity recognition tasks. Recent techniques based on deep learning not only significantly reduce the hand crafted feature engineering phase but also determined relevant improvements in the BioNER task. However, such systems are still facing challenges. One of them is the limited availability of annotated text data. Multi-task learning approaches tackle this problem by training different related tasks simultaneously. This enables multi-task models to learn common features among different tasks where they share some layers. To explore the advantages of the multi-task learning, we propose a model based on convolution neural networks, long-short term memories, and conditional random fields. The model we propose shows comparable results to state-of-the-art approaches. Moreover, we present an empirical analysis considering the impact of different word input representations (word embedding, character embedding, and case embedding) on the model performance.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call