Abstract

Natural language processing tasks in the health domain often deal with limited amount of labeled data. Pre-trained language models show us a promising way to compensate for the lake of training data, such as Bidirectional Encoder Representations from Transformers (BERT). However, previous downstream tasks often used training data at such a large scale that is unlikely to obtain in health domain. In this work, We conducted a learning curve analysis on a disease classification task to study the behavior of BERT and baseline models can still benefit downstream tasks when training data are relatively small in the context of health NLP. 1

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.