Abstract

Deep learning has emerged as a versatile tool for a wide range of NLP tasks, due to its superior capacity in representation learning. But its applicability is limited by the reliance on annotated examples, which are difficult to produce at scale. Indirect supervision has emerged as a promising direction to address this bottleneck, either by introducing labeling functions to automatically generate noisy examples from unlabeled text, or by imposing constraints over interdependent label decisions. A plethora of methods have been proposed, each with respective strengths and limitations. Probabilistic logic offers a unifying language to represent indirect supervision, but end-to-end modeling with probabilistic logic is often infeasible due to intractable inference and learning. In this paper, we propose deep probabilistic logic (DPL) as a general framework for indirect supervision, by composing probabilistic logic with deep learning. DPL models label decisions as latent variables, represents prior knowledge on their relations using weighted first-order logical formulas, and alternates between learning a deep neural network for the end task and refining uncertain formula weights for indirect supervision, using variational EM. This framework subsumes prior indirect supervision methods as special cases, and enables novel combination via infusion of rich domain and linguistic knowledge. Experiments on biomedical machine reading demonstrate the promise of this approach.

Highlights

  • Deep learning has proven successful in a wide range of NLP tasks (Bahdanau et al, 2014; Bengio et al, 2003; Clark and Manning, 2016; Hermann et al, 2015; Sutskever et al, 2014)

  • We propose deep probabilistic logic (DPL) as a unifying framework for indirect supervision (Figure 1)

  • Using cross-sentence relation extraction and entity linking as case studies, we show that distant supervision, data programming, and joint inference can be seamlessly combined in DPL to substantially improve machine reading accuracy, without requiring any manually labeled examples

Read more

Summary

Introduction

Deep learning has proven successful in a wide range of NLP tasks (Bahdanau et al, 2014; Bengio et al, 2003; Clark and Manning, 2016; Hermann et al, 2015; Sutskever et al, 2014). Joint inference incurs greater modeling complexity and often requires specialized learning and inference procedures Since these methods draw on diverse and often orthogonal sources of indirect supervision, combining them may help address their limitations and amplify their strengths. We show that all popular forms of indirect supervision can be represented in DPL by generalizing virtual evidence (Subramanya and Bilmes, 2007; Pearl, 2014) These diverse methods can be combined within a single framework for mutual amplification. Using cross-sentence relation extraction and entity linking as case studies, we show that distant supervision, data programming, and joint inference can be seamlessly combined in DPL to substantially improve machine reading accuracy, without requiring any manually labeled examples.

Related Work
Deep Probabilistic Logic
Biomedical Machine Reading
Cross-sentence relation extraction
Entity linking
Precision Recall
Joint entity and relation extraction
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call