Abstract
Relation extraction independently verifies all entity pairs in a sentence to identify predefined relationships between named entities. Because these entity pairs share the same contextual features of a sentence, they lead to a complicated semantic structure. To distinguish semantic expressions between relation instances, manually designed rules or elaborate deep architectures are usually applied to learn task-relevant representations. In this paper, a discrete convolutional network is proposed to incorporate discrete linguistic interactions and deep feature weighting. This network applies a discretization strategy to fix parameters of convolutional kernels into ternary values. Then, these discretized kernels are used to learn discrete semantic structures from vectorized token representations. Our approach leverages the ability of discrete CNNs to capture discrete linguistic patterns of a sentence, thereby maintaining model expressiveness and improving performance in the relation extraction task. Furthermore, our method has the advantages of reducing the overfitting problem caused by depending on prior knowledge and decreasing the computational complexity by reducing the number of trainable parameters. Our model is evaluated on five widely used benchmark datasets. It achieves state-of-the-art performance, outperforming all compared related works. Experimental results also demonstrate that, compared with traditional CNN networks, it achieves an average improvement of 14.66% in F1-score and accelerates training by an average of 17.46%, highlighting the efficiency and effectiveness of our model in the relation extraction task.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have