Abstract

Event detection (ED) task aims to classify events by identifying key event trigger words embedded in a piece of text. Previous research have proved the validity of fusing syntactic dependency relations into Graph Convolutional Networks(GCN). While existing GCN-based methods explore latent node-to-node dependency relations according to a stationary adjacency tensor, an attention-based dynamic tensor, which can pay much attention to the key node like event trigger or its neighboring nodes, has not been developed. Simultaneously, suffering from the phenomenon of graph information vanishing caused by the symmetric adjacency tensor, existing GCN models can not achieve higher overall performance. In this paper, we propose a novel model Self-Attention Graph Residual Convolution Networks (SA-GRCN) to mine node-to-node latent dependency relations via self-attention mechanism and introduce Graph Residual Network (GResNet) to solve graph information vanishing problem. Specifically, a self-attention module is constructed to generate an attention tensor, representing the dependency attention scores of all words in the sentence. Furthermore, a graph residual term is added to the baseline SA-GCN to construct a GResNet. Considering the syntactically connection of the network input, we initialize the raw adjacency tensor without processed by the self-attention module as the residual term. We conduct experiments on the ACE2005 dataset and the results show significant improvement over competitive baseline methods.

Highlights

  • Existing GCN-based Event detection (ED) methods(Liu et al, 2018) updates the graph by an adjacency mation vanishing caused by the symmetric adjacency tensor, existing GCN models can not achieve higher overall performance

  • 2020b) proposed to explore latent dependency rela- asymmetric adjacency tensor is initialized to repretions by aggregating information from neighbors of sent the graph, considering the directionality of the each node through specific edge, it will edge- dependency labels

  • A self-attention module is like “punct”, “det”connected with key nodes will constructed to update the self-attention tensor of interfere the event detection and dependency rela- the whole graph

Read more

Summary

Introduction

Existing GCN-based ED methods(Liu et al, 2018) updates the graph by an adjacency mation vanishing caused by the symmetric adjacency tensor, existing GCN models can not achieve higher overall performance. We propose a novel model Self-Attention Graph Residual Convolution Networks (SAGRCN) to mine node-to-node latent dependency relations via self-attention mechanism and introduce Graph Residual Network Such a graph structure only pay attention to the directly connected nodes. A graph residual term is added to the baseline SA-GCN to construct a GResNet. Considering the syntactically connection of the network input, we initialize the raw adjacency tensor recognized even though it’s connected to the “dobj” node “deaths” with “nmod” (noun compound modifier) dependency. Considering the syntactically connection of the network input, we initialize the raw adjacency tensor recognized even though it’s connected to the “dobj” node “deaths” with “nmod” (noun compound modifier) dependency Such an observation indicates that for multiple ED tasks, some event triggers may be ignored if they are indirectly connected to the root node. As the entity labeled by the ACE2005, both "Center" and "hospi-

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call