AI helps you reading Science

AI generates interpretation videos

AI extracts and analyses the key points of the paper to generate videos automatically


pub
Go Generating

AI Traceability

AI parses the academic lineage of this thesis


Master Reading Tree
Generate MRT

AI Insight

AI extracts a summary of this paper


Weibo:
We demonstrate how gating mechanisms, gate diversity, and graph structure can be used to integrating syntactic information and improve the hidden vectors for event detection models

Event Detection: Gate Diversity and Syntactic Importance Scores for Graph Convolution Neural Networks

EMNLP 2020, pp.5405-5411, (2020)

Cited by: 0|Views231
Full Text
Bibtex
Weibo

Abstract

Recent studies on event detection (ED) have shown that the syntactic dependency graph can be employed in graph convolution neural networks (GCN) to achieve state-of-the-art performance. However, the computation of the hidden vectors in such graph-based models is agnostic to the trigger candidate words, potentially leaving irrelevant infor...More

Code:

Data:

0
Introduction
  • Event Detection (ED) is an important task in Information Extraction of Natural Language Processing.
  • The main goal of this task is to identify event instances presented in text.
  • The event detection task, precisely speaking, seeks to identify the event triggers and classify them into some types of interest.
  • (1) They’ll be fired on at the crossing.
  • (2) She is on her way to get fired.
  • An ideal ED system should be able to recognize the two words “fired” in the sentences as the triggers of the event types “Attack” and “End-Position”
Highlights
  • Event Detection (ED) is an important task in Information Extraction of Natural Language Processing
  • We propose to filter the noisy information from the hidden vectors of graph convolution neural net-works (GCN) so that only the relevant information for the trigger candidate is preserved
  • Datasets: We evaluate our proposed model on two event detection (ED) datasets, i.e., ACE2005 and Litbank
  • The first class includes the models with noncontextualized embedding, i.e., CNN: a CNN model (Nguyen and Grishman, 2015), NCNN: non-consecutive CNN model: (Nguyen and Grishman, 2016), and GCN-ED: a GCN model (Nguyen and Grishman, 2018)
  • We demonstrate how gating mechanisms, gate diversity, and graph structure can be used to integrating syntactic information and improve the hidden vectors for ED models
  • The proposed model achieves state-of-the-art performance on two ED datasets
Methods
  • Datasets: The authors evaluate the proposed model on two ED datasets, i.e., ACE2005 and Litbank.
  • Once the authors re-implement this model and apply it to the data version pre-processed and provided by the prior work (Nguyen and Grishman, 2015, 2018), the authors are only able to achieve an F1 score of 76.2% on the test set.
  • In this work, the authors employ the exact data version that has been pre-processed and released by the early work on ED for ACE-2005 in (Nguyen and Grishman, 2015, 2018) and for Litbank in (Sims et al, 2019)
Results
  • The authors compare the model with two classes of baselines on ACE-2005. The first class includes the models with noncontextualized embedding, i.e., CNN: a CNN model (Nguyen and Grishman, 2015), NCNN: non-consecutive CNN model: (Nguyen and Grishman, 2016), and GCN-ED: a GCN model (Nguyen and Grishman, 2018).
  • The second class of baselines concern the models with the contextualized embeddings, i.e., DMBERT: a model with dynamic pooling (Wang et al, 2019) and BERT+MLP: a MLP model with BERT (Yang et al, 2019)
  • These models currently have the best-reported performance for ED on ACE-2005.
  • For Litbank, the authors use the following baselines reported in the original paper (Sims et al, 2019): BiLSTM: a BiLSTM model with word2vec, BERT+BiLSTM: a BiLSTM model with BERT, and DMBERT (Wang et al, 2019)
Conclusion
  • The authors demonstrate how gating mechanisms, gate diversity, and graph structure can be used to integrating syntactic information and improve the hidden vectors for ED models.
  • The proposed model achieves state-of-the-art performance on two ED datasets.
  • The authors plan to apply the proposed model for the related tasks and other settings of ED, including new type extension (Nguyen et al, 2016b; Lai and Nguyen, 2019), and few-shot learning (Lai et al, 2020a,b)
Tables
  • Table1: Performance on the ACE-2005 test set
  • Table2: Performance on the Litbank test set
  • Table3: Ablation study on the ACE-2005 dev set
Download tables as Excel
Related work
Funding
  • This research has been supported in part by Vingroup Innovation Foundation (VINIF) in project code VINIF.2019.DA18 and Adobe Research Gift
  • This research is also based upon work supported in part by the Office of the Director of National Intelligence (ODNI), Intelligence Advanced Research Projects Activity (IARPA), via IARPA Contract No 2019-19051600006 under the Better Extraction from Text Towards Enhanced Retrieval (BETTER) Program
Reference
  • David Ahn. 2006. The stages of event extraction. In Proceedings of the Workshop on Annotating and Reasoning about Time and Events.
    Google ScholarLocate open access versionFindings
  • Yubo Chen, Shulin Liu, Xiang Zhang, Kang Liu, and Jun Zhao. 2017. Automatically labeled data generation for large scale event extraction. In ACL.
    Google ScholarFindings
  • Yubo Chen, Liheng Xu, Kang Liu, Daojian Zeng, and Jun Zhao. 2015. Event extraction via dynamic multipooling convolutional neural networks. In ACLIJCNLP.
    Google ScholarFindings
  • Yubo Chen, Hang Yang, Kang Liu, Jun Zhao, and Yantao Jia. 2018. Collective event detection via a hierarchical and bias tagging networks with gated multilevel attention mechanisms. In EMNLP.
    Google ScholarFindings
  • Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In NAACL:HLT, pages 4171–4186.
    Google ScholarLocate open access versionFindings
  • Xiaocheng Feng, Lifu Huang, Duyu Tang, Heng Ji, Bing Qin, and Ting Liu. 201A languageindependent neural network for event detection. In ACL (Volume 2: Short Papers), volume 2, pages 66– 71.
    Google ScholarLocate open access versionFindings
  • Yu Hong, Jianfeng Zhang, Bin Ma, Jianmin Yao, Guodong Zhou, and Qiaoming Zhu. 2011. Using cross-entity inference to improve event extraction. In ACL.
    Google ScholarFindings
  • Abhyuday N Jagannatha and Hong Yu. 2016. Bidirectional rnn for medical event detection in electronic health records. In NAACL.
    Google ScholarFindings
  • Heng Ji and Ralph Grishman. 2008. Refining event extraction through cross-document inference. In ACL.
    Google ScholarFindings
  • Thomas N. Kipf and Max Welling. 2017. Semisupervised classification with graph convolutional networks. In ICLR.
    Google ScholarFindings
  • Viet Dac Lai, Franck Dernoncourt, and Thien Huu Nguyen. 2020a. Exploiting the matching information in the support set for few shot event classification. In Proceedings of the 24th Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD).
    Google ScholarLocate open access versionFindings
  • Viet Dac Lai, Franck Dernoncourt, and Thien Huu Nguyen. 2020b. Extensively matching for few-shot learning event detection. In Proceedings of the 1st Joint Workshop on Narrative Understanding, Storylines, and Events (NUSE) at ACL 2020.
    Google ScholarLocate open access versionFindings
  • Viet Dac Lai and Thien Nguyen. 2019. Extending event detection to new types with learning from keywords. In Proceedings of the 5th Workshop on Noisy User-generated Text (W-NUT 2019) at EMNLP 2019.
    Google ScholarLocate open access versionFindings
  • Qi Li, Heng Ji, and Liang Huang. 2013. Joint event extraction via structured prediction with global features. In ACL.
    Google ScholarFindings
  • Shulin Liu, Yubo Chen, Kang Liu, and Jun Zhao. 2017. Exploiting argument information to improve event detection via supervised attention mechanisms. In ACL.
    Google ScholarFindings
  • Xiao Liu, Zhunchen Luo, and He-Yan Huang. 2018. Jointly multiple events extraction via attentionbased graph information aggregation. In EMNLP, pages 1247–1256.
    Google ScholarLocate open access versionFindings
  • Diego Marcheggiani and Ivan Titov. 20Encoding sentences with graph convolutional networks for semantic role labeling. In EMNLP.
    Google ScholarFindings
  • Teruko Mitamura, Zhengzhong Liu, and Eduard Hovy. 2015. Overview of tac kbp 2015 event nugget track. In TAC.
    Google ScholarFindings
  • Thien Huu Nguyen, Kyunghyun Cho, and Ralph Grishman. 2016. Joint event extraction via recurrent neural networks. In NAACL.
    Google ScholarFindings
  • Thien Huu Nguyen, Lisheng Fu, Kyunghyun Cho, and Ralph Grishman. 2016b. A two-stage approach for extending event detection to new types via neural networks. In Proceedings of the 1st ACL Workshop on Representation Learning for NLP (RepL4NLP).
    Google ScholarLocate open access versionFindings
  • Thien Huu Nguyen and Ralph Grishman. 2015. Event detection and domain adaptation with convolutional neural networks. In ACL-IJCNLP.
    Google ScholarFindings
  • Thien Huu Nguyen and Ralph Grishman. 2016. Modeling skip-grams for event detection with convolutional neural networks. In EMNLP.
    Google ScholarFindings
  • Thien Huu Nguyen and Ralph Grishman. 2018. Graph convolutional networks with argument-aware pooling for event detection. In AAAI.
    Google ScholarFindings
  • Thien Huu Nguyen, Adam Meyers, and Ralph Grishman. 2016g. New york university 2016 system for kbp event nugget: A deep learning approach. In Proceedings of Text Analysis Conference (TAC).
    Google ScholarLocate open access versionFindings
  • Trung Minh Nguyen and Thien Huu Nguyen. 2019. One for all: Neural joint modeling of entities and events. In AAAI.
    Google ScholarFindings
  • Matthew Sims, Jong Ho Park, and David Bamman. 2019. Literary event detection. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 3623–3634.
    Google ScholarLocate open access versionFindings
  • Amir Pouran Ben Veyseh, Thien Huu Nguyen, and Dejing Dou. 2019. Graph based neural networks for event factuality prediction using syntactic and semantic structures. In ACL.
    Google ScholarFindings
  • Christopher Walker, Stephanie Strassel, Julie Medero, and Kazuaki Maeda. 2006. Ace 2005 multilingual training corpus. Linguistic Data Consortium, Philadelphia, 57.
    Google ScholarLocate open access versionFindings
  • Xiaozhi Wang, Xu Han, Zhiyuan Liu, Maosong Sun, and Peng Li. 2019. Adversarial training for weakly supervised event detection. In NAACL-HLT, pages 998–1008.
    Google ScholarLocate open access versionFindings
  • Haoran Yan, Xiaolong Jin, Xiangbin Meng, Jiafeng Guo, and Xueqi Cheng. 2019. Event detection with multi-order graph convolution and aggregated attention. In EMNLP-IJCNLP, pages 5770–5774.
    Google ScholarLocate open access versionFindings
  • Sen Yang, Dawei Feng, Linbo Qiao, Zhigang Kan, and Dongsheng Li. 2019. Exploring pre-trained language models for event extraction and generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5284– 5294.
    Google ScholarLocate open access versionFindings
Author
Tuan Ngo Nguyen
Tuan Ngo Nguyen
Thien Huu Nguyen
Thien Huu Nguyen
Your rating :
0

 

Tags
Comments
数据免责声明
页面数据均来自互联网公开来源、合作出版商和通过AI技术自动分析结果,我们不对页面数据的有效性、准确性、正确性、可靠性、完整性和及时性做出任何承诺和保证。若有疑问,可以通过电子邮件方式联系我们:report@aminer.cn
小科