skip to main content
short-paper

Chinese Event Extraction via Graph Attention Network

Published: 19 January 2022 Publication History

Abstract

Event extraction plays an important role in natural language processing (NLP) applications, including question answering and information retrieval. Most of the previous state-of-the-art methods were lack of ability in capturing features in long range. Recent methods applied dependency tree via dependency-bridge and attention-based graph. However, most of the automatic processing tools used in those methods show poor performance on Chinese texts due to mismatching between word segmentation and labels, which results in error propagation. In this article, we propose a novel character-level Chinese event extraction framework via graph attention network (CAEE). We build our model upon the sequence labeling model, but enhance it with word information by incorporating the word lexicon into the character representations. We further exploit the inter-dependencies between event triggers and argument by building a word-character-based graph network via syntactic shortcut arcs with dependency-parsing. The architecture of the graph minimizes error propagation, which is the result of the error detection of the word boundaries in the processing of Chinese texts. To demonstrate the effectiveness of our work, we build a large-scale real-world corpus consisting of announcements of Chinese financial news without golden entities. Experiments on the corpus show that our approach achieves competitive results compared with previous work in the field of Chinese texts.

References

[1]
Deng Cai and Wai Lam. 2020. Graph transformer for graph-to-sequence learning. In AAAI. 7464–7471.
[2]
Chen Chen and Vincent Ng. 2012. Joint modeling for Chinese event extraction with rich linguistic features. In COLING. 529–544.
[3]
Yubo Chen, Liheng Xu, Kang Liu, Daojian Zeng, and Jun Zhao. 2015. Event extraction via dynamic multi-pooling convolutional neural networks. In ACL. The Association for Computer Linguistics, 167–176.
[4]
Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014).
[5]
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
[6]
Tao Gui, Yicheng Zou, Qi Zhang, Minlong Peng, Jinlan Fu, Zhongyu Wei, and Xuan-Jing Huang. 2019. A lexicon-based graph neural network for Chinese NER. In EMNLP-IJCNLP. 1039–1049.
[7]
Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural Computat. 9, 8 (1997), 1735–1780.
[8]
Thomas N. Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016).
[9]
Shen Li, Zhe Zhao, Renfen Hu, Wensi Li, Tao Liu, and Xiaoyong Du. 2018. Analogical reasoning on Chinese morphological and semantic relations. In ACL. Association for Computational Linguistics, 138–143. Retrieved from http://aclweb.org/anthology/P18-2023.
[10]
Hongyu Lin, Yaojie Lu, Xianpei Han, and Le Sun. 2018. Nugget proposal networks for Chinese event detection. arXiv preprint arXiv:1805.00249 (2018).
[11]
Shulin Liu, Yubo Chen, Shizhu He, Kang Liu, and Jun Zhao. 2016. Leveraging FrameNet to improve automatic event detection. In ACL. The Association for Computer Linguistics.
[12]
Shulin Liu, Kang Liu, Shizhu He, and Jun Zhao. 2016. A probabilistic soft logic based approach to exploiting latent and global information in event classification. In AAAI. AAAI Press, 2993–2999.
[13]
Xiao Liu, Zhunchen Luo, and Heyan Huang. 2018. Jointly multiple events extraction via attention-based graph information aggregation. In EMNLP. Association for Computational Linguistics, 1247–1256.
[14]
Ruotian Ma, Minlong Peng, Qi Zhang, Zhongyu Wei, and Xuanjing Huang. 2020. Simplify the usage of lexicon in Chinese NER. In ACL. Association for Computational Linguistics, 5951–5960.
[15]
Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781 (2013).
[16]
Thien Huu Nguyen, Kyunghyun Cho, and Ralph Grishman. 2016. Joint event extraction via recurrent neural networks. In HLT-NAACL. The Association for Computational Linguistics, 300–309.
[17]
Thien Huu Nguyen and Ralph Grishman. 2015. Event detection and domain adaptation with convolutional neural networks. In ACL/IJCNLP. Association for Computational Linguistics, Beijing, China, 365–371. DOI:https://doi.org/10.3115/v1/P15-2060
[18]
Lei Sha, Feng Qian, Baobao Chang, and Zhifang Sui. 2018. Jointly extracting event triggers and arguments by dependency-bridge RNN and tensor-based argument interaction. In AAAI. AAAI Press, 5916–5923.
[19]
Linfeng Song, Yue Zhang, Zhiguo Wang, and Daniel Gildea. 2018. N-ary relation extraction using graph state LSTM. arXiv preprint arXiv:1808.09101 (2018).
[20]
Dianbo Sui, Yubo Chen, Kang Liu, Jun Zhao, and Shengping Liu. 2019. Leverage lexical knowledge for Chinese named entity recognition via collaborative graph network. In EMNLP/IJCNLP. Association for Computational Linguistics, 3828–3838.
[21]
Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. 2017. Graph attention networks. arXiv preprint arXiv:1710.10903 (2017).
[22]
Bishan Yang and Tom M. Mitchell. 2016. Joint extraction of events and entities within a document context. In HLT-NAACL. The Association for Computational Linguistics, 289–299.
[23]
Ying Zeng, Honghui Yang, Yansong Feng, Zheng Wang, and Dongyan Zhao. 2016. A convolution BiLSTM neural network model for Chinese event extraction. In Natural Language Understanding and Intelligent Applications. Springer, 275–287.
[24]
Shuai Zhang, Lijie Wang, Ke Sun, and Xinyan Xiao. 2020. A Practical Chinese Dependency Parser Based on a Large-scale Dataset. arxiv:cs.CL/2009.00901.
[25]
Yue Zhang and Jie Yang. 2018. Chinese NER using lattice LSTM. In ACL. Association for Computational Linguistics, 1554–1564.
[26]
Shun Zheng, Wei Cao, Wei Xu, and Jiang Bian. 2019. Doc2EDAG: An end-to-end document-level framework for Chinese financial event extraction. In EMNLP/IJCNLP. Association for Computational Linguistics, 337–346.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Transactions on Asian and Low-Resource Language Information Processing
ACM Transactions on Asian and Low-Resource Language Information Processing  Volume 21, Issue 4
July 2022
464 pages
ISSN:2375-4699
EISSN:2375-4702
DOI:10.1145/3511099
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 19 January 2022
Accepted: 01 October 2021
Revised: 01 September 2021
Received: 01 January 2021
Published in TALLIP Volume 21, Issue 4

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Event extraction
  2. graph neural network

Qualifiers

  • Short-paper
  • Refereed

Funding Sources

  • National Natural Science Foundation of China
  • Sichuan Science and Technology Program

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)141
  • Downloads (Last 6 weeks)19
Reflects downloads up to 06 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Full Text

View this article in Full Text.

Full Text

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media