AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
Article Link
Collect
Submit Manuscript
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Regular Paper

Label-Aware Chinese Event Detection with Heterogeneous Graph Attention Network

Institute of Information Engineering, Chinese Academy of Sciences, Beijing 100190, China
School of Cyber Security, University of Chinese Academy of Sciences, Beijing 100049, China
Cyberspace Institute of Advanced Technology, Guangzhou University, Guangzhou 510006, China
School of Cyber Security, Beijing University of Posts and Telecommunications, Beijing 100088, China
Show Author Information

Abstract

Event detection (ED) seeks to recognize event triggers and classify them into the predefined event types. Chinese ED is formulated as a character-level task owing to the uncertain word boundaries. Prior methods try to incorporate word-level information into characters to enhance their semantics. However, they experience two problems. First, they fail to incorporate word-level information into each character the word encompasses, causing the insufficient word-character interaction problem. Second, they struggle to distinguish events of similar types with limited annotated instances, which is called the event confusing problem. This paper proposes a novel model named Label-Aware Heterogeneous Graph Attention Network (L-HGAT) to address these two problems. Specifically, we first build a heterogeneous graph of two node types and three edge types to maximally preserve word-character interactions, and then deploy a heterogeneous graph attention network to enhance the semantic propagation between characters and words. Furthermore, we design a pushing-away game to enlarge the predicting gap between the ground-truth event type and its confusing counterpart for each character. Experimental results show that our L-HGAT model consistently achieves superior performance over prior competitive methods.

Electronic Supplementary Material

Download File(s)
JCST-2104-11541-Highlights.pdf (390.6 KB)

References

[1]
Filatova E, Hatzivassiloglou V. Event-based extractive summarization. In Proc. the 42nd Annual Meeting of the Association for Computational Linguistics Workshop on Summarization, Jul. 2004, pp.104–111.
[2]
Mitamura T, Liu Z Z, Hovy E H. Events detection, coreference and sequencing: What’s next? Overview of the TAC KBP 2017 event track. In Proc. the 2017 Text Analysis Conference, Nov. 2017.
[3]
Ji H, Grishman R. Knowledge base population: Successful approaches and challenges. In Proc. the 49th Annual Meeting of the Association for Computational Linguistics, Jun. 2011, pp.1148–1158.
[4]
Basile P, Caputo A, Semeraro G, Siciliani L. Extending an information retrieval system through time event extraction. In Proc. the 8th International Workshop on Information Filtering and Retrieval Co-Located with XⅢ AI*IA Symposium on Artificial Intelligence, Dec. 2014, pp.36–47.
[5]
Yang H, Chua T S, Wang S G, Koh C K. Structured use of external knowledge for event-based open domain question answering. In Proc. the 26th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Jul. 2003, pp.33–40. DOI: 10.1145/860435.860444.
[6]
Chen Y B, Xu L H, Liu K, Zeng D J, Zhao J. Event extraction via dynamic multi-pooling convolutional neural networks. In Proc. the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Jul. 2015, pp.167–176. DOI: 10.3115/v1/P15-1017.
[7]
Nguyen T H, Grishman R. Event Detection and domain adaptation with convolutional neural networks. In Proc. the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Jul. 2015, pp.365–371. DOI: 10.3115/v1/P15-2060.
[8]
Nguyen T H, Grishman R. Modeling skip-grams for event detection with convolutional neural networks. In Proc. the 2016 Conference on Empirical Methods in Natural Language Processing, Nov. 2016, pp.886–891. DOI: 10.18653/v1/D16-1085.
[9]
Liu S L, Chen Y B, Liu K, Zhao J. Exploiting argument information to improve event detection via supervised attention mechanisms. In Proc. the 55th Annual Meeting of the Association for Computational Linguistics, Jul. 2017, pp.1789–1798. DOI: 10.18653/v1/P17-1164.
[10]
Nguyen T, Grishman R. Graph convolutional networks with argument-aware pooling for event detection. In Proc. the 32nd AAAI Conference on Artificial Intelligence, the 30th Innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence, Feb. 2018, pp.5900–5907. DOI: 10.1609/aaai.v32i1.12039.
[11]
Liu X, Luo Z C, Huang H Y. Jointly multiple events extraction via attention-based graph information aggregation. In Proc. the 2018 Conference on Empirical Methods in Natural Language Processing, Oct. 2018, pp.1247–1256. DOI: 10.18653/v1/D18-1156.
[12]
Chen Y B, Yang H, Liu K, Zhao J, Jia Y T. Collective event detection via a hierarchical and bias tagging networks with gated multi-level attention mechanisms. In Proc. the 2018 Conference on Empirical Methods in Natural Language Processing, Oct. 2018, pp.1267–1276. DOI: 10.18653/v1/D18-1158.
[13]
Yan H R, Jin X L, Meng X B, Guo J F, Cheng X Q. Event detection with multi-order graph convolution and aggregated attention. In Proc. the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Nov. 2019, pp.5766–5770. DOI: 10.18653/v1/D19-1582
[14]
Cui S Y, Yu B W, Liu T W, Zhang Z Y, Wang X B, Shi J Q. Edge-enhanced graph convolution networks for event Detection with syntactic relation. In Proc. the 2020 Findings of the Association for Computational Linguistics: EMNLP 2020, Nov. 2020, pp.2329–2339. DOI: 10.18653/v1/2020.findings-emnlp.211.
[15]
Lin H Y, Lu Y J, Han X P, Sun L. Nugget proposal networks for Chinese event detection. In Proc. the 56th Annual Meeting of the Association for Computational Linguistics, Jul. 2018, pp.1565–1574. DOI: 10.18653/v1/P18-1145.
[16]
Ding N, Li Z R, Liu Z Y, Zheng H T, Lin Z B. Event detection with trigger-aware lattice neural network. In Proc. the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Nov. 2019, pp.347–356. DOI: 10.18653/v1/D19-1033.
[17]
Sui D, Chen Y B, Liu K, Zhao J, Liu S P. Leverage lexical knowledge for Chinese named entity recognition via collaborative graph network. In Proc. the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Nov. 2019, pp.3830–3840. DOI: 10.18653/v1/D19-1396.
[18]
Xue M G, Yu B W, Liu T W, Zhang Y, Meng E L, Wang B. Porous lattice transformer encoder for Chinese NER. In Proc. the 28th International Conference on Computational Linguistics, Dec. 2020, pp.3831–3841. DOI: 10.18653/v1/2020.coling-main.340.
[19]
Xi X Y, Zhang T, Ye W, Zhang J L, Xie R, Zhang S K. A hybrid character representation for Chinese event detection. In Proc. the 2019 International Joint Conference on Neural Networks (IJCNN), Jul. 2019. DOI: 10.1109/IJCNN.2019.8851786.
[20]
Ahn D. The stages of event extraction. In Proc. the 2006 Workshop on Annotating and Reasoning about Time and Events, Jul. 2006.
[21]
Ji H, Grishman R. Refining event extraction through cross-document inference. In Proc. the 46th Annual Meeting of the Association for Computational Linguistics, Jun. 2008, pp.254–262.
[22]
Patwardhan S, Riloff E. A unified model of phrasal and sentential evidence for information extraction. In Proc. the 2009 Conference on Empirical Methods in Natural Language Processing, Aug. 2009, pp.151–160.
[23]
Hong Y, Zhang J F, Ma B, Yao J M, Zhou G D, Zhu Q M. Using cross-entity inference to improve event extraction. In Proc. the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Jun. 2011, pp.1127–1136.
[24]
McClosky D, Surdeanu M, Manning C. Event extraction as dependency parsing. In Proc. the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, Jun. 2011, pp.1626–1635.
[25]
Huang R H, Riloff E. Modeling textual cohesion for event extraction. In Proc. the 26th AAAI Conference on Artificial Intelligence, Jul. 2012, pp.1664–1670.
[26]
Li P F, Zhu Q M, Zhou G D. Argument inference from relevant event mentions in Chinese argument extraction. In Proc. the 51st Annual Meeting of the Association for Computational Linguistics, Aug. 2013, pp.1477–1487.
[27]
Li Q, Ji H, Huang L. Joint event extraction via structured prediction with global features. In Proc. the 51st Annual Meeting of the Association for Computational Linguistics, Aug. 2013, pp.73–82.
[28]
Chen Z, Ji H. Language specific issue and feature exploration in Chinese event extraction. In Proc. the 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Companion Volume: Short Papers, Jun. 2009, pp.209–212.
[29]

Qin B, Zhao Y Y, Ding X, Liu T, Zhai G F. Event type recognition based on trigger expansion. Tsinghua Science and Technology , 2010, 15(3): 251–258. DOI: 10.1016/S1007- 0214(10)70058-4.

[30]
Li P F, Zhou G D, Zhu Q M, Hou L B. Employing compositional semantics and discourse consistency in Chinese event extraction. In Proc. the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Jul. 2012, pp.1006–1016.
[31]
Li P F, Zhou G D. Employing morphological structures and sememes for Chinese event extraction. In Proc. the 24th International Conference on Computational Linguistics, Dec. 2012, pp.1619–1634.
[32]
Kipf T N, Welling M. Semi-supervised classification with graph convolutional networks. In Proc. the 5th International Conference on Learning Representations (ICLR), Apr. 2017.
[33]
Zhang C X, Song D J, Huang C, Swami A, Chawla N V. Heterogeneous graph neural network. In Proc. the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Jul. 2019, pp.793–803. DOI: 10.1145/3292500.3330961.
[34]
Wang X, Ji H Y, Shi C, Wang B, Ye Y F, Cui P, Yu P S. Heterogeneous graph attention network. In Proc. the 2019 Web Conference, May 2019, pp.2022–2032. DOI: 10.1145/3308558.3313562.
[35]
Hu Z N, Dong Y X, Wang K S, Sun Y Z. Heterogeneous graph transformer. In Proc. the 2020 Web Conference, Apr. 2020, pp.2704–2710. DOI: 10.1145/3366423.3380027.
[36]
Tu M, Wang G T, Huang J, Tang Y, He X D, Zhou B W. Multi-hop reading comprehension across multiple documents by reasoning over heterogeneous graphs. In Proc. the 57th Annual Meeting of the Association for Computational Linguistics, Jul. 2019, pp.2704–2713. DOI: 10.18653/v1/P19-1260.
[37]
Hu L M, Yang T C, Shi C, Ji H Y, Li X L. Heterogeneous graph attention networks for semi-supervised short text classification. In Proc. the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Nov. 2019, pp.4821–4830. DOI: 10.18653/v1/D19-1488.
[38]
Wang D Q, Liu P F, Zheng Y N, Qiu X P, Huang X J. Heterogeneous graph neural networks for extractive document summarization. In Proc. the 58th Annual Meeting of the Association for Computational Linguistics, Jul. 2020, pp.6209–6219. DOI: 10.18653/v1/2020.acl-main.553.
[39]
Jia R P, Cao Y N, Tang H Z, Fang F, Cao C, Wang S. Neural extractive summarization with hierarchical attentive heterogeneous graph network. In Proc. the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Nov. 2020, pp.3622–3631. DOI: 10.18653/v1/2020.emnlp-main.295.
[40]
Wang G Y, Li C Y, Wang W L, Zhang Y Z, Shen D H, Zhang X Y, Henao R, Carin L. Joint embedding of words and labels for text classification. In Proc. the 56th Annual Meeting of the Association for Computational Linguistics, Jul. 2018, pp.2321–2331. DOI: 10.18653/v1/P18-1216.
[41]
Zhang H L, Xiao L Q, Chen W Q, Wang Y K, Jin Y H. Multi-task label embedding for text classification. In Proc. the 2018 Conference on Empirical Methods in Natural Language Processing, Oct. 31–Nov. 4, 2018, pp.4545–4553. DOI: 10.18653/v1/D18-1484.
[42]
Du C X, Chen Z Z, Feng F L, Zhu L, Gan T, Nie L Q. Explicit interaction model towards text classification. In Proc. the 33rd AAAI Conference on Artificial Intelligence, the 31st Innovative Applications of Artificial Intelligence Conference, and the 9th AAAI Symposium on Educational Advances in Artificial Intelligence, Jul. 2019, pp.6359–6366. DOI: 10.1609/aaai.v33i01.33016359.
[43]
Huang L F, Ji H, Cho K, Dagan I, Riedel S, Voss C. Zero-shot transfer learning for event extraction. In Proc. the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Jul. 2018, pp.2160–2170. DOI: 10.18653/v1/p18-1201.
[44]
Lai V D, Nguyen T. Extending event detection to new types with learning from keywords. In Proc. the 5th Workshop on Noisy User-generated Text (W-NUT 2019), Nov. 2019, pp.243–248. DOI: 10.18653/v1/d19-5532.
[45]
Du X Y, Cardie C. Event extraction by answering (almost) natural questions. In Proc. the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Nov. 2020, pp.671–683. DOI: 10.18653/v1/2020.emnlp-main.49.
[46]
Velickovic P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y. Graph attention networks. In Proc. the 6th International Conference on Learning Representations, Apr. 30–May 3, 2018.
[47]

Viterbi A. Error bounds for convolutional codes and an asymptotically optimum decoding algorithm. IEEE Trans . Information Theory , 1967, 13(2): 260–269. DOI: 10.1109/tit. 1967.1054010.

[48]
Walker C, Strassel S, Medero J, Maeda K. ACE 2005 multilingual training corpus. Technical Report LDC2006T06, Linguistic Data Consortium, 2006. https://catalog.ldc.upenn.edu/LDC2006T06, Jan. 2024 .
[49]
Feng X C, Huang L F, Tang D Y, Ji H, Qin B, Liu T. A language-independent neural network for event detection. In Proc. the 54th Annual Meeting of the Association for Computational Linguistics, Aug. 2016, pp.66–71. DOI: 10.18653/v1/P16-2011.
[50]
Devlin J, Chang M W, Lee K, Toutanova K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proc. the 2019 Conference of the North American Chapter of the Association for Computational Linguistics, Jun. 2019, pp.4171–4186. DOI: 10.18653/v1/N19-1423.
[51]
Kingma D P, Ba J. Adam: A method for stochastic optimization. In Proc. the 3rd International Conference on Learning Representations, May 2015.
[52]
Chen C, Ng V. Joint modeling for Chinese event extraction with rich linguistic features. In Proc. the 24th International Conference on Computational Linguistics, Dec. 2012, pp.529–544.
[53]
Makarov P, Clematide S. UZH at TAC KBP 2017: Event nugget detection via joint learning with Softmax-margin objective. In Proc. the 2017 Text Analysis Conference, Nov. 2017.
[54]
Zeng Y, Yang H H, Feng Y S, Wang Z, Zhao D Y. A convolution BiLSTM neural network model for Chinese event extraction. In Proc. the 5th CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2016, and the 24th International Conference on Computer Processing of Oriental Languages, Natural Language Understanding and Intelligent Applications, Dec. 2016, pp.275–287. DOI: 10.1007/978-3-319-50496-4_23.
Journal of Computer Science and Technology
Pages 227-242
Cite this article:
Cui S-Y, Yu B-W, Cong X, et al. Label-Aware Chinese Event Detection with Heterogeneous Graph Attention Network. Journal of Computer Science and Technology, 2024, 39(1): 227-242. https://doi.org/10.1007/s11390-023-1541-6

176

Views

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Altmetrics

Received: 23 April 2021
Accepted: 02 January 2023
Published: 25 January 2024
© Institute of Computing Technology, Chinese Academy of Sciences 2024
Return