TY - GEN
T1 - Employing Glyphic Information for Chinese Event Extraction with Vision-Language Model
AU - Bao, Xiaoyi
AU - Gu, Jinghang
AU - Wang, Zhongqing
AU - Qiang, Mingjie
AU - Huang, Chu-Ren
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024/11
Y1 - 2024/11
N2 - As a complex task that requires rich information input, features from various aspects have been utilized in event extraction. However, most of the previous works ignored the value of glyph, which could contain enriched semantic information and can not be fully expressed by the pre-trained embedding in hieroglyphic languages like Chinese. We argue that, compared with combining the sophisticated textual features, glyphic information from visual modality could provide us with extra and straight semantic information in extracting events. Motivated by this, we propose a glyphic multi-modal Chinese event extraction model with hieroglyphic images to capture the intra- and inter-character morphological structure from the sequence. Extensive experiments build a new state-of-the-art performance in the ACE2005 Chinese and KBP Eval 2017 dataset, which underscores the effectiveness of our proposed glyphic event extraction model, and more importantly, the glyphic feature can be obtained at nearly zero cost. Code and data can be found at https://github.com/HoraceXIaoyiBao/GlyphicVLM-for-ChineseEE.
AB - As a complex task that requires rich information input, features from various aspects have been utilized in event extraction. However, most of the previous works ignored the value of glyph, which could contain enriched semantic information and can not be fully expressed by the pre-trained embedding in hieroglyphic languages like Chinese. We argue that, compared with combining the sophisticated textual features, glyphic information from visual modality could provide us with extra and straight semantic information in extracting events. Motivated by this, we propose a glyphic multi-modal Chinese event extraction model with hieroglyphic images to capture the intra- and inter-character morphological structure from the sequence. Extensive experiments build a new state-of-the-art performance in the ACE2005 Chinese and KBP Eval 2017 dataset, which underscores the effectiveness of our proposed glyphic event extraction model, and more importantly, the glyphic feature can be obtained at nearly zero cost. Code and data can be found at https://github.com/HoraceXIaoyiBao/GlyphicVLM-for-ChineseEE.
UR - http://www.scopus.com/inward/record.url?scp=85217618818&partnerID=8YFLogxK
U2 - 10.18653/v1/2024.findings-emnlp.58
DO - 10.18653/v1/2024.findings-emnlp.58
M3 - Conference article published in proceeding or book
AN - SCOPUS:85217618818
T3 - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024
SP - 1068
EP - 1080
BT - EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2024
A2 - Al-Onaizan, Yaser
A2 - Bansal, Mohit
A2 - Chen, Yun-Nung
PB - Association for Computational Linguistics (ACL)
T2 - 2024 Conference on Empirical Methods in Natural Language Processing, EMNLP 2024
Y2 - 12 November 2024 through 16 November 2024
ER -