TY - GEN
T1 - Zero-shot Cross-lingual NER via Mitigating Language Difference: An Entity-aligned Translation Perspective
AU - Zhang, Zhihao
AU - Lee, Sophia Yat Mei
AU - Zhang, Dong
AU - Li, Shoushan
AU - Zhou, Guodong
N1 - Publisher Copyright:
©2025 Association for Computational Linguistics.
PY - 2025/11
Y1 - 2025/11
N2 - Cross-lingual Named Entity Recognition (CL-NER) aims to transfer knowledge from high-resource languages to low-resource languages. However, existing zero-shot CL-NER (ZCL-NER) approaches primarily focus on Latin script language (LSL), where shared linguistic features facilitate effective knowledge transfer. In contrast, for non-Latin script language (NSL), such as Chinese and Japanese, performance often degrades due to deep structural differences. To address these challenges, we propose an entity-aligned translation (EAT) approach †. Leveraging large language models (LLMs), EAT employs a dual-translation strategy to align entities between NSL and English. In addition, we fine-tune LLMs using multilingual Wikipedia data to enhance the entity alignment from source to target languages. Extensive experiments demonstrate that EAT outperforms prior methods on NSL by bridging language gaps through entity-aware translation.
AB - Cross-lingual Named Entity Recognition (CL-NER) aims to transfer knowledge from high-resource languages to low-resource languages. However, existing zero-shot CL-NER (ZCL-NER) approaches primarily focus on Latin script language (LSL), where shared linguistic features facilitate effective knowledge transfer. In contrast, for non-Latin script language (NSL), such as Chinese and Japanese, performance often degrades due to deep structural differences. To address these challenges, we propose an entity-aligned translation (EAT) approach †. Leveraging large language models (LLMs), EAT employs a dual-translation strategy to align entities between NSL and English. In addition, we fine-tune LLMs using multilingual Wikipedia data to enhance the entity alignment from source to target languages. Extensive experiments demonstrate that EAT outperforms prior methods on NSL by bridging language gaps through entity-aware translation.
UR - https://www.scopus.com/pages/publications/105028997601
U2 - 10.18653/v1/2025.findings-emnlp.244
DO - 10.18653/v1/2025.findings-emnlp.244
M3 - Conference article published in proceeding or book
AN - SCOPUS:105028997601
T3 - EMNLP 2025 - 2025 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2025
SP - 4541
EP - 4557
BT - EMNLP 2025 - 2025 Conference on Empirical Methods in Natural Language Processing, Findings of EMNLP 2025
A2 - Christodoulopoulos, Christos
A2 - Chakraborty, Tanmoy
A2 - Rose, Carolyn
A2 - Peng, Violet
PB - Association for Computational Linguistics (ACL)
T2 - 30th Conference on Empirical Methods in Natural Language Processing, EMNLP 2025
Y2 - 4 November 2025 through 9 November 2025
ER -