TY - GEN
T1 - AoE
T2 - 62nd Annual Meeting of the Association for Computational Linguistics, ACL 2024
AU - Li, Xianming
AU - Li, Jing
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024/8
Y1 - 2024/8
N2 - Text embedding is pivotal in semantic textual similarity (STS) tasks, which are crucial components in Large Language Model (LLM) applications. STS learning largely relies on the cosine function as the optimization objective to reflect semantic similarity. However, the cosine has saturation zones rendering vanishing gradients and hindering learning subtle semantic differences in text embeddings. To address this issue, we propose a novel Angle-optimized Embedding model, AoE. It optimizes angle differences in complex space to explore similarity in saturation zones better. To set up a comprehensive evaluation, we experimented with existing short-text STS, our newly collected long-text STS, and downstream task datasets. Extensive experimental results on STS and MTEB benchmarks show that AoE significantly outperforms popular text embedding models neglecting cosine saturation zones. It highlights that AoE can produce high-quality text embeddings and broadly benefit downstream tasks. The code is available at: https://github.com/SeanLee97/AnglE.
AB - Text embedding is pivotal in semantic textual similarity (STS) tasks, which are crucial components in Large Language Model (LLM) applications. STS learning largely relies on the cosine function as the optimization objective to reflect semantic similarity. However, the cosine has saturation zones rendering vanishing gradients and hindering learning subtle semantic differences in text embeddings. To address this issue, we propose a novel Angle-optimized Embedding model, AoE. It optimizes angle differences in complex space to explore similarity in saturation zones better. To set up a comprehensive evaluation, we experimented with existing short-text STS, our newly collected long-text STS, and downstream task datasets. Extensive experimental results on STS and MTEB benchmarks show that AoE significantly outperforms popular text embedding models neglecting cosine saturation zones. It highlights that AoE can produce high-quality text embeddings and broadly benefit downstream tasks. The code is available at: https://github.com/SeanLee97/AnglE.
UR - https://www.scopus.com/pages/publications/85204492105
U2 - 10.18653/v1/2024.acl-long.101
DO - 10.18653/v1/2024.acl-long.101
M3 - Conference article published in proceeding or book
AN - SCOPUS:85204492105
VL - 1
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 1825
EP - 1839
BT - Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics
A2 - Ku, Lun-Wei
A2 - Martins, Andre F. T.
A2 - Srikumar, Vivek
PB - Association for Computational Linguistics (ACL)
Y2 - 11 August 2024 through 16 August 2024
ER -