TY - GEN
T1 - Investigating Aspect Features in Contextualized Embeddings with Semantic Scales and Distributional Similarity
AU - Li, Yuxi
AU - Chersoni, Emmanuele
AU - Hsu, Yu Yin
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024/6
Y1 - 2024/6
N2 - Aspect, a linguistic category describing how actions and events unfold over time, is traditionally characterized by three semantic properties: stativity, durativity and telicity. In this study, we investigate whether and to what extent these properties are encoded in the verb token embeddings of the contextualized spaces of two English language models – BERT and GPT-2. First, we propose an experiment using semantic projections to examine whether the values of the vector dimensions of annotated verbs for stativity, durativity and telicity reflect human linguistic distinctions. Second, we use distributional similarity to replicate the notorious Imperfective Paradox described by Dowty (1977), and assess whether the embedding models are sensitive to capture contextual nuances of the verb telicity. Our results show that both models encode the semantic distinctions for the aspect properties of stativity and telicity in most of their layers, while durativity is the most challenging feature. As for the Imperfective Paradox, only the embedding similarities computed with the vectors from the early layers of the BERT model align with the expected pattern.
AB - Aspect, a linguistic category describing how actions and events unfold over time, is traditionally characterized by three semantic properties: stativity, durativity and telicity. In this study, we investigate whether and to what extent these properties are encoded in the verb token embeddings of the contextualized spaces of two English language models – BERT and GPT-2. First, we propose an experiment using semantic projections to examine whether the values of the vector dimensions of annotated verbs for stativity, durativity and telicity reflect human linguistic distinctions. Second, we use distributional similarity to replicate the notorious Imperfective Paradox described by Dowty (1977), and assess whether the embedding models are sensitive to capture contextual nuances of the verb telicity. Our results show that both models encode the semantic distinctions for the aspect properties of stativity and telicity in most of their layers, while durativity is the most challenging feature. As for the Imperfective Paradox, only the embedding similarities computed with the vectors from the early layers of the BERT model align with the expected pattern.
UR - http://www.scopus.com/inward/record.url?scp=85207822733&partnerID=8YFLogxK
U2 - 10.18653/v1/2024.starsem-1.7
DO - 10.18653/v1/2024.starsem-1.7
M3 - Conference article published in proceeding or book
AN - SCOPUS:85207822733
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 80
EP - 92
BT - Proceedings of The 13th Joint Conference on Lexical and Computational Semantics (*SEM 2024)
A2 - Bollegala, Danushka
A2 - Shwartz, Vered
PB - Association for Computational Linguistics (ACL)
T2 - 13th Joint Conference on Lexical and Computational Semantics, StarSEM 2024
Y2 - 20 June 2024 through 21 June 2024
ER -