TY - GEN
T1 - How Do Transformers Integrate Meanings? An Investigation Using Interpretable Brain-Based Componential Semantics in Two-Word Phrases
AU - Wang, Shaonan
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025/4/20
Y1 - 2025/4/20
N2 - Transformer models have proven to be an efficient architecture for large language models, enabling them to exhibit human-like behaviors. However, it remains unclear how they combine the properties of individual words. To investigate this, we focus on the simplest compositional unit—two-word phrases—and employ four simple, parameter-free composition operations: addition, multiplication, first word, and second word. This approach serves as the first step toward understanding the composition mechanisms in transformer modules. We propose straightforward interpretation methods based on brain-based componential semantics. Specifically, we map the distributed vector space in transformers to an interpretable brain-based componential space to explore the intrinsic properties of representation and their semantic compositionality. Our findings show that, like humans, phrase types and semantic features influence the combination process in transformers. However, unlike humans, most phrase types paired with semantic features require more complex combination operations than simple addition or multiplication.
AB - Transformer models have proven to be an efficient architecture for large language models, enabling them to exhibit human-like behaviors. However, it remains unclear how they combine the properties of individual words. To investigate this, we focus on the simplest compositional unit—two-word phrases—and employ four simple, parameter-free composition operations: addition, multiplication, first word, and second word. This approach serves as the first step toward understanding the composition mechanisms in transformer modules. We propose straightforward interpretation methods based on brain-based componential semantics. Specifically, we map the distributed vector space in transformers to an interpretable brain-based componential space to explore the intrinsic properties of representation and their semantic compositionality. Our findings show that, like humans, phrase types and semantic features influence the combination process in transformers. However, unlike humans, most phrase types paired with semantic features require more complex combination operations than simple addition or multiplication.
KW - Brain-based componential semantics
KW - Composition operations
KW - Transformer
KW - Word representations
UR - https://www.scopus.com/pages/publications/105003622896
U2 - 10.1007/978-981-96-4001-0_29
DO - 10.1007/978-981-96-4001-0_29
M3 - Conference article published in proceeding or book
AN - SCOPUS:105003622896
SN - 9789819640003
T3 - Communications in Computer and Information Science
SP - 407
EP - 419
BT - Human Brain and Artificial Intelligence - 4th International Workshop, HBAI 2024, Proceedings
A2 - Liu, Quanying
A2 - Qu, Youzhi
A2 - Wu, Haiyan
A2 - Qi, Yu
A2 - Zeng, An
A2 - Pan, Dan
PB - Springer Science and Business Media Deutschland GmbH
T2 - 4th International Workshop on Human Brain and Artificial Intelligence, HBAI 2024
Y2 - 3 August 2024 through 3 August 2024
ER -