TY - GEN
T1 - KANFeel
T2 - 2025 CHI Conference on Human Factors in Computing Systems, CHI EA 2025
AU - Fang, Le
AU - Chai, Bo
AU - Xu, Yingqing
AU - Wang, Stephen Jia
N1 - Publisher Copyright:
© 2025 Copyright held by the owner/author(s).
PY - 2025/4/26
Y1 - 2025/4/26
N2 - Multimodal emotion recognition (MER) has become a significant interdisciplinary research area in HCI by analyzing various data modalities. Recent studies have shown promising results in MER by fusing features across modalities. However, efficiency in real-world scenarios remains a challenge. To improve the model efficiency, Kolmogorov-Arnold Networks (KANs) have been proposed as efficient alternatives to Multi-Layer Perceptrons (MLPs). However, employing KANs in MER is still underexplored. This paper introduces KANFeel, a novel KAN-based MER framework that processes multimodal inputs to predict emotional states. Furthermore, we adopt the KAN-based model to replace the attention mechanism in the transformer, referred to as KANFeel-Attent, to achieve enhanced recognition performance. Comprehensive evaluations across three public datasets are conducted to analyze the efficiency improvements of KANFeel by comparing model parameters and training speeds. Finally, emotion recognition enhancements, measured through accuracy and F1-score, are validated using the KANFeel and KANFeel-Attent models against baseline and state-of-the-art methods.
AB - Multimodal emotion recognition (MER) has become a significant interdisciplinary research area in HCI by analyzing various data modalities. Recent studies have shown promising results in MER by fusing features across modalities. However, efficiency in real-world scenarios remains a challenge. To improve the model efficiency, Kolmogorov-Arnold Networks (KANs) have been proposed as efficient alternatives to Multi-Layer Perceptrons (MLPs). However, employing KANs in MER is still underexplored. This paper introduces KANFeel, a novel KAN-based MER framework that processes multimodal inputs to predict emotional states. Furthermore, we adopt the KAN-based model to replace the attention mechanism in the transformer, referred to as KANFeel-Attent, to achieve enhanced recognition performance. Comprehensive evaluations across three public datasets are conducted to analyze the efficiency improvements of KANFeel by comparing model parameters and training speeds. Finally, emotion recognition enhancements, measured through accuracy and F1-score, are validated using the KANFeel and KANFeel-Attent models against baseline and state-of-the-art methods.
KW - Deep Learning
KW - Human-Computer Interaction
KW - Kolmogorov-Arnold Networks
KW - Multimodal Emotion Recognition
UR - https://www.scopus.com/pages/publications/105005758049
U2 - 10.1145/3706599.3720217
DO - 10.1145/3706599.3720217
M3 - Conference article published in proceeding or book
AN - SCOPUS:105005758049
T3 - Conference on Human Factors in Computing Systems - Proceedings
BT - CHI EA 2025 - Extended Abstracts of the 2025 CHI Conference on Human Factors in Computing Systems
PB - Association for Computing Machinery
Y2 - 26 April 2025 through 1 May 2025
ER -