TY - GEN
T1 - STRec: Sparse Transformer for Sequential Recommendations
AU - Li, Chengxi
AU - Wang, Yejing
AU - Liu, Qidong
AU - Zhao, Xiangyu
AU - Wang, Wanyu
AU - Wang, Yiqi
AU - Zou, Lixin
AU - Fan, Wenqi
AU - Li, Qing
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/9/14
Y1 - 2023/9/14
N2 - With the rapid evolution of transformer architectures, researchers are exploring their application in sequential recommender systems (SRSs) and presenting promising performance on SRS tasks compared with former SRS models. However, most existing transformer-based SRS frameworks retain the vanilla attention mechanism, which calculates the attention scores between all item-item pairs. With this setting, redundant item interactions can harm the model performance and consume much computation time and memory. In this paper, we identify the sparse attention phenomenon in transformer-based SRS models and propose Sparse Transformer for sequential Recommendation tasks (STRec) to achieve the efficient computation and improved performance. Specifically, we replace self-attention with cross-attention, making the model concentrate on the most relevant item interactions. To determine these necessary interactions, we design a novel sampling strategy to detect relevant items based on temporal information. Extensive experimental results validate the effectiveness of STRec, which achieves the state-of-the-art accuracy while reducing 54% inference time and 70% memory cost. We also provide massive extended experiments to further investigate the property of our framework.
AB - With the rapid evolution of transformer architectures, researchers are exploring their application in sequential recommender systems (SRSs) and presenting promising performance on SRS tasks compared with former SRS models. However, most existing transformer-based SRS frameworks retain the vanilla attention mechanism, which calculates the attention scores between all item-item pairs. With this setting, redundant item interactions can harm the model performance and consume much computation time and memory. In this paper, we identify the sparse attention phenomenon in transformer-based SRS models and propose Sparse Transformer for sequential Recommendation tasks (STRec) to achieve the efficient computation and improved performance. Specifically, we replace self-attention with cross-attention, making the model concentrate on the most relevant item interactions. To determine these necessary interactions, we design a novel sampling strategy to detect relevant items based on temporal information. Extensive experimental results validate the effectiveness of STRec, which achieves the state-of-the-art accuracy while reducing 54% inference time and 70% memory cost. We also provide massive extended experiments to further investigate the property of our framework.
KW - efficient transformer
KW - recommendation system
KW - sequential recommendation
UR - http://www.scopus.com/inward/record.url?scp=85174493165&partnerID=8YFLogxK
U2 - 10.1145/3604915.3608779
DO - 10.1145/3604915.3608779
M3 - Conference article published in proceeding or book
AN - SCOPUS:85174493165
T3 - Proceedings of the 17th ACM Conference on Recommender Systems, RecSys 2023
SP - 101
EP - 111
BT - Proceedings of the 17th ACM Conference on Recommender Systems, RecSys 2023
PB - Association for Computing Machinery, Inc
T2 - 17th ACM Conference on Recommender Systems, RecSys 2023
Y2 - 18 September 2023 through 22 September 2023
ER -