TY - GEN
T1 - Large Language Model-Enhanced Algorithm Selection: Towards Comprehensive Algorithm Representation
AU - Wu, Xingyu
AU - Zhong, Yan
AU - Wu, Jibin
AU - Jiang, Bingbing
AU - Tan, Kay Chen
N1 - Publisher Copyright:
© 2024 International Joint Conferences on Artificial Intelligence. All rights reserved.
PY - 2024/8
Y1 - 2024/8
N2 - Algorithm selection, a critical process of automated machine learning, aims to identify the most suitable algorithm for solving a specific problem pri- or to execution. Mainstream algorithm selection techniques heavily rely on problem features, while the role of algorithm features remains largely unexplored. Due to the intrinsic complexity of algorithms, effective methods for universally extracting algorithm information are lacking. This paper takes a significant step towards bridging this gap by introducing Large Language Models (LLMs) into algorithm selection for the first time. By comprehending the code text, LLM not only captures the structural and semantic aspects of the algorithm, but also demonstrates contextual awareness and library function understanding. The high-dimensional algorithm representation extracted by LLM, after undergoing a feature selection module, is combined with the problem representation and passed to the similarity calculation module. The selected algorithm is determined by the matching degree between a given problem and different algorithms. Extensive experiments validate the performance superiority of the proposed model and the efficacy of each key module. Furthermore, we present a theoretical upper bound on model complexity, showcasing the influence of algorithm representation and feature selection modules. This provides valuable theoretical guidance for the practical implementation of our method.
AB - Algorithm selection, a critical process of automated machine learning, aims to identify the most suitable algorithm for solving a specific problem pri- or to execution. Mainstream algorithm selection techniques heavily rely on problem features, while the role of algorithm features remains largely unexplored. Due to the intrinsic complexity of algorithms, effective methods for universally extracting algorithm information are lacking. This paper takes a significant step towards bridging this gap by introducing Large Language Models (LLMs) into algorithm selection for the first time. By comprehending the code text, LLM not only captures the structural and semantic aspects of the algorithm, but also demonstrates contextual awareness and library function understanding. The high-dimensional algorithm representation extracted by LLM, after undergoing a feature selection module, is combined with the problem representation and passed to the similarity calculation module. The selected algorithm is determined by the matching degree between a given problem and different algorithms. Extensive experiments validate the performance superiority of the proposed model and the efficacy of each key module. Furthermore, we present a theoretical upper bound on model complexity, showcasing the influence of algorithm representation and feature selection modules. This provides valuable theoretical guidance for the practical implementation of our method.
UR - http://www.scopus.com/inward/record.url?scp=85198098829&partnerID=8YFLogxK
U2 - 10.24963/ijcai.2024/579
DO - 10.24963/ijcai.2024/579
M3 - Conference article published in proceeding or book
T3 - IJCAI International Joint Conference on Artificial Intelligence
SP - 5235
EP - 5244
BT - Proceedings of the 33rd International Joint Conference on Artificial Intelligence, IJCAI 2024
A2 - Larson, Kate
ER -