TY - GEN
T1 - Large Language Model for Humanoid Cognition in Proactive Human-Robot Collaboration
AU - Li, Shufei
AU - Wang, Zuoxu
AU - Yan, Zhijie
AU - Gao, Yiping
AU - Jiang, Han
AU - Zheng, Pai
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024/8
Y1 - 2024/8
N2 - Proactive Human-Robot Collaboration (HRC), which aims to achieve mutual-cognitive, predictable, and self-organizing collaboration between multiple humans and robots, is crucial for today's human-centric smart manufacturing. To enable Proactive HRC, various methods have been explored, including deep neural networks for visual detection, scene graph for decision-making, and reinforcement learning for robot execution. However, these methods often require re-training with domain-specific datasets in different scenarios, lacking generalizability and transferability for diverse manufacturing activities. The advent of Large Language Model (LLM) technology offers a promising solution for comprehending diverse tasks, modelling human intentions, and planning robot operations using natural vision-language instructions. This ability closely resembles human intelligence, specifically humanoid cognition, which allows flexible knowledge acquisition of the surrounding environment and exerting physical influence on tasks. Therefore, this paper delves into the concept of humanoid cognition in Proactive HRC and evaluates relevant LLM methods from the perspectives of task explainability, human-centricity, and robot executability. Based on the testing results, the authors provide discussions and future prospects for successfully integrating LLM approaches into Proactive HRC in the manufacturing domain.
AB - Proactive Human-Robot Collaboration (HRC), which aims to achieve mutual-cognitive, predictable, and self-organizing collaboration between multiple humans and robots, is crucial for today's human-centric smart manufacturing. To enable Proactive HRC, various methods have been explored, including deep neural networks for visual detection, scene graph for decision-making, and reinforcement learning for robot execution. However, these methods often require re-training with domain-specific datasets in different scenarios, lacking generalizability and transferability for diverse manufacturing activities. The advent of Large Language Model (LLM) technology offers a promising solution for comprehending diverse tasks, modelling human intentions, and planning robot operations using natural vision-language instructions. This ability closely resembles human intelligence, specifically humanoid cognition, which allows flexible knowledge acquisition of the surrounding environment and exerting physical influence on tasks. Therefore, this paper delves into the concept of humanoid cognition in Proactive HRC and evaluates relevant LLM methods from the perspectives of task explainability, human-centricity, and robot executability. Based on the testing results, the authors provide discussions and future prospects for successfully integrating LLM approaches into Proactive HRC in the manufacturing domain.
UR - http://www.scopus.com/inward/record.url?scp=85208236368&partnerID=8YFLogxK
U2 - 10.1109/CASE59546.2024.10711379
DO - 10.1109/CASE59546.2024.10711379
M3 - Conference article published in proceeding or book
AN - SCOPUS:85208236368
SN - 9798350358520
T3 - IEEE International Conference on Automation Science and Engineering
SP - 540
EP - 545
BT - 2024 IEEE 20th International Conference on Automation Science and Engineering, CASE 2024
PB - IEEE Computer Society
T2 - 20th IEEE International Conference on Automation Science and Engineering, CASE 2024
Y2 - 28 August 2024 through 1 September 2024
ER -