TY - JOUR
T1 - Synergetic-informed deep reinforcement learning for sustainable management of transportation networks with large action spaces
AU - Lai, Li
AU - Dong, You
AU - Andriotis, Charalampos P.
AU - Wang, Aijun
AU - Lei, Xiaoming
N1 - Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2024/4
Y1 - 2024/4
N2 - Effective transportation network management systems should consider safety and sustainability objectives. Existing research on large-scale transportation network management often employs the assumption that bridges can be considered individually under these objectives. However, this simplification misses accurate system-level representations, induced by multiple components, network topology, and global maintenance actions. To address these limitations, this paper presents a deep reinforcement learning (DRL) framework that draws inspiration from biological learning behaviors to determine optimal life-cycle management policies. It incorporates synergetic branches and hierarchical rewards, factorizing the action space and, thereby, diminishing system complexity from exponential to linear with respect to the number of bridges. Extensive experiments based on a realistic case study demonstrate that the proposed method outperforms expert maintenance strategies and state-of-the-art decision-making methods. Overall, the proposed DRL framework can assist engineers by offering adaptive solutions to maintenance planning. It also provides solutions that address large action spaces within complex systems.
AB - Effective transportation network management systems should consider safety and sustainability objectives. Existing research on large-scale transportation network management often employs the assumption that bridges can be considered individually under these objectives. However, this simplification misses accurate system-level representations, induced by multiple components, network topology, and global maintenance actions. To address these limitations, this paper presents a deep reinforcement learning (DRL) framework that draws inspiration from biological learning behaviors to determine optimal life-cycle management policies. It incorporates synergetic branches and hierarchical rewards, factorizing the action space and, thereby, diminishing system complexity from exponential to linear with respect to the number of bridges. Extensive experiments based on a realistic case study demonstrate that the proposed method outperforms expert maintenance strategies and state-of-the-art decision-making methods. Overall, the proposed DRL framework can assist engineers by offering adaptive solutions to maintenance planning. It also provides solutions that address large action spaces within complex systems.
UR - http://www.scopus.com/inward/record.url?scp=85183948524&partnerID=8YFLogxK
U2 - 10.1016/j.autcon.2024.105302
DO - 10.1016/j.autcon.2024.105302
M3 - Journal article
AN - SCOPUS:85183948524
SN - 0926-5805
VL - 160
JO - Automation in Construction
JF - Automation in Construction
M1 - 105302
ER -