Large Language Model-Assisted Deep Reinforcement Learning from Human Feedback for Job Shop Scheduling

Uloženo v:
Podrobná bibliografie
Vydáno v:Machines vol. 13, no. 5 (2025), p. 361
Hlavní autor: Zeng Yuhang
Další autoři: Lou, Ping, Hu, Jianmin, Fan Chuannian, Liu, Quan, Hu, Jiwei
Vydáno:
MDPI AG
Témata:
On-line přístup:Citation/Abstract
Full Text + Graphics
Full Text - PDF
Tagy: Přidat tag
Žádné tagy, Buďte první, kdo vytvoří štítek k tomuto záznamu!
Popis
Abstrakt:The job shop scheduling problem (JSSP) is a classical NP-hard combinatorial optimization challenge that plays a crucial role in manufacturing systems. Deep reinforcement learning has shown great potential in solving this problem. However, it still has challenges in reward function design and state feature representation, which makes it suffer from slow policy convergence and low learning efficiency in complex production environments. Therefore, a human feedback-based large language model-assisted deep reinforcement learning (HFLLMDRL) framework is proposed to solve this problem, in which few-shot prompt engineering by human feedback is utilized to assist in designing instructive reward functions and guiding policy convergence. Additionally, a self-adaptation symbolic visualization Kolmogorov–Arnold Network (KAN) is integrated as the policy network in DRL to enhance state feature representation, thereby improving learning efficiency. Experimental results demonstrate that the proposed framework significantly boosts both learning performance and policy convergence, presenting a novel approach to the JSSP.
ISSN:2075-1702
DOI:10.3390/machines13050361
Zdroj:Engineering Database