Large Language Model-Assisted Deep Reinforcement Learning from Human Feedback for Job Shop Scheduling

محفوظ في:
التفاصيل البيبلوغرافية
الحاوية / القاعدة:Machines vol. 13, no. 5 (2025), p. 361
المؤلف الرئيسي: Zeng Yuhang
مؤلفون آخرون: Lou, Ping, Hu, Jianmin, Fan Chuannian, Liu, Quan, Hu, Jiwei
منشور في:
MDPI AG
الموضوعات:
الوصول للمادة أونلاين:Citation/Abstract
Full Text + Graphics
Full Text - PDF
الوسوم: إضافة وسم
لا توجد وسوم, كن أول من يضع وسما على هذه التسجيلة!
الوصف
مستخلص:The job shop scheduling problem (JSSP) is a classical NP-hard combinatorial optimization challenge that plays a crucial role in manufacturing systems. Deep reinforcement learning has shown great potential in solving this problem. However, it still has challenges in reward function design and state feature representation, which makes it suffer from slow policy convergence and low learning efficiency in complex production environments. Therefore, a human feedback-based large language model-assisted deep reinforcement learning (HFLLMDRL) framework is proposed to solve this problem, in which few-shot prompt engineering by human feedback is utilized to assist in designing instructive reward functions and guiding policy convergence. Additionally, a self-adaptation symbolic visualization Kolmogorov–Arnold Network (KAN) is integrated as the policy network in DRL to enhance state feature representation, thereby improving learning efficiency. Experimental results demonstrate that the proposed framework significantly boosts both learning performance and policy convergence, presenting a novel approach to the JSSP.
تدمد:2075-1702
DOI:10.3390/machines13050361
المصدر:Engineering Database