论文标题
深厚的强化学习指导了工作店计划的启发式启发式启发式
Deep Reinforcement Learning Guided Improvement Heuristic for Job Shop Scheduling
论文作者
论文摘要
最近使用深入增强学习(DRL)来解决工作店调度问题(JSSP)的最新研究重点是建筑启发式方法。但是,它们的性能远非最佳性,主要是因为基础图表方案不适合在每个施工步骤中对部分解决方案进行建模。本文提出了一种新颖的DRL引导改进启发式启发式启发式,用于求解JSSP,其中使用图表来编码完整的解决方案。我们设计了一个基于图形神经网络的表示方案,该方案由两个模块组成,可有效捕获在改进过程中遇到的图形中动态拓扑和不同类型的节点的信息。为了加快在改进过程中的解决方案评估,我们提出了一种新的消息,可以同时评估多个解决方案。我们证明,我们方法的计算复杂性随问题大小线性缩放。经典基准测试的实验表明,我们方法所学到的改进策略优于最先进的DRL方法。
Recent studies in using deep reinforcement learning (DRL) to solve Job-shop scheduling problems (JSSP) focus on construction heuristics. However, their performance is still far from optimality, mainly because the underlying graph representation scheme is unsuitable for modelling partial solutions at each construction step. This paper proposes a novel DRL-guided improvement heuristic for solving JSSP, where graph representation is employed to encode complete solutions. We design a Graph Neural-Network-based representation scheme, consisting of two modules to effectively capture the information of dynamic topology and different types of nodes in graphs encountered during the improvement process. To speed up solution evaluation during improvement, we present a novel message-passing mechanism that can evaluate multiple solutions simultaneously. We prove that the computational complexity of our method scales linearly with problem size. Experiments on classic benchmarks show that the improvement policy learned by our method outperforms state-of-the-art DRL-based methods by a large margin.