计算机工程与应用 ›› 2025, Vol. 61 ›› Issue (12): 344-351.DOI: 10.3778/j.issn.1002-8331.2407-0326

• 工程与应用 • 上一篇    下一篇

结合图神经网络的DDQN算法的动态车间调度问题研究

杨蓝,毕利,杨众   

  1. 宁夏大学 信息工程学院,银川 750021
  • 出版日期:2025-06-15 发布日期:2025-06-13

Research on Dynamic Job Shop Scheduling Problem Using DDQN Algorithm Combined with Graph Neural Network

YANG Lan, BI Li, YANG Zhong   

  1. College of Information Engineering, Ningxia University, Yinchuan 750021, China
  • Online:2025-06-15 Published:2025-06-13

摘要: 针对紧急插单事件的动态作业车间调度问题,以最小化所有工件的提前与延迟完工时间为目标,创建了动态作业车间环境模型。将调度问题转换为马尔可夫过程,并采用结合图神经网络的DDQN(double deep Q-network,DDQN)深度强化学习算法进行求解。通过图神经网络对车间状态析取图做特征提取从而避免状态定义依赖人工经验设计的问题,引入了注意力机制能够增强强化学习智能体对状态信息的获取能力,并将六组规则调度作为智能体的决策空间,定义了全新的奖励方法,加强其对智能体学习的指导能力。通过多组对照实验说明了所构建的求解模型的有效性和可行性。

关键词: 动态作业车间调度, 马尔可夫过程, 图神经网络, 深度强化学习, 注意力机制

Abstract: For the dynamic job shop scheduling problem with emergency order insertion event, this paper creates a dynamic job shop environment model with the objective of minimizing the early and late completion time of all workpieces. The scheduling problem is converted into a Markov process and solved by a DDQN (double deep Q-network) deep reinforcement learning algorithm combined with a graph neural network. Firstly, the graph neural network is used to do feature extraction on the state analysis map of the workshop so as to avoid the problem of relying on artificial experience to design the state definition, the attention mechanism is introduced to enhance the ability of the reinforcement learning intelligences to acquire the state information, and six sets of rule scheduling are used as the decision space of the intelligences, and secondly, a brand-new reward method is defined to strengthen the ability of guiding the learning of the intelligences. Finally, the effectiveness and feasibility of the solution model constructed in this paper are illustrated by multi-group controlled experiments.

Key words: dynamic job shop scheduling, Markov process, graph neural network, deep reinforcement learning, attention mechanism