A Reinforcement Learning Approach to Robust Scheduling of Semiconductor Manufacturing Facilities

As semiconductor manufacturers, recently, have focused on producing multichip products (MCPs), scheduling semiconductor manufacturing operations become complicated due to the constraints related to reentrant production flows, sequence-dependent setups, and alternative machines. At the same time, the scheduling problems need to be solved frequently to effectively manage the variabilities in production requirements, available machines, and initial setup status. To minimize the makespan for an MCP scheduling problem, we propose a setup change scheduling method using reinforcement learning (RL) in which each agent determines setup decisions in a decentralized manner and learns a centralized policy by sharing a neural network among the agents to deal with the changes in the number of machines. Furthermore, novel definitions of state, action, and reward are proposed to address the variabilities in production requirements and initial setup status. Numerical experiments demonstrate that the proposed approach outperforms the rule-based, metaheuristic, and other RL methods in terms of the makespan while incurring shorter computation time than the metaheuristics considered. Note to Practitioners—This article studies a scheduling problem for die attach and wire bonding stages of a semiconductor packaging line. Due to the variabilities in production requirements, the number of available machines, and initial setup status, it is challenging for a scheduler to produce high-quality schedules within a specific time limit using existing approaches. In this article, a new scheduling method using reinforcement learning is proposed to enhance the robustness against the variabilities while achieving performance improvements. To verify the robustness of the proposed method, neural networks (NNs) trained on small-scale scheduling problems are used to solve large-scale scheduling problems. Experimental results show that the proposed method outperforms the existing approaches while requiring a short computation time. Furthermore, the trained NN performs well in solving unseen real-world scale problems even under stochastic processing time, suggesting the viability of the proposed method for real-world semiconductor packaging lines.

[1]  Jonghun Park,et al.  Fast Scheduling of Semiconductor Manufacturing Facilities Using Case-Based Reasoning , 2016, IEEE Transactions on Semiconductor Manufacturing.

[2]  Yi-Chi Wang,et al.  Application of reinforcement learning for agent-based production scheduling , 2005, Eng. Appl. Artif. Intell..

[3]  Martin A. Riedmiller,et al.  Distributed policy search reinforcement learning for job-shop scheduling tasks , 2012 .

[4]  Chao-Ton Su,et al.  Real-time scheduling for a smart factory using a reinforcement learning approach , 2018, Comput. Ind. Eng..

[5]  Jonghun Park,et al.  Setup Change Scheduling for Semiconductor Packaging Facilities Using a Genetic Algorithm With an Operator Recommender , 2014, IEEE Transactions on Semiconductor Manufacturing.

[6]  Frederick R. Forst,et al.  On robust estimation of the location parameter , 1980 .

[7]  Zhibin Jiang,et al.  Simulation-based optimization of dispatching rules for semiconductor wafer fabrication system scheduling by the response surface methodology , 2009 .

[8]  Weiping Wang,et al.  Minimizing mean weighted tardiness in unrelated parallel machine scheduling with reinforcement learning , 2012, Comput. Oper. Res..

[9]  Yoshua Bengio,et al.  Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..

[10]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[11]  Philip Bachman,et al.  Deep Reinforcement Learning that Matters , 2017, AAAI.

[12]  Zhe Xu,et al.  Efficient Large-Scale Fleet Management via Multi-Agent Deep Reinforcement Learning , 2018, KDD.

[13]  R. Haupt,et al.  A survey of priority rule-based scheduling , 1989 .

[14]  Jin Young Choi,et al.  A GSPN-Based Approach to Stacked Chips Scheduling Problem , 2010, IEEE Transactions on Semiconductor Manufacturing.

[15]  Weiping Wang,et al.  Flow shop Scheduling with Reinforcement Learning , 2013, Asia Pac. J. Oper. Res..

[16]  Yuehwern Yih,et al.  Selection of dispatching rules on multiple dispatching decision points in real-time scheduling of a semiconductor wafer fabrication system , 2003 .

[17]  Li Zheng,et al.  Dynamic parallel machine scheduling with mean weighted tardiness objective by Q-Learning , 2007 .

[18]  Jonghun Park,et al.  Learning to Dispatch Operations with Intentional Delay for Re-Entrant Multiple-Chip Product Assembly Lines , 2018, Sustainability.

[19]  Der-Jiunn Deng,et al.  Smart Manufacturing Scheduling With Edge Computing Using Multiclass Deep Q Network , 2019, IEEE Transactions on Industrial Informatics.

[20]  Chi-Bin Cheng,et al.  Efficient Due-Date Quoting and Production Scheduling for Integrated Circuit Packaging With Reentrant Processes , 2018, IEEE Transactions on Components, Packaging and Manufacturing Technology.

[21]  Lenz Belzner,et al.  Deep reinforcement learning for semiconductor production scheduling , 2018, 2018 29th Annual SEMI Advanced Semiconductor Manufacturing Conference (ASMC).

[22]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[23]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[24]  Shimon Whiteson,et al.  Learning to Communicate with Deep Multi-Agent Reinforcement Learning , 2016, NIPS.

[25]  Shimon Whiteson,et al.  Multi-Agent Common Knowledge Reinforcement Learning , 2018, NeurIPS.

[26]  Tie-Yan Liu,et al.  A Cooperative Multi-Agent Reinforcement Learning Framework for Resource Balancing in Complex Logistics Network , 2019, AAMAS.

[27]  L. Li,et al.  Adaptive Dispatching Rule for Semiconductor Wafer Fabrication Facility , 2013, IEEE Transactions on Automation Science and Engineering.

[28]  Stéphane Dauzère-Pérès,et al.  Solving the flexible job shop scheduling problem with sequence-dependent setup times , 2018, Eur. J. Oper. Res..

[29]  James T. Lin,et al.  Simulation optimization with GA and OCBA for semiconductor back-end assembly scheduling , 2015, 2015 International Conference on Industrial Engineering and Operations Management (IEOM).

[30]  Daniela Fischer,et al.  Digital Design And Computer Architecture , 2016 .

[31]  Mohammad Saidi-Mehrabad,et al.  Flexible job shop scheduling with tabu search algorithms , 2007 .

[32]  Li Li,et al.  Data-based scheduling framework and adaptive dispatching rule of complex manufacturing systems , 2013 .

[33]  Mingyuan Chen,et al.  A parallel genetic algorithm for a flexible job-shop scheduling problem with sequence dependent setups , 2010 .

[34]  Mykel J. Kochenderfer,et al.  Cooperative Multi-agent Control Using Deep Reinforcement Learning , 2017, AAMAS Workshops.

[35]  Rainer Kolisch,et al.  Semi-active, active, and non-delay schedules for the resource-constrained project scheduling problem , 1995 .

[36]  Sang-Jin Lee,et al.  Scheduling a multi-chip package assembly line with reentrant processes and unrelated parallel machines , 2008, 2008 Winter Simulation Conference.

[37]  A. S. Xanthopoulos,et al.  Intelligent controllers for bi-objective dynamic scheduling on a single machine with sequence-dependent setups , 2013, Appl. Soft Comput..

[38]  MengChu Zhou,et al.  Scheduling Semiconductor Testing Facility by Using Cuckoo Search Algorithm With Reinforcement Learning and Surrogate Modeling , 2019, IEEE Transactions on Automation Science and Engineering.

[39]  Bernd Scholz-Reiter,et al.  Generating dispatching rules for semiconductor manufacturing to minimize weighted tardiness , 2010, Proceedings of the 2010 Winter Simulation Conference.

[40]  V. Vinoda,et al.  Scheduling a dynamic job shop production system with sequence-dependent setups : An experimental study , 2007 .

[41]  Oscar Castillo,et al.  A state of the art review of intelligent scheduling , 2018, Artificial Intelligence Review.

[42]  Martin A. Riedmiller,et al.  ADAPTIVE REACTIVE JOB-SHOP SCHEDULING WITH REINFORCEMENT LEARNING AGENTS , 2008 .

[43]  Andrew L. Maas Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .

[44]  Jonathan F. Bard,et al.  A performance analysis of dispatch rules for semiconductor assembly & test operations , 2019, J. Simulation.

[45]  Fei Qiao,et al.  A Lot Dispatching Strategy Integrating WIP Management and Wafer Start Control , 2007, IEEE Transactions on Automation Science and Engineering.

[46]  Longbo Huang,et al.  A Deep Reinforcement Learning Framework for Rebalancing Dockless Bike Sharing Systems , 2018, AAAI.

[47]  Jonathan F. Bard,et al.  Hierarchy machine set-up for multi-pass lot scheduling at semiconductor assembly and test facilities , 2019 .