Dynamic assembly sequence selection using reinforcement learning

Determining the most appropriate sequence for assembling products requires assessment of the process, product, and the technology applied. Most production engineers apply constraint based evaluation and history to identify the solution sequence. What if their solution is sub-optimal? In this paper a self-learning technique for selecting a sequence and dynamically changing the sequence is presented, selection is based on the history of assemblies. The evaluation is dependent on part properties rather than parts and their relationships, thus no previous knowledge of parts and their interaction is required in the decision making process. The method assumes assembly is without constraint, for example, a highly flexible robotic assembly cell. This maximises the ability of the algorithm to select sequences for new products and optimise them. The heart of the algorithm is a reinforcement-learning model, which punishes failed assembly steps, this facilitates feedback sequence selection, when: current methods are merely feedforward. This feedback approach addresses combinatorial explosion that can cripple assembly planners.

[1]  Jan Wolter,et al.  A structure-oriented approach to assembly sequence planning , 1997, IEEE Trans. Robotics Autom..

[2]  Sukhan Lee,et al.  Assembly planning based on geometric reasoning , 1990, Comput. Graph..

[3]  Daniel E Whitney,et al.  DEVELOPMENT AND MANUFACTURING , 2000 .

[4]  Arthur C. Sanderson,et al.  A correct and complete algorithm for the generation of mechanical assembly sequences , 1989, Proceedings, 1989 International Conference on Robotics and Automation.

[5]  Maria Joao Abrantes,et al.  Expert-system Generation of Assembly Sequences , 1993 .

[6]  Bruce Romney,et al.  On the concurrent design of assembly sequences and fixture , 1997 .

[7]  A. H. Redford,et al.  Design for Assembly , 1983, Methods and Tools for Computer Integrated Manufacturing.