Case study: Performance analysis and development of robotized screwing application with integrated vision sensing system for automotive industry

Industrial cameras starting to play a significant role in current industrial environment and they represent a strong tool for robotics mainly in cases when they are combined with high-speed robots. However, there are still some difficulties in vision system integration. The capability of such system (e.g. assembly or technological system) depends on several factors, for instance, the camera position, lightning conditions, pattern recognition algorithms, precise setup, as well as well-trained programmer and engineer. The skills of the engineering staff and the precise analysis of conditions and process requirements seem to be crucial for successful solution, what was proved also in our experimental test. The main aim of the article is development and complex performance analysis of robotized screwing application with integrated vision system, concretely the case study of automated assembly system in automotive industry—bolting tightening robotized station as a part of car seat assembly process. The main key elements of the designed workplace are industrial robot FANUC M-20iA/20M with integrated iRVision system containing the industrial camera Sony XC-56. The real influence of inaccuracies during the design process, gradual “step-by-step” refinement, and final influence of all changings on the final quality and efficiency of designed system was demonstrated. At the end, we reached the point where the total number of screwing operations with NOK results (not OK / negative results) is about six single negative results per day, which represents less than 0.1% of all recognitions (overall reliability is higher than 99.9%).

[1]  Pero Raos,et al.  The essence and applications of machine vision , 2014 .

[2]  Wei Zhang,et al.  A Unified Framework for Street-View Panorama Stitching , 2016, Sensors.

[3]  Gordon Christie,et al.  Fast inspection for size-based analysis in aggregate processing , 2014, Machine Vision and Applications.

[4]  Jelena Kocić,et al.  Sensors and Sensor Fusion in Autonomous Vehicles , 2018, 2018 26th Telecommunications Forum (TELFOR).

[5]  Petar Ćurković,et al.  Context-Aware System Applied in Industrial Assembly Environment , 2012 .

[6]  Lei He,et al.  Fast depth extraction from a single image , 2016 .

[7]  Miguel Hernando,et al.  Home Camera-Based Fall Detection System for the Elderly , 2017, Sensors.

[8]  Karl Lenhardt Optical Systems in Machine Vision , 2007 .

[9]  Tatiana Kelemenová,et al.  Information Contents of a Signal at Repeated Positioning Measurements of the Coordinate Measuring Machine (CMM) by Laser Interferometer , 2016 .

[10]  Alexander Ferrein,et al.  Calibration of a Rotating or Revolving Platform with a LiDAR Sensor , 2019 .

[11]  Bandar Alghamdi,et al.  An Integrated Robotic System: 2D-Vision Based Inspection Robot with Automated PLC Conveyor System , 2017 .

[12]  Radovan Holubek,et al.  The Simulation of Conveyor Control System Using the Virtual Commissioning and Virtual Reality , 2018, Advances in Science and Technology Research Journal.

[13]  Dong-Joong Kang,et al.  Real-time 2D height mapping method for an unmanned vehicle using a stereo camera and laser sensor fusion , 2012 .

[14]  Aleš Janota,et al.  Control system for the haptic paddle used in mobile robotics , 2017 .

[15]  Frederick M. Waltz,et al.  Machine vision for industrial applications , 2001 .

[16]  Khairul Salleh Mohamed Sahari,et al.  Detecting negative obstacle using Kinect sensor , 2017 .

[17]  Jong-Eun Ha,et al.  Detection of Calibration Patterns for Camera Calibration with Irregular Lighting and Complicated Backgrounds , 2008 .

[18]  Pavol Bozek,et al.  The calculations of Jordan curves trajectory of the robot movement , 2016 .

[19]  Ivan Kuric,et al.  VISION GUIDED PARALLEL ROBOT AND ITS APPLICATION FOR AUTOMATED ASSEMBLY TASK , 2018 .

[20]  Václav Krys,et al.  Kinect v2 infrared images correction , 2018 .

[21]  Feng Gao,et al.  Performance Analysis and Evaluation of Geometric Parameters in Stereo Deflectometry , 2018, Engineering.

[22]  Ertugrul Bayraktar,et al.  A hybrid image dataset toward bridging the gap between real and simulation environments for robotics , 2018, Machine Vision and Applications.

[23]  František Duchoň,et al.  Localization of mobile robot using visual system , 2017 .

[24]  Dong-Joong Kang,et al.  Machine learning-based imaging system for surface defect inspection , 2016, International Journal of Precision Engineering and Manufacturing-Green Technology.

[25]  Pavol Božek,et al.  Trends in Simulation and Planning of Manufacturing Companies , 2016 .

[26]  Javaid Iqbal,et al.  On the Improvement of Multi-Legged Locomotion over Difficult Terrains Using a Balance Stabilization Method: , 2012 .

[27]  W. Echelmeyer,et al.  Development of a robot-based system for automated unloading of variable packages out of transport units and containers , 2008, 2008 IEEE International Conference on Automation and Logistics.

[28]  Ivan Virgala,et al.  Miniature Mobile Bristled In-Pipe Machine , 2014 .

[29]  Hong-Hyun Kim,et al.  Multi-task convolutional neural network system for license plate recognition , 2017, International Journal of Control, Automation and Systems.

[30]  Dong-Joong Kang,et al.  Unified convolutional neural network for direct facial keypoints detection , 2018, The Visual Computer.

[31]  Karol Velíšek,et al.  Design of a robotized workstation making use of the integration of CAD models and Robotic Simulation software as way of pairing and comparing real and virtual environments , 2017 .

[32]  Raquel Frizera Vassallo,et al.  Hybrid Collaborative Stereo Vision System for Mobile Robots Formation , 2009 .

[33]  I. Welber,et al.  Factory of the future , 1987, IEEE Control Systems Magazine.

[34]  Zhen Liu,et al.  Calibration of Binocular Vision Sensors Based on Unknown-Sized Elliptical Stripe Images , 2017, Sensors.

[35]  Ivan Kuric,et al.  WORKSHOP PROGRAMMING AS PART OF TECHNOLOGICAL PREPARATION OF PRODUCTION , 2017 .