Partial Observer Decision Process Model for Crane-Robot Action

The most common use of robots is to effectively decrease the human’s effort with desirable output. In the human-robot interaction, it is essential for both parties to predict subsequent actions based on their present actions so as to well complete the cooperative work. A lot of effort has been devoted in order to attain cooperative work between human and robot precisely. In case of decision making , it is observed from the previous studies that short-term or midterm forecasting have long time horizon to adjust and react. To address this problem, we suggested a new vision-based interaction model. The suggested model reduces the error amplification problem by applying the prior inputs through their features, which are repossessed by a deep belief network (DBN) though Boltzmann machine (BM) mechanism. Additionally, we present a mechanism to decide the possible outcome (accept or reject). The said mechanism evaluates the model on several datasets. Hence, the systems would be able to capture the related information using the motion of the objects. And it updates this information for verification, tracking, acquisition, and extractions of images in order to adapt the situation. Furthermore, we have suggested an intelligent purifier filter (IPF) and learning algorithm based on vision theories in order to make the proposed approach stronger. Experiments show the higher performance of the proposed model compared to the state-of-the-art methods.

[1]  A. Treisman,et al.  Perception of objects in natural scenes: is it really attention free? , 2005, Journal of experimental psychology. Human perception and performance.

[2]  P. Lazaridis,et al.  Software Component Selection Based on Quality Criteria Using the Analytic Network Process , 2014 .

[3]  M. Giese,et al.  Norm-based face encoding by single neurons in the monkey inferotemporal cortex , 2006, Nature.

[4]  Jiancheng Lv,et al.  A motion classification model with improved robustness through deformation code integration , 2018, Neural Computing and Applications.

[5]  Jayachandra Chilukamari,et al.  A computational model of visual attention , 2017 .

[6]  Christof Koch,et al.  Modeling attention to salient proto-objects , 2006, Neural Networks.

[7]  Jianping Li,et al.  Comparative Analysis of the Classification Performance of Machine Learning Classifiers and Deep Neural Network Classifier for Prediction of Parkinson Disease , 2018, 2018 15th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP).

[8]  Zhang Peng,et al.  Detecting Salient Regions Based on Location Shift and Extent Trace , 2004 .

[9]  Marisa Carrasco,et al.  Speed of visual processing increases with eccentricity , 2003, Nature Neuroscience.

[10]  Asad Waqar Malik,et al.  Vision-Based Inceptive Integration for Robotic Control , 2019 .

[11]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[12]  Jian-Ping Li,et al.  Vision prehension with CBIR for cloud robo , 2014, 2014 11th International Computer Conference on Wavelet Actiev Media Technology and Information Processing(ICCWAMTIP).

[13]  Hongxia Wang,et al.  Reversible data hiding in homomorphically encrypted image using interpolation technique , 2019, J. Inf. Secur. Appl..

[14]  King Ngi Ngan,et al.  Dynamic Bit Allocation for Multiple Video Object Coding , 2006, IEEE Transactions on Multimedia.

[15]  Sher Afzal Khan,et al.  A Novel Rules Based Approach for Estimating Software Birthmark , 2015, TheScientificWorldJournal.

[16]  Asad Malik,et al.  Reversible Data Hiding with Multiple Data for Multiple Users in an Encrypted Image , 2019, Int. J. Digit. Crime Forensics.

[17]  Lala Septem Riza,et al.  Birthmark-Based Software Classification Using Rough Sets , 2017 .

[18]  Claudio M. Privitera,et al.  Algorithms for Defining Visual Regions-of-Interest: Comparison with Eye Fixations , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[20]  Haleem Farman,et al.  Estimation of software features based birthmark , 2017, Cluster Computing.

[21]  Thomas Serre,et al.  Robust Object Recognition with Cortex-Like Mechanisms , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  King Ngi Ngan,et al.  Unsupervised extraction of visual attention objects in color images , 2006, IEEE Transactions on Circuits and Systems for Video Technology.

[23]  Asad Malik,et al.  Feature Selection Based on L1-Norm Support Vector Machine and Effective Recognition System for Parkinson’s Disease Using Voice Recordings , 2019, IEEE Access.

[24]  J ValdésJulio,et al.  2006 Special issue , 2006 .

[25]  Jianping Li,et al.  A novel integrated diagnosis method for breast cancer detection , 2020, J. Intell. Fuzzy Syst..

[26]  C. Koch,et al.  A saliency-based search mechanism for overt and covert shifts of visual attention , 2000, Vision Research.

[27]  Edward H. Adelson,et al.  The Laplacian Pyramid as a Compact Image Code , 1983, IEEE Trans. Commun..

[28]  Laurent Itti,et al.  Ieee Transactions on Pattern Analysis and Machine Intelligence 1 Rapid Biologically-inspired Scene Classification Using Features Shared with Visual Attention , 2022 .

[29]  Laurent Itti,et al.  Automatic foveation for video compression using a neurobiological model of visual attention , 2004, IEEE Transactions on Image Processing.

[30]  Jianfeng Feng,et al.  Cue-guided search: a computational model of selective attention , 2005, IEEE Transactions on Neural Networks.

[31]  S Ullman,et al.  Shifts in selective visual attention: towards the underlying neural circuitry. , 1985, Human neurobiology.

[32]  Andrew Zisserman,et al.  Two-Stream Convolutional Networks for Action Recognition in Videos , 2014, NIPS.

[33]  Patrick Le Callet,et al.  A coherent computational approach to model bottom-up visual attention , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.