ROBO: Robust, Fully Neural Object Detection for Robot Soccer

Deep Learning has become exceptionally popular in the last few years due to its success in computer vision and other fields of AI. However, deep neural networks are computationally expensive, which limits their application in low power embedded systems, such as mobile robots. In this paper, an efficient neural network architecture is proposed for the problem of detecting relevant objects in robot soccer environments. The ROBO model's increase in efficiency is achieved by exploiting the peculiarities of the environment. Compared to the state-of-the-art Tiny YOLO model, the proposed network provides approximately 35 times decrease in run time, while achieving superior average precision, although at the cost of slightly worse localization accuracy.

[1]  Fei-Fei Li,et al.  Deep visual-semantic alignments for generating image descriptions , 2015, CVPR.

[2]  Visvanathan Ramesh,et al.  Large-Scale Stochastic Scene Generation and Semantic Annotation for Deep Convolutional Neural Network Training in the RoboCup SPL , 2017, RoboCup.

[3]  Nicolás Cruz,et al.  Using Convolutional Neural Networks in Robots with Limited Computational Resources: Detecting NAO Robots while Playing Soccer , 2017, RoboCup.

[4]  Stephan K. Chalup,et al.  Visual Mesh: Real-time Object Detection Using Constant Sample Density , 2018, RoboCup.

[5]  Marcus M. Scheunemann,et al.  Deep Learning for Semantic Segmentation on Minimal Hardware , 2018, RoboCup.

[6]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Timo Aila,et al.  Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.

[8]  Marton Szemenyei,et al.  Real-Time Scene Understanding Using Deep Neural Networks for RoboCup SPL , 2018, RoboCup.

[9]  Wei Liu,et al.  SSD: Single Shot MultiBox Detector , 2015, ECCV.

[10]  Jacky Baltes,et al.  Humanoid Robot Detection Using Deep Learning: A Speed-Accuracy Tradeoff , 2017, RoboCup.

[11]  Danilo Comminiello,et al.  Group sparse regularization for deep neural networks , 2016, Neurocomputing.

[12]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[13]  Ali Farhadi,et al.  YOLO9000: Better, Faster, Stronger , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[15]  Xin Jiang,et al.  Neural Generative Question Answering , 2015, IJCAI.

[16]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  Wonyong Sung,et al.  Structured Pruning of Deep Convolutional Neural Networks , 2015, ACM J. Emerg. Technol. Comput. Syst..

[18]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[19]  Kaiming He,et al.  Mask R-CNN , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[20]  Oliver Urbann,et al.  A Robust and Calibration-Free Vision System for Humanoid Soccer Robots , 2015, RoboCup.

[21]  Alexei A. Efros,et al.  Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Peter Stone,et al.  Fast and Precise Black and White Ball Detection for RoboCup Soccer , 2017, RoboCup.

[23]  Joseph Redmon,et al.  YOLOv3: An Incremental Improvement , 2018, ArXiv.

[24]  Simon O'Keeffe,et al.  A Benchmark Data Set and Evaluation of Deep Learning Architectures for Ball Detection in the RoboCup SPL , 2017, RoboCup.

[25]  Sven Behnke,et al.  Learning Visual Obstacle Detection Using Color Histogram Features , 2011, RoboCup.

[26]  Alex Graves,et al.  Neural Turing Machines , 2014, ArXiv.

[27]  Ali Farhadi,et al.  YOLOv3: An Incremental Improvement , 2018, ArXiv.

[28]  Geoffrey E. Hinton,et al.  Dynamic Routing Between Capsules , 2017, NIPS.