Digital reality: a model-based approach to supervised learning from synthetic data

Hierarchical neural networks with large numbers of layers are the state of the art for most computer vision problems including image classification, multi-object detection and semantic segmentation. While the computational demands of training such deep networks can be addressed using specialized hardware, the availability of training data in sufficient quantity and quality remains a limiting factor. Main reasons are that measurement or manual labelling are prohibitively expensive, ethical considerations can limit generating data, or a phenomenon in questions has been predicted, but not yet observed. In this position paper, we present the Digital Reality concept are a structured approach to generate training data synthetically. The central idea is to simulate measurements based on scenes that are generated by parametric models of the real world. By investigating the parameter space defined of such models, training data can be generated in a controlled way compared to data that was captured from real world situations. We propose the Digital Reality concept and demonstrate its potential in different application domains, including industrial inspection, autonomous driving, smart grid, and microscopy research in material science and engineering.

[1]  H Liers,et al.  Creation of pre-crash simulations in global traffic accident scenarios based on the iGLAD database , 2015 .

[2]  David B. Williams,et al.  Transmission Electron Microscopy , 1996 .

[3]  Heinz Graafsma,et al.  In-situ Materials Characterization , 2014 .

[4]  Deborah Silver,et al.  Feature Visualization , 1994, Scientific Visualization.

[5]  Antonio M. López,et al.  The SYNTHIA Dataset: A Large Collection of Synthetic Images for Semantic Segmentation of Urban Scenes , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Michael Goesele,et al.  Reconstructing scenes with mirror and glass surfaces , 2018, ACM Trans. Graph..

[7]  R. Krupke,et al.  Understanding the graphitization and growth of free-standing nanocrystalline graphene using in situ transmission electron microscopy. , 2017, Nanoscale.

[8]  Lin Ma,et al.  Image-level to Pixel-wise Labeling: From Theory to Practice , 2018, IJCAI.

[9]  Christian Müller,et al.  Toward a Methodology for Training with Synthetic Data on the Example of Pedestrian Detection in a Frame-by-Frame Semantic Segmentation Task , 2018, 2018 IEEE/ACM 1st International Workshop on Software Engineering for AI in Autonomous Systems (SEFAIAS).

[10]  Peter A. Crozier,et al.  In situ and operando transmission electron microscopy of catalytic materials , 2015 .

[11]  Krishna Rajan,et al.  The future of atom probe tomography , 2012 .

[12]  Giuseppe Claudio Guarnera,et al.  BRDF Representation and Acquisition , 2016, Comput. Graph. Forum.

[13]  Harry Shum,et al.  Motion texture: a two-level statistical model for character motion synthesis , 2002, ACM Trans. Graph..

[14]  Jane McGonigal,et al.  Keynote: Jane McGonigal , 2012, ACM Transactions on Graphics.

[15]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Networks , 2013, ECCV.

[16]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[17]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[18]  Nataska Statham,et al.  Use of Photogrammetry in Video Games: A Historical Overview , 2018, Games Cult..

[19]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[20]  Nicolas Pronost,et al.  Interactive Character Animation Using Simulated Physics: A State‐of‐the‐Art Review , 2012, Comput. Graph. Forum.

[21]  Javier Alonso-Mora,et al.  Planning and Decision-Making for Autonomous Vehicles , 2018, Annu. Rev. Control. Robotics Auton. Syst..

[22]  Ole Winther,et al.  A Deep Learning Approach to Identify Local Structures in Atomic‐Resolution Transmission Electron Microscopy Images , 2018, Advanced Theory and Simulations.

[23]  Michael K Miller,et al.  Invited review article: Atom probe tomography. , 2007, The Review of scientific instruments.

[24]  Hod Lipson,et al.  Understanding Neural Networks Through Deep Visualization , 2015, ArXiv.

[25]  Andrew Zisserman,et al.  Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.

[26]  K. Novoselov,et al.  A roadmap for graphene , 2012, Nature.

[27]  Mohan M. Trivedi,et al.  Looking at Humans in the Age of Self-Driving and Highly Automated Vehicles , 2016, IEEE Transactions on Intelligent Vehicles.

[28]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[29]  George Yannis,et al.  A critical assessment of pedestrian behaviour models , 2009 .

[30]  Taku Komura,et al.  Phase-functioned neural networks for character control , 2017, ACM Trans. Graph..

[31]  Lucas Kovar,et al.  Motion graphs , 2002, SIGGRAPH Classes.

[32]  David Bommes,et al.  Boundary element octahedral fields in volumes , 2017, TOGS.

[33]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[34]  Andrew Y. Ng,et al.  Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .

[35]  Taku Komura,et al.  A Deep Learning Framework for Character Motion Synthesis and Editing , 2016, ACM Trans. Graph..

[36]  Steve Marschner,et al.  Matching Real Fabrics with Micro-Appearance Models , 2015, ACM Trans. Graph..

[37]  Carolina Wählby,et al.  Automated Training of Deep Convolutional Neural Networks for Cell Segmentation , 2017, Scientific Reports.

[38]  Tom Quirk,et al.  There’s Plenty of Room at the Bottom , 2006, Size Really Does Matter.

[39]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[40]  Arvind Satyanarayan,et al.  The Building Blocks of Interpretability , 2018 .

[41]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[42]  Slobodan Ilic,et al.  Framework for Generation of Synthetic Ground Truth Data for Driver Assistance Applications , 2013, GCPR.

[43]  Fengli Wang,et al.  Catalyst-free growth of nanocrystalline graphene/graphite patterns from photoresist. , 2013, Chemical communications.

[44]  Frédo Durand,et al.  On the Importance of Label Quality for Semantic Segmentation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[45]  Germán Ros,et al.  CARLA: An Open Urban Driving Simulator , 2017, CoRL.

[46]  Hiroshi Ishikawa,et al.  Mastering Sketching , 2017, ACM Trans. Graph..

[47]  Lucas Kovar,et al.  Motion Graphs , 2002, ACM Trans. Graph..

[48]  Aaron Hertzmann,et al.  Style-based inverse kinematics , 2004, SIGGRAPH 2004.

[49]  Jian-Jun Zhang,et al.  Adaptive motion synthesis for virtual characters: a survey , 2014, The Visual Computer.

[50]  Ashish Kapoor,et al.  AirSim: High-Fidelity Visual and Physical Simulation for Autonomous Vehicles , 2017, FSR.

[51]  Marcelo Kallmann,et al.  An Analysis of Motion Blending Techniques , 2012, MIG.

[52]  Jinxiang Chai,et al.  Motion graphs++ , 2012, ACM Trans. Graph..

[53]  Tomohiko Mukai,et al.  Geostatistical motion interpolation , 2005, SIGGRAPH 2005.

[54]  Wolfgang Wenzel,et al.  Nanocrystalline graphene at high temperatures: insight into nanoscale processes , 2019, Nanoscale advances.

[55]  Thomas Brox,et al.  Striving for Simplicity: The All Convolutional Net , 2014, ICLR.