Robiot: A Design Tool for Actuating Everyday Objects with Automatically Generated 3D Printable Mechanisms

Users can now easily communicate digital information with an Internet of Things; in contrast, there remains a lack of support to automate physical tasks that involve legacy static objects, e.g. adjusting a desk lamp's angle for optimal brightness, turning on/off a manual faucet when washing dishes, sliding a window to maintain a preferred indoor temperature. Automating these simple physical tasks has the potential to improve people's quality of life, which is particularly important for people with a disability or in situational impairment. We present Robiot -- a design tool for generating mechanisms that can be attached to, motorized, and actuating legacy static objects to perform simple physical tasks. Users only need to take a short video manipulating an object to demonstrate an intended physical behavior. Robiot then extracts requisite parameters and automatically generates 3D models of the enabling actuation mechanisms by performing a scene and motion analysis of the 2D video in alignment with the object's 3D model. In an hour-long design session, six participants used Robiot to actuate seven everyday objects, imbuing them with the robotic capability to automate various physical tasks.

[1]  Antonis A. Argyros,et al.  Hobbit , a care robot supporting independent living at home : First prototype and lessons learned , 2015 .

[2]  C. Martin 2015 , 2015, Les 25 ans de l’OMC: Une rétrospective en photos.

[3]  Pedro Lopes,et al.  TrussFormer: 3D Printing Large Kinetic Structures , 2018, UIST.

[4]  Takeo Kanade,et al.  An Iterative Image Registration Technique with an Application to Stereo Vision , 1981, IJCAI.

[5]  Ivan Poupyrev,et al.  Touché: enhancing touch interaction on humans, screens, liquids, and everyday objects , 2012, CHI.

[6]  Daniel Ashbrook,et al.  Towards Augmented Fabrication: Combining Fabricated and Existing Objects , 2016, CHI Extended Abstracts.

[7]  Ruta Desai,et al.  Assembly-aware Design of Printable Electromechanical Devices , 2018, UIST.

[8]  Bernhard Thomaszewski,et al.  LinkEdit: interactive linkage editing using symbolic kinematics , 2015, ACM Trans. Graph..

[9]  Xiang 'Anthony' Chen,et al.  Facade: Auto-generating Tactile Interfaces to Appliances , 2016, ASSETS.

[10]  Dimitrios Tzovaras,et al.  RAMCIP Robot: A Personal Robotic Assistant; Demonstration of a Complete Framework , 2018, ECCV Workshops.

[11]  Wojciech Matusik,et al.  Computational design of mechanical characters , 2013, ACM Trans. Graph..

[12]  Patrick Baudisch,et al.  Grafter: Remixing 3D-Printed Machines , 2018, CHI.

[13]  Tovi Grossman,et al.  Dream Lens: Exploration and Visualization of Large-Scale Generative Design Datasets , 2018, CHI.

[14]  Wojciech Matusik,et al.  AutoConnect , 2015, ACM Trans. Graph..

[15]  Tovi Grossman,et al.  RetroFab: A Design Tool for Retrofitting Physical Interfaces using Actuators, Sensors and 3D Printing , 2016, CHI.

[16]  Ana Paiva,et al.  What My Eyes Can't See, A Robot Can Show Me: Exploring the Collaboration Between Blind People and Robots , 2018, ASSETS.

[17]  Alex Olwal,et al.  shiftIO: Reconfigurable Tactile Elements for Dynamic Affordances and Mobile Interaction , 2017, CHI.

[18]  Dongxiao Li,et al.  A Depth Extraction Method Based on Motion and Geometry for 2D to 3D Conversion , 2009, 2009 Third International Symposium on Intelligent Information Technology Application.

[19]  Xiang 'Anthony' Chen,et al.  Forte: User-Driven Generative Design , 2018, CHI.

[20]  Vladimir A. Kulyukin,et al.  Robot-assisted shopping for the blind: issues in spatial cognition and product selection , 2008, Intell. Serv. Robotics.

[21]  Robert Kovacs,et al.  Patching Physical Objects , 2015, UIST.

[22]  Desney S. Tan,et al.  An ultra-low-power human body motion sensor using static electric field sensing , 2012, UbiComp.

[23]  Yunyi Jia,et al.  Facilitating Human–Robot Collaborative Tasks by Teaching-Learning-Collaboration From Human Demonstrations , 2019, IEEE Transactions on Automation Science and Engineering.

[24]  Chao Liu,et al.  Depth map estimation from motion for 2D to 3D conversion , 2012, 2012 IEEE International Conference on Electro/Information Technology.

[25]  Rubaiat Habib Kazi,et al.  DreamSketch: Early Stage 3D Design Explorations with Sketching and Generative Design , 2017, UIST.

[26]  Mark D. Gross,et al.  Learning about Complexity with Modular Robots , 2008, 2008 Second IEEE International Conference on Digital Game and Intelligent Toy Enhanced Learning.

[27]  Andrey Kurenkov DeepCrop : Directed Object Segmentation with Deep Learning , 2016 .

[28]  Mariangela Manti,et al.  Soft assistive robot for personal care of elderly people , 2016, 2016 6th IEEE International Conference on Biomedical Robotics and Biomechatronics (BioRob).

[29]  N. Altman An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression , 1992 .

[30]  Alex S. Taylor,et al.  Mechanical hijacking: how robots can accelerate UbiComp deployments , 2011, UbiComp '11.

[31]  Wendy Ju,et al.  Designing robots with movement in mind , 2014, Journal of Human-Robot Interaction.

[32]  Hiroshi Ishii,et al.  TRANSFORM as Adaptive and Dynamic Furniture , 2015, CHI Extended Abstracts.

[33]  Xiang 'Anthony' Chen,et al.  Reprise: A Design Tool for Specifying, Generating, and Customizing 3D Printable Adaptations on Everyday Objects , 2016, UIST.

[34]  Florence March,et al.  2016 , 2016, Affair of the Heart.

[35]  Xiang 'Anthony' Chen,et al.  Encore: 3D printed augmentation of everyday objects with printed-over, affixed and interlocked attachments , 2015, SIGGRAPH Posters.

[36]  Jay David Bolter,et al.  DART: a toolkit for rapid design exploration of augmented reality experiences , 2005, SIGGRAPH 2005.

[37]  Aaron Steinfeld,et al.  Using Robot Manipulation to Assist Navigation by People Who Are Blind or Low Vision , 2017, HRI.

[38]  Tek-Jin Nam,et al.  SketchStudio: Experience Prototyping with 2.5-Dimensional Animated Design Scenarios , 2018, Conference on Designing Interactive Systems.

[39]  Hod Lipson,et al.  Molecubes: An Open-Source Modular Robotics Kit , 2007 .

[40]  Henk Nijmeijer,et al.  Robot Programming by Demonstration , 2010, SIMPAR.

[41]  Andrew Blake,et al.  "GrabCut" , 2004, ACM Trans. Graph..

[42]  Markus H. Gross,et al.  Interactive design of 3D-printable robotic creatures , 2015, ACM Trans. Graph..

[43]  Stefan Schaal,et al.  Robot Programming by Demonstration , 2009, Springer Handbook of Robotics.

[44]  Patrick Baudisch,et al.  RoMA: Interactive Fabrication with Augmented Reality and a Robotic 3D Printer , 2018, CHI.

[45]  Carlo Tomasi,et al.  Good features to track , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[46]  Zoe Doulgeri,et al.  Grasping Flat Objects by Exploiting Non-Convexity of the Object and Support Surface , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[47]  R. Dillmann,et al.  TEACHING SERVICE ROBOTS COMPLEX TASKS : PROGRAMMING BY DEMONSTRATION FOR WORKSHOP AND HOUSEHOLD ENVIRONMENTS , 2001 .