A brief glimpse at a haptic target is sufficient for multisensory integration in reaching movements

Goal-directed aiming movements toward visuo-haptic targets (i.e., seen and handheld targets) are generally more precise than those toward visual only or haptic only targets. This multisensory advantage stems from a continuous inflow of haptic and visual target information during the movement planning and execution phases. However, in everyday life, multisensory movements often occur without the support of continuous visual information. Here we investigated whether and to what extent limiting visual information to the initial stage of the action still leads to a multisensory advantage. Participants were asked to reach a handheld target while vision was briefly provided during the movement planning phase (50 ms, 100 ms, 200 ms of vision before movement onset), or during the planning and early execution phases (400 ms of vision), or during the entire movement. Additional conditions were performed in which only haptic target information was provided, or, only vision was provided either briefly (50 ms, 100 ms, 200 ms, 400 ms) or throughout the entire movement. Results showed that 50 ms of vision before movement onset were sufficient to trigger a direction-specific visuo-haptic integration process that increased movement precision. We conclude that, when a continuous support of vision is not available, movement precision is determined by the less recent, but most reliable multisensory information rather than by the latest unisensory (haptic) inputs.

[1]  Michel Desmurget,et al.  The effect of viewing the static hand prior to movement onset on pointing kinematics and variability , 1994, Experimental Brain Research.

[2]  Robert Volcic,et al.  Grasping movements toward seen and handheld objects , 2019, Scientific Reports.

[3]  M. Latash,et al.  Effects of altering initial position on movement direction and extent. , 2003, Journal of neurophysiology.

[4]  R Core Team,et al.  R: A language and environment for statistical computing. , 2014 .

[5]  Fulvio Domini,et al.  A framework for the study of vision in active observers , 2014, Electronic Imaging.

[6]  Robert J. van Beers,et al.  How humans combine simultaneous proprioceptive and visual position information , 1996, Experimental Brain Research.

[7]  R. Sainburg,et al.  The effect of target modality on visual and proprioceptive contributions to the control of movement distance , 2006, Experimental Brain Research.

[8]  R. Johansson,et al.  Eye–Hand Coordination in Object Manipulation , 2001, The Journal of Neuroscience.

[9]  Jiqiang Guo,et al.  Stan: A Probabilistic Programming Language. , 2017, Journal of statistical software.

[10]  Paul-Christian Bürkner,et al.  brms: An R Package for Bayesian Multilevel Models Using Stan , 2017 .

[11]  Robert Volcic,et al.  Integration of haptics and vision in human multisensory grasping , 2020, Cortex.

[12]  D. Westwood,et al.  No Evidence for Accurate Visuomotor Memory: Systematic and Variable Error in Memory-Guided Reaching , 2003, Journal of motor behavior.

[13]  W. G. Cochran Problems arising in the analysis of a series of similar experiments , 1937 .

[14]  D. Wolpert,et al.  When Feeling Is More Important Than Seeing in Sensorimotor Adaptation , 2002, Current Biology.

[15]  Luc Tremblay,et al.  An optimal velocity for online limb-target regulation processes? , 2016, Experimental Brain Research.

[16]  Fulvio Domini,et al.  On-line visual control of grasping movements , 2016, Experimental Brain Research.

[17]  C. Galletti,et al.  Contribution of visual and proprioceptive information to the precision of reaching movements , 2010, Experimental Brain Research.

[18]  HighWire Press The journal of neuroscience : the official journal of the Society for Neuroscience. , 1981 .

[19]  K. Nagai,et al.  SCN output drives the autonomic nervous system: with special reference to the autonomic function related to the regulation of glucose metabolism. , 1996, Progress in brain research.

[20]  Robert Volcic,et al.  MOTOM toolbox: MOtion Tracking via Optotrak and Matlab , 2018, Journal of Neuroscience Methods.

[21]  R. J. van Beers,et al.  Integration of proprioceptive and visual position-information: An experimentally supported model. , 1999, Journal of neurophysiology.

[22]  D. Elliott,et al.  The Influence of Premovement Visual Information on Manual Aiming , 1987, The Quarterly journal of experimental psychology. A, Human experimental psychology.

[23]  Robert L. Sainburg,et al.  Differential influence of vision and proprioception on control of movement distance , 2006, Experimental Brain Research.

[24]  E. Cressman,et al.  Sensory integration during reaching: the effects of manipulating visual target availability , 2014, Experimental Brain Research.

[25]  Robert Volcic,et al.  Grasping adjustments to haptic, visual and visuo-haptic object perturbations are contingent on the sensory modality. , 2019, Journal of neurophysiology.

[26]  F. Wolohan Experimental Brain Research , 2013 .

[27]  E A Roy,et al.  The accuracy of reaching movements in brief delay conditions. , 2001, Canadian journal of experimental psychology = Revue canadienne de psychologie experimentale.

[28]  M. Murray,et al.  Multisensory Integration: Flexible Use of General Operations , 2014, Neuron.

[29]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[30]  Luc Tremblay,et al.  Online Vision as a Function of Real-Time Limb Velocity: Another Case for Optimal Windows , 2015, Journal of motor behavior.

[31]  Anne C. Sittig,et al.  Localization of a seen finger is based exclusively on proprioception and on vision of the finger , 1999, Experimental Brain Research.

[32]  Michael I. Jordan,et al.  Viewing the hand prior to movement improves accuracy of pointing performed toward the unseen contralateral hand , 1997, Experimental Brain Research.

[33]  James L. Lyons,et al.  Optimal Control Strategies Under Different Feedback Schedules: Kinematic Evidence , 2002, Journal of motor behavior.

[34]  Brendan D. Cameron,et al.  Target modality affects visually guided online control of reaching , 2015, Vision Research.

[35]  Inmaculada Álvarez,et al.  A Panel Data Toolbox for MATLAB , 2017 .

[36]  C. Bard,et al.  On-line versus off-line control of rapid aiming movements. , 1993, Journal of motor behavior.

[37]  M. Posner,et al.  Processing of visual feedback in rapid movements. , 1968, Journal of experimental psychology.

[38]  Lifeng Lin,et al.  Performing Arm-Based Network Meta-Analysis in R with the pcnetmeta Package. , 2017, Journal of statistical software.

[39]  Gavin P. Lawrence,et al.  Inferring online and offline processing of visual feedback in target-directed movements from kinematic data , 2006, Neuroscience & Biobehavioral Reviews.

[40]  Marieke Rohde,et al.  Statistically Optimal Multisensory Cue Integration: A Practical Tutorial. , 2016, Multisensory research.

[41]  Michael I. Jordan,et al.  Computational models of sensorimotor integration , 1997 .