PurposeInterventional radiology is performed in a sterile environment where speech and motion control of image review is needed to simplify and expedite routine procedures. The requirements and limitations were defined by testing an interventional radiology test bed speech and motion control system.MethodsMotion control software was implemented using the Microsoft$$^{\circledR }$$ Kinect$$^{\circledR }$$ (Microsoft Corp., USA) framework. The system was tested by 10 participants using a predefined set of six voice and six gesture commands under different lighting conditions to assess the influence of illumination on command recognition. The participants rated the convenience of the application and its possible use in everyday clinical routine. A basic set of voice or gesture commands required for interventional radiology were identified.ResultsThe majority (93 %) of commands were recognized successfully. Speech commands were less prone to errors than gesture commands. Unwanted side effects occurred (e.g., accidentally issuing a gesture command) in about 30 % of cases. Dimmed lighting conditions did not have a measurable effect on the recognition rate. Six out of 10 participants would consider using the application in everyday routine. The necessary voice/gesture commands for interventional radiology were identified and integrated into the control system.ConclusionSpeech and motion control of image review provides a new man–machine interface for radiological image handling that is especially useful in sterile environments due to no-touch navigation. Command recognition rates were high and remained stable under different lighting conditions. However, the rate of accidental triggering due to unintended commands should be reduced.
[1]
Antonis A. Argyros,et al.
Efficient model-based 3D tracking of hand articulations using Kinect
,
2011,
BMVC.
[2]
H. Stern,et al.
A gesture-based tool for sterile browsing of radiology images.
,
2008,
Journal of the American Medical Informatics Association : JAMIA.
[3]
Antonis A. Argyros,et al.
Markerless and Efficient 26-DOF Hand Pose Recovery
,
2010,
ACCV.
[4]
Suvranu De,et al.
Microsoft Kinect Based Head Tracking for Life Size Collaborative Surgical Simulation Environments (LS-CollaSSLE)
,
2013,
MMVR.
[5]
Guilherme C. S. Ruppert,et al.
Touchless gesture user interface for interactive image visualization in urological surgery
,
2012,
World Journal of Urology.
[6]
J.K. Aggarwal,et al.
Human activity analysis
,
2011,
ACM Comput. Surv..
[7]
Zhan Yu,et al.
A Portable Immersive Surgery Training System Using RGB-D Sensors
,
2013,
MMVR.
[8]
Thomas B. Moeslund,et al.
View invariant gesture recognition using 3D motion primitives
,
2008,
2008 IEEE International Conference on Acoustics, Speech and Signal Processing.
[9]
Junsong Yuan,et al.
Depth camera based hand gesture recognition and its applications in Human-Computer-Interaction
,
2011,
2011 8th International Conference on Information, Communications & Signal Processing.
[10]
William W. Boonn,et al.
Using the Microsoft Kinect for Patient Size Estimation and Radiation Dose Normalization: Proof of Concept and Initial Validation
,
2012,
Journal of Digital Imaging.