Automatic annotation of hip anatomy in fluoroscopy for robust and efficient 2D/3D registration

Purpose Fluoroscopy is the standard imaging modality used to guide hip surgery and is therefore a natural sensor for computer-assisted navigation. In order to efficiently solve the complex registration problems presented during navigation, human-assisted annotations of the intraoperative image are typically required. This manual initialization interferes with the surgical workflow and diminishes any advantages gained from navigation. In this paper, we propose a method for fully automatic registration using anatomical annotations produced by a neural network. Methods Neural networks are trained to simultaneously segment anatomy and identify landmarks in fluoroscopy. Training data are obtained using a computationally intensive, intraoperatively incompatible, 2D/3D registration of the pelvis and each femur. Ground truth 2D segmentation labels and anatomical landmark locations are established using projected 3D annotations. Intraoperative registration couples a traditional intensity-based strategy with annotations inferred by the network and requires no human assistance. Results Ground truth segmentation labels and anatomical landmarks were obtained in 366 fluoroscopic images across 6 cadaveric specimens. In a leave-one-subject-out experiment, networks trained on these data obtained mean dice coefficients for left and right hemipelves, left and right femurs of 0.86, 0.87, 0.90, and 0.84, respectively. The mean 2D landmark localization error was 5.0 mm. The pelvis was registered within $$1^{\circ }$$ 1 ∘  for 86% of the images when using the proposed intraoperative approach with an average runtime of 7 s. In comparison, an intensity-only approach without manual initialization registered the pelvis to $$1^{\circ }$$ 1 ∘  in 18% of images. Conclusions We have created the first accurately annotated, non-synthetic, dataset of hip fluoroscopy. By using these annotations as training data for neural networks, state-of-the-art performance in fluoroscopic segmentation and landmark localization was achieved. Integrating these annotations allows for a robust, fully automatic, and efficient intraoperative registration during fluoroscopic navigation of the hip.

[1]  Yue Shi,et al.  A modified particle swarm optimizer , 1998, 1998 IEEE International Conference on Evolutionary Computation Proceedings. IEEE World Congress on Computational Intelligence (Cat. No.98TH8360).

[2]  M. Powell The BOBYQA algorithm for bound constrained optimization without derivatives , 2009 .

[3]  Nikolaus Hansen,et al.  Completely Derandomized Self-Adaptation in Evolution Strategies , 2001, Evolutionary Computation.

[4]  Eric M Slotkin,et al.  Accuracy of Fluoroscopic Guided Acetabular Component Positioning During Direct Anterior Total Hip Arthroplasty. , 2015, The Journal of arthroplasty.

[5]  Todd C. Kelley,et al.  Role of navigation in total hip arthroplasty. , 2009, The Journal of bone and joint surgery. American volume.

[6]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[8]  R Burgkart,et al.  Intraoperative, fluoroscopy‐based planning for complex osteotomies of the proximal femur , 2005, The international journal of medical robotics + computer assisted surgery : MRCAS.

[9]  Sotirios A. Tsaftaris,et al.  Medical Image Computing and Computer Assisted Intervention , 2017 .

[10]  Klaus Radermacher,et al.  Fluoroscopic navigation system for hip surface replacement , 2007, Computer aided surgery : official journal of the International Society for Computer Aided Surgery.

[11]  Jia Deng,et al.  Stacked Hourglass Networks for Human Pose Estimation , 2016, ECCV.

[12]  Nassir Navab,et al.  Learning to detect anatomical landmarks of the pelvis in X-rays from arbitrary views , 2019, International Journal of Computer Assisted Radiology and Surgery.

[13]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[14]  Bostjan Likar,et al.  A review of 3D/2D registration methods for image-guided interventions , 2012, Medical Image Anal..

[15]  Andreas Maier,et al.  Multi-task Localization and Segmentation for X-ray Guided Planning in Knee Surgery , 2019, MICCAI.

[16]  Mathias Unberath,et al.  Towards Fully Automatic X-Ray to CT Registration , 2019, MICCAI.

[17]  Seyed-Ahmad Ahmadi,et al.  V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation , 2016, 2016 Fourth International Conference on 3D Vision (3DV).

[18]  Theo van Walsum,et al.  Fully Automatic and Real-Time Catheter Segmentation in X-Ray Fluoroscopy , 2017, MICCAI.

[19]  Rainer Storn,et al.  Differential Evolution – A Simple and Efficient Heuristic for global Optimization over Continuous Spaces , 1997, J. Glob. Optim..

[20]  A. Troelsen,et al.  Surgical advances in periacetabular osteotomy for treatment of hip dysplasia in adults , 2009, Acta orthopaedica. Supplementum.

[21]  Joachim Grifka,et al.  Visual intraoperative estimation of cup and stem position is not reliable in minimally invasive hip arthroplasty , 2016, Acta orthopaedica.

[22]  Tania S. Douglas,et al.  Review of 2-D/3-D Reconstruction Using Statistical Shape and Intensity Models and X-Ray Image Synthesis: Toward a Unified Framework , 2019, IEEE Reviews in Biomedical Engineering.

[23]  Andreas K. Maier,et al.  Intraoperative stent segmentation in X-ray fluoroscopy for endovascular aortic repair , 2018, International Journal of Computer Assisted Radiology and Surgery.

[24]  Russell H. Taylor,et al.  Patch-Based Image Similarity for Intraoperative 2D/3D Pelvis Registration During Periacetabular Osteotomy , 2018, OR 2.0/CARE/CLIP/ISIC@MICCAI.

[25]  Branislav Jaramaz,et al.  Description of Anatomic Coordinate Systems and Rationale for Use in an Image-Guided Total Hip Replacement System , 2000, MICCAI.

[26]  Russell H. Taylor,et al.  Fast and automatic periacetabular osteotomy fragment pose estimation using intraoperatively implanted fiducials and single-view fluoroscopy , 2019, Physics in medicine and biology.

[27]  Russell H. Taylor,et al.  Localizing dexterous surgical tools in X-ray for image-based navigation , 2019, ArXiv.

[28]  Nassir Navab,et al.  Enabling machine learning in X-ray-based procedures via realistic simulation of image formation , 2019, International Journal of Computer Assisted Radiology and Surgery.

[29]  Yoshinobu Sato,et al.  Construction and Application of Large-Scale Image Database in Orthopedic Surgery , 2018 .

[30]  Trevor Darrell,et al.  Fully Convolutional Networks for Semantic Segmentation , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[31]  Russell H. Taylor,et al.  Pose Estimation of Periacetabular Osteotomy Fragments With Intraoperative X-Ray Navigation , 2019, IEEE Transactions on Biomedical Engineering.

[32]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[33]  Cong Gao,et al.  Fiducial-free 2D/3D registration of the proximal femur for robot-assisted femoroplasty , 2020, Medical Imaging: Image-Guided Procedures.

[34]  Nassir Navab,et al.  Concurrent Segmentation and Localization for Tracking of Surgical Instruments , 2017, MICCAI.

[35]  Z. Jane Wang,et al.  A CNN Regression Approach for Real-Time 2D/3D Registration , 2016, IEEE Transactions on Medical Imaging.

[36]  Charl P. Botha,et al.  A fluoroscopy-based planning and guidance software tool for minimally invasive hip refixation by cement injection , 2015, International Journal of Computer Assisted Radiology and Surgery.