A similarity measure method fusing deep feature for mammogram retrieval.

BACKGROUND Breast cancer is one of the most important malignant tumors among women causing a serious impact on women's lives and mammography is one the most important methods for breast examination. When diagnosing the breast disease, radiologists sometimes may consult some previous diagnosis cases as a reference. But there are many previous cases and it is important to find which cases are the similar cases, which is a big project costing lots of time. Medical image retrieval can provide objective reference information for doctors to diagnose disease. The method of fusing deep features can improve the retrieval accuracy, which solves the "semantic gap" problem caused by only using content features and location features. METHODS A similarity measure method combining deep feature for mammogram retrieval is proposed in this paper. First, the images are pre-processed to extract the low-level features, including content features and location features. Before extracting location features, registration with the standard image is performed. Then, the Convolutional Neural Network, the Stacked Auto-encoder Network, and the Deep Belief Network are built to extract the deep features, which are regarded as high-level features. Next, content similarity and deep similarity are calculated separately using the Euclidean distance between the query image and the dataset images. The location similarity is obtained by calculating the ratio of intersection to union of the mass regions. Finally, content similarity, location similarity, and deep similarity are fused to form the image fusion similarity. According to the similarity, the specified number of the most similar images can be returned. RESULTS In the experiment, 740 MLO mammograms are used, which are from women in Northeast China. The content similarity, location similarity, and deep similarity are fused by different weight coefficients. When only considering low-level features, the results are better with fusing 60% content feature similarity and 40% lesion location feature similarity. On this basis, CNN deep similarity, DBN deep similarity, and SAE deep similarity are fused separately. The experiments show that when fusing 60% DBN deep feature similarity and 40% low-level feature similarity, the results have obvious advantages. At this time, the precision is 0.745, recall is 0.850, comprehensive evaluation index is 0.794. CONCLUSIONS We propose a similarity measure method fusing deep feature, content feature, and location feature. The retrieval results show that the precision and recall of this method have obvious advantage, compared with the content-based image retrieval and location-based image retrieval.

[1]  Yuanjie Zheng,et al.  Scalable Mammogram Retrieval Using Composite Anchor Graph Hashing With Iterative Quantization , 2017, IEEE Transactions on Circuits and Systems for Video Technology.

[2]  Geoffrey E. Hinton Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.

[3]  Andriy Myronenko,et al.  Point Set Registration: Coherent Point Drift , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Networks , 2013, ECCV.

[5]  Ursula Gather,et al.  Repeated median and hybrid filters , 2006, Comput. Stat. Data Anal..

[6]  Vijay V. Raghavan,et al.  A critical investigation of recall and precision as measures of retrieval system performance , 1989, TOIS.

[7]  D. Abraham Chandy,et al.  Texture feature extraction using gray level statistical matrix for content-based mammogram retrieval , 2013, Multimedia Tools and Applications.

[8]  Roland Holland,et al.  Image-detected breast cancer: state of the art diagnosis and treatment. , 2005, Journal of the American College of Surgeons.

[9]  Xiaohui Liu,et al.  Mammogram retrieval on similar mass lesions , 2012, Comput. Methods Programs Biomed..

[10]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[11]  H. Greenspan,et al.  Automated retrieval of CT images of liver lesions on the basis of image similarity: method and preliminary results. , 2010, Radiology.

[12]  Ming-Kuei Hu,et al.  Visual pattern recognition by moment invariants , 1962, IRE Trans. Inf. Theory.

[13]  Qian Chen,et al.  Image enhancement based on equal area dualistic sub-image histogram equalization method , 1999, IEEE Trans. Consumer Electron..

[14]  Robert M. Haralick,et al.  Textural Features for Image Classification , 1973, IEEE Trans. Syst. Man Cybern..

[15]  N. Otsu A threshold selection method from gray level histograms , 1979 .

[16]  Hideyuki Tamura,et al.  Textural Features Corresponding to Visual Perception , 1978, IEEE Transactions on Systems, Man, and Cybernetics.

[17]  C. Maggio State of the art of current modalities for the diagnosis of breast lesions , 2004, European Journal of Nuclear Medicine and Molecular Imaging.

[18]  Jianzhong Wu,et al.  Stacked Sparse Autoencoder (SSAE) for Nuclei Detection on Breast Cancer Histopathology Images , 2016, IEEE Transactions on Medical Imaging.

[19]  Yang Li,et al.  A similarity measure method combining location feature for mammogram retrieval. , 2018, Journal of X-ray science and technology.

[20]  Thierry Pun,et al.  A new method for grey-level picture thresholding using the entropy of the histogram , 1980 .

[21]  Carla E. Brodley,et al.  ASSERT: A Physician-in-the-Loop Content-Based Retrieval System for HRCT Image Databases , 1999, Comput. Vis. Image Underst..

[22]  Dipesh Patel,et al.  Improvement in Performance of Image Retrieval using Various Features in CBIR System , 2016 .

[23]  Arnaldo de Albuquerque Araújo,et al.  MammoSys: A content-based image retrieval system using breast density patterns , 2010, Comput. Methods Programs Biomed..