Background:Early computer-aided detection systems for mammography have failed to improve the performance of radiologists. With the remarkable success of deep learning, some recent studies have described computer systems with similar or even superior performance to that of human experts. Among them, Shen et al. (Nature Sci. Rep., 2019) present a promising “end-to-end” training approach. Instead of training a convolutional net with whole mammograms, they first train a “patch classifier” that recognizes lesions in small subimages. Then, they generalize the patch classifier to “whole image classifier” using the property of fully convolutional networks and the end-to-end approach. Using this strategy, the authors have obtained a per-image AUC of 0.87 [0.84, 0.90] in the CBIS-DDSM dataset. Standard mammography consists of two views for each breast: bilateral craniocaudal (CC) and mediolateral oblique (MLO). The algorithm proposed by Shen et al. processes only single-view mammography. We extend their work, presenting the end-to-end training of convolutional net for two-view mammography.
Methods:First, we reproduced Shen et al.'s work, using the CBIS-DDSM dataset. We trained a ResNet50-based net for classifying patches with 224x224 pixels using segmented lesions. Then, the weights of the patch classifier were transferred to the whole image single-view classifier, obtained by removing the dense layers from the patch classifier and stacking one ResNet block at the top. This single-view classifier was trained using full images from the same dataset. Trying to replicate Shen et al.'s work, we obtained an AUC of 0.8524±0.0560, less than 0.87 reported in the original paper. We attribute this worsening to the fact that we are using only 2260 images with two views, instead of 2478 images from the original work.
Finally, we built the two-view classifier that receives CC and MLO views as input. This classifier has inside two copies of the patch classifier, loaded with the weights from the single-view classifier. The features extracted by the two patch classifiers are concatenated and submitted to the ResNet block. The two-view classifier is end-to-end trained using full images, refining all its weights, including those inside the two patch classifiers.
Results:The two-view classifier yielded an AUC of 0.9199±0.0623 in 5-fold cross-validation to classify mammographies into malignant/non-malignant, using single-model and without test-time data augmentation. This is better than the Shen et al.'s AUC (0.87), our single-view AUC (0.85). Zhang et al. (Plos One, 2020) present another two-view algorithm (without end-to-end training) with AUC of 0.95. However, this work cannot directly be compared with ours, as it was tested on a different set of images.
Conclusions:We presented end-to-end training of convolutional net for two-view mammography. Our system's AUC was 0.92, better than the 0.87 obtained by the previous single-view system.
Citation Format: Daniel G. Petrini, Carlos Shimizu, Gabriel V. Valente, Guilherme Folgueira, Guilherme A. Novaes, Maria L. Katayama, Pedro Serio, Rosimeire A. Roela, Tatiana C. Tucunduva, Maria Aparecida A. Folgueira, Hae Y. Kim. End-to-end training of convolutional network for breast cancer detection in two-view mammography [abstract]. In: Proceedings of the American Association for Cancer Research Annual Meeting 2021; 2021 Apr 10-15 and May 17-21. Philadelphia (PA): AACR; Cancer Res 2021;81(13_Suppl):Abstract nr 183.