4 CONCLUSIONS
The first goal of the study was to investigate the im-
pact of different architectures on the resulting accu-
racy of the models. The best-performing architecture
was ResNet101, but other architectures achieved very
comparable accuracies.
The second goal aimed to explore the transferabil-
ity of knowledge from a model trained on digitalized
films (indirect digital mammography) to direct digital
mammography images and vice versa. It was con-
firmed that models are not transferable to data ob-
tained using different technology. Combining these
training data into a unified dataset significantly con-
tributed to the overall improvement of model accu-
racy. Such a model achieved an accuracy of 76.2
The final part involved examining patches with
incorrect predictions, specifically focusing on those
where the prediction was incorrect across all tested
models. The results were discussed with radiologists,
confirming that many patches incorrectly classified as
malignant pose a significant challenge even for med-
ical professionals and cannot be classified without a
tissue biopsy.
During the experiments, it was observed that the
decision-making in some patches involved the area
around the finding, which did not contain abnormali-
ties. This behavior could potentially be addressed, for
example, by adding a third class containing patches
from healthy tissue. Adding such a class will be the
subject of our next study.
There is a relatively wide scope for improving re-
sults, including better hyperparameter optimization,
adding augmented data, or incorporating regulariza-
tion methods. However, the primary intent of this
work was to explore the questions outlined in the
stated goals.
It is important to note that the created models may
be biased, as all training/validation data used had un-
dergone a biopsy. This means they represent findings
where doctors were uncertain whether the abnormal-
ity was benign or malignant.
ACKNOWLEDGEMENTS
This research was supported by the Ministry of Ed-
ucation, Science, Research and Sport of the Slovak
Republic under the contract No. VEGA 1/0525/23.
REFERENCES
Allred, D. (2010). Ductal carcinoma in situ: Terminology,
classification, and natural history. Journal of the Na-
tional Cancer Institute. Monographs, 2010:134–8.
Chhatwal, J., Alagoz, O., and Burnside, E. S. (2010). Opti-
mal breast biopsy decision-making based on mammo-
graphic features and demographic factors. Operations
research, 58(6):1577–1591.
Geras, K. J., Wolfson, S., Kim, S. G., Moy, L., and Cho,
K. (2017). High-resolution breast cancer screening
with multi-view deep convolutional neural networks.
ArXiv, abs/1703.07047.
Gildenblat, J. and contributors (2021). Pytorch li-
brary for cam methods. https://github.com/jacobgil/
pytorch-grad-cam.
Glorot, X. and Bengio, Y. (2010). Understanding the dif-
ficulty of training deep feedforward neural networks.
Journal of Machine Learning Research - Proceedings
Track, 9:249–256.
Grimm, L. J., Rahbar, H., Abdelmalak, M., Hall, A. H., and
Ryser, M. D. (2022). Ductal carcinoma in situ: State-
of-the-art review. Radiology, 302(2):246–255. PMID:
34931856.
Halling-Brown, M. D., Warren, L. M., Ward, D., Lewis,
E., Mackenzie, A., Wallis, M. G., Wilkinson, L. S.,
Given-Wilson, R. M., McAvinchey, R., and Young,
K. C. (2021). Optimam mammography image
database: A large-scale resource of mammography
images and clinical data. Radiology: Artificial Intelli-
gence, 3(1):e200103. PMID: 33937853.
He, K., Zhang, X., Ren, S., and Sun, J. (2015). Deep
residual learning for image recognition. CoRR,
abs/1512.03385.
Huang, G., Liu, Z., and Weinberger, K. Q. (2016).
Densely connected convolutional networks. CoRR,
abs/1608.06993.
Lee, R., Gimenez, F., Hoogi, A., Miyake, K., Gorovoy, M.,
and Rubin, D. (2017). A curated mammography data
set for use in computer-aided detection and diagnosis
research. Scientific Data, 4:170–177.
Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J.,
Chanan, G., Killeen, T., Lin, Z., Gimelshein, N.,
Antiga, L., Desmaison, A., Kopf, A., Yang, E., De-
Vito, Z., Raison, M., Tejani, A., Chilamkurthy, S.,
Steiner, B., Fang, L., Bai, J., and Chintala, S. (2019).
Pytorch: An imperative style, high-performance deep
learning library. In Advances in Neural Information
Processing Systems 32, pages 8024–8035. Curran As-
sociates, Inc.
Shen, L., Margolies, L., Rothstein, J., Fluder, E., McBride,
R., and Sieh, W. (2019). Deep learning to improve
breast cancer detection on screening mammography.
Scientific Reports, 9:1–12.
Simonyan, K. and Zisserman, A. (2015). Very deep convo-
lutional networks for large-scale image recognition.
Sung, H., Ferlay, J., Siegel, R. L., Laversanne, M., Soerjo-
mataram, I., Jemal, A., and Bray, F. (2021). Global
cancer statistics 2020: Globocan estimates of in-
cidence and mortality worldwide for 36 cancers in
BIOINFORMATICS 2024 - 15th International Conference on Bioinformatics Models, Methods and Algorithms
534