Breast cancer remains the most common cancer among African women, contributing to high mortality largely due to late-stage diagnosis. More than 70% of cases in many African countries are detected at advanced stages, when treatment options are limited and survival rates are significantly reduced. Barriers such as limited access to screening, a shortage of radiologists, and socio-economic constraints further delay early detection. Artificial intelligence (AI)-based diagnostic tools offer an opportunity to strengthen screening systems and improve timely diagnosis; however, the lack of publicly available African mammography datasets remains a major challenge. This study introduces MAMMOAI, an AI-driven framework designed to enhance early breast cancer detection with potential applicability to African contexts. Due to the scarcity of large-scale annotated African mammography datasets a critical barrier identified in current literature we adopted a pragmatic approach combining a multi-task deep learning model integrating convolutional neural networks (CNN) with Transformer layers to perform simultaneous risk assessment, cancer detection, staging, risk factor analysis, and differential diagnosis using mammograms. The model was trained using the King AbdulAziz University Breast Cancer Mammogram Dataset (KAU-BCMD) from Saudi Arabia, supplemented with preliminary data collected from Kenyan healthcare facilities. This multi-regional approach was necessitated by insufficient African data volume for robust deep learning model training. The framework employs a multi-task learning approach that integrates a ResNet50–Transformer backbone for spatial feature extraction, feeding five specialized branches for risk assessment, cancer detection, staging, risk factor analysis, and differential diagnosis. All images underwent standardized preprocessing, including resizing to 224×224 pixels, normalization, contrast enhancement, and extensive data augmentation. Weighted categorical cross-entropy losses supported joint optimization across tasks. Model interpretability was ensured using Grad-CAM–based heatmaps and uncertainty estimation, and predictions were compiled into automated, clinician-friendly HTML reports. The model achieved overall accuracy of 98% on the majority class (BI-RADS 1), with macro-averaged F1-scores of 0.61-0.62 across all branches, reflecting challenges in detecting minority classes. Critically, the model failed to identify any BI-RADS 5 (highly malignant) cases, misclassifying all as BI-RADS 4. Grad-CAM visualizations provided interpretable insights, supporting clinical decision-making. While these results demonstrate technical feasibility and the potential of hybrid architectures, they also underscore critical limitations: severe class imbalance, inadequate minority class performance, and unproven generalizability to diverse African populations. Future work must prioritize collection of larger, balanced, multi-center African datasets and external validation across diverse Sub-Saharan African healthcare settings before clinical deployment can be considered.
| Published in | American Journal of Artificial Intelligence (Volume 10, Issue 1) |
| DOI | 10.11648/j.ajai.20261001.11 |
| Page(s) | 1-13 |
| Creative Commons |
This is an Open Access article, distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution and reproduction in any medium or format, provided the original work is properly cited. |
| Copyright |
Copyright © The Author(s), 2025. Published by Science Publishing Group |
Breast Cancer, Deep Learning, Mammography, MammoAI
| [1] | International Agency for Research on Cancer (IARC), Breast cancer overall survival, annual risks of death, and survival gap apportionment in sub-Saharan Africa (ABC-DO): Seven-year follow-up of a prospective cohort study. The Lancet Global Health; 2025. Available from: |
| [2] | World Health Organization (WHO). Assessment of breast cancer control capacities in the WHO African Region. Brazzaville: WHO Regional Office for Africa; 2022. Available from: |
| [3] | Demographic and Health Surveys (DHS) Study. Barriers to clinical breast examination in sub-Saharan Africa: Financial, social, and access challenges. African Researchers Network; 2025. Available from: |
| [4] | Bayatmakou F, Taleei R, Amir Toutounchian M, Mohammadi A. Integrating AI for human-centric breast cancer diagnostics: A multi-scale and multi-view Swin Transformer framework. arXiv; 2025. Available from: |
| [5] | Mehmood A, Hu Y, Khan SH, et al. A novel channel-boosted residual CNN–Transformer with regional-boundary learning for breast cancer detection. arXiv; 2025. Available from: |
| [6] | Anyigba, C. A., Awandare, G. A., & Paemka, L. (2021). Breast cancer in sub-Saharan Africa: The current state and uncertain future. Experimental biology and medicine (Maywood, N. J.), 246(12), 1377–1387. |
| [7] | Martei, Y. M., Dauda, B., & Vanderpuye, V. (2022). Breast cancer screening in sub-Saharan Africa: a systematic review and ethical appraisal. BMC cancer, 22(1), 203. |
| [8] | Brahmareddy, A., Selvan, M. P. TransBreastNet a CNN transformer hybrid deep learning framework for breast cancer subtype classification and temporal lesion progression analysis. Sci Rep 15, 35106 (2025). |
| [9] | Yala, A., Lehman, C., Schuster, T., Portnoi, T., Barzilay, R. A Deep Learning Mammography-based Model for Improved Breast Cancer Risk Prediction. Radiology. 2019, 292(1), 60–66. |
| [10] | Abdikenov, B., Zhaksylyk, T., Imasheva, A., & Rakishev, D. (2025). From Mammogram Analysis to Clinical Integration with Deep Learning in Breast Cancer Diagnosis. Informatics, 12(4), 106. |
| [11] | Francesco Manigrasso, Rosario Milazzo, Alessandro Sebastian Russo, Fabrizio Lamberti, Fredrik Strand, Andrea Pagnani, Lia Morra, Mammography classification with multi-view deep learning techniques: Investigating graph and transformer-based architectures, Medical Image Analysis, Volume 99, 2025, 103320, |
| [12] | Tardy Mickael, Mateus Diana, Leveraging Multi-Task Learning to Cope With Poor and Missing Labels of Mammograms, Frontiers in Radiology, Volume 1 – 2021, 2022, |
| [13] | Chowdary J, Yogarajah P, Chaurasia P, Guruviah V. A Multi-Task Learning Framework for Automated Segmentation and Classification of Breast Tumors From Ultrasound Images. Ultrasonic Imaging. 2022; 44(1): 3-12. |
| [14] | Talaat, F. M., Gamel, S. A., El-Balka, R. M., Shehata, M., & ZainEldin, H. (2024). Grad-CAM Enabled Breast Cancer Classification with a 3D Inception-ResNet V2: Empowering Radiologists with Explainable Insights. Cancers, 16(21), 3668. |
| [15] | Mellado, D., Mayeta-Revilla, L., Sotelo, J., Querales, M., Godoy, E., Lever, S., Pardo, F., Chabert, S., & Salas, R. (2025). Identifying clinically relevant findings in breast cancer using deep learning and feature attribution on local views from high-resolution mammography. Frontiers in oncology, 15, 1601929. |
| [16] | Mendes, J., Oliveira, B., Araújo, C., Galrão, J., Mota, A. M., Garcia, N. C., & Matela, N. (2025). Deep learning in breast cancer risk prediction: a review of recent applications in full-field digital mammography. Frontiers in oncology, 15, 1656842. |
| [17] | Urooj, A., Dapamede, T., Patel, B., Ayoub, C., Arsanjani, R., O'Neill, W. C., Trivedi, H., & Banerjee, I. (2025). A Multi-Task Learning Approach for Segmentation of Breast Arterial Calcifications in Screening Mammograms. AMIA... Annual Symposium proceedings. AMIA Symposium, 2024, 1139–1148. |
| [18] | Sharma, S., Singh, Y. & Choudhury, T. Advanced deep learning architectures for enhanced mammography classification: a comparative study of CNNs and ViT. Discov Artif Intell 5, 187 (2025). |
| [19] | Mingshuang Fang, Binxiong Xu, Transformer-based multi-modal learning for breast cancer screening: Merging imaging and genetic data, Journal of Radiation Research and Applied Sciences, Volume 18, Issue 3, 2025, 101586, |
| [20] | Tekin, A. & Toktay, B. & Günay, A. C. & Yazgan, H. & İnan, N. G. & Kocadağlı, O. (2024). Bi-Rads Classification in Mammography using Deep Learning. In: Akoğul, S. & Tuna, E. (eds.), Academic Studies with Current Econometric and Statistical Applications. Özgür Publications. |
| [21] | Luo, Y., Wei, J., Gu, Y., Zhu, C., & Xu, F. (2025). Predicting molecular subtype in breast cancer using deep learning on mammography images. Frontiers in oncology, 15, 1638212. |
| [22] | Jeba Prasanna Idas, S., Hemalatha, K., Naveenkumar, J. et al. Recent trends on mammogram breast density analysis using deep learning models: neoteric review. Artif Intell Rev 58, 240 (2025). |
| [23] | K. K. Sreekala, Jayakrushna Sahoo, Enhancing mammogram classification using explainable Conditional Self-Attention Generative Adversarial Network, Expert Systems with Applications, Volume 293, 2025, 128640, |
| [24] | Michael G. Kawooya, Richard Malumba, Alfred Jatho, Andrew Katumba, Joyce Nakatumba-Nabende, Denis Musinguzi, Simon Allan Achuka, Maruf Adewole, Udunna C Anazodo, medRxiv 2025.08.26.25334483; |
| [25] | Al-Hejri, A. M., Sable, A. H., Al-Tam, R. M. et al. A hybrid explainable federated-based vision transformer framework for breast cancer prediction via risk factors. Sci Rep 15, 18453 (2025). |
| [26] | D'Orsi, C. J., Sickles, E. A., Mendelson, E. B., & Morris, E. A. (2013). ACR BI-RADS Atlas, Breast Imaging Reporting and Data System. Reston, VA: American College of Radiology. |
| [27] | Alsolami, A. S., Shalash, W., Alsaggaf, W., Ashoor, S., Refaat, H., & Elmogy, M. (2021). King Abdulaziz University Breast Cancer Mammogram Dataset (KAU-BCMD). Data, 6(11), 111. |
| [28] |
Alsolami, A. (2021). King Abdulaziz University Mammogram Dataset [Data set]. Kaggle.
https://www.kaggle.com/datasets/asmaasaad/king-abdulaziz-university-mammogram-dataset |
| [29] | Pisano, E. D., Zong, S., Hemminger, B. M., DeLuca, M., Johnston, R. E., Muller, K.,... & Pizer, S. M. (1998). Contrast Limited Adaptive Histogram Equalization image processing to improve the detection of simulated spiculations in dense mammograms. Journal of Digital Imaging, 11(4), 193-200. |
| [30] | Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. (2017). Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision. (p. 618-626. |
| [31] | Gal, Y., & Ghahramani, Z. (2016). Dropout as a Bayesian approximation: Representing model uncertainty in deep learning. In International Conference on Machine Learning. pp. 1050-1059. |
| [32] | He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. pp. 770-778. |
APA Style
Chambulilo, K., Wako, P., Bassa, D., Kisonga, A., Shija, E. (2025). Hybrid CNN-Transformer Model for Early Detection and Diagnosis of Breast Cancer: A Multi-Regional Dataset Study with Implications for African Healthcare Settings. American Journal of Artificial Intelligence, 10(1), 1-13. https://doi.org/10.11648/j.ajai.20261001.11
ACS Style
Chambulilo, K.; Wako, P.; Bassa, D.; Kisonga, A.; Shija, E. Hybrid CNN-Transformer Model for Early Detection and Diagnosis of Breast Cancer: A Multi-Regional Dataset Study with Implications for African Healthcare Settings. Am. J. Artif. Intell. 2025, 10(1), 1-13. doi: 10.11648/j.ajai.20261001.11
AMA Style
Chambulilo K, Wako P, Bassa D, Kisonga A, Shija E. Hybrid CNN-Transformer Model for Early Detection and Diagnosis of Breast Cancer: A Multi-Regional Dataset Study with Implications for African Healthcare Settings. Am J Artif Intell. 2025;10(1):1-13. doi: 10.11648/j.ajai.20261001.11
@article{10.11648/j.ajai.20261001.11,
author = {Kasim Chambulilo and Paul Wako and Davis Bassa and Asha Kisonga and Emmanuel Shija},
title = {Hybrid CNN-Transformer Model for Early Detection and Diagnosis of Breast Cancer: A Multi-Regional Dataset Study with Implications for African Healthcare Settings},
journal = {American Journal of Artificial Intelligence},
volume = {10},
number = {1},
pages = {1-13},
doi = {10.11648/j.ajai.20261001.11},
url = {https://doi.org/10.11648/j.ajai.20261001.11},
eprint = {https://article.sciencepublishinggroup.com/pdf/10.11648.j.ajai.20261001.11},
abstract = {Breast cancer remains the most common cancer among African women, contributing to high mortality largely due to late-stage diagnosis. More than 70% of cases in many African countries are detected at advanced stages, when treatment options are limited and survival rates are significantly reduced. Barriers such as limited access to screening, a shortage of radiologists, and socio-economic constraints further delay early detection. Artificial intelligence (AI)-based diagnostic tools offer an opportunity to strengthen screening systems and improve timely diagnosis; however, the lack of publicly available African mammography datasets remains a major challenge. This study introduces MAMMOAI, an AI-driven framework designed to enhance early breast cancer detection with potential applicability to African contexts. Due to the scarcity of large-scale annotated African mammography datasets a critical barrier identified in current literature we adopted a pragmatic approach combining a multi-task deep learning model integrating convolutional neural networks (CNN) with Transformer layers to perform simultaneous risk assessment, cancer detection, staging, risk factor analysis, and differential diagnosis using mammograms. The model was trained using the King AbdulAziz University Breast Cancer Mammogram Dataset (KAU-BCMD) from Saudi Arabia, supplemented with preliminary data collected from Kenyan healthcare facilities. This multi-regional approach was necessitated by insufficient African data volume for robust deep learning model training. The framework employs a multi-task learning approach that integrates a ResNet50–Transformer backbone for spatial feature extraction, feeding five specialized branches for risk assessment, cancer detection, staging, risk factor analysis, and differential diagnosis. All images underwent standardized preprocessing, including resizing to 224×224 pixels, normalization, contrast enhancement, and extensive data augmentation. Weighted categorical cross-entropy losses supported joint optimization across tasks. Model interpretability was ensured using Grad-CAM–based heatmaps and uncertainty estimation, and predictions were compiled into automated, clinician-friendly HTML reports. The model achieved overall accuracy of 98% on the majority class (BI-RADS 1), with macro-averaged F1-scores of 0.61-0.62 across all branches, reflecting challenges in detecting minority classes. Critically, the model failed to identify any BI-RADS 5 (highly malignant) cases, misclassifying all as BI-RADS 4. Grad-CAM visualizations provided interpretable insights, supporting clinical decision-making. While these results demonstrate technical feasibility and the potential of hybrid architectures, they also underscore critical limitations: severe class imbalance, inadequate minority class performance, and unproven generalizability to diverse African populations. Future work must prioritize collection of larger, balanced, multi-center African datasets and external validation across diverse Sub-Saharan African healthcare settings before clinical deployment can be considered.},
year = {2025}
}
TY - JOUR T1 - Hybrid CNN-Transformer Model for Early Detection and Diagnosis of Breast Cancer: A Multi-Regional Dataset Study with Implications for African Healthcare Settings AU - Kasim Chambulilo AU - Paul Wako AU - Davis Bassa AU - Asha Kisonga AU - Emmanuel Shija Y1 - 2025/12/29 PY - 2025 N1 - https://doi.org/10.11648/j.ajai.20261001.11 DO - 10.11648/j.ajai.20261001.11 T2 - American Journal of Artificial Intelligence JF - American Journal of Artificial Intelligence JO - American Journal of Artificial Intelligence SP - 1 EP - 13 PB - Science Publishing Group SN - 2639-9733 UR - https://doi.org/10.11648/j.ajai.20261001.11 AB - Breast cancer remains the most common cancer among African women, contributing to high mortality largely due to late-stage diagnosis. More than 70% of cases in many African countries are detected at advanced stages, when treatment options are limited and survival rates are significantly reduced. Barriers such as limited access to screening, a shortage of radiologists, and socio-economic constraints further delay early detection. Artificial intelligence (AI)-based diagnostic tools offer an opportunity to strengthen screening systems and improve timely diagnosis; however, the lack of publicly available African mammography datasets remains a major challenge. This study introduces MAMMOAI, an AI-driven framework designed to enhance early breast cancer detection with potential applicability to African contexts. Due to the scarcity of large-scale annotated African mammography datasets a critical barrier identified in current literature we adopted a pragmatic approach combining a multi-task deep learning model integrating convolutional neural networks (CNN) with Transformer layers to perform simultaneous risk assessment, cancer detection, staging, risk factor analysis, and differential diagnosis using mammograms. The model was trained using the King AbdulAziz University Breast Cancer Mammogram Dataset (KAU-BCMD) from Saudi Arabia, supplemented with preliminary data collected from Kenyan healthcare facilities. This multi-regional approach was necessitated by insufficient African data volume for robust deep learning model training. The framework employs a multi-task learning approach that integrates a ResNet50–Transformer backbone for spatial feature extraction, feeding five specialized branches for risk assessment, cancer detection, staging, risk factor analysis, and differential diagnosis. All images underwent standardized preprocessing, including resizing to 224×224 pixels, normalization, contrast enhancement, and extensive data augmentation. Weighted categorical cross-entropy losses supported joint optimization across tasks. Model interpretability was ensured using Grad-CAM–based heatmaps and uncertainty estimation, and predictions were compiled into automated, clinician-friendly HTML reports. The model achieved overall accuracy of 98% on the majority class (BI-RADS 1), with macro-averaged F1-scores of 0.61-0.62 across all branches, reflecting challenges in detecting minority classes. Critically, the model failed to identify any BI-RADS 5 (highly malignant) cases, misclassifying all as BI-RADS 4. Grad-CAM visualizations provided interpretable insights, supporting clinical decision-making. While these results demonstrate technical feasibility and the potential of hybrid architectures, they also underscore critical limitations: severe class imbalance, inadequate minority class performance, and unproven generalizability to diverse African populations. Future work must prioritize collection of larger, balanced, multi-center African datasets and external validation across diverse Sub-Saharan African healthcare settings before clinical deployment can be considered. VL - 10 IS - 1 ER -