Early detection of breast cancer is critical for improving patient outcomes, and deep learning (DL) has shown promise in mammographic analysis. However, many existing systems tackle a single task (e.g., malignancy classification) and provide limited interpretability, which hinders clinical adoption. We present a multi-task, attention-guided framework that jointly predicts malignancy (benign vs. malignant) and BI-RADS breast density (A–D) from a single mammogram while producing interpretable attention maps. The architecture couples a pretrained convolutional backbone with lightweight channel–spatial attention and two task-specific heads. We evaluate six backbones (ResNet-50, DenseNet-121, EfficientNet-B0/B3/B7, MobileNetV3-Large) under a rigorous evaluation protocol with stratified hold-out split (64%/16%/20%) and 5-fold cross-validation. To support reliable deployment, we report comprehensive metrics including discrimination measures (ROC-AUC, PR-AUC, F1-score, sensitivity, specificity), calibration metrics (Brier score, Expected Calibration Error), and ablation studies contrasting single-task vs. multi-task learning and models with vs. without attention mechanisms. Across all evaluation protocols, multi-task learning consistently outperforms single-task baselines by 2.8% in pathology accuracy and 3.5% in density estimation, indicating that density cues provide complementary information for malignancy assessment. Among backbones, EfficientNet-B3 attains the strongest discrimination (AUC= 0.962, accuracy= 93.6%) on hold-out test with excellent calibration (Brier= 0.010), while MobileNetV3-Large achieves competitive performance (AUC= 0.940, accuracy= 91.2%) with only 4.26M parameters—an optimal accuracy–efficiency trade-off for resource-constrained deployment. Cross-validation results demonstrate robust generalization (AUC = 0.954 ± 0.003 , accuracy = 93.0 ± 0.3% ). Attention maps highlight diagnostically relevant regions including architectural distortions and suspicious calcifications, offering interpretable visual rationale aligned with radiological assessment. We discuss limitations including dataset size (410 images), class imbalance in rare density categories, and calibration challenges (ECE= 0.365–0.419), and outline directions for external validation on multi-institutional cohorts and integration of uncertainty quantification methods.
Multi-Task Attention-Guided Deep Learning for Simultaneous Breast Cancer Detection and Density Estimation in Mammography
Laura la Paglia;
2025
Abstract
Early detection of breast cancer is critical for improving patient outcomes, and deep learning (DL) has shown promise in mammographic analysis. However, many existing systems tackle a single task (e.g., malignancy classification) and provide limited interpretability, which hinders clinical adoption. We present a multi-task, attention-guided framework that jointly predicts malignancy (benign vs. malignant) and BI-RADS breast density (A–D) from a single mammogram while producing interpretable attention maps. The architecture couples a pretrained convolutional backbone with lightweight channel–spatial attention and two task-specific heads. We evaluate six backbones (ResNet-50, DenseNet-121, EfficientNet-B0/B3/B7, MobileNetV3-Large) under a rigorous evaluation protocol with stratified hold-out split (64%/16%/20%) and 5-fold cross-validation. To support reliable deployment, we report comprehensive metrics including discrimination measures (ROC-AUC, PR-AUC, F1-score, sensitivity, specificity), calibration metrics (Brier score, Expected Calibration Error), and ablation studies contrasting single-task vs. multi-task learning and models with vs. without attention mechanisms. Across all evaluation protocols, multi-task learning consistently outperforms single-task baselines by 2.8% in pathology accuracy and 3.5% in density estimation, indicating that density cues provide complementary information for malignancy assessment. Among backbones, EfficientNet-B3 attains the strongest discrimination (AUC= 0.962, accuracy= 93.6%) on hold-out test with excellent calibration (Brier= 0.010), while MobileNetV3-Large achieves competitive performance (AUC= 0.940, accuracy= 91.2%) with only 4.26M parameters—an optimal accuracy–efficiency trade-off for resource-constrained deployment. Cross-validation results demonstrate robust generalization (AUC = 0.954 ± 0.003 , accuracy = 93.0 ± 0.3% ). Attention maps highlight diagnostically relevant regions including architectural distortions and suspicious calcifications, offering interpretable visual rationale aligned with radiological assessment. We discuss limitations including dataset size (410 images), class imbalance in rare density categories, and calibration challenges (ECE= 0.365–0.419), and outline directions for external validation on multi-institutional cohorts and integration of uncertainty quantification methods.| File | Dimensione | Formato | |
|---|---|---|---|
|
2025 Marat Multi-Task_Attention-Guided_Deep_Learning_for_Simultaneous_Breast_Cancer_Detection_and_Density_Estimation_in_Mammography.pdf
non disponibili
Tipologia:
Versione Editoriale (PDF)
Licenza:
Dominio pubblico
Dimensione
2.43 MB
Formato
Adobe PDF
|
2.43 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.


