A CNN–Transformer Fusion Approach Integrating Texture Encoding and Cross-Patch Attention for Efficient Histopathological Breast Cancer Classification - 10/03/26
, Jenish S. Alden b
, Ramamurthy Karthik c, ⁎
, Kulanthaivelu Suganthi c 
Abstract |
Objectives |
Breast cancer remains one of the most prevalent and life-threatening malignancies worldwide, wherein accurate and early diagnosis plays a pivotal role in improving patient outcomes. The screening is commonly done by traditional means, which tend to be time-consuming and labor-intensive. The detection relies on expert review, sometimes leading to issues of subjectivity, delayed diagnosis, and treatment.
Methods |
This study presents a novel Deep Learning (DL) approach, designed to improve texture-structure understanding for breast cancer classification. In the proposed approach, Res-MorphNet synergistically combines ResNet with a Morphological Texture Encoding (MTE) module to extract intricate morphologic textures, while Swin-CPSANet leverages the Swin Transformer and a Cross-Patch Spatial Aggregation (CPSA) block to enhance global context understanding and feature interaction. To the best of our knowledge, this is the first attempt to integrate morphological texture encoding with cross-patch attention within a unified CNN–Transformer fusion framework for histopathological breast cancer classification.
Results |
Evaluation on the BACH dataset indicates that the proposed model achieves an accuracy of 95%, a precision of 95.49%, a recall of 95%, and an F1 score of 95.05%. These results outperform traditional CNN- and transformer-based baseline models, demonstrating the effectiveness of combining morphological texture encoding with cross-patch attention in a unified framework.
Conclusion |
The findings indicate that the proposed architecture achieved efficient and well-balanced classification performance across multiple breast tissue classes, underscoring its potential as a valuable tool to assist in clinical diagnosis.
Le texte complet de cet article est disponible en PDF.Graphical abstract |
Highlights |
• | A novel deep learning network with Res-MorphNet and Swin-CPSANet model is proposed. |
• | MTE block extracts local textures, while CPSA block captures global context. |
• | Cross and collaborative attention unify dual-branch features. |
• | The model achieves 95.00% accuracy on the BACH dataset. |
Keywords : CNN, Deep learning, Breast cancer classification, Histopathology images, ResNet, Transformer
Plan
Vol 47 - N° 2
Article 100937- avril 2026 Retour au numéroBienvenue sur EM-consulte, la référence des professionnels de santé.
L’accès au texte intégral de cet article nécessite un abonnement.
Déjà abonné à cette revue ?
