Hybrid deep learning for interpretable lung cancer recognition across computed tomography and histopathological imaging modalities

Mahmud, Mohammad Rasel and Fardin, Hasib and Siddiqui, Md Ismail Hossain and Sakib, Anamul Haque and Sakib, Abdullah Al (2025) Hybrid deep learning for interpretable lung cancer recognition across computed tomography and histopathological imaging modalities. International Journal of Science and Research Archive, 15 (1). pp. 1798-1810. ISSN 2582-8185

[thumbnail of IJSRA-2025-1161.pdf] Article PDF
IJSRA-2025-1161.pdf - Published Version
Available under License Creative Commons Attribution Non-commercial Share Alike.

Download ( 1MB)

Abstract

Lung cancer is one of the deadliest cancers globally, primarily due to its silent development and difficulties with late diagnoses. Traditional diagnostic methods like manual CT scans and histopathological slide analysis face challenges such as observer variability, limited sensitivity, and difficulties in handling large volumes of data. Convolutional neural networks (CNNs) can automate image classification, but their limited receptive fields hinder complex tissue structure analysis. Vision Transformers (ViTs) provide a solution but typically need large datasets and significant computing power, making them impractical in clinical settings. Our study introduces a hybrid deep learning (DL) framework using the LEViT architecture to enhance lung cancer classification. We utilized two public datasets: IQ-OTH/NCCD, containing 1,097 CT images categorized into normal, benign, and malignant, and another dataset with 25,000 histopathological images across five tissue types. Our methodology included a multi-stage preprocessing pipeline to resize, reduce noise, enhance contrast, normalize, and augment data to tackle class imbalance and improve generalization. We evaluated our model using metrics like accuracy, F1 score, specificity, PR AUC, and Matthews Correlation Coefficient (MCC) through 10-fold stratified cross-validation. Our LEViT-based model surpassed top models such as CoAtNet and CrossViT, achieving 99.43% accuracy and 98.36% MCC on the IQ-OTH/NCCD dataset, and 99.02% accuracy with 97.97% MCC on the other dataset. Additionally, we developed a real-time web application for clinicians to upload images and receive visual explanations via Grad-CAM, promoting transparency in decision-making. This work provides a scalable, accurate, and explainable AI solution for lung cancer recognition, connecting high-performance algorithms with clinical practice.

Item Type: Article
Official URL: https://doi.org/10.30574/ijsra.2025.15.1.1161
Uncontrolled Keywords: Lung cancer; Vision transformer; Medical imaging; Explainable AI; Diagnostic tool
Depositing User: Editor IJSRA
Date Deposited: 22 Jul 2025 23:28
Related URLs:
URI: https://eprint.scholarsrepository.com/id/eprint/1716