Javascript is required
Agneya, D. A., Shekar, M. S., Bharadwaj, A., Vineeth, N., & Neelima, M. L. (2024). Deep learning in medical image analysis: A survey. In 2024 International Conference on Innovation and Novelty in Engineering and Technology (INNOVA), Vijayapura, India (pp. 1–5). [Google Scholar] [Crossref]
Ahmad, J., Akram, S., Jaffar, A., Rashid, M., & Bhatti, S. M. (2023). Breast cancer detection using deep learning: An investigation using the DDSM dataset and a customized AlexNet and support vector machine. IEEE Access, 11, 108386–108397. [Google Scholar] [Crossref]
Amran, G. A., Alsharam, M. S., Blajam, A. O. A., Hasan, A. A., Alfaifi, M. Y., Amran, M. H., Gumaei, A., & Eldin, S. M. (2022). Brain tumor classification and detection using hybrid deep tumor network. Electronics, 11(21), 3457. [Google Scholar] [Crossref]
Ayana, G., Dese, K., Abagaro, A. M., Jeong, K. C., Yoon, S. D., & Choe, S. W. (2024). Multistage transfer learning for medical images. Artif. Intell. Rev., 57(9), 232. [Google Scholar] [Crossref]
Bhati, D., Neha, F., & Amiruzzaman, M. (2024). A survey on explainable artificial intelligence (XAI) techniques for visualizing deep learning models in medical imaging. J. Imaging, 10(10), 239. [Google Scholar] [Crossref]
Cai, L., Gao, J., & Zhao, D. (2020). A review of the application of deep learning in medical image classification and segmentation. Ann. Transl. Med., 8(11), 713. [Google Scholar] [Crossref]
Chouhan, N., Khan, A., Shah, J. Z., Hussnain, M., & Khan, M. W. (2021). Deep convolutional neural network and emotional learning based breast cancer detection using digital mammography. Comput. Biol. Med., 132, 104318. [Google Scholar] [Crossref]
Deheyab, A. O. A., Alwan, M. H., Rezzaqe, I. K. A., Mahmood, O. A., Hammadi, Y. I., Kareem, A. N., & Ibrahim, M. (2022). An overview of challenges in medical image processing. In Proceedings of the 6th International Conference on Future Networks & Distributed Systems, New York, USA (pp. 511–516). [Google Scholar] [Crossref]
Esteva, A., Kuprel, B., Novoa, R. A., Ko, J., Swetter, S. M., Blau, H. M., & Thrun, S. (2017). Dermatologist-level classification of skin cancer with deep neural networks. Nature, 542(7639), 115–118. [Google Scholar] [Crossref]
He, Y. (2024). Diving deep: The role of deep learning in medical image analysis, today and tomorrow. In 2024 9th International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS) (pp. 537–540). [Google Scholar] [Crossref]
Lehmann, T. M., Gonner, C., & Spitzer, K. (2002). Survey: Interpolation methods in medical image processing. IEEE Trans. Med. Imaging, 18(11), 1049–1075. [Google Scholar] [Crossref]
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, USA (pp. 1135–1144). [Google Scholar] [Crossref]
Saikia, P. & Kalita, S. K. (2024). Alzheimer disease detection using MRI: Deep learning review. SN Comput. Sci., 5(5), 507. [Google Scholar] [Crossref]
Selvaraju, R. R., Cogswell, M., Abhishek, D., Ramakrishna, V., Devi, P., & Dhruv, B. (2020). Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis., 128(2), 336–359. [Google Scholar] [Crossref]
Shamshirband, S., Fathi, M., Dehzangi, A., Chronopoulos, A. T., & Alinejad-Rokny, H. (2021). A review on deep learning approaches in healthcare systems: Taxonomies, challenges, and open issues. J. Biomed. Inform., 113, 103627. [Google Scholar] [Crossref]
Singh, A., Sengupta, S., & Lakshminarayanan, V. (2020). Explainable deep learning models in medical image analysis. J. Imaging, 6(6), 52. [Google Scholar] [Crossref]
Son, J., Shin, J. Y., Kim, H. D., Jung, K. H., Park, K. H., & Park, S. J. (2020). Development and validation of deep learning models for screening multiple abnormal findings in retinal fundus images. Ophthalmology, 127(1), 85–94. [Google Scholar] [Crossref]
Stadlhofer, A. & Mezhuyev, V. (2023). Approach to provide interpretability in machine learning models for image classification. Ind. Artif. Intell., 1(1), 10. [Google Scholar] [Crossref]
Sunil Kumar Aithal, A. & Rajashree. (2023). Deep learning based automated pneumonia detection from X-ray images. In 2023 7th International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India (pp. 664–668). [Google Scholar] [Crossref]
Van der Velden, B. H., Kuijf, H. J., Gilhuijs, K. G., & Viergever, M. A. (2022). Explainable artificial intelligence (XAI) in deep learning-based medical image analysis. Med. Image Anal., 79, 102470. [Google Scholar] [Crossref]
Wang, J., Lu, L., Zhang, Z., & Slam, N. (2022). A novel deep convolution neural network model for CT image classification based on Covid-19. In 2022 3rd International Conference on Computer Vision, Image and Deep Learning & International Conference on Computer Engineering and Applications (CVIDL & ICCEA) (pp. 15–20). [Google Scholar] [Crossref]
Xie, C., Wang, Z., Yang, C., Liu, J., & Liang, H. (2024). Machine learning for detecting atrial fibrillation from ECGs: Systematic review and meta-analysis. Rev. Cardiovasc. Med., 25(1). [Google Scholar] [Crossref]
Zakaria, N., Mohamed, F., Abdelghani, R., & Sundaraj, K. (2021). VGG16, ResNet-50, and GoogLeNet deep learning architecture for breathing sound classification: A comparative study. In 2021 International Conference on Artificial Intelligence for Cyber Security Systems and Privacy (AI-CSP) (pp. 1–6). [Google Scholar] [Crossref]
Zeiler, M. D. & Fergus, R. (2014). Visualizing and understanding convolutional networks. In European Conference on Computer Vision (ECCV) (pp. 818–833). [Google Scholar] [Crossref]
Zhang, H. & Ogasawara, K. (2023). Grad-CAM-based explainable artificial intelligence related to medical text processing. Bioengineering, 10(9), 1070. [Google Scholar] [Crossref]
Zhu, J., Bolsterlee, B., Song, Y., & Meijering, E. (2025). Improving cross-domain generalizability of medical image segmentation using uncertainty and shape-aware continual test-time domain adaptation. Med. Image Anal., 101, 103422. [Google Scholar] [Crossref]
Search
Open Access
Research article

Hybrid Deep Learning Architecture for Automated Chest X-ray Disease Detection with Explainable Artificial Intelligence

B. Vivekanandam1,
Kambala Vijaya Kumar2,
Jagadeeswara Rao Annam3*
1
School of Artificial Intelligence, Computing and Multimedia, Lincoln University College, 47301 Selangor, Malaysia
2
Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, 522502 Andhra Pradesh, India
3
Department of Computer Science and Engineering (Artificial Intelligence and Machine Learning), CVR College of Engineering, 501510 Telangana, India
Healthcraft Frontiers
|
Volume 3, Issue 2, 2025
|
Pages 86-96
Received: 03-13-2025,
Revised: 04-21-2025,
Accepted: 05-04-2025,
Available online: 05-08-2025
View Full Article|Download PDF

Abstract:

Deep learning (DL) has increasingly been adopted to support automated medical diagnosis, particularly in radiological imaging where rapid and reliable interpretation is essential. In this study, a hybrid architecture integrating convolutional neural network (CNN), residual networks (ResNet), and densely connected networks (DenseNet) was developed to improve automated disease recognition in chest X-ray images. This unified framework was designed to capture shallow, residual, and densely connected representations simultaneously, thereby strengthening feature diversity and improving classification robustness relative to conventional single-model or dual-model approaches. The model was trained and evaluated using the ChestX-ray14 dataset, comprising more than 100,000 X-ray images representing 14 thoracic disease classes. Performance was assessed using established metrics, including accuracy, precision, recall, F1-score, and the area under the receiver operating characteristic curve (AUC-ROC). A classification accuracy of 92.5% was achieved, representing an improvement over widely used machine learning (ML) and contemporary DL baselines. To promote transparency and clinical interpretability, Gradient-weighted Class Activation Mapping (Grad-CAM) was incorporated, enhancing clinician confidence in model decisions. The findings demonstrate that DL-based diagnostic support systems can reduce diagnostic uncertainty, alleviate clinical workload, and facilitate timely decision-making in healthcare environments. The proposed hybrid model illustrates the potential of advanced feature-integration strategies to improve automated radiographic interpretation and underscores the importance of explainable artificial intelligence (XAI) in promoting trustworthy deployment of medical artificial intelligence (AI) technologies.
Keywords: Deep learning, Medical diagnosis, AI-based image recognition, Chest X-ray, XAI, Grad-CAM

1. Introduction

AI has helped healthcare by giving new opportunities to recognize medical images. Thanks to DL in ML, it is now possible to spot important features in medical imaging data. With AI, doctors can be confident that they diagnose patients correctly and quickly, alleviating their workload. In this study, the application of DL for automated medical diagnosis, with a focus on AI treatment of X-rays, magnetic resonance imaging (MRI) scans, and computed tomography (CT) scans, is discussed. The initiative aims to present today's advancements in DL in healthcare, explore a more effective approach to diagnosis, and raise awareness of trust and transparency by highlighting XAI.

Medical imaging has long been employed to obtain essential insights into patient health and to support the diagnosis and monitoring of a wide range of diseases. It is best used when guided by the expertise of radiologists and clinicians but can be time-consuming, prone to errors, and require the use of trained professionals. As more medical images are being created, automatic systems must process a large amount of data without compromising their accuracy. Medical imaging classification and detection succeed mostly thanks to DL methods such as CNNs. Many applications in medical imaging have found that CNNs in DL tend to be very successful (H​e​,​ ​2​0​2​4). Because these models can identify patterns in an image that form a hierarchy, they enable the detection of issues such as tumors or fractures by analyzing X-ray and MRI results (W​a​n​g​ ​e​t​ ​a​l​.​,​ ​2​0​2​2). Earlier work on ImageNet enables fine-tuned models to aid in the diagnosis of various medical disorders (Z​a​k​a​r​i​a​ ​e​t​ ​a​l​.​,​ ​2​0​2​1).

Although DL is making significant progress in analyzing medical images, several challenges remain. The absence of well-marked medical data presents numerous challenges. Due to difficulties in privacy and medical imaging, it isn't easy to find large and clearly labeled datasets (L​e​h​m​a​n​n​ ​e​t​ ​a​l​.​,​ ​2​0​0​2). Because DL models are opaque, people are less willing to use them in clinical fields. In addition, the reason why the model has reached its conclusions cannot be explained. Therefore, clinicians struggle to put faith in the diagnoses AI provides when mistakes are made (R​i​b​e​i​r​o​ ​e​t​ ​a​l​.​,​ ​2​0​1​6). Both improving models and making them clear and easy to understand, which can be achieved by XAI methods, are necessary to handle these challenges (Z​e​i​l​e​r​ ​&​ ​F​e​r​g​u​s​,​ ​2​0​1​4).

This study is intended to develop an AI-enabled image recognition system that enhances the reliability of medical image analysis and provides interpretable explanations designed to strengthen clinician confidence in automated diagnostic outputs. In this study, a hybrid model constructed from CNN and ResNet was employed, with features incorporated to explain the model’s actions, such as the use of saliency maps and Grad-CAM. With this process, the system is expected to make accurate predictions and present them in a way that is easy for healthcare providers to explain (S​e​l​v​a​r​a​j​u​ ​e​t​ ​a​l​.​,​ ​2​0​2​0). While XAI has been explored in medical imaging, its role in enhancing clinician trust within chest X-ray diagnosis remains insufficiently addressed. This study contributes by integrating interpretable visualization methods that highlight decision-critical regions, thereby strengthening transparency in automated diagnosis.

1.1 Background

In the medical field, AI and DL have demonstrated success in image classification tasks. Previously, image classification was performed by machines via support vector machines (SVMs) and decision trees (DTs). However, making complex features by hand often prevented these methods from performing well. Thanks to CNNs, DL has now made it easier to identify useful features in images and, in turn, increased the diagnostic accuracy of imaging tools (S​h​a​m​s​h​i​r​b​a​n​d​ ​e​t​ ​a​l​.​,​ ​2​0​2​1). Many medical imaging problems, such as cancer, fractures, and diseases, have been addressed using CNNs, which have achieved results that match or exceed those of human radiologists (A​g​n​e​y​a​ ​e​t​ ​a​l​.​,​ ​2​0​2​4). It has been found that DL can catch lung cancer, breast cancer and brain tumors from CT scans and MRIs in radiology (C​a​i​ ​e​t​ ​a​l​.​,​ ​2​0​2​0). The study by E​s​t​e​v​a​ ​e​t​ ​a​l​.​ ​(​2​0​1​7​) illustrates that computers can classify skin cancer more accurately than dermatologists working with images. Similarly, S​o​n​ ​e​t​ ​a​l​.​ ​(​2​0​2​0​) developed an AI system for accurately identifying diabetic retinopathy in retinal fundus photographs. These advancements have encouraged the continued investigation of DL techniques across a broader range of medical imaging modalities.

On the other hand, DL has not been adopted in healthcare as much as expected due to several problems. Currently, there is a lack of high-quality training data to build reliable systems in DL (D​e​h​e​y​a​b​ ​e​t​ ​a​l​.​,​ ​2​0​2​2) effectively. Medical datasets are often limited in size, incomplete, and affected by substantial class imbalance, which can cause deep learning models to overfit and generalize poorly during clinical deployment. As a result, researchers have been utilizing data augmentation and transfer learning to construct models that can effectively operate with smaller datasets and accurately adapt to various settings and scanning devices (V​a​n​ ​d​e​r​ ​V​e​l​d​e​n​ ​e​t​ ​a​l​.​,​ ​2​0​2​2). A further issue is the comprehensibility of DL models. DL models are recognized for their excellent performance in image recognition; however, it is challenging to understand what enables these models to make their predictions (S​i​n​g​h​ ​e​t​ ​a​l​.​,​ ​2​0​2​0). Because incorrect healthcare diagnoses may be fatal, it is very concerning that AI doesn’t explain itself well. To understand this issue, XAI approaches have been introduced, and saliency maps, class activation maps (CAMs) and Grad-CAM are examples used to highlight which areas of an image matter most to the model (B​h​a​t​i​ ​e​t​ ​a​l​.​,​ ​2​0​2​4). They provide an essential understanding of the model’s decision-making process, which can help clinicians make better medical choices.

Section 2 of this study summarizes relevant studies that utilize DL for medical image recognition. In Section 3, the approach is described, listing the data employed, the hybrid model structure and tools to improve both model accuracy and understanding. The results presented in Section 4 include performance metrics and comparisons with existing models. In the final section, findings are summarized, possible limitations are highlighted, and ideas for further research are presented.

2. Related Work

This prospective observational study was conducted to evaluate the efficacy of empiric antibiotic therapies in DFIs and to develop a visual risk stratification model correlating antibiotic response with potential amputation risk. Ethical approval was obtained from the Postgraduate Medical Institute (PGMI), Hayatabad Peshawar. Additional institutional permissions were granted by the medical directors, deputy medical superintendents (DMS), heads of endocrinology wards, and heads of microbiology sections across participating hospitals. Written informed consent was obtained from all patients in accordance with ethical standards for research involving human participants.

The past few years have seen DL grow rapidly in medical image recognition, and researchers are now applying CNNs and similar models to help automate medical diagnosis. CNNs help identify and recognize diseases in medical images, but concerns remain due to their generalizability, limited data, and lack of explanation. This section summarizes the latest progress in using DL to address challenges in medical imaging. Medical experts are using DL more often to find diseases such as cancer, trouble with the brain and illnesses of the heart. Several DL models have been built to help detect breast cancer using mammography pictures. The method proposed by C​h​o​u​h​a​n​ ​e​t​ ​a​l​.​ ​(​2​0​2​1​) enabled a deep CNN to compete effectively with human radiologists in reviewing mammograms. Thanks to this technique, the model could detect cancerous tumors, making it a valuable tool in informing clinical decisions. In terms of breast cancer screening, the model proposed by A​h​m​a​d​ ​e​t​ ​a​l​.​ ​(​2​0​2​3​) achieved a high area under the curve (AUC) of 0.94 when detecting breast cancer on digital mammograms.

Using DL, doctors have made significant advancements in detecting lung illnesses, particularly pneumonia in X-ray images. S​u​n​i​l​ ​K​u​m​a​r​ ​A​i​t​h​a​l​ ​&​ ​R​a​j​a​s​h​r​e​e​ ​(​2​0​2​3​) used a deep CNN to distinguish between chest X-rays that were normal, pneumonia-positive, or showed signs of tuberculosis. The accuracy for this model was 92%, which is better than many other traditional measurement tools offer. It was found that by applying DL, many diseases could be recognized and immediate diagnoses given, despite not having specialist radiologists present at those sites. In terms of Alzheimer’s, DL has been applied to multiple sclerosis and brain tumors to look at brain MRI scans. S​a​i​k​i​a​ ​&​ ​K​a​l​i​t​a​ ​(​2​0​2​4​) utilized a CNN on MRI images organized into structures to detect Alzheimer’s disease with a success rate of 95%. Findings from brain imaging allowed the model to separate those with Alzheimer’s from healthy people, showing that DL could detect the disease earlier than other approaches. A​m​r​a​n​ ​e​t​ ​a​l​.​ ​(​2​0​2​2​) developed a new method that utilizes DL to detect brain tumors in MRIs. Combining CNNs and recurrent neural networks (RNNs) made it easier for them to detect repeated events and identify tumors, which improved their performance in both segmentation and classification.

Doctors can detect heart problems using DL and identify certain diseases by analyzing echocardiograms and CT scans. X​i​e​ ​e​t​ ​a​l​.​ ​(​2​0​2​4​) developed a novel DL system for detecting atrial fibrillation, a common heart rhythm disorder, by analyzing electrocardiogram (ECG) signals. The traditional method of performing ECGs did not achieve the same sensitivity and specificity as the proposed ECG model. It further showed that DL tools can understand patient signals and help improve the accuracy of doctors’ diagnoses. While DL performs well for many medical image tasks, it remains challenging to ensure that these models function correctly across all medical communities, hospitals, and various types of imaging devices. To enhance the strength and utility of DL models, several studies have employed data augmentation, transfer learning, and domain adaptation. Thanks in large part to data augmentation introduced by Z​h​u​ ​e​t​ ​a​l​.​ ​(​2​0​2​5​), the proposed model could better handle new hospitals and imaging circumstances. In addition, A​y​a​n​a​ ​e​t​ ​a​l​.​ ​(​2​0​2​4​) employed transfer learning to fine-tune pre-trained CNNs for specific medical image classification tasks using a limited amount of data.

In addition to making predictions more accurate, researchers are making sure DL models can be understood by anyone who interprets them. With models in DL being difficult to explain, ongoing research has focused on developing methods that provide meaningful insight into the decision-making processes of AI systems in healthcare. According to S​t​a​d​l​h​o​f​e​r​ ​&​ ​M​e​z​h​u​y​e​v​ ​(​2​0​2​3​), the prediction of any ML model can be described using a new method called Local Interpretable Model-agnostic Explanations (LIME). It allows for precise, local explanations that are very important for clinicians to trust medical systems. Interestingly, Z​h​a​n​g​ ​&​ ​O​g​a​s​a​w​a​r​a​ ​(​2​0​2​3​) utilized Grad-CAM in medical image classification models, identifying the parts of the image with the most significant impact on the model’s outcome. Such XAI methods benefit the clinical community by providing insight into how AI models arrive at their conclusions. Advances in DL for medical image analysis have faced challenges in obtaining sufficient data, utilizing it effectively in other contexts, and understanding its underlying mechanisms. Additionally, healthcare professionals must utilize these models after they have been approved and proven reliable through proper testing. Research on these problems helps ensure that AI-based diagnostic systems can be easily adopted into clinicians’ work processes with confidence.

3. Methodology

This section explains the establishment of a neural network for automatic medical diagnosis in medical images using DL techniques. The methodology is intended to increase the accuracy of models in healthcare while also making their behavior more transparent. The methodology contains four parts: data selection, design of the DL network structure, formulation of the equations, and the model training and evaluation procedures. The unique aspect of this research is the combination of modern DL and XAI approaches to achieve both accuracy and transparency.

3.1 Dataset Description

This research was conducted using the ChestX-ray14 dataset, which is publicly available for detecting pneumonia, tuberculosis, and various diseases in chest X-ray images. The collection of over 100,000 frontal-view chest X-rays, all accompanied by disease information, makes it the ideal choice for testing the proposed DL algorithm. The data includes labels for 14 different diseases, such as pneumonia, tuberculosis and several other respiratory diseases. All images were normalized using dataset-specific mean and standard deviation values, ensuring consistent pixel intensity scaling. Preprocessing also included resizing to 224 × 224 pixels and intensity standardization to reduce variability across imaging devices. Table 1 shows the dataset parameters.

Table 1. Dataset parameters

Parameter

Value

Image count

112,120 images

Disease labels

14 different disease labels

Image dimensions

1024 × 1024 pixels (resized)

Image format

JPEG

Annotation type

Multi-label classification

The data was split into training, validation, and test sets, with proportions of 70%, 15%, and 15%, respectively. As a result, the model was trained with a large number of images, and its performance was checked using unseen data to prevent overfitting. Minority class samples were added to the dataset through data augmentation to address class imbalance. To improve robustness across heterogeneous clinical settings, the dataset preprocessing pipeline was designed to account for variations in imaging devices, exposure levels, and patient positioning. Normalization and augmentation procedures mimic real-world imaging differences, helping the model generalize across diverse acquisition environments.

The dataset contains imbalanced distribution across the 14 disease categories. To mitigate this, augmentation techniques were employed, including ± 15° random rotations, horizontal flips with 0.5 probability, and controlled brightness and contrast adjustments. These steps increased representation of minority disease classes. Below is a representation of chest X-ray images from the dataset.

Table 2 shows the chest X-rays from three categories: pneumonia, tuberculosis and normal chest X-rays. The pneumonia image highlights red circles to indicate where the infection has occurred in the lungs. The image of tuberculosis reveals the usual signs, including abnormal conditions in the lungs. On a normal chest X-ray, the lungs are healthy and show no signs of abnormality. Many DL models in medical diagnostics were trained with these images.

Table 2. Chest X-ray images from the dataset

Pneumonia Image

Tuberculosis Image

Normal Chest X-ray

3.2 DL Architecture

The combination of ResNet and DenseNet was chosen due to their complementary strengths: residual connections preserve gradient flow in deep layers, while dense connectivity enhances feature reuse and representation richness. Together, these structures improve feature extraction efficiency for complex thoracic patterns. The proposed model was constructed with a hybrid approach that includes CNN from ResNet and DenseNet. The hybrid architecture enables the model to identify features more efficiently, with minimal impact on computer power. The design was selected to achieve high performance with reasonable computational demands. The architecture consists of the following components:

• Preprocessing layer: This layer is responsible for resizing the input images to 224 × 224 pixels, normalization, and data augmentation (random rotations, flips, and zooms).

• Base CNN layers: The convolutional layers in the base network extract low-level features such as edges and textures. This layer consists of a stack of convolutional layers with increasing filter sizes (3 × 3, 5 × 5, etc.) and ReLU activation functions.

• Residual blocks: This strategy, which comes from ResNet, helps avoid the gradient problem and passes information smoothly into deeper layers. A block in this network includes two convolutional layers and skips the input to go right into the block’s output.

• DenseNet layers: DenseNet blocks are added after the residual blocks to improve the efficiency of feature reuse. In a DenseNet block, each layer receives input from all previous layers, allowing the model to learn more compact and effective representations.

• Fully connected layer: After the convolutional and dense blocks, the feature maps are flattened and passed through a fully connected layer with dropout regularization to reduce overfitting.

• Output layer: The output layer uses a sigmoid activation function to provide probabilities for each of the 14 disease classes (multi-label classification).

Figure 1 illustrates the architecture of a medical diagnosis system comprising multiple layers. The process begins with a layer that processes the input data. After that, the base CNN extracts essential features from the input. Residual blocks help training become faster and avoid the problem of overfitting. DenseNet layers enable us to share features more efficiently and facilitate the creation of deeper networks. Each feature is combined in the fully connected layer, allowing the output layer to display the diagnosis. Its structure supports purposes such as detecting objects in images and categorizing them appropriately in healthcare. The model treats each disease label as an independent sigmoid output while sharing a common feature backbone. This design prevents unintended coupling of unrelated labels while indirectly capturing correlations through shared feature representations learned by the hybrid architecture.

Figure 1. DL architecture for medical diagnosis

The hybrid architecture is designed as follows:

• Input layer: 224 × 224 × 3 (RGB image input)

• Convolutional layer 1: 64 filters, 3 × 3 kernel, stride 1, and ReLU activation

• Residual block 1: 128 filters, 3 × 3 kernel, and skip connection

• DenseNet block 1: 256 filters, 3 × 3 kernel, and dense connections

• Flattening layer

• Fully connected layer: 1024 units and dropout rate = 0.5

• Output layer: 14 units (for 14 disease categories) and sigmoid activation

As shown in Table 3, the learning rate of 0.0001 and batch size of 32 were selected after iterative tuning experiments using a stepwise search. These values provided the best balance between convergence stability and validation performance when compared with higher and lower parameter ranges. The model employs a multi-label classification approach to predict the presence of multiple diseases simultaneously. The key mathematical formulation of the proposed model is as follows: $X \in R^{224 \times 224 \times 3}$ is the input image; the output of the model, $Y \in R^{14}$, is a vector of probabilities, where each element corresponds to the probability of the presence of a specific disease class.

Table 3. Model hyperparameters

Hyperparameter

Value

Learning rate

0.0001

Batch size

32

Optimizer

Adam

Epochs

50

Loss function

Binary cross-entropy

The model was trained by minimizing the binary cross-entropy loss function:

$\mathrm{L}(Y, \widehat{Y})=-\sum_{i=1}^{14}\left[y_i \log \left(\widehat{y}_i\right)+\left(1-y_i\right) \log \left(1-\widehat{y}_i\right)\right]$
(1)

where, Y is the ground truth label vector (with binary values indicating the presence or absence of diseases), $\widehat{Y}$ is the predicted probability vector for each class, yi is the true label for the i-th disease class, and is the predicted probability for the i-th disease class.

Adam was used as the optimizer, which adapted the learning rate based on the first and second moments of the gradient, and the learning rate was set to 10-4 to ensure stable convergence during training. The following steps outline the training and evaluation process:

Algorithm

1. Data Preprocessing:

Load and resize the chest X-ray images to 224 × 224 pixels.

Apply random transformations such as rotations, flips, and zooms to augment the dataset and reduce overfitting.

Normalize pixel values to the range [0, 1] by dividing by 255.

2. Model Training:

Initialize the model with pre-trained weights from ImageNet using transfer learning.

Fine-tune the model by training it on the medical dataset for 50 epochs with a batch size of 32.

Use the Adam optimizer and binary cross-entropy loss for optimization.

During transfer learning, early convolutional layers inherited from ImageNet were kept frozen in the initial training stage to preserve general visual feature representations. Deeper layers were gradually unfrozen in subsequent epochs, allowing finer adaptation to chest X-ray–specific structures.

Training regularization included early stopping based on validation loss and adaptive learning rate decay, which reduced overfitting and improved model stability during fine-tuning.

3. Model Evaluation:

After completing training, evaluate the model on the validation set using accuracy, precision, recall, F1-score, and AUC-ROC for each type of disease included.

Compute CAMs and Grad-CAMs to illustrate the reasons behind the model’s predictions.

4. Interpretability and Visualization:

Rely on saliency maps and CAMs to discover the essential parts of images that the model looks at when it forms a conclusion.

5. Model Deployment:

• The deployed AI model utilizes chest X-ray images to generate real-time predictions about various diseases.

4. Results and Discussion

This section illustrates the outcomes from the proposed model using chest X-ray images. The model was built to achieve excellent accuracy in detecting multiple diseases from medical images and to explain its predictions in ways that are easy to understand using XAI methods. Key performance indicators were studied, and the achievements were checked against the latest model. The illustrations of the results make them more transparent.

4.1 Evaluation Criteria

Several performance metrics were examined to assess how effectively the proposed model identified diseases from chest X-rays. The evaluation criteria consist of:

• Accuracy: The proportion of correctly classified samples.

• Precision: The proportion of true positive predictions among all positive predictions.

• Recall: The proportion of true positive predictions among all actual positive cases.

• F1-score: The harmonic mean of precision and recall, providing a balance between the two.

• AUC-ROC: This metric assesses the model’s ability to distinguish between different classes.

• Inference time: The time taken for the model to process and provide a prediction for a given image.

These measurements reveal the overall accuracy, precision and generalization of the model for several different diseases.

4.2 Performance Results

Table 4 summarizes the results obtained by the proposed model on the test dataset (15% of the total dataset), which includes 14 disease categories (multi-label classification):

Table 4. Overall model performance

Metric

Value

Accuracy

92.5%

Precision

0.91

Recall

0.93

F1-score

0.92

AUC-ROC

0.96

Inference time

150 ms/image

According to Table 4, the model achieved an accuracy of 92.5% and an F1-score of 0.92, indicating its good performance across each disease category. With an AUC-ROC of 0.96, the model demonstrates excellent performance in identifying positive and negative cases for medical image classification. For a more granular understanding, the performance for each of the 14 disease categories was assessed. The results are presented in Table 5.

Table 5. Per-class performance

Disease

Precision

Recall

F1-score

AUC-ROC

Pneumonia

0.91

0.94

0.92

0.96

Tuberculosis

0.88

0.91

0.89

0.94

Lung cancer

0.94

0.93

0.94

0.97

Cardiomegaly

0.89

0.87

0.88

0.92

Consolidation

0.92

0.90

0.91

0.95

Atelectasis

0.87

0.88

0.87

0.91

Effusion

0.91

0.93

0.92

0.95

Infiltration

0.85

0.89

0.87

0.90

Edema

0.93

0.92

0.92

0.96

Emphysema

0.88

0.89

0.88

0.93

Fibrosis

0.92

0.90

0.91

0.94

Pleural thickening

0.86

0.87

0.86

0.90

Mass

0.90

0.92

0.91

0.95

Nodule

0.89

0.91

0.90

0.94

These per-class metrics show that the model performs particularly well in detecting conditions like lung cancer, pleural effusion, and consolidation, with AUC-ROC values exceeding 0.95 for many of the disease categories. Performance variations across diseases arise from multiple factors, including limited samples for certain categories, subtle disease manifestations, and overlapping radiographic patterns. Conditions such as infiltration and pleural thickening show lower AUC due to inherent visual similarity with other thoracic findings, reflecting challenges related to subtle radiographic characteristics and overlapping visual signatures, which naturally make these classes harder to distinguish.

4.3 Comparison with Existing Models

The performance of the proposed model was evaluated by comparing it with leading models used in medical image classification, including SVM, Random Forest (RF) and DL models like VGG16, ResNet50 and InceptionV3, all trained using the same ChestX-ray14 dataset. Table 6 summarizes the comparison.

Table 6. Comparison with existing models

Model

Accuracy

Precision

Recall

F1-score

AUC-ROC

Proposed model (CNN + ResNet + DenseNet)

92.5%

0.91

0.93

0.92

0.96

VGG16

85.3%

0.83

0.84

0.83

0.89

ResNet50

88.6%

0.85

0.86

0.85

0.91

InceptionV3

89.4%

0.87

0.88

0.87

0.92

SVM

83.2%

0.80

0.81

0.80

0.85

RF

84.1%

0.82

0.83

0.82

0.86

The results in the table demonstrate that the proposed hybrid model outperforms SVM, RF, VGG16, ResNet50, and InceptionV3. More accurate, precise, helpful, and reliable results are noted, revealing how a hybrid CNN-ResNet-DenseNet architecture, combined with XAI, works well.

4.4 Visual Results and Model Interpretability

In healthcare, the ease with which a model can be interpreted is essential, as clarifying its decisions can increase clinicians’ trust and support for the model. This section presents several visualizations from the proposed model, including precision-recall (PR) curves, receiver operating characteristic (ROC) curves, and Grad-CAM heatmaps. They allow the model’s performance to be evaluated across different disease categories and reveal the image regions that the model attends to when generating predictions.

4.4.1 PR curve

A PR curve is used to examine how the balance between precision and recall changes with varying thresholds. Because the model is built for multiple categories, a PR curve was created for each disease group it predicts. As shown in Figure 2, a PR plot is provided for the disease pneumonia as an example.

Figure 2. PR curve for pneumonia detection
4.4.2 ROC curve

Sensitivity against 1-specificity (false positive rate) can be seen on an ROC curve. It is an excellent method for judging how well the model can separate different categories. The AUC-ROC can be determined, which delivers a single number that allows us to compare the model’s results. Figure 3 shows the ROC curve for pneumonia detection.

Figure 3. ROC curve for pneumonia detection
4.4.3 Evaluation of the results with charts

In addition to the performance metrics, visual charts are provided to summarize the model’s classification accuracy and performance across different diseases. The confusion matrix (Figure 4) provides insight into how well the model classifies each disease and helps identify specific diseases where the model may be making more errors.

Figure 4. Confusion matrix for pneumonia detection

5. Conclusion

A DL model was designed in this study to automatically diagnose medical conditions, focusing on AI for image recognition in chest X-rays. With CNN, ResNet, and DenseNet, the model achieved an accuracy of 92.5%, surpassing the accuracies of VGG16 and ResNet50. Moreover, the model achieved impressive results in many disease categories, yielding high precision, recall, and F1-score values. The use of Grad-CAM and other XAI techniques makes the model predictions more straightforward and clinicians more confident in the model's decisions. Nevertheless, the study has a few limitations despite these positive outcomes. First, since the training data dealt only with chest X-ray images, its performance may not be well-suited for other image types commonly used in healthcare. Relying heavily on large, annotated datasets is problematic because data collection on a widespread scale can be challenging due to privacy concerns in specific locations. The model is also limited by its complexity, which could be challenging when trying to use it in real time on devices with few resources.

Improving the model's ability to work well with other data, as well as with hospitals and imaging methods, will be a primary focus moving forward. Incorporation of diverse images in the model, such as those from various locations and people, will help it perform more effectively. Computational efficiency is particularly essential for real-world deployment. Further study on how different forms of data can be combined (such as X-ray images and patient information) may help doctors make faster and more accurate diagnoses. Future work may incorporate multimodal inputs such as patient demographics, clinical history, and laboratory markers. Integrating these features through attention- or fusion-based architectures can enhance diagnostic reliability by combining radiological and clinical evidence. The model maintains an average inference time of approximately 150 ms per image, making it suitable for deployment on standard GPU workstations. The computational footprint supports real-time use in hospital environments without requiring high-end hardware.

Author Contributions

Conceptualization, A.K.P.; methodology, A.K.P.; software, A.K.P.; data curation, A.K.P.; investigation, A.K.P. and B.V.; validation, A.K.P. and B.V.; formal analysis, A.K.P.; resources, B.V.; writing—original draft preparation, A.K.P.; writing—review and editing, A.K.P. and K.V.K.; supervision, K.V.K.; project administration, K.V.K.; medical domain guidance, K.V.K.; All authors have read and agreed to the published version of the manuscript.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References
Agneya, D. A., Shekar, M. S., Bharadwaj, A., Vineeth, N., & Neelima, M. L. (2024). Deep learning in medical image analysis: A survey. In 2024 International Conference on Innovation and Novelty in Engineering and Technology (INNOVA), Vijayapura, India (pp. 1–5). [Google Scholar] [Crossref]
Ahmad, J., Akram, S., Jaffar, A., Rashid, M., & Bhatti, S. M. (2023). Breast cancer detection using deep learning: An investigation using the DDSM dataset and a customized AlexNet and support vector machine. IEEE Access, 11, 108386–108397. [Google Scholar] [Crossref]
Amran, G. A., Alsharam, M. S., Blajam, A. O. A., Hasan, A. A., Alfaifi, M. Y., Amran, M. H., Gumaei, A., & Eldin, S. M. (2022). Brain tumor classification and detection using hybrid deep tumor network. Electronics, 11(21), 3457. [Google Scholar] [Crossref]
Ayana, G., Dese, K., Abagaro, A. M., Jeong, K. C., Yoon, S. D., & Choe, S. W. (2024). Multistage transfer learning for medical images. Artif. Intell. Rev., 57(9), 232. [Google Scholar] [Crossref]
Bhati, D., Neha, F., & Amiruzzaman, M. (2024). A survey on explainable artificial intelligence (XAI) techniques for visualizing deep learning models in medical imaging. J. Imaging, 10(10), 239. [Google Scholar] [Crossref]
Cai, L., Gao, J., & Zhao, D. (2020). A review of the application of deep learning in medical image classification and segmentation. Ann. Transl. Med., 8(11), 713. [Google Scholar] [Crossref]
Chouhan, N., Khan, A., Shah, J. Z., Hussnain, M., & Khan, M. W. (2021). Deep convolutional neural network and emotional learning based breast cancer detection using digital mammography. Comput. Biol. Med., 132, 104318. [Google Scholar] [Crossref]
Deheyab, A. O. A., Alwan, M. H., Rezzaqe, I. K. A., Mahmood, O. A., Hammadi, Y. I., Kareem, A. N., & Ibrahim, M. (2022). An overview of challenges in medical image processing. In Proceedings of the 6th International Conference on Future Networks & Distributed Systems, New York, USA (pp. 511–516). [Google Scholar] [Crossref]
Esteva, A., Kuprel, B., Novoa, R. A., Ko, J., Swetter, S. M., Blau, H. M., & Thrun, S. (2017). Dermatologist-level classification of skin cancer with deep neural networks. Nature, 542(7639), 115–118. [Google Scholar] [Crossref]
He, Y. (2024). Diving deep: The role of deep learning in medical image analysis, today and tomorrow. In 2024 9th International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS) (pp. 537–540). [Google Scholar] [Crossref]
Lehmann, T. M., Gonner, C., & Spitzer, K. (2002). Survey: Interpolation methods in medical image processing. IEEE Trans. Med. Imaging, 18(11), 1049–1075. [Google Scholar] [Crossref]
Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, USA (pp. 1135–1144). [Google Scholar] [Crossref]
Saikia, P. & Kalita, S. K. (2024). Alzheimer disease detection using MRI: Deep learning review. SN Comput. Sci., 5(5), 507. [Google Scholar] [Crossref]
Selvaraju, R. R., Cogswell, M., Abhishek, D., Ramakrishna, V., Devi, P., & Dhruv, B. (2020). Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis., 128(2), 336–359. [Google Scholar] [Crossref]
Shamshirband, S., Fathi, M., Dehzangi, A., Chronopoulos, A. T., & Alinejad-Rokny, H. (2021). A review on deep learning approaches in healthcare systems: Taxonomies, challenges, and open issues. J. Biomed. Inform., 113, 103627. [Google Scholar] [Crossref]
Singh, A., Sengupta, S., & Lakshminarayanan, V. (2020). Explainable deep learning models in medical image analysis. J. Imaging, 6(6), 52. [Google Scholar] [Crossref]
Son, J., Shin, J. Y., Kim, H. D., Jung, K. H., Park, K. H., & Park, S. J. (2020). Development and validation of deep learning models for screening multiple abnormal findings in retinal fundus images. Ophthalmology, 127(1), 85–94. [Google Scholar] [Crossref]
Stadlhofer, A. & Mezhuyev, V. (2023). Approach to provide interpretability in machine learning models for image classification. Ind. Artif. Intell., 1(1), 10. [Google Scholar] [Crossref]
Sunil Kumar Aithal, A. & Rajashree. (2023). Deep learning based automated pneumonia detection from X-ray images. In 2023 7th International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India (pp. 664–668). [Google Scholar] [Crossref]
Van der Velden, B. H., Kuijf, H. J., Gilhuijs, K. G., & Viergever, M. A. (2022). Explainable artificial intelligence (XAI) in deep learning-based medical image analysis. Med. Image Anal., 79, 102470. [Google Scholar] [Crossref]
Wang, J., Lu, L., Zhang, Z., & Slam, N. (2022). A novel deep convolution neural network model for CT image classification based on Covid-19. In 2022 3rd International Conference on Computer Vision, Image and Deep Learning & International Conference on Computer Engineering and Applications (CVIDL & ICCEA) (pp. 15–20). [Google Scholar] [Crossref]
Xie, C., Wang, Z., Yang, C., Liu, J., & Liang, H. (2024). Machine learning for detecting atrial fibrillation from ECGs: Systematic review and meta-analysis. Rev. Cardiovasc. Med., 25(1). [Google Scholar] [Crossref]
Zakaria, N., Mohamed, F., Abdelghani, R., & Sundaraj, K. (2021). VGG16, ResNet-50, and GoogLeNet deep learning architecture for breathing sound classification: A comparative study. In 2021 International Conference on Artificial Intelligence for Cyber Security Systems and Privacy (AI-CSP) (pp. 1–6). [Google Scholar] [Crossref]
Zeiler, M. D. & Fergus, R. (2014). Visualizing and understanding convolutional networks. In European Conference on Computer Vision (ECCV) (pp. 818–833). [Google Scholar] [Crossref]
Zhang, H. & Ogasawara, K. (2023). Grad-CAM-based explainable artificial intelligence related to medical text processing. Bioengineering, 10(9), 1070. [Google Scholar] [Crossref]
Zhu, J., Bolsterlee, B., Song, Y., & Meijering, E. (2025). Improving cross-domain generalizability of medical image segmentation using uncertainty and shape-aware continual test-time domain adaptation. Med. Image Anal., 101, 103422. [Google Scholar] [Crossref]

Cite this:
APA Style
IEEE Style
BibTex Style
MLA Style
Chicago Style
GB-T-7714-2015
Vivekanandam, B., Kumar, K. V., & Annam, J. R. (2025). Hybrid Deep Learning Architecture for Automated Chest X-ray Disease Detection with Explainable Artificial Intelligence. Healthcraft. Front., 3(2), 86-96. https://doi.org/10.56578/hf030203
B. Vivekanandam, K. V. Kumar, and J. R. Annam, "Hybrid Deep Learning Architecture for Automated Chest X-ray Disease Detection with Explainable Artificial Intelligence," Healthcraft. Front., vol. 3, no. 2, pp. 86-96, 2025. https://doi.org/10.56578/hf030203
@research-article{Vivekanandam2025HybridDL,
title={Hybrid Deep Learning Architecture for Automated Chest X-ray Disease Detection with Explainable Artificial Intelligence},
author={B. Vivekanandam and Kambala Vijaya Kumar and Jagadeeswara Rao Annam},
journal={Healthcraft Frontiers},
year={2025},
page={86-96},
doi={https://doi.org/10.56578/hf030203}
}
B. Vivekanandam, et al. "Hybrid Deep Learning Architecture for Automated Chest X-ray Disease Detection with Explainable Artificial Intelligence." Healthcraft Frontiers, v 3, pp 86-96. doi: https://doi.org/10.56578/hf030203
B. Vivekanandam, Kambala Vijaya Kumar and Jagadeeswara Rao Annam. "Hybrid Deep Learning Architecture for Automated Chest X-ray Disease Detection with Explainable Artificial Intelligence." Healthcraft Frontiers, 3, (2025): 86-96. doi: https://doi.org/10.56578/hf030203
VIVEKANANDAM B, KUMAR K V, ANNAM J R. Hybrid Deep Learning Architecture for Automated Chest X-ray Disease Detection with Explainable Artificial Intelligence[J]. Healthcraft Frontiers, 2025, 3(2): 86-96. https://doi.org/10.56578/hf030203
cc
©2025 by the author(s). Published by Acadlore Publishing Services Limited, Hong Kong. This article is available for free download and can be reused and cited, provided that the original published version is credited, under the CC BY 4.0 license.