Skip to main content

PulmoNet: a novel deep learning based pulmonary diseases detection model

Abstract

Pulmonary diseases are various pathological conditions that affect respiratory tissues and organs, making the exchange of gas challenging for animals inhaling and exhaling. It varies from gentle and self-limiting such as the common cold and catarrh, to life-threatening ones, such as viral pneumonia (VP), bacterial pneumonia (BP), and tuberculosis, as well as a severe acute respiratory syndrome, such as the coronavirus 2019 (COVID-19). The cost of diagnosis and treatment of pulmonary infections is on the high side, most especially in developing countries, and since radiography images (X-ray and computed tomography (CT) scan images) have proven beneficial in detecting various pulmonary infections, many machine learning (ML) models and image processing procedures have been utilized to identify these infections. The need for timely and accurate detection can be lifesaving, especially during a pandemic. This paper, therefore, suggested a deep convolutional neural network (DCNN) founded image detection model, optimized with image augmentation technique, to detect three (3) different pulmonary diseases (COVID-19, bacterial pneumonia, and viral pneumonia). The dataset containing four (4) different classes (healthy (10,325), COVID-19 (3,749), BP (883), and VP (1,478)) was utilized as training/testing data for the suggested model. The model’s performance indicates high potential in detecting the three (3) classes of pulmonary diseases. The model recorded average detection accuracy of 94%, 95.4%, 99.4%, and 98.30%, and training/detection time of about 60/50 s. This result indicates the proficiency of the suggested approach when likened to the traditional texture descriptors technique of pulmonary disease recognition utilizing X-ray and CT scan images. This study introduces an innovative deep convolutional neural network model to enhance the detection of pulmonary diseases like COVID-19 and pneumonia using radiography. This model, notable for its accuracy and efficiency, promises significant advancements in medical diagnostics, particularly beneficial in developing countries due to its potential to surpass traditional diagnostic methods.

Peer Review reports

Introduction

Recently, the human race has been threatened by a high mortality rate due to pulmonary diseases. In the year 2020 alone, over 3.5 million pulmonary diseases such as Covid-19-related deaths were recorded worldwide, making it the third source of demise worldwide, with 80% of these demises, predominantly in developing and low to middle-income countries [1, 2]. Pulmonary disease is deadly because it affects the respiratory tissues and organs, especially the lungs, making breathing challenging for animals. Virtually all varieties of pulmonary disease share similar indications, including fever, dry cough, fatigue, and rapid shallow breathing [3].

Lung diseases posed enormous risk by causing around 4 million premature deaths annually until late 2019, when the risk factor doubled due to the outburst of the new coronavirus illness (Covid-19) in Wuhan, China [4, 5]. Since then, over 202 million cases have been reported, 4.2 million deaths, and an average daily cases of about 750,000 as of August 2021 [6]. Considering the transience degree owing to lung ailment infections and the contagious nature of some of these diseases, coupled with time-consuming and expensive diagnosis procedures, there is an urgent need for responsive diagnostic and detection systems, as early detection could save many lives by preventing premature deaths.

One promising approach to the automated diagnosis of illnesses recently is artificial intelligence (AI) [7,8,9]. It has been revolutionizing the medical and healthcare industries in many ways, such as drug research and medical imagery, with a reasonable level of accuracy [10]. Two major techniques embody AI: ML and deep learning (DL). It has shown its efficacy by solving problems ranging from natural language processing to image classification using different DL and ML models. It makes predictions and inferences by analyzing the large quantity of input data, performs intelligent tasks such as feature detection, pattern recognition, translation, and perceptron on the data, and then makes the relevant decision [11,12,13]. Because of the high performance and other proven efficient qualities of AI in image identification and classification tasks, many researchers have conducted studies linking AI-founded techniques to the classification and prediction of different pulmonary diseases using either CT or X-ray imageries [14,15,16,17].

The challenge of possible overlap of symptoms of pulmonary diseases makes diagnosis difficult, especially when there is a shortage of experienced personnel or an absence of a patient’s health record at hand. This scenario prompted for creation of an automated diagnosis process that can accurately detect the presence of any of the three pulmonary diseases earlier mentioned. To the authors’ knowledge, this investigation is the first to utilize AI techniques to classify and detect three pulmonary diseases.

For this reason, this investigation, therefore, suggested a DCNN-based image detection model, specifically, a CNN optimized with an image augmentation approach, to detect three (3) different pulmonary diseases (COVID-19, BP, and VP). CT and X-ray imageries were utilized to have more image datasets, enough to make concrete inferences and enhance the model’s detection ability, as CT can detect abnormal features in the chest images even before symptoms appear.

The core contribution of this study is as follows:

  1. 1.

    Proposed a novel PulmoNet model based on the DCNN model capable of performing multiclass image classification utilizing CT scans and X-ray images.

  2. 2.

    Use of an image augmentation technique to improve the proposed DCNN model

  3. 3.

    Use of DCNN models to classify and detect three pulmonary diseases.

  4. 4.

    Classification of pulmonary diseases into four (4) classes, three (3) classes, two (2) classes (Covid_19 and Healthy), and two (2) classes (Pneumonia and Healthy).

  5. 5.

    Evaluation of the proposed system using confusion matrix performance measures such as precision, recall, and f1-score

  6. 6.

    Comparison of the suggested PulmoNet model with existing systems

The research aimed to create and evaluate a specialized deep convolutional neural network (DCNN) model for accurately detecting various lung illnesses by analyzing radiographic imaging methods, such as X-rays and CT scans. The focus of our study was on carefully constructing and perfecting this model. The process included thorough assessments of the model’s accuracy in detecting illnesses such as COVID-19, bacterial pneumonia, and viral pneumonia. In addition, the research sought to compare the diagnostic capabilities of the DCNN model with those of traditional approaches to acquire a thorough grasp of its effectiveness and future enhancements in the area of medical diagnostics.

The remainder of this article is prearranged thus: “Related literature” section emphasizes the analysis of related literature; “Methodology” section discusses the research methods, including the materials, various datasets, and their repositories. “Results and discussion” section demonstrates the experimented outcomes and discussion, and lastly, “Discussion” section concludes the investigation.

Related literature

Several research works have experimented with AI in healthcare to improve the conventional healthcare system and industry [18]. The need for an affordable, efficient and accurate diagnostic system that can serve as an alternative to the existing traditional time-consuming and expensive diseases diagnosis, especially pulmonary diseases, triggered an upsurge in AI-in-healthcare research [19,20,21]. To investigate the relationship between AI schemes and diseases detection, researchers have conducted investigations ranging from predicting future diseases using patient’s previous historical health records to detecting and identifying possible ailments using CT and X-ray imageries [10, 15, 22, 23].

Due to the similarity of symptoms of pulmonary diseases, experts sometimes misdiagnose one for the other [24]. For instance, when the coronavirus 2019 (COVID-19) pandemic broke out officially in Wuhan, China, around December 2019 [1, 25,26,27], it was initially thought to be severe acute pneumonia, probably severe acute respiratory syndrome (SARS) or middle east respiratory syndrome (MERS) until it was established as novel COVID-19. Because of its novelty, there were challenges in testing and diagnosis as it is highly infectious. Many health workers have been unconsciously infected while the medical services provisioning and the testing kits were either unavailable or expensive and time-consuming [14]. This necessitates an alternative method of virus detection, of which AI has shown a great prospect, as pointed out by Haenssle et al. [25]; Mar & Soyer [28].

Radiological images obtained from CT and X-rays have similarly been a great source of data for training AI-founded disease recognition and analysis systems to achieve swift pulmonary disease diagnosis and management [29]. CT and X-ray imageries have been identified as sensitive means of detecting pneumonia in COVID-19 patients and can be exploited as screening equipment sideways with reverse transcription-polymerase chain reaction (RT-PCR), the utmost popular COVID-19 recognition technique [30]. In Pan et al. [30], changes were observed in COVID-19 pneumonia survivors’ lungs as revealed by CT scan imageries ten days after the start of the indications. Chan et al. [31] also pointed out that changes were observed in CT and X-ray imageries of the chest long before clinical symptoms of COVID-19 appeared. These changes can vary from right infrahilar space opacity, as in Kong & Agarwal [32]), to round lung opacities Zu et al. [33] and ground-glass opacities (GGO) or mixed GGO in so many patients [33]. Furthermore, it was identified that one out of three patients studied had one nodular opacity at the lung’s lower left region, whereas others had up to four and five irregular opacities in the two lungs [34,35,36].

Of all the AI schemes that have been adopted in healthcare systems, DL-based techniques have been promising [37,38,39]. One DL-based method popularly adopted for various disease recognition exploiting CT scans and/or X-ray imageries is CNN [40, 41]. It is biologically inspired as its neuron connectivity pattern resembles the animal visual cortex [42, 43]. CNN is the best commonly used technique for image categorization because it relatively optimizes the kernels via automated learning and requires little preprocessing [44, 45].

CNN Pulmonary disease detection systems have been well-researched in recent years due to the COVID-19 global pandemic. Although so many excellent research works on using CT [46] and X-ray imageries to detect pulmonary diseases have been published, they are not yet an alternative to the traditional methods of diagnosis. However, they have promised to be a helpful association with the traditional diagnosis methods, thereby creating a huge avenue for research and amelioration before commercialization. (Some of the research works focusing on DL-based pulmonary diseases detection using features from CT imageries and X-ray imageries of in-patients include the work of [47]. In their work, a deep feature network, specifically ResNet50, was combined with a support vector machine (SVM) to identify patients infected with COVID-19 and classify normal patients from infected ones. Although their work achieved an accuracy of about 95.4%, it only focused on a single class detection (COVID-19). Similarly, Hemdan et al. [48] proposed a DL-based classifier (COVIDX net) to categorize two classes of 50 clinical X-ray imageries of COVID-19 and normal chest images. Although their model accomplished about 90% accuracy, the dataset utilized to train and test the suggested techniques is not enough for the model to generalize.

In another work, Song et al. [49] established a classifier (DRE-Net) for COVID-19 and non-COVID-19 categorization utilizing CT imageries. The resulting technique has an accuracy of about 86% but only focuses on two-class classification with only 88 CT scan images. Also, in Wang et al. [50], a two-class classification system was suggested using CT scan images. An M-inception model was modified and tested on the obtained pathogen-confirmed COVID-19 images, and the resulting model had an average of about 84.4% accuracy.

Apart from these previous works, other research focuses on more than two classes of pulmonary disease detection and classification. Some of these works include Narin et al. [51], which suggested a DCNN ResNet-50 technique to categorize four classes of (COVID-19, normal (healthy), VP, and BP). The technique accomplished an accuracy of about 98% for two-class classification. Also, Ozturk et al. [52] proposed that the DarkCovidNet model performs two and three-class classifications, and the technique achieved an average accuracy of about 90.6% for both class classifications. Khan et al. [53] proposed that Xception-base Coronet was exploited for multiclass categorization by exploiting X-ray imageries.

This study focuses on building, training, and evaluating the effectiveness of the suggested PulmoNet founded on the CNN model and capable of performing multiclass image classification exploiting CT scans and X-ray imageries. It comprises a combination of convolutional, dense, max pooling, and flattened layers combined with four (4) different activation functions: ReLU, Sigmoid, Leaky ReLU, and Tanh.

The current body of research in artificial intelligence (AI) and healthcare primarily emphasizes the creation of cost-effective, streamlined, and precise diagnostic tools, with a particular emphasis on pulmonary disorders. Research encompasses a wide range of activities, including predicting illnesses based on historical health information and using CT and X-ray images for disease diagnosis. There is a strong focus on the need for alternative approaches because of the incorrect identification of comparable symptoms in respiratory disorders, which was brought to attention during the COVID-19 pandemic. Research has used radiological images to teach artificial intelligence systems, demonstrating their efficacy in rapid diagnosis. Deep Learning (DL) methods, particularly Convolutional Neural Networks (CNN), have been significant in these investigations, with CNN being a favoured option because of its practical picture classification skills. Several studies have investigated deep learning (DL) algorithms for identifying lung illnesses utilizing CT and X-ray images, with specific emphasis on detecting COVID-19. These investigations have shown encouraging outcomes. However, they have not reached comparability with conventional diagnostic techniques. This suggests that further study and development are necessary.

Methodology

The methodology employed to achieve the aim and objectives of this work is in three stages, which include:

  1. a.

    Data collection

  2. b.

    Data preprocessing and splitting

  3. c.

    Model design, training, and evaluation

Data collection

The dataset for this investigation was obtained from the combination and modification of different Covid-19, VP, BP, and healthy data repositories. These include the Actualmed-covid-chest x-ray dataset [54], SARS-COV-2 CT-scan dataset [55], x-ray dataset of Covid-19 Pneumonia detector and Covid-19 Radiography Database on Kaggle [56, 57], Covid-19 image dataset of Cohen et al. [55] and so on. In the end, 16,435 image datasets comprised of 883 bacterial pneumonia images, 1,478 viral pneumonia images, 3,749 covid-19 images, and 10,325 healthy images were generated. Sample imageries for each of the classes are presented in Figs. 1, 2, 3 and 4.

Fig. 1
figure 1

Sample of chest X-ray image of a BP patient

Fig. 2
figure 2

Sample of chest X-ray image of a non-Covid VP patient

Fig. 3
figure 3

Sample of chest X-ray image of a Covid-19 patient

Fig. 4
figure 4

Sample of chest X-ray image of a healthy person

Figure 1 presents an x-ray imageries of a BP patient. The figure shows a focal opacity that indicates consolidation without cavitation on the right lung’s upper lobe. The cardio-phrenic and costo-phrenic angles can be seen to be free of air-fluid level.

Figure 2 presents an x-ray image of a non-Covid VP patient. The image demonstrates extended two-sided ground glass opacities (GGO), which indicates extended septal thickenings and alveolar mutilation. Similarly, in Fig. 3, the chest x-ray of a Covid-19 patient is presented. The radiography image shows hyperlucent lung fields, which specify lung hyperinflation owing to such small airway obstruction caused by bronchitis. Other features, such as manifold patchy opacities in both lung fields and blunting of costo-phrenic angles, can also be observed in the image.

Figure 4 presents the x-ray representation of a healthy person. The radiography chest image shows the normal shape and size of the chest wall in a normal situation and with a normal presence of the trachea, mediastinum, and heart. No focal opacity, air-fluid level, abnormal bronco-vascular markings, or hyperlucency is observed. Both cardio-phrenic and costo-phrenic angles are also unrestrained.

Description of dataset

The dataset used for the study consists of 16,435 image datasets which are divided into four categories: bacterial pneumonia, viral pneumonia, COVID-19, and healthy cases. Here are the details of the dataset:

  • Bacterial Pneumonia (BP): 883 images.

  • Viral Pneumonia (VP): 1,478 images.

  • COVID-19: 3,749 images.

  • Healthy: 10,325 images.

These images were sourced from various public data repositories and modified to create a comprehensive dataset for the study. Notably, the repositories included the Actualmed-COVID-chest x-ray dataset, SARS-COV-2 CT-scan dataset, a COVID-19 Pneumonia detector x-ray dataset, and the COVID-19 Radiography Database available on Kaggle. Additionally, the COVID-19 image dataset from Cohen et al. was also utilized.

The images for each class were presented in the study to demonstrate typical features associated with each condition. For example, the BP patient images showed focal opacity without cavitation in the right lung’s upper lobe, indicative of bacterial pneumonia.

For the model training and testing, the dataset was divided into 85% for training and 15% for testing. The dataset’s distribution across various classes was instrumental in training the model to distinguish between the different pulmonary diseases and healthy cases, demonstrating the effectiveness of the model in terms of accuracy and potential for real-world application.

The comprehensive dataset was vital in achieving the high detection accuracies reported in the study: 94% for 4-class classification, 95.4% for 3-class, 99.4% for COVID-19 vs. Healthy, and 98.3% for Pneumonia vs. Healthy. These results indicate the model’s proficiency compared to traditional pulmonary disease recognition methods using x-ray and CT scan images.

Data preprocessing and splitting

After the dataset is generated, the sizes of x-ray images are reduced by rescaling to obtain a faster model training process. The rescaled images are then converted to greyscale. The obtained preprocessed dataset is then subdivided into three subsets for 4 class, 3 class, and 2 class classifications correspondingly. Tables 1, 2 and 3 show the detail of each class classification and the sum of images utilized for training and testing each class. Finally, the dataset is split into 85% training and 15% testing samples.

Table 1 Summary of the parameters used for the PulmoNet model
Table 2 Accuracy per class (4 classes)
Table 3 Accuracy per class (3 classes)

Significant disparity in the number of input images for each class have been handled using image augmentation method as it is a valuable strategy for addressing data imbalance in CNN models. By artificially expanding the training dataset with augmented samples, it helps to alleviate the challenges posed by imbalanced class distributions. Various image augmentation techniques employed to introduce diversity and variability into the training data and in turn improve the detection ability of our model include:

  1. 1.

    Rotation: Rotating images within a certain range to simulate different angles of view.

  2. 2.

    Translation: Shifting images horizontally or vertically to simulate different positions within the frame.

  3. 3.

    Scaling: Rescaling images to different sizes, allowing the model to learn robustness to variations in object sizes. And

  4. 4.

    Flipping: Mirroring images horizontally or vertically to create additional variations.

Model design, training, and evaluation

This study’s proposed model (PulmoNet) is an improved 26-layer CNN-based model motivated by a wide residual network (WRN) architecture. This version of CNN was considered because it takes a shorter training time due to its shallow nature. WRN solves problems faster by adding shortcuts between its various layers, thereby preventing distortion that can occur as the network becomes complex. The model was implemented using Keras, Tensorflow, and Jupyter notebook. The training was done using an NVIDIA K80 graphics processing unit (GPU) with and batch size of 16, an epoch of 50, a learning rate of 0.001, and Adam optimizer were set as training parameters. Batch normalization was employed to stabilize the technique by standardizing the input size. Various hyperparameter combinations experimented with convolution, and max pool layers gradually increased as data shuffling was enabled at each epoch. The effectiveness of the examined approach was observed and recorded each time these layers were increased until a better and more stable outcome was recorded. All these were done to obtain a sensitive approach that can obtain any slight change in features of the input images.

In the end, the final model consists of 26 layers in total, including convolutional layers (‘Conv2D’), max-pooling layers (‘MaxPooling2D’), a flatten layer, and densely connected layers (‘Dense’). The activation function used is ReLU, except for the output layer where softmax activation is used for multi-class classification. The model is compiled with the Adam optimizer and categorical cross-entropy loss. Summary of parameters of proposed PulmoNet model is presented in Table 1. It can be demonstrated that the model input images with input shape (64, 64, 1) and loss function used for the implementation is Categorical Cross_entropy. A learning rate of 0.001 was set for the experiment with adam optimizer. The activation functions used was ReLU and Softmax at the dense layer. While training the model the metric used was accuracy.

An n-fold cross-validation technique was exploit to assess the effectiveness of the suggested classification technique by randomly dividing the training dataset into three (3) equal parts. Two were utilized for training, and the remaining was utilized for validation. Shifting of the testing and validation sets was strategically repeated continuously Five (5) times. The overall performance of the suggested approach was assessed by computing the average values from each fold.

Finally, the suggested approach’s effectiveness was evaluated by exploiting a system of measurement which includes accuracy (Acc), precision (Prec), recall (Rec), f1-score, and confusion matrix.

PulmoNet mathematical background

PulmoNet is a deep CNN-based model developed to classify multiclass images using CT scans and X-ray images. It is a 26 layers model motivated by WRN architecture which introduces the concept of widening the network by increasing the number of channels in the convolutional layers compared to traditional Residual Networks (ResNets). The key idea is to have a wider network with an increased capacity to capture more complex patterns. An overview of the mathematical representation of PulmoNet is presented thus:

Assuming input images of size (W, H, C), where W is the width, H is the height, and C is the number of color channels.

And convolutional layers performing convolution operation such that:

  • Each convolutional layer applies a set of filters to the input image.

  • Mathematically, the convolution operation can be defined as:

$$C\_i=A\_i\left(Conv2D\left(F\_i,K\_i\right)\left(C\_\left\{i-1\right\}\right)\right)$$
(1)

where C_i is the output feature map of the i-th convolutional layer, F_i is the filter/kernel of size K_i, and A_i is the activation function applied to the output.

The main building blocks of the WRN are the Wide Residual Blocks, which contain multiple convolutional layers and skip connections.

Each Wide Residual Block consists of the following operations:

  • Batch Normalization which normalizes the activations of the previous layer.

  • Activation which applies an activation function, such as ReLU.

  • Convolution which applies a series of convolutional layers with increased channel size. And

  • Skip Connection which adds the input of the block to the output of the block, creating a shortcut connection.

Mathematically, the Wide Residual Block is represented thus:

$$B\_i=SkipConnection\left(BatchNormalization\left(Activation\left(Convolution\left(F\_i,K\_i\right)\right)\right)\right)+C\_\left\{i-1\right\}$$
(2)

Where B_i is the output of the i-th Wide Residual Block and \(C\_\left\{i-1\right\}\) is the input of the block.

After the Wide Residual Blocks, a global average pooling operation is performed to aggregate the spatial information into a single vector.

Mathematically, the global average pooling operation is represented as:

$$GAP=GlobalAveragePooling2D()\left(B\_n\right)$$
(3)

Finally, a fully connected layer is applied to process the aggregated features.

It is Mathematically represented as thus:

$$D=Dense\left(U,activation=A\right)\left(GAP\right)$$
(4)

The softmax activation function is applied to the output layer to obtain class probabilities and it is represented mathematically as:

$$S=Soft{\text{max}}\left(D\right)$$
(5)

Categorical cross-entropy loss function is then applied and represented mathematically as thus:

$$L=CategoricalCrossentropy\left(S,True\_Labels\right)$$
(6)

During training, the model aims to minimize the loss by adjusting its weights and biases and Adam optimizer is applied to update the model’s parameters.

Mathematically, the parameter update is represented as:

$$\theta \_i=\theta \_i-learning\_rate*\partial L/\partial \theta \_i$$
(7)

Where \(\theta \_i\) represents the trainable parameters in the i-th layer, learning_rate is the learning rate for the optimizer, and \(\partial L/\partial \theta \_i\) is the derivative of the loss with respect to \(\theta \_i\).

The provided description outlines the mathematics behind the operations of a Convolutional Neural Network (CNN), particularly focusing on Wide Residual Networks (WRN). These equations and operations collectively constitute the forward and backward passes of training a CNN, enabling it to learn from data and make predictions.

  1. 1.

    Convolution Operation (Eq. 1): This describes the process where a set of learnable filters is applied to the input image to create feature maps. Each filter detects different features, and an activation function like ReLU introduces non-linearity, helping the network learn complex patterns.

  2. 2.

    Wide Residual Blocks (Eq. 2): WRNs improve upon standard residual networks by increasing the width (number of channels) instead of the depth (number of layers). These blocks use batch normalization for stabilizing learning, and skip connections for alleviating the vanishing gradient problem by allowing direct gradients flow.

  3. 3.

    Global Average Pooling (Eq. 3): This operation reduces each feature map to a single number by averaging out the spatial dimensions, reducing the total number of parameters and computation in the network.

  4. 4.

    Fully Connected Layer (Eq. 4): Here, the pooled features are flattened and connected to as many neurons as there are classes, to prepare for classification.

  5. 5.

    Softmax Function (Eq. 5): This activation function is used in the output layer to calculate the probabilities of each class, ensuring they sum up to one.

  6. 6.

    Categorical Cross-Entropy (Eq. 6): It’s a loss function suitable for multi-class classification problems, measuring the difference between the predicted probabilities and the actual class.

  7. 7.

    Parameter Update (Eq. 7): Represents the optimization process, where the Adam optimizer is typically used to adjust the weights and biases to minimize the loss function.

Algorithm of the proposed model

Algorithm: PulmoNet for Pulmonary Disease Detection

Input: Set of labeled radiographic imaging data (X-rays/CT scans)

Output: Disease classification (e.g., Healthy, Bacterial Pneumonia, Covid_19, Viral Pneumonia)

Begin

  1. 1.

    Data Collection:

    • Collect diverse datasets of labeled images.

  2. 2.

    Data Preprocessing:

    • Perform image augmentation to enhance dataset size and variability.

    • Normalize images and convert to grayscale if necessary.

    • Split data into training, validation, and test sets.

  3. 3.

    Model Design (PulmoNet using CNN):

    • Define a CNN architecture with layers suitable for image analysis.

    • Apply convolutional layers with filters to extract features.

    • Utilize activation functions like ReLU for non-linearity.

    • Incorporate Wide Residual Blocks for deeper learning with fewer issues.

    • Use Global Average Pooling to reduce dimensionality.

    • Add Fully Connected Layers for classification.

    • Implement softmax activation for output layer probability distribution.

  4. 4.

    Model Training:

    • Use categorical cross-entropy as the loss function.

    • Apply Adam optimizer for efficient training.

    • Train model using backpropagation and mini-batch gradient descent.

  5. 5.

    Model Evaluation:

    • Evaluate model using accuracy, precision, recall, and F1 score on validation set.

    • Adjust hyperparameters and model structure based on performance.

  6. 6.

    Model Testing:

    • Test the final model on unseen data.

    • Generate a confusion matrix to understand classification performance.

End

Results and discussion

The outcomes of implementing the proposed approach for the 4 class, 3 class, and 2 class classifications are calculated and presented. Performance of the model for each class after training and validation procedures are evaluated in terms of accuracy (acc), precision (prec), recall (rec), f1 score, and confusion matrix.

The suggested approach acc was calculated as the ratio of the sum of images categorized correctly by the model to the entire sum of all images. Mathematically,

$$Acc= \frac{Number\ of\ images\ classified\ correctly}{Total\ number\ of\ all\ images}$$
(8)

Results of accuracy per class for all three (3) classes were presented in Table 1, 2, 3 and 4.

Table 4 Accuracy per class (2 classes)

The tables in this section present the accuracy of the PulmoNet model in classifying pulmonary diseases across different class configurations. Table 2 presents the diagnostic accuracy of the PulmoNet model for four different classes based on the data. The diagnostic tool has exceptional precision in differentiating between various lung diseases, with a 95% accuracy rate for bacterial pneumonia, 93% for COVID-19, and 91% for viral pneumonia. The model demonstrates the highest level of precision in correctly identifying persons in good health, with an accuracy rate of 97%. The varied samples for each class show the model’s resilience across diverse dataset sizes, ranging from 131 for bacterial pneumonia to 1,548 for healthy.

The PulmoNet model has robust diagnostic accuracy for three categories, as shown in Table 3. The model accurately detects bacterial pneumonia in 95% of cases out of 131 samples, identifies COVID-19 in 96% of cases out of 561 samples, and correctly diagnoses healthy instances with an accuracy of 95.3% out of 395 samples. This demonstrates consistent and dependable performance in both illness and healthy conditions, suggesting that the model might be efficiently used within these three categories for screening purposes.

Table 4 demonstrates the remarkable efficacy of the PulmoNet model in a binary classification test. It achieved a 99% accuracy in detecting COVID-19 from a dataset of 561 samples and an almost perfect accuracy of 99.8% in recognizing healthy patients from a dataset of 395 samples. These statistics indicate that the model is proficient in differentiating between COVID-19 and healthy lung conditions, demonstrating its potential as a highly dependable instrument for rapid screening in medical environments.

Table 5 displays the accuracy of the model in a binary classification test, specifically in differentiating between occurrences of pneumonia and healthy patients. The model has a remarkable discriminative ability in this binary classification scenario, accurately detecting pneumonia with a 97% precision from a sample size of 131 and identifying healthy patients with a 99% precision from a sample size of 395.

Table 5 Accuracy per class (2 classes)

As shown in the Tables, different prediction accuracy was recorded per class. In Table 2, our model recorded an accuracy of 97% for healthy, 95% for bacterial pneumonia, 93% for covid-19, and 91% for viral pneumonia in the 4-class classification. In Table 3, the accuracy for healthy images reduced to 95.3%, and that of covid_19 improved to 96%, while that of bacterial pneumonia is 95% for the 3-class classification. Tables 4 and 5 are for 2 class classifications but for different class combinations. It can be seen that accuracies of 99% and 99.8% were recorded for both covid_19 and healthy, while 97% and 99% were recorded for pneumonia and healthy.

To further demonstrate how well our model performed, graphs of model accuracy and loss per epoch for each class classification are presented in Figs. 5, 6, 7, and 8. Furthermore, the confusion matrix for each class classification is also obtainable in Figs. 9, 10, 11, and 12. These Figures indicate that the suggested model is well-trained without over-fitting or under-fitting, and this is further justified by the obtained results of precision, recall, and f1 score, respectively.

Fig. 5
figure 5

Model Accuracy per epoch for 4 class classification

Fig. 6
figure 6

Model Accuracy per epoch for 3 class classification

Fig. 7
figure 7

Model Accuracy per epoch for 2 class classifications (Covid_19 and Healthy)

Fig. 8
figure 8

Model Accuracy per epoch for 2 class classifications (Pneumonia and Healthy)

Fig. 9
figure 9

Model loss per epoch for 4 class classification

Fig. 10
figure 10

Model loss per epoch for 3 class classification

Fig. 11
figure 11

Model loss per epoch for 2 class classifications (Covid_19 and Healthy)

Fig. 12
figure 12

Model loss per epoch for 2 class classifications (Pneumonia and Healthy)

The Fig. 5 illustrates the accuracy trends across epochs throughout the training and validation phases of a deep learning model. The training accuracy (acc) is represented by the blue line, while the validation accuracy (test acc) is shown by the orange line. Both accuracies exhibit fast improvement during the early epochs, suggesting that the model is effectively acquiring knowledge from the data. The model reaches a state of convergence near the end, indicating that it no longer gains significant knowledge from further training. The proximity of the two lines indicates strong generalization, since the validation accuracy closely corresponds to the training accuracy, indicating that the model is not excessively fitting to the training data.

Figure 6 depicts the progression of a deep learning model’s performance throughout the training process over a fixed number of epochs for a job that involves classifying data into three distinct categories. The two lines, denoting training accuracy (acc) and validation accuracy (test acc) indicate the model’s proficiency in learning and generalizing to novel data. The elevated numbers suggest the model performs strongly on the training and unseen validation data. The precise alignment of the lines with minimal deviation indicates that the model does not excessively fit the data and has successfully acquired the ability to generalize the distinguishing characteristics of the three classes. This shows that the model is well-calibrated and has the potential to be dependable in a real-world scenario for identifying the provided classes, which may include different types of pulmonary disorders.

Figure 7 illustrates the progression of a deep learning model in accurately classifying two classes. The accuracy trends for the training set (acc) and validation set (test acc) increase rapidly in the beginning epochs and subsequently level off, remaining at high levels around 1.0, indicating almost flawless accuracy. This plateau demonstrates that the model has successfully acquired the ability to differentiate between COVID-19 and healthy categories. The proximity between the training and validation accuracy suggests that the model effectively applies to unfamiliar data, a crucial attribute for reliable diagnostic tools in clinical environments.

Figure 8 displays the accuracy of a machine learning model during training (acc) and validation (test acc) as it learns to differentiate between pictures of lungs with pneumonia and healthy lungs. The training and validation accuracies of the model reach a high level and are stable early in the training phase, indicating a rapid capacity to learn and generalize. The consistent and steady accuracy in the training and validation phases suggests that the model is appropriately calibrated and does not suffer from overfitting. This is crucial for ensuring the model’s reliability in real-world medical diagnostics. The graph indicates that the model has the potential to be a valuable tool for screening and diagnosing pneumonia based on radiographic data.

Figure 9 shows the loss of the model for the training data (loss) and validation data (test loss) across 50 epochs. The blue line corresponds to the training loss, exhibiting an initial steep decrease, demonstrating rapid learning during the earliest phases of model training. Subsequently, the data stabilizes, indicating that the model is approaching a point of minimal loss.

The orange line, which represents the validation loss, exhibits more volatility. The variability observed may signify that the model is facing challenges in effectively applying its learned knowledge to the validation dataset. This might potentially indicate overfitting if the pattern persists or becomes more pronounced. Nevertheless, given there is no significant rise in validation loss over time, it seems that the model can maintain its capacity to generalize rather well. The objective is to reduce the disparity between the training and validation loss, which signifies a model that exhibits effective learning and generalization capabilities towards novel, unobserved input.

Figure 10 illustrates the progression of a model’s training in categorizing data into three distinct groups. The training loss, shown by the blue line, exhibits a rapid decline followed by a period of stability. This characteristic pattern indicates that the model efficiently acquires knowledge from the training data. The test loss, shown by the orange line, declines in parallel with the training loss but displays intermittent variations, which may be attributed to the model’s performance on the validation dataset. The graph demonstrates a learning process in which the model effectively adapts to the training data and maintains a satisfactory performance on the validation data that has not been previously observed. This pattern is beneficial as it indicates that the model has successfully acquired generalizable patterns from the data without suffering from overfitting.

Figure 11 illustrates the progression of training loss and validation loss during 50 epochs during the training of a machine learning model. The training loss, shown by the blue line, exhibits a sharp decrease, suggesting fast acquisition of knowledge, and then stabilizes, indicating that the model has begun to reach a state of convergence. Nevertheless, the validation loss (shown by the orange line) displays instability and a notable sudden increase, suggesting the presence of an abnormality in the validation data or the possibility of overfitting at that particular time. However, the validation loss consistently decreases, indicating that the model maintains good generalization to new data even after the first spike. Such training aims to ensure that the validation loss drops in parallel with the training loss, reducing the disparity between them. This indicates a model consistently performing well on familiar and unfamiliar data.

Figure 12 illustrates the loss measure of a deep learning model throughout 50 epochs. The blue line shows the training loss offers a substantial decline after the early epochs, indicating the model’s rapid learning progress. Subsequently, it reaches a stable state, suggesting the model is approaching a minimum loss. The test loss, shown in orange, initially shows a similar drop but is distinguished by intermittent spikes that indicate fluctuations in the model’s performance on the validation set. Following spikes, the test loss becomes stable, although it does not reach the same level as the training loss. This implies that the model may not effectively adapt to new data, indicating a potential lack of generalization. This may mean more adjustments to enhance the model’s ability to apply to various situations or examine the data to comprehend the cause of instability.

Furthermore, prec, rec, and f1 score for each class classification are also calculated as follows [58]:

  1. 1.

    Precision (Prec): This is another model performance evaluation metric employed in this study. It can be defined as the quantification of the “True Positive” (TP) results predicted by the model. It is expressed as:

$$Precision= \frac{TP}{TP+FP} \times 100\%$$
(9)

where TP signifies true positive and FP signifies false positive

  1. 2.

    Recall (Sensitivity): Recall is the metric that measures how sensitive a predictive model performs. It measures the proportion of the positive instances that are truly predicted as positive. It is usually expressed as:where TP implies true positive, and FN signifies false negative

    $$Recall= \frac{TP}{TP+FN} \times100$$
    (10)
  1. 3.

    F1-Score: F1-Score is a metric that expresses stability between Prec and Rec. It is the harmonic average of both Rec and Prec.

    F1-score is expressed as:

$$F1\ score=2 \times \frac{Precision \times Recall}{Precision+Recall}$$
(11)

The average prec, rec, and f1 scores for each class classification are obtainable in Table 6. Table 6 presents a concise overview of the performance indicators of the PulmoNet model in several categorization situations. The model demonstrates a notable equilibrium between accuracy (94%) and recall (95.01%) in the 4-class classification, indicating its proficiency in accurately detecting affirmative instances within the four categories. The model has a commendable accuracy (90.52%) and F1 score (88.65%), suggesting its strong performance. However, there is potential for improvement in achieving a better equilibrium between precision and recall.

Table 6 Average acc, prec, rec, and f1 scores for each class classification

In the case of the 3-class scenario, the model demonstrates enhancement in all measures, notably achieving a recall of 98.54% and an F1 score of 97.34%. These results indicate a remarkable level of model sensitivity and a well-balanced precision-recall trade-off.

The model performs exceptionally in binary classifications, with accuracy, precision, recall, and F1 scores over 98%. This implies that when the model is given the duty of differentiating between just two categories, its capacity to recognize and categorize each one accurately is quite dependable.

Figure 13 visually illustrates the accuracy of the PulmoNet model in categorizing instances as Bacterial Pneumonia, COVID-19, Viral Pneumonia, or Healthy. The matrix displays the quantity of forecasts generated by the model compared to the actual categories. Specifically, the cases anticipated to be Bacterial Pneumonia 125 were accurately diagnosed, whereas a small number of cases were mistakenly labeled as Covid_19, Viral Pneumonia, or Healthy. Likewise, a significant proportion of Covid_19 cases were correctly categorized, although there was some ambiguity in distinguishing them from healthy instances. Some misdiagnosis is seen in cases of Viral Pneumonia, especially regarding Healthy individuals. The Healthy class has the most significant number of accurate classifications, showing a robust capability of the model to identify persons in good health, but with some occasional misunderstanding with other illness states. The matrix is essential for comprehending the model’s diagnostic capabilities and identifying areas needing refinement.

Fig. 13
figure 13

Confusion matrix for 4 class classification

The confusion matrix shown in Fig. 14 illustrates the model’s efficacy in differentiating between the Bacterial Pneumonia, Covid-19, and Healthy classes in a 3-class classification scenario. The matrix demonstrates high precision in categorizing Covid-19, with 536 accurate forecasts and only a minimal number of instances misclassified as Bacterial Pneumonia or Healthy. Bacterial Pneumonia has a high accuracy rate of categorization (125 out of 131) but with significant misclassification in the Healthy group. The Healthy class has a high level of accuracy, with 373 correct predictions. However, there are a few situations when Healthy occurrences are incorrectly labelled as Bacterial Pneumonia or Covid-19. This graphic facilitates comprehension of the model’s diagnostic precision and may inform enhancements in the accuracy of predictions for certain classes.

Fig. 14
figure 14

Confusion matrix for 3 class classification

Figure 15 displays the results of categorizing samples into either the COVID-19 or Healthy classes. The dark blue squares on the diagonal indicate accurate categorizations, consisting of 522 instances correctly identified as COVID-19 positive and 380 samples correctly identified as Healthy negative. The lighter blue squares indicate misclassifications, composed of 39 false negatives, where COVID-19 patients were erroneously categorized as Healthy, and 15 false positives, where Healthy cases were wrongly recognized as COVID-19. The elevated values along the diagonal indicate a robust model efficacy in differentiating COVID-19 patients from healthy persons.

Fig. 15
figure 15

Confusion matrix for 2 class classifications (Covid_19 and Healthy)

Figure 16 illustrates a confusion matrix for a binary classification issue that distinguishes between instances of pneumonia and healthy individuals. The matrix accurately identifies 127 pneumonia and 391 cases of healthy controls by the model, as seen by the substantial numbers along the diagonal from the top left to the bottom right. The occurrence of misclassification is minimal, with a mere four instances of pneumonia being erroneously labelled as healthy and four healthy cases misclassified as pneumonia. The matrix exhibits notable accuracy and precision in the model’s predictions, demonstrating a well-balanced performance across both classes.

Fig. 16
figure 16

Confusion matrix for 2 class classifications (Pneumonia and Healthy)

Figures 13, 14, 15 and 16 shows the confusion matrix of the classifications such as 4 class, 3 class, and 2 classes. In summary, the suggested technique realized an average accuracy of 94%, 95%, 95%, and 98% for 4 classes, 3 classes, 2 classes (Covid_19 and Healthy), and 2 classes (Pneumonia and Healthy) classifications. Prec, rec, and f1 score values recorded for each class show good performance of the model as a higher recall value indicates an encouraging low false negative case.

Experimental outcomes attained by the suggested technique are further likened to some selected related research conducted in the past. This is done to validate further the experimental outcomes realized by our model. A summary of the comparative investigation is obtainable in Table 7.

Table 7 Comparative examination of the performance of PulmoNet with other selected models

Table 7 displays a comparative examination of performance measures for different models in various class classification tasks. PulmoNet demonstrates exceptional performance, especially in the 2-class classifications for COVID-19/Healthy and Pneumonia/Healthy, achieving accuracies of 98%. This demonstrates PulmoNet’’s particular aptitude in binary categorizations. Furthermore, it shows strong competitiveness in multi-class settings, exhibiting performance measures equivalent or superior to those of other models. This suggests its resilience and ability to generalize. This comparison highlights the potential of PulmoNet as a dependable tool in clinical settings for diagnosing lung illnesses.

As presented in the table, the performance accuracy of Khan et al., [53] is 89.6% for 4 classes, 95% for 3 classes, and 99% for two classes (covid_19 and healthy) classifications and (Hussain et al. [59] recorded 91.2%, 94.2%, and 99.1% respectively for 4 class, 3 class, and 2 class (covid_19 and healthy) classifications compared to 94%, 95.4%, 99.4% obtained by PulmoNet for 4 class, 3 class, and 2 class (covid_19 and healthy) classifications. Another point worthy of mention here is the second 2 class (pneumonia and healthy) classification which is not reported by both Khan et al., [53] or Hussain et al., [59] in their works. This class classification recorded an accuracy of 98.3% in this work. Finally, higher precision, recall, and f1 score recorded by our work in most class classifications indicate better performance by our model.

The study key findings

The study on the PulmoNet model represents a significant advancement in the application of artificial intelligence for the diagnosis of pulmonary diseases. The model’s performance was rigorously evaluated across various classification scenarios, yielding insightful results that demonstrate its efficacy and potential clinical applicability.

Performance in multi-class and binary classifications:

  • 4-Class Configuration (Table 2): PulmoNet exhibited high diagnostic accuracy across different pulmonary diseases and healthy states. Specifically, it achieved 95% accuracy for bacterial pneumonia, 93% for COVID-19, 91% for viral pneumonia, and an impressive 97% for healthy cases. These results underscore the model’s capability to accurately distinguish between complex disease categories, a crucial requirement in clinical diagnostics.

  • 3-Class Configuration (Table 3): The model maintained consistent performance, particularly showing improvement in COVID-19 detection (96% accuracy). This indicates the model’s adaptability and precision in scenarios with fewer classification categories.

  • Binary Classifications (Tables 4 and 5): In binary classifications, PulmoNet demonstrated exceptional accuracy, especially notable in distinguishing COVID-19 (99% accuracy) and healthy cases (99.8% accuracy). Such high performance in binary classifications suggests the model’s potential utility in preliminary screening and rapid diagnostics.

Comparative Analysis with Existing Models:

  • Comparative Evaluation (Table 7): When compared with other studies (Khan et al., Hussain et al.), PulmoNet showed competitive or superior performance in several aspects. This comparison not only validates the effectiveness of PulmoNet but also positions it as a potential benchmark in the field of AI-driven pulmonary disease detection.

Training and Validation Insights:

  • Model Training Dynamics (Figs. 5, 6, 7 and 8): The accuracy and loss graphs for each class configuration provided insights into the model’s training dynamics. The close alignment of training and validation accuracy indicates that PulmoNet is well-trained, effectively learning from the training data without overfitting.

  • Confusion Matrix Analysis (Figs. 13, 14, 15 and 16): The confusion matrices for each classification task offer a detailed view of the model’s diagnostic accuracy, illustrating its strengths in correctly identifying cases and areas where improvements could be made.

Clinical Implications and Future Directions:

  • Clinical Relevance: The high accuracy and reliability of PulmoNet suggest its significant potential in clinical environments for enhancing diagnostic procedures for pulmonary diseases.

  • Future Enhancements: Acknowledging the limitations regarding dataset diversity and size, future studies are planned to test the model on a larger and more varied dataset. This will help in further validating the model’s robustness and generalization capabilities.

In summary, the PulmoNet model, through this study, demonstrates promising capabilities in revolutionizing AI-driven illness diagnosis, particularly in the realm of pulmonary diseases. Its high accuracy across various class configurations and strong performance compared to existing models highlight its potential as a valuable tool in medical diagnostics. Future enhancements and broader clinical testing are anticipated to further solidify its applicability and effectiveness in healthcare settings.

Discussion

The PulmoNet model demonstrated exceptional performance in a 4-class classification task, with accuracies of 97% for healthy cases, 95% for bacterial pneumonia cases, 93% for COVID-19 cases, and 91% for viral pneumonia cases. The accuracies for the three classes in a 3-class configuration were as follows: 95.3% for healthy, 96% for COVID-19, and 95% for bacterial pneumonia. The model attained a 99% accuracy for classifying COVID-19 cases and a 99.8% accuracy for classifying healthy cases in a 2-class classification task, demonstrating its proficiency in binary classifications. The measurements, including accuracy, recall, and f1-scores, highlight the promise of PulmoNet in revolutionizing AI-driven illness diagnosis and its suitability in clinical environments.

The PulmoNet model demonstrated remarkable precision in identifying lung illnesses in different categorization circumstances. The system’s performance in a 4-class configuration was outstanding, with excellent accuracy rates for each illness category, notably excelling in binary classifications. The model’s resilience was further validated by crucial performance indicators such as accuracy, recall, and f1-score, vital for minimizing incorrect diagnoses in medical environments.

PulmoNet exhibited more incredible skills than current models, indicating the potential to redefine AI-driven pulmonary illness identification benchmarks. The training and validation procedure, as shown by the balanced accuracy and loss graphs and confusion matrices, suggested a proficiently trained model without any problems of overfitting or underfitting.

These findings emphasize the significant influence of artificial intelligence (AI) and deep learning in medical diagnosis, especially for pulmonary illnesses. The high accuracy and dependability of PulmoNet make it very applicable in clinical settings, where it can potentially improve diagnostic processes and patient outcomes. Potential advancements may enhance the model, expand its clinical use, and enhance healthcare diagnostics.

Limitations of the study

The study’s limitation mostly pertain to the dataset and methodology. The dataset, sourced from many public sources, may lack the necessary diversity and size to guarantee the model’s generalisation capacity across different populations. Furthermore, the model explicitly targets three distinct lung disorders, potentially restricting its relevance to other respiratory ailments. The exclusive dependence on X-ray and CT scan pictures may also provide problems since these imaging modalities include inherent limits. Additionally, the research necessitates clinical validation to evaluate the real-time applicability and efficacy in realistic healthcare environments. Prospects include investigating picture segmentation and hyperparameter adjustment to improve accuracy.

Conclusions

Ever since the COVID-19 pandemic broke out in late 2019, pulmonary diseases—which are already the third biggest killer on Earth—have been a major problem for public health throughout the planet. Because of this, finding reliable diagnostic tools quickly is more important than ever. This paper presents PulmoNet, a state-of-the-art deep learning (DL) model that uses radiographic images like X-rays and CT scans to improve the diagnosis of a range of pulmonary disorders, including bacterial pneumonia, COVID-19, and viral pneumonia.

A dataset consisting of thousands of photos taken from freely accessible databases was heavily used to build PulmoNet—the extensive dataset allowed for thorough model training, guaranteeing its efficacy in many diagnostic situations. Compared with two other popular methods, PulmoNet performed far better in accuracy, precision, recall, and f1-score. A promising diagnostic tool, PulmoNet showed remarkable competence in both multi-class and binary classifications.

The relevance of PulmoNet is that it can provide healthcare providers with a significant tool for understanding lung disorders, including COVID-19, and making correct diagnoses promptly. Offering a solution to the urgent demand for effective illness detection techniques in the face of recurring health crises, its excellent accuracy in diagnosing numerous lung disorders represents a considerable contribution to medical diagnostics.

The research finds several ways to improve PulmoNet’s effectiveness in the future. To further enhance accuracy in 4 and 3 class setups, future research will concentrate on investigating supplementary approaches like as picture segmentation and hyperparameter tuning. Plans are also in the works for the model’s specialist clinical testing and validation, which will guarantee its dependability and practicality. These procedures are essential for bringing the model into practical use, where it may greatly assist in dealing with the difficulties caused by lung disorders. As a result of these planned developments, PulmoNet should become an even more effective weapon in the fight against pulmonary disorders, which should lead to improved healthcare results on a worldwide scale.

In the future we have proposed to test our model on a more extensive dataset. This will help to further evaluate the model’s generalization capabilities and ensure that it maintains high accuracy and reliability when exposed to a broader range of data. We have also propose re-evaluating the models using a standard dataset or delineating the differences in test sets and settings when discussing the results.

Availability of data and materials

The data that support the findings of this study are available on Kaggle. The data can also be accessed at https://www.kaggle.com/datasets/abdulahiabdulrahman/pulmonet-dataset.

References

  1. (WHO) W. H. O. Top 10 causes of death. 2020. https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death.

  2. Lukman AF, Rauf RI, Abiodun O, Oludoun O, Ayinde K, Ogundokun RO. COVID-19 prevalence estimation: four most affected African countries. Infect Dis Model. 2020;5:827–38.

    PubMed  PubMed Central  Google Scholar 

  3. Daniyal M, Ogundokun RO, Abid K, Khan MD, Ogundokun OE. Predictive modeling of COVID-19 death cases in Pakistan. Infect Dis Model. 2020;5:897–904.

    PubMed  PubMed Central  Google Scholar 

  4. Chandrasekaran B, Fernandes S. Since January 2020 Elsevier has created a COVID-19 resource center with free information in English and Mandarin on the novel coronavirus. The COVID-19 resource center is hosted on Elsevier Connect, the company’s public news and information website. Diabetes Metab Syndr. 2020;14(4):337–9.

    Google Scholar 

  5. Ogundokun RO, Lukman AF, Kibria GB, Awotunde JB, Aladeitan BB. Predictive modelling of COVID-19 confirmed cases in Nigeria. Infect Dis Model. 2020;5:543–8.

    PubMed  PubMed Central  Google Scholar 

  6. Ayinde K, Bello HA, Rauf RI, Attah OM, Nwosu UI, Bodunwa OK, ... Olanrewaju SO. Modeling Covid-19 cases in West African countries: a comparative analysis of quartic curve estimation models and estimators. In: modeling, control and drug development for COVID-19 outbreak prevention. Cham: Springer; 2022. p. 359-454.

  7. Lu S, Yang J, Yang B, Li X, Yin Z, Yin L,... Zheng W. Surgical instrument posture estimation and tracking based on LSTM. ICT Express. 2024. https://doi.org/10.1016/j.icte.2024.01.002.

  8. Liu F, Zhao X, Zhu Z, Zhai Z, Liu Y. Dual-microphone active noise cancellation paved with Doppler assimilation for TADS. Mech Syst Signal Process. 2023;184:109727. https://doi.org/10.1016/j.ymssp.2022.109727.

    Article  Google Scholar 

  9. Wang W, Qi F, Wipf DP, Cai C, Yu T, Li Y,... Wu W. Sparse Bayesian Learning for End-to-End EEG Decoding. IEEE Trans Pattern Anal Mach Intell. 2023;45(12):15632-15649. https://doi.org/10.1109/TPAMI.2023.3299568.

  10. Hassantabar S, Ahmadi M, Sharifi A. Diagnosis and detection of infected tissue of COVID-19 patients based on lung x-ray image using convolutional neural network approaches. Chaos Solitons Fractals. 2020;140:110170. https://doi.org/10.1016/j.chaos.2020.110170.

    Article  MathSciNet  PubMed  PubMed Central  Google Scholar 

  11. Arowolo MO, Ogundokun RO, Misra S, Agboola BD, Gupta B. Machine learning-based IoT system for COVID-19 epidemics. Computing. 2023;105(4):831–47.

    Article  Google Scholar 

  12. Ogundokun RO, Adelodun AB, Adebiyi AA, Daniyal M, Odusami M, Application of pre-trained CNN methods to identify COVID-19 pneumonia from Chest X-Ray. In: 2023 International conference on Science, Engineering and Business for Sustainable Development Goals (SEB-SDG) (Vol. 1). New York: IEEE. 2023. p. 1–6.

  13. Arowolo MO, Ogundokun RO, Misra S, Kadri AF, Aduragba TO. Machine learning approach using KPCA-SVMs for predicting COVID-19. Healthcare informatics for fighting COVID-19 and future epidemics. 2022. p. 193–209.

  14. Ibrahim DM, Elshennawy NM, Sarhan AM. Deep-chest: multi-classification deep learning model for diagnosing COVID-19, pneumonia, and lung cancer chest diseases. Comput Biol Med. 2021;132:104348. https://doi.org/10.1016/j.compbiomed.2021.104348.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  15. Ausawalaithong W, Thirach A, Marukatat S, & Wilaiprasitporn T. Automatic Lung cancer prediction from Chest X-ray images using the deep learning approach. BMEiCON 2018 - 11th Biomedical Engineering International Conference. 2019. https://doi.org/10.1109/BMEICON.2018.8609997.

  16. Sun W, Zheng B, Qian W. Automatic feature learning using multichannel ROI based on deep structured algorithms for computerized lung cancer diagnosis. Comput Biol Med. 2017;89:530–9. https://doi.org/10.1016/J.COMPBIOMED.2017.04.006.

    Article  PubMed  Google Scholar 

  17. Hao S, Jiali P, Xiaomin Z, Xiaoqin W, Lina L, Xin Q,... Qin L. Group identity modulates bidding behavior in repeated lottery contest: neural signatures from event-related potentials and electroencephalography oscillations. Front Neurosci. 2023;17. https://doi.org/10.3389/fnins.2023.1184601.

  18. Shambhu S, Koundal D,  Das P. Edge-based segmentation for accurate detection of malaria parasites in microscopic blood smear images: a novel approach using FCM and MPP algorithms. In: 2023 2nd International Conference on Smart Technologies and Systems for Next Generation Computing (ICSTSN). New York: IEEE. 2023. p. 1–6.

  19. Cui Y, Wang X, Lin F, Li W, Zhao Y, Zhu F, Yang H, Rao M, Li Y, Liang H, Dai M, Liu B, Chen L, Han D, Lu R, Peng W, Zhang Y, Song C, Luo Y, Pan P. MiR-29a-3p improves acute lung injury by reducing alveolar epithelial cell PANoptosis. Aging Dis. 2022;13(3):899–909.

    Article  PubMed  PubMed Central  Google Scholar 

  20. Shambhu S, Koundal D. Recent trends in image processing using granular computing. In: International conference on advanced communication and computational technology. Singapore: Springer Nature Singapore. 2019. p. 469-479.

  21. Hao S, Xin Q, Xiaomin Z, Jiali P, Xiaoqin W, Rong Y,... Cenlin Z. (2023). Group membership modulates the hold-up problem: an event-related potentials and oscillations study. Soc Cogn Affect Neurosci. 2023;18(1). https://doi.org/10.1093/scan/nsad071.

  22. Gu Y, Lu X, Yang L, Zhang B, Yu D, Zhao Y, Gao L, Wu L, Zhou T. Automatic lung nodule detection using a 3D deep convolutional neural network combined with a multi-scale prediction strategy in chest CTs. Comput Biol Med. 2018;103:220–31. https://doi.org/10.1016/J.COMPBIOMED.2018.10.011.

    Article  PubMed  Google Scholar 

  23. Vinay R, Soujanya KLS, Singh P. Disease prediction by using deep learning based on patient treatment history. Int J Recent Technol Eng. 2019;7(6):745–54.

    Google Scholar 

  24. Tort-Nasarre G, Alvarez B, Galbany-Estragués P, Subías-Miquel M, Vázquez-Segura E, Marre D, Romeu-Labayen M. Front-line nurses’ responses to organisational changes during the COVID-19 in Spain: a qualitative rapid appraisal. J Nurs Manag. 2021;29(7):1983–91.

    Article  PubMed  PubMed Central  Google Scholar 

  25. Haenssle HA, Fink C, Schneiderbauer R, Toberer F, Buhl T, Blum A, Kalloo A, Ben Hadj Hassen A, Thomas L, Enk A, Uhlmann L, Alt C, Arenbergerova M, Bakos R, Baltzer A, Bertlich I, Blum A, Bokor-Billmann T, Bowling J, … Zalaudek I. Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann Oncol. 2018;29(8):1836–1842. https://doi.org/10.1093/ANNONC/MDY166.

  26. Huang H, Wu N, Liang Y, Peng X, Shu J. SLNL: a novel method for gene selection and phenotype classification. Int J Intell Syst. 2022;37(9):6283–304. https://doi.org/10.1002/int.22844.

    Article  Google Scholar 

  27. Yu Y, Wang L, Ni S, Li D, Liu J, Chu HY,... Zhang G. Targeting loop3 of sclerostin preserves its cardiovascular protective action and promotes bone formation. Nat Commun. 2022;13(1):4241. https://doi.org/10.1038/s41467-022-31997-8.

  28. Mar VJ, Soyer HP. Artificial intelligence for melanoma diagnosis: how can we deliver on the promise? Ann Oncol. 2018;29(8):1625–8. https://doi.org/10.1093/ANNONC/MDY19.

    Article  CAS  PubMed  Google Scholar 

  29. Lee EYP, Ng MY, Khong PL. COVID-19 pneumonia: what has CT taught us? Lancet Infect Dis. 2020;20(4):384–5. https://doi.org/10.1016/S1473-3099(20)30134-1.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  30. Pan F, Ye T, Sun P, Gui S, Liang B, Li L, Zheng D, Wang J, Hesketh RL, Yang L, Zheng C. Time course of lung changes on chest CT during recovery from 2019 novel coronavirus (COVID-19) pneumonia. Radiology. 2020;295(3):715–21. https://doi.org/10.1148/RADIOL.2020200370.

    Article  PubMed  Google Scholar 

  31. Chan JFW, Yuan S, Kok KH, To KKW, Chu H, Yang J, Xing F, Liu J, Yip CCY, Poon RWS, Tsoi HW, Lo SKF, Chan KH, Poon VKM, Chan WM, Ip JD, Cai JP, Cheng VCC, Chen H, … Yuen KY. A familial cluster of pneumonia associated with the 2019 novel coronavirus indicating person-to-person transmission: a study of a family cluster. Lancet. 2020;395(10223):514–523. https://doi.org/10.1016/S0140-6736(20)30154-9.

  32. Kong W, Agarwal PP. Chest imaging appearance of COVID-19 infection. Radiol Cardiothorac Imaging. 2020;2(1):e200028. https://doi.org/10.1148/RYCT.2020200028.

    Article  PubMed  PubMed Central  Google Scholar 

  33. Zu ZY, Jiang MD, Xu PP, Chen W, Ni QQ, Lu GM, Zhang LJ. Coronavirus disease 2019 (COVID-19): a perspective from China. Radiology. 2020;296(2):E15–25. https://doi.org/10.1148/RADIOL.2020200490.

    Article  PubMed  Google Scholar 

  34. Zhao W, Zhong Z, Xie X, Yu Q, Liu J. Relation between chest CT findings and clinical conditions of Coronavirus Disease (COVID-19) pneumonia: A Multicenter Study. AJR Am J Roentgenol. 2020;214(5):1072–7. https://doi.org/10.2214/AJR.20.22976.

    Article  PubMed  Google Scholar 

  35. Yoon SH, Lee KH, Kim JY, Lee YK, Ko H, Kim KH, Park CM, Kim Y-H. Chest radiographic and CT findings of the 2019 novel coronavirus Disease (COVID-19): analysis of nine patients treated in Korea. Korean J Radiol. 2020;21(4):494–500. https://doi.org/10.3348/KJR.2020.0132.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Fukushima K. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol Cybern. 1980;36(4):193–202.

    Article  CAS  PubMed  Google Scholar 

  37. Wang H, Yang T, Wu J, Chen D, Wang W. Unveiling the mystery of SUMO-activating enzyme subunit 1: a groundbreaking biomarker in the early detection and advancement of hepatocellular carcinoma. Transpl Proc. 2023;55(4):945–51. https://doi.org/10.1016/j.transproceed.2023.03.042.

    Article  CAS  Google Scholar 

  38. Han M, He W, He Z, Yan X, Fang X. Anatomical characteristics affecting the surgical approach of oblique lateral lumbar interbody fusion: an MR-based observational study. J Orthop Surg Res. 2022;17(1):426. https://doi.org/10.1186/s13018-022-03322-y.

    Article  PubMed  PubMed Central  Google Scholar 

  39. Shambhu S, Koundal D, Das P, Sharma C. Binary classification of covid-19 ct images using cnn: covid diagnosis using ct. Int J E-Health Med Commun. 2021;13(2):1–13.

    Article  Google Scholar 

  40. He B, Lu Q, Lang J, Yu H, Peng C, Bing P,... Tian G. A new method for CTC images recognition based on machine learning. Front Bioeng Biotechnol. 2020;8. https://doi.org/10.3389/fbioe.2020.00897.

  41. Zhao Y, Hu M, Jin Y, Chen F, Wang X, Wang B,... Ren H. Predicting the transmission trend of respiratory viruses in new regions via geospatial similarity learning. Int J Appl Earth Observ Geoinform. 2023;125:103559. https://doi.org/10.1016/j.jag.2023.103559.

  42. Chen X, Liao Y, Long D, Yu T, Shen F,... Lin X. The Cdc2/Cdk1 inhibitor, purvalanol A, enhances the cytotoxic effects of taxol through Op18/stathmin in non-small cell lung cancer cells in vitro. Int J Mol Med. 2017;40(1):235-242. https://doi.org/10.3892/ijmm.2017.2989.

  43. Yi X, Guan X, Chen C, Zhang Y, Zhang Z, Li M, Liu P, Yu A, Long X, Liu L, Chen BT, Zee C. Adrenal incidentaloma: machine learning-based quantitative texture analysis of unenhanced CT can effectively differentiate sPHEO from lipid-poor adrenal adenoma. J Cancer. 2018;9(19):3577–82. https://doi.org/10.7150/jca.26356.

    Article  PubMed  PubMed Central  Google Scholar 

  44. Matsugu M, Mori K, Mitari Y, Kaneda Y. Subject-independent facial expression recognition with robust face detection using a convolutional neural network. Neural Netw. 2003;16(5–6):555–9. https://doi.org/10.1016/S0893-6080(03)00115-1.

    Article  PubMed  Google Scholar 

  45. Kumar Sethy P, Kumari Behera S, Kumar Ratha P, Biswas P. Detection of coronavirus Disease (COVID-19) based on deep features and support vector machine. 2020. www.preprints.org.

  46. Chen S, Zhao Y, Shen F, Long D, Yu T,... Lin X. Introduction of exogenous wild‑type p53 mediates the regulation of oncoprotein 18/stathmin signaling via nuclear factor‑κB in non‑small cell lung cancer NCI‑H1299 cells. Oncol Rep. 2019;41(3):2051-2059. https://doi.org/10.3892/or.2019.6964.

  47. Lu H, Stratton CW, Tang Y-W. The outbreak of pneumonia of unknown etiology in Wuhan, China: The mystery and the miracle. J Med Virol. 2020;92:401–2. https://doi.org/10.1002/jmv.25678.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  48. Hemdan EED, Shouman MA, Karar ME. COVIDX-net: a framework of deep learning classifiers to diagnose COVID-19 in X-Ray images. 2020. http://arxiv.org/abs/2003.11055.

  49. Song Y, Zheng S, Li L, Zhang X, Zhang X, Huang Z, ... Yang Y. (2021). Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images. IEEE/ACM Trans Comput Biol Bioinform. 2021;18(6):2775–2780.

  50. Wang S, Kang B, Ma J, Zeng X, Xiao M, Guo J, Cai M, Yang J, Li Y, Meng X, Xu B. A deep learning algorithm using CT images to screen for Coronavirus disease (COVID-19). Eur Radiol. 2021;31(8):6096. https://doi.org/10.1007/S00330-021-07715-1.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  51. Narin A, Kaya C, Pamuk Z. Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks. Pattern Anal Appl. 2021;24(3):1207–20. https://doi.org/10.1007/s10044-021-00984-y.

    Article  PubMed  PubMed Central  Google Scholar 

  52. Ozturk T, Talo M, Yildirim EA, Baloglu UB, Yildirim O, Rajendra Acharya U. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput Biol Med. 2020;2020:103792. https://doi.org/10.1016/j.compbiomed.2020.103792.

    Article  CAS  Google Scholar 

  53. Khan AI, Shah JL, Bhat MM. CoroNet: a deep neural network for detection and diagnosis of COVID-19 from chest x-ray images. Comput Methods Prog Biomed. 2020;196:105581. https://doi.org/10.1016/j.cmpb.2020.105581.

    Article  Google Scholar 

  54. Agchung. Actualmed-COVID-chestxray-dataset. 2020. https://github.com/agchung/Actualmed-COVID-chestxray-dataset.

  55. Cohen JP, Morrison P, Dao L. COVID-19 image data collection. 2020. http://arxiv.org/abs/2003.11597.

  56. Ab Rahman Z, Kashim MIAM, Mohd Noor AY, Saari CZ, Hasan AZ, Ridzuan AR, Norhayati Rafida AR. Critical review of religion in coping against the COVID-19 pandemic by former COVID-19 muslim patients in Malaysia. J Crit Rev. 2020;7(5):1145–54.

    Google Scholar 

  57. Nouri S, Khoong EC, Lyles CR, Karliner L. Addressing equity in telemedicine for chronic disease management during the Covid-19 pandemic. NEJM Catal Innov Care Deliv. 2020;1(3):1–13.

    Google Scholar 

  58. Arowolo MO, Ogundokun RO, Misra S, Kadri AF, Aduragba TO. Machine learning approach using KPCA-SVMs for predicting COVID-19. In: Healthcare informatics for fighting COVID-19 and future epidemics. Cham: Springer; 2022. p. 193-209.

  59. Hussain E, Hasan M, Rahman MA, Lee I, Tamanna T, Parvez MZ. CoroDet: a deep learning-based classification for COVID-19 detection using chest X-ray images. Chaos Solitons Fractals. 2021;142:110495. https://doi.org/10.1016/J.CHAOS.2020.1104.

    Article  MathSciNet  PubMed  Google Scholar 

Download references

Acknowledgements

Not applicable.

Code availability

https://github.com/Toshman-hub/PulmoNet.

Inform consent

The research does not require gathering datasets or using humans, so therefore, this statement is not applicable to this study.

Experimental protocols

Not applicable and all methods were carried out as supposed.

Data Availability

The dataset and the implementation codes employed in this study can be seen on GitHub with the link: https://www.kaggle.com/datasets/abdulahiabdulrahman/pulmonet-dataset and https://github.com/Toshman-hub/PulmoNet.

Funding

The research received no funding grant from any funding agency in the public, commercial, or not-for-profit sectors.

Author information

Authors and Affiliations

Authors

Contributions

AbdulRahman Tosho Abdulahi: Conceptualization, Methodology, Implementation, Visualization, Investigation Roseline Oluwaseun Ogundokun: Conceptualization, Supervision, Validation, Investigation, Reviewing and Editing Ajiboye Raimot Adenike: Investigation, Data curation, Writing- Original draft preparation Mohd Asif Shah: Guidance, Conceptualization, Supervision, Reviewing and Editing Yusuf Kola Ahmed: Supervision, Validation, Reviewing and Editing.

Corresponding author

Correspondence to Mohd Asif Shah.

Ethics declarations

Ethics approval and consent to participate

The dataset used in the study was obtained from a public repository (Kaggle), therefore ethical approval from the ethical committee is not applicable here.

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Abdulahi, A.T., Ogundokun, R.O., Adenike, A.R. et al. PulmoNet: a novel deep learning based pulmonary diseases detection model. BMC Med Imaging 24, 51 (2024). https://doi.org/10.1186/s12880-024-01227-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12880-024-01227-2

Keywords