Skip to main content

An improved method for diagnosis of Parkinson’s disease using deep learning models enhanced with metaheuristic algorithm

Abstract

Parkinson's disease (PD) is challenging for clinicians to accurately diagnose in the early stages. Quantitative measures of brain health can be obtained safely and non-invasively using medical imaging techniques like magnetic resonance imaging (MRI) and single photon emission computed tomography (SPECT). For accurate diagnosis of PD, powerful machine learning and deep learning models as well as the effectiveness of medical imaging tools for assessing neurological health are required. This study proposes four deep learning models with a hybrid model for the early detection of PD. For the simulation study, two standard datasets are chosen. Further to improve the performance of the models, grey wolf optimization (GWO) is used to automatically fine-tune the hyperparameters of the models. The GWO-VGG16, GWO-DenseNet, GWO-DenseNet + LSTM, GWO-InceptionV3 and GWO-VGG16 + InceptionV3 are applied to the T1,T2-weighted and SPECT DaTscan datasets. All the models performed well and obtained near or above 99% accuracy. The highest accuracy of 99.94% and AUC of 99.99% is achieved by the hybrid model (GWO-VGG16 + InceptionV3) for T1,T2-weighted dataset and 100% accuracy and 99.92% AUC is recorded for GWO-VGG16 + InceptionV3 models using SPECT DaTscan dataset.

Peer Review reports

Introduction

Parkinson's disease, also known as neurodegeneration, is a long-term, neurological, and progressive motor illness [1] characterised by the progressive death of dopamine-producing brain cells. Dopamine is an organic substance produced by neurons that serves as a neurotransmitter in the brain, facilitating communication between neurons. Parkinson's disease results from impaired neuronal communication due to insufficient dopamine production in the brain. The substantia nigra, a small region where the neurons of the human brain are affected due to Parkinson’s disease.

A new United Nations research claims that nearly 1 billion people worldwide, or approximately one in six, suffer from neurological conditions like epilepsy, migraine, brain injuries, and neuro-infections like Alzheimer's, PD as well as stroke, and multiple sclerosis. Each year, 6.8 million of these sufferers pass away (https://news.un.org/en/story/2007/02/210312). Although the actual cause of Parkinson's disease is unknown, it is believed that a combination of inherited and environmental factors is responsible for it [2]. In the modern world, PD is affecting 2% to 3% of people who are at the age of 65 and older [3]. Parkinson's disease progresses differently in every patient, and it is impossible to anticipate how quickly the disease may progress in any specific person. While some people may have only minor symptoms for years, others may do so quite fast as they progress to more severe problems. Parkinson's disease often starts with minor tremors or other motor symptoms on one side of the body and progresses slowly over a number of years. The disease's symptoms could extend throughout the body and get worse, possibly affecting both sides of it. Even though Parkinson's disease is an ongoing and advancing condition, there are medicines that can help to manage symptoms and improve the standard of living. Parkinson's disease does not yet have an appropriate early diagnosis or treatment. Medication, physical therapy, and lifestyle modifications are some of its treatments. Parkinson's disease progression can be slowed down or stopped, even though there is presently no known cure for it.

These days, artificial intelligence (AI) approaches—machine learning (ML) and state of art deep learning (DL) are greatly assisting medical professionals in the early diagnosis of illnesses. Due to this, research has recently been done to automatically identify Parkinson’s disease using MRI images utilizing a variety of AI and ML algorithms. Many different diseases and ailments have been diagnosed using deep learning, and the findings frequently outperform traditional benchmarks [4].

Deep learning models have become powerful and are mostly used in image classification problems. With their ability to learn intricate patterns and features from images, they can often surpass traditional machine-learning approaches in accuracy. It automatically extracts relevant features from images, hence causing the elimination of manual feature engineering. This feature extraction capability permits the model to learn complicated representations and capture both low-level and high-level features present in the images. It can handle large-scale datasets efficiently. They can learn from vast amounts of labelled data, which is essential for training accurate image classifiers. Deep learning frameworks and libraries are designed to leverage parallel computing resources like GPUs to accelerate training and inference processes.

Over the past two decades, meta-heuristic optimization techniques have gained a lot of popularity. A few of these include particle swarm optimization (PSO) [5], grey wolf optimization (GWO) [6], ant colony optimization (ACO) [7], artificial bee colony optimization (ABC) [8], etc. Hyperparameter tuning is one of the tedious jobs to manually fine-tune the parameters to obtain the best optimal values. The population-based metaheuristic algorithm known as grey wolf optimization (GWO) is influenced by the way grey wolves hunt. It searches for the best answers in a problem area by combining exploration (diversification) with exploitation (intensification). It is used to automatically fine-tune the parameters which mimic the social behaviour of grey wolves, including their leadership hierarchy and group hunting. The improved capacity of GWO prevents results from being stuck in the local optimal value [9]. It also finds the best solution with a quick convergence rate.

The benefits of deep learning and GWO in image classification include higher accuracy, autonomous feature extraction, scalability, transfer learning skills, robustness to fluctuations, finding the optimal solutions, automatic hyperparameter tunning and continuous progress through continuing research and development. A variety of AI approaches using ML and DL models have been created in the past. In this study, a new framework is employed by combining grey wolf optimization (GWO) with four deep learning models known as VGG16 [10], DenseNet [11], InceptionV3 [12], DenseNet-LSTM [13] and a hybrid model VGG16 + InceptionV3.

The following is a concise explanation of the paper's main contribution.

  1. (I)

    Number of images created empty tuples, from which the performance of deep learning models degrades. These empty tuples are removed to obtain better performance by using the Python function.

  2. (II)

    Proposed four deep learning models with hyperparameter optimization by GWO known as GWO-VGG16, GWO-DenseNet, GWO-InceptionV3, and GWO-DenseNet-LSTM.

  3. (III)

    Proposed hybrid model using GWO-VGG16 + InceptionV3.

  4. (IV)

    The proposed models are compared with the existing models using various performance metrics.

Following is the format for the remaining section: The earlier studies are covered in “Related literature” section. “Materials and methods” section explains the preprocessing of MRI images and the development of methodologies. Experimental results and discussions, comparisons between the existing models and proposed models are discussed in “Results and discussion” section. In “Conclusion and future work” section, conclusions and future scope are discussed briefly.

Related literature

In the past few years, various studies have been created and published by academics worldwide to help in Parkinson's disease diagnosis. Many of these researchers have used various AI methods to analyse and classify the MRI brain images in order to detect various diseases related to Parkinson’s disease. Deep learning techniques are the most often used method for classifying MRI images, due to their capacity to deliver superior results than those obtained by more conventional machine learning techniques. This particular section explains the research using ML and DL methods to diagnose patients with Parkinson’s disease.

Related review literature using T1, T2-weighted dataset

Camacho et al., (2023) [14] have developed a robust explainable deep learning classifier trained for the classification of Parkinson’s disease using T1-weighted MRI dataset. A total of 1,024 PD and 1,017 subjects from matched controls (HC) of the same age and gender are gathered from 2,041 MRI data i.e. T1-weighted MRI datasets from 13 separate investigations. The datasets undergone a skull-stripping process, isotropic resampling, bias field correction, and nonlinear registration to the MNI PD25 atlas. Convolutional Neural Network is trained to categorise PD and HC participants using the Jacobian maps produced from the fields of deformation and fundamental clinical data. The authors provide improved knowledge of the clinical variables associated with Impulse Control Behaviors (ICB) and structural and functional brain abnormalities in PD patients. They have measured grey and white matter brain volume, and graph topological features using multimodal MRI data [15]. A new technique is introduced for categorising a person's 3-D magnetic resonance scans as a diagnostic tool for Parkinson's disease by using one of the largest Parkinson’s Progressive Marker Initiative (PPMI) MRI datasets from a patient group with the condition and healthy controls. Due to the fact that gender has a substantial impact on neurobiology and PD cases are developed more likely in males than women, it is advantageous that different research is conducted for men and women [16]. The viability and usefulness of employing multi-modal MRI datasets to automatically distinguish between PD, PSP-RS, and HC subjects are examined. For this investigation, there are 45 PD, 20 PSP-RS, and 38 HC subjects with available T1-weighted MRI datasets, T2-weighted MRI datasets, and diffusion-tensor (DTI) MRI datasets [17]. Brain morphology using T1-weighted, brain iron metabolism using T2-weighted, and microstructural integrity using DTI dataset regional values are determined by an atlas-based approach. These values are used to choose features, and then classification is performed using a variety of well-known machine learning approaches. A 3D CNN architecture is proposed after data pre-processing in order to learn the complex patterns in MRI images for the identification of Parkinson's Disease. From the baseline visit, 406 individuals including 203 in good health and 203 with Parkinson's disease are selected for the experiment [18].

A novel method is used which trains a deep neural network model using data from new patients, specifically with T1 MRI and DaTscan datasets. The information utilized to model the knowledge retrieved from the PPMI database contains a set of vectors that represent the clustering centers of these representations, along with the matching Deep Neural Network (DNN) structure. The ability of the unified model created using these many datasets to predict Parkinson's disease in an effective and transparent manner has then been demonstrated [19]. Two new deep learning techniques are proposed for ensemble learning-based Parkinson's disease detection. Instead of using the entire MRI image, authors focused on the Grey and White Matter areas which greatly improved detection accuracy and obtained 94.7% accuracy [20]. To discover which brain regions are important in the decision-making process for architecture is performed by occlusion analysis as well. Multiple parcellated brain areas are used to train a CNN. The idea is to create a complicated model by combining the models from various locations using a greedy algorithm. Three retrospective investigations included 305 PD patients (59.9–9.7 years of age) and 227 HC patients (61.0–7.4 years of age). Based on the Automatic Anatomic Labelling template, fractional anisotropy and mean diffusivity are determined and then divided into 90 different brain regions of interest (ROIs) [21].

The authors have suggested CNN with eight layers deep for 3D T1-weighted MRI images to differentiate between PD and HC individuals. The proposed model additionally made use of the information provided by the individuals' ages and genders. In addition, batch and group normalization are applied to the designed model, increasing the accuracy up to 100% [22]. An autonomous diagnosis approach that distinguishes the PD and HC with high accuracy. Benchmark T2-weighted MRI scans for both PD and HC are made available to the public by the PPMI. Image registration technique is used to choose and align the middle 500 slices of a T2-weighted MRI scan [23].

The study evaluates the viability of machine learning techniques for classifying patients with Parkinson's disease (PD) and non-proliferative osteoporosis (NPOD) using 30 patients data. It demonstrates that PD patients can be distinguished from NPOD patients by (a) using T1-weighted axial magnetic resonance imaging (MRI) scans, or (b) using morphometric measurements such as cortical thickness, cortical surface area, and volumetric measurements of the brain's subcortical and cortical areas division [24].

Based on the analysis performed using the brain MRI slices, authors [25] has suggested machine-learning-technique (MLT) to assess and classify the tumour locations into low/high grade using 30 patients. A series of operations, including pre, post and classification procedures, are carried out by MLT. The Social Group Optimization (SGO) method in conjunction with Fuzzy-Tsallis thresholding improves the tumour section during pre-processing. For the mining of the tumour area Level-Set Segmentation (LSS) is utilized in the post-processing step.

Related review literature using SPECT DaTscan dataset

Thakur et al. (2022) [11] have constructed a CNN model that can accurately pinpoint the ROIs after feature extraction. In the study, 1,390 groups of DaTscan images with PD and normal classes are examined. The final classification layer includes a soft-attention block which makes use of the DenseNet-121 design. After classifying the images, Soft Attention Maps and feature map representation are reutilized to visually analyse the region of interest (ROI). The work sought to establish an ensemble deep learning technique with three stages for PD patient prognosis. Retrospective information on 198 Parkinson's disease (PD) patients is obtained from the PPMI database and then randomly 118 patients are assigned to training, and 40–40 patients are assigned to both validation and test sets. The features are extracted from the DaTscan dataset and clinical assessments of motor symptoms in steps 1 and 2. In step 3, an ensemble of DNN are trained to predict 4 years of patient outcome [26]. A CNN model is created that can distinguish between PD patients and HC patients based on SPECT images. In this study, 2723 images of the SPECT dataset are used out of which 1364 samples from the PD group and 1359 samples from the HC group. The image normalization method is used to improve the regions of interest (ROIs) required for the network to learn attributes that set them apart from other regions of interest (ROIs). In order to assess the effectiveness of the network model, tenfold cross validation is used [27]. Six well-known interpretation techniques and four deep-convolutional neural network are designed [28]. Also, the authors suggest a mechanism for evaluating interpretation performance as well as a way to use interpreted input to aid in model selection. It is suggested to develop a computer learning model that accurately identifies whether every given DaTscan has PD or not while offering a logical justification for the prediction. Visual indicators are created utilizing Local Interpretable Model-Agnostic Explainer (LIME) approaches. Further, transfer learning is used to train DaTscans on a CNN (VGG16) from the PPMI database, and the resulting models have 95.2% accuracy. Finally, the paper concludes that the suggested approach may successfully assist medical professionals in PD detection because of its measured interpretability and accuracy11. To analyse pictures from dopamine transporter single-photon emission computed tomography (DAT-SPECT) has been suggested utilizing an ANN. With the use of an active contour model, striatal regions are segmented and utilized as the data performing transfer learning on the artificial neural network which is pre-trained to distinguish Parkinson’s disease. To serve as a benchmark, the support vector machine is trained to use semi-quantitative measurement metrics including the specific binding ratio (SBR) and asymmetry index [29].

The active contour model is utilized to segment the striatal regions in the images. These segmented regions are then employed as the dataset for an already-trained ANN to do transfer learning. The goal is to separate PD from Parkinsonism associated with other diseases. Artificial neural networks (ANN) and image processing techniques are proposed to identify Parkinson's disease in its early stages [30]. The images used are 200 SPECT scans from the PPMI dataset, out of which 130 are of normal participants and 70 are of Parkinson's disease (PD) patients. Using the sequential grass fire algorithm, the caudate and putamen areas of the images are determined. To distinguish healthy and Parkinson's disease-infected people, these above features are loaded into an ANN. A novel approach is introduced for the medical treatment of neurodegenerative disorders, like Parkinson's, that utilizes trained DNNs to extract and utilize latent information. The paper uses transfer learning along with k-means clustering, K-NN classification, and DNN trained representations to enhance disease prediction using MRI data [31]. In the recent past, authors have presented a model for the early identification of PD which combines image processing with ANN in order to improve the imaging diagnosis of PD. The caudate and putamen serve as the study's region of interest (ROI), and the model identified them by analysing 200 SPECT images from the PPMI database, out of which 100 are of healthy people and 100 are of PD people. The ANN is then fed with the ROI area data, with a thought it will recognise patterns similar to how a human observer would do [32]. A novel method is suggested that uses 3-dimensional convolutional neural networks (CNNs) to differentiate between PD and healthy control. In order to reduce overfitting and boost the neural network's generalisation abilities, the training set as well as the data from this set's sagittal plane using a straightforward data augmentation technique is given as input to the model [33].

One of the difficult challenges all are facing is determining Parkinson's disease in the early stages. To conduct research on the early detection of PD using MRI images, various authors developed numerous computer-based machine learning and deep learning methods as described above.

In this study, authors have proposed four deep learning models whose hyperparameters are optimized using GWO, namely GWO-VGG16, GWO-DenseNet, GWO-DenseNet + LSTM, GWO-IncepionV3, and a hybrid model (GWO-VGG16 + InceptionV3) which is the novelty of this paper. No authors earlier used these models with T1, T2-weighted and SPECT DaTscan for PD detection. Here, a number of images are creating empty tuples, from which the performance of the deep learning models degrades. These empty tuples are properly handled and removed to obtain better performance. This problem has also never been addressed by any authors previously in the literature.

Materials and methods

This section illustrates the proposed methodology, preprocessing of MRI images, and model development. After that, the data is divided into two sets using an 80:20 ratio for the train and test sets. Again, the train set is divided into train and validation sets. The 80% of input images are fed to the proposed model for training and then the models are validated using 20% from the train set samples. Finally, models are tested using the remaining 20% of the data. The distribution is depicted in Fig. 1 below.

Fig. 1
figure 1

Splitting of MRI datasets

Proposed methodology

In the proposed methodology, steps are given in the following:

  • Step 1: Firstly, T1, T2-weighted and SPECT DaTscan MRI datasets are collected from the PPMI website.

  • Step 2: MRI images are then pre-processed using preprocessing techniques such as the conversion of DICOM file to.jpg format, cropped images using MicroDICOM Viewer desktop application, removing empty tuples and finally skull stripping is done using the python package (simple ITK). Normalization is also done using batch normalization for scaling.

  • Step 3: Datasets are divided into train and test sets using the holdout method (80:20 ratio) Again train set is divided into (80:20 ratio) two sets i.e. train and validation set.

  • Step 4: Four deep learning models are proposed whose hyperparameters are optimized by GWO, known as GWO-VGG16, GWO-DenseNet, GWO-DenseNet-LSTM, GWO-InceptionV3, with one hybrid model GWO-VGG16 + InceptionV3.

  • Step 5: Finally, results are evaluated using various performance measures such as accuracy (acc), sensitivity (sen), specificity (spe), precision (pre), f1_score (f1-scr) and AUC score. The proposed methodology is also graphically presented in Fig. 2

Fig. 2
figure 2

Proposed methodology for early detection of PD using GWO and deep learning models

MRI data collection

The MRI data are extracted from PPMI website [34]. The PPMI dataset is a large-scale longitudinal investigation of Parkinson's Disease (PD) conducted by the Michael J. Fox Foundation for the research of Parkinson's. The objective of the study is to find biomarkers that can aid in predicting the onset and progression of PD and to create new treatments for the condition. The PPMI dataset contains a variety of information, including clinical evaluations, genetic information, biospecimen samples (blood and CSF), and brain imaging data (MRI and DaTscan). Researchers from all across the world can analyse and do research on the dataset.

One of the distinguishing characteristics of the PPMI dataset is its longitudinal nature, which monitors patients over a number of years. This feature enables researchers to examine changes in disease development and find potential biomarkers for the illness. The dataset also includes a large control group of healthy individuals, which provides a baseline for comparison. T1, T2-weighted MRI [35] and SPECT DaTscan [34] datasets used in this study are collected from the PPMI website.

MRI data samples

T1,T2-weighted and SPECT DaTscan dataset from PPMI are chosen for this investigation. A 1.5—3 Tesla scanner was used to create these pictures. The entire scan takes about twenty to thirty minutes. Three distinct views—axial, sagittal, and coronal—were used to acquire the T1, T2-weighted MRI images as a three-dimensional sequence with a slice thickness of 1.5 mm or less. The description of MRI images of both datasets is given below in Tables 1 and 2.

Table 1 Description of MRI images of T1,T2-weighted dataset
Table 2 Description of MRI images of SPECT DaTscan dataset

In this study, two datasets are used i.e. T1,T2-weighted and SPECT DaTscan. A total of 30 subjects are included in T1,T2-weighted MRI dataset from which 15 subjects (Male-7, Female-8) are Parkinson’s disease (PD) and 15 subjects (Male-7, Female-8) are healthy control (HC) which contains a total number of 9070 MRI images of different sizes. Out of 9070 MRI images, 3620 are PD subjects and 5450 are HC subjects. A total of 36 subjects are included in SPECT DaTscan dataset from which 18 subjects (Male-9, Female-9) are suffering from Parkinson’s disease (PD) and 18 subjects (Male-9, Female-9) are healthy control (HC) which contains a total of 20,096 MRI images. Out of 20,096 MRI images, 5752 are PD subjects, and 14,344 are HC subjects. The sample size is distributed as shown in Table 3.

Table 3 MRI data sample distribution

Inclusion criteria

Those patients are included in the study whose age is between 55 and 75 years. Only PD and HC subjects are included.

Exclusion criteria

Patients whose age is less than 55 and greater than 75 are excluded from this study. Other category subjects are excluded, such as SWEDD, PRODROMAL, etc.

Image pre-processing

MRI images are available in DICOM (Digital Imaging and Communications in Medicine) (https://www.microdicom.com/dicom-viewer-user-manual/) file format which is used to store and send medical pictures like X-rays, CT scans, and MRIs. A lot of image-related metadata, including patient data, information on the image's acquisition, and other medical data, is included in DICOM files. However, the DICOM file format is difficult to deal with when employing these pictures for machine learning tasks.

Many machine learning libraries and frameworks don't natively support DICOM files, which is one of the reasons DICOM images are generally transformed to other image formats, like png or jpg, before being used for image classification. Although Python has libraries for reading and manipulating DICOM files, it can often be simpler to convert the images to a more widely used format, such as png or jpg, and then use conventional image processing packages to work with the images.

Another reason for converting DICOM images to jpg is that DICOM images have different pixel representations and bit depths, depending on the specific equipment and software used to generate them. Jpg images, on the other hand, have a standardized pixel representation and bit depth, making them more consistent and easier to work with.

Finally, unlike some other picture formats, png, jpg images don't lose any information when they are compressed, which might be crucial in the area of medical imaging, where even minor data loss can have serious repercussions.

In this study, all the DICOM (.dic) file format images are first converted into the .jpg format using MicroDICOM Viewer desktop application. The original image size is 256 × 256 × 3. The images which generate empty tuples are removed from the selected images. Empty tuples are those that create the null arrays for which the machine learning models create a huge number of misclassifications. These images are removed based on the threshold value of 30 pixels. Then images are cropped and stripped using Python library functions. Then, images are normalized using batch normalization. After preprocessing, the final size of the MRI images is 224 × 224 × 3, which is given as input to the models. The original MRI images are shown in Fig. 3a and b.

Fig. 3
figure 3

Original MRI brain images of T1, T2-weighted dataset (a) PD subject (b) HC subject

After pre-processing the images are shown in Fig. 4a and b.

Fig. 4
figure 4

Image resizing and skull stripping of T1,T2-weighted dataset of both the PD and HC subjects in (a) gray scale and in (b) color

Model development

Four deep learning models with the combination of grey wolf optimization technique GWO-VGG16, GWO-DenseNet, GWO-DenseNet-LSTM, GWO-InceptionV3 and a hybrid model GWO-VGG16 + InceptionV3 have been proposed in this study for detection of PD accurately. All the proposed models are explained briefly below:

  • VGG16: VGG16 (Visual Geometry Group 16) [10] is a deep CNN architecture that was suggested by the University of Oxford's Visual Geometry Group in 2014. It is created for image classification problems and has accomplished state-of-the-art performance on various benchmarks, including the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) dataset. Thirteen Conv (convolutional) layers, 3 fully connected dense layers, and other layers made up the 16-layer, VGG16. The input layer accepts an image as input of size 224 × 224 × 3. Each of the 13 convolutional layers is having 3 × 3 filters with a stride of (1). After each max pooling layer, the number of filters doubles i.e. 6 × 6 with a stride of 2, starting with the first convolutional layer that includes 64 filters. The max pooling layers help to decrease the number of model parameters and avoid overfitting by reducing the spatial dimensions of the output by a factor of 2. Padding is a technique that is used by all convolutional layers to guarantee that the output's spatial dimensions match those of the inputs. Rectified linear unit (ReLU) is one of the activation functions that introduces nonlinearity into the model comes after each convolutional layer. It has 2 fully connected layers, each with 256, and 128 neurons respectively. There are 128 neurons in the output layer, corresponding to the two classes in the T1,T2-weighted and SPECT DaTscan datasets. In order to output a VGG16 algorithm is renowned for its ease of use and capacity to extract intricate information from images. However, it can be expensive to train and utilize computationally probability distribution over the classes, it uses a “sigmoid” activation function. VGG16 is a very deep network with huge parameters.

  • DenseNet: DenseNet [11], short for Dense CNN, is a deep learning architecture that Huang et al. have first presented in 2016. It is designed to address the vanishing gradient problem and encourage deep neural networks that reuse features. It creates connections that are dense between all layers. Each layer in this architecture receives feature maps from all levels below it as input. Gradient flow throughout the network is made possible by this connection structure, which provides direct access to features at various depths.

DenseNet is made up of dense blocks, each of which has several levels. Each layer in a dense block is connected to all layers before it. The overall network design is created by gradually connecting dense units. Convolutional and pooling layers are employed as transition layers to shorten the distance between packed blocks. They contribute to preserving connections while lowering computational complexity and feature map sizes. The key advantages of DenseNet are feature reuse, parameter efficiency, and mitigating the vanishing gradient problem.

DenseNet is widely used and has produced state of art outcomes for a number of computer vision applications, such as semantic segmentation, image classification and object recognition. It is now a well-liked option among deep learning researchers and practitioners.

  • DenseNet-LSTM: DenseNet with LSTM [13] refers to a network that combines Long Short-Term Memory (LSTM) networks with the DenseNet network. The strengths of LSTM's modelling of sequential data and ability to detect temporal relationships are combined with DenseNet's feature extraction skills in this hybrid architecture.

The DenseNet component serves as the feature extraction backbone. The dense connections and hierarchical structure aid in the efficient acquisition of both local and global image features. At various degrees of abstraction, the DenseNet layers process the input image or sequence to extract significant information.

Afterwards, an LSTM network receives the output from the DenseNet layers. The LSTM is a form of recurrent neural network (RNN) that excels at modelling sequential data because it preserves long-term dependencies and detects temporal patterns. Memory cells are present in the network, allowing it to selectively recall or forget information over time. DenseNet and LSTM are used in various applications such as video action recognition, natural language processing, sentiment analysis, etc. in order to identify actions or activities, DenseNet captures features from individual frames, while LSTM processes the sequence of features.

  • InceptionV3: InceptionV3 [13] is a variant of the Inception architecture that is introduced by Christian Szegedy et al. in 2015. InceptionV3 is a deep neural network that is created for image classification and object detection tasks. It consists of an input layer, stem network, inception modules, auxiliary classifiers, average pooling, fully connected dense layers and a final (output) layer.

The images are given as input to the input layer, typically of size 224 × 224 × 3. The stem network extracts feature from the input images using three convolutional layers. With a 3 × 3 kernel, the first, second and third layers consist of 32, 32 and 64 filters respectively. The max pooling layer, which follows the stem network, has a 3 × 3 filter with a stride of 2.

There are several inception modules in InceptionV3 that are responsible for doing feature extraction at various scales. Each inception module is made up of a number of convolutional layers with pooling layers and of various filters of sizes (1 × 1, 3 × 3, and 5 × 5) concatenated along the channel dimension. Compared to conventional convolutional layers, Inception modules are computationally inexpensive. Two auxiliary classifiers are included in InceptionV3 after the 5th and 9th inception modules. The auxiliary classifiers are made up of a dropout layer, a softmax activation function, a ReLU activation function, a fully connected layer with 1024 neurons, and a global average pooling layer. The auxiliary classifiers' role includes supplying the network with more training data and minimizing the vanishing gradient issue.

After the last inception module, InceptionV3 utilizes a global average pooling layer to shrink the output's spatial dimensions to a 1 × 1 feature map. A fully connected layer with 128 neurons is fed with the output of the global average pooling layer, which corresponds to the two classes in the T1,T2-weighted and SPECT DaTscan datasets. The fully connected layer outputs a probability distribution over the classes using a sigmoid activation function.

Proposed hybrid model (VGG16 + InceptionV3)

The proposed hybrid model is the fusion of VGG16 [10] and two blocks of InceptionV3 [36] as illustrated in Fig. 5. All sixteen layers of VGG16 makes the Block-1 followed by two blocks Block-2 and Block-3 of inception-reduction. Then in Block-4 global average pooling, fully connected and sigmoid layers are used. The size of the input to the Block-1 (VGG16) is 224 × 224 × 3 and output matrix size from it is 7 × 7 × 512. This output is passed as input to the inception-reduction block (named as Block-2) and output matrix obtained is of the size 7 × 7 × 640. Same process is repeated in Block-3 and size of the output is 3 × 3 × 832. Finally, in Block-4, global average pooling is executed and its output is fed to the fully connected layer to detect the patient as PD or HC.

Fig. 5
figure 5

Architecture of the proposed Hybrid Model (VGG16 + InceptionV3) for PD detection

Figure 6a and b present the detailed structure of Block-2 and Block-3 which are made up of inception-reduction blocks. Four (1 × 1) convolution, three (3 × 3) convolution, and maxpooling of kernel size (3 × 3) are present in the inception block. In comparison to the (3 × 3) convolution, the (1 × 1) convolution has a smaller coefficient which can decrease the number of input channels, and speed up the training process [37]. To extract the image's low-level features, such as edges, lines and corners, the (1 × 1) and (3 × 3) convolution layers are used. These are concatenated and routed to the reduction block. To prevent a representational blockage, the reduction block is employed which is made up of three (3 × 3) convolution, one (1 × 1) convolution and maxpooling layers. The advantage of using this Block-2 is that it lowers the cost and increases the efficiency of the network. The Block-3 is exhibited in Fig. 6b, which is also contains one inception and one reduction blocks. The generated output from the Block-2 is passed as input to the Block-3. To extract high-level features like events and objects, a convolution with a kernel of size (7 × 7) is employed. In place of (7 × 7) convolution, a (7 × 1) and (1 × 7) convolutions are used. The inception block consists of four (1 × 1) convolution and three sets of [(7 × 1) and (1 × 7)], as well as (3 × 3) average pooling layers [36]. Comparing the model to a single (7 × 7) convolution, factorization reduces the model's cost. Afterwards, the reduction block receives all of these layers concatenated together. The reduction block consists of two (1 × 1) convolution, two (3 × 3) convolution, one set of [(7 × 1) and (1 × 7)] and one (3 × 3) maxpooling layer. The output from the Block-3 is passed as input to the Block-4, global average pooling layer which determines the image's overall feature average. After that, the output of global average pooling layer is passed to the fully connected layer. The detailed output of each convolution layer is presented in Table 4. Finally, the predicted class is determined by selecting the class with the highest probability, which is represented by the Sigmoid layer.

Fig. 6
figure 6

Architecture of (a) Block-2 and (b) Block-3 of Inception V3 models

Table 4 The detailed output of each convolution layer of proposed hybrid model

The detailed output size of each layer is given below in Table 4.

Grey Wolf Optimization (GWO)

Seyedali Mirjalili introduced GWO in 2014 by imitating the social conduct, hierarchy of leadership, and hunting on the communal land of grey wolves [6]. Canidae is the family that includes the grey wolf (Canis lupus). As the top predators in the food chain, grey wolves are known as apex predators. The majority of grey wolves prefer to live in packs. The typical size of the group is between 5 and 12 people. Alpha, Beta, Delta and Omega are four different species denoted by (α), (β), (δ) and (ω).

The step-by-step procedure of grey wolf hunting is.

  1. 1.

    tracking, chasing, and approaching the prey.

  2. 2.

    As soon as the target starts moving, it is pursued, hounded, and surrounded.

  3. 3.

    attacking the prey or assaulting it.

In this section, social hierarchy, encircling, and attacking is mathematically represented as follows

  • Social hierarchy: Alpha is the best solution (α) to mathematically express the social hierarchy, followed by (β) and (δ) as the next two best options. The remaining candidate solution is the (ω). α, β, and δ serve as the hunting (or optimization) cues in the GWO algorithm. The remaining ω wolves come after these α, β, and δ wolves.

  • Encircling/Surrounding Prey: Grey wolves circle their prey during hunting. The encircling behavior is mathematically represented as

$$V=\left|\overrightarrow{S}.{\overrightarrow{T}}_{x}\left(t\right)-\overrightarrow{T}(t)\right|$$
(1)
$$\overrightarrow{T}\left(t+1\right)={\overrightarrow{T}}_{x}\left(t\right)-\overrightarrow{U}.\overrightarrow{V}$$
(2)

where the current iteration is denoted by t, coefficient vectors are denoted by S and U, the position of the prey is denoted by Tx, and the grey wolf's position is denoted by T. The vector \(\overrightarrow{S}\) and \(\overrightarrow{U}\) are represented as

$$\overrightarrow{U}=2\overrightarrow{p}.\overrightarrow{{q}_{1}}-\overrightarrow{p}$$
(3)
$$\overrightarrow{S}= 2.\overrightarrow{{q}_{2}}$$
(4)

where q1, q2 are arbitrary vectors with a range of [0, 1] and the components of \(\overrightarrow{p}\) decrease linearly from the value 2 to 0 throughout the course of iterations.

  • Hunting the prey: Grey wolves have the ability to track down and encircle their prey. Typically, the alpha leads the hunt. Hunting may occasionally be done by the beta and delta. It is assumed that the most promising candidate solution, alpha, delta, improve knowledge of the potential prey’s location in order to mathematically replicate the hunting behaviour of the wolves. The three best candidate solutions are mathematically represented to update their position as follows –

Alpha Wolf, Beta Wolf, Delta Wolf

$$\overrightarrow{{V}_{\alpha }}=\left|\overrightarrow{{S}_{1}}.\overrightarrow{{T}_{\alpha }}-\overrightarrow{T}\right|$$
(5)
$$\overrightarrow{{V}_{\beta }}=\left|\overrightarrow{{S}_{2}}.\overrightarrow{{T}_{\beta }}-\overrightarrow{T}\right|$$
(6)
$$\overrightarrow{{V}_{\delta }}=\left|\overrightarrow{{S}_{3}}.\overrightarrow{{T}_{\delta }}-\overrightarrow{T}\right|$$
(7)
$$\overrightarrow{{T}_{1}}=\overrightarrow{{T}_{\alpha }}-\overrightarrow{{U}_{1}}.\left(\overrightarrow{{V}_{\alpha }}\right)$$
(8)
$$\overrightarrow{{T}_{2}}=\overrightarrow{{T}_{\beta }}-\overrightarrow{{U}_{2}}.\left(\overrightarrow{{V}_{\beta }}\right)$$
(9)
$$\overrightarrow{{T}_{3}}=\overrightarrow{{T}_{\delta }}-\overrightarrow{{U}_{3}}.\left(\overrightarrow{{V}_{\delta }}\right)$$
(10)
$$\overrightarrow{T}(t+1)=\frac{\overrightarrow{{T}_{1}}+\overrightarrow{{T}_{2}}+\overrightarrow{{T}_{3}}}{3}$$
(11)

The wolves in motion attack the prey when it stops. \(\overrightarrow{U}\) is a randomly chosen number between -2r and 2r, while r2 is a number between -1 and 1. The search agent's next position is a position that falls somewhere between the object's most recent location and its preyer position. Thus, the attacking state is appropriate when \(\left|\overrightarrow{U}\right|<1\) [38]. The behaviour of wolves is used to depict the process of finding the best solution. Following is the pseudocode of grey wolf optimization-

figure a

Algorithm 1. Hyperparameters optimization of deep learning models using GWO –

Step-by-step procedure:

  • Step-1: Set the ranges of hyperparameter values. The ranges are given in Table 6.

  • Step-2: Set the population size of the grey wolves.

  • Step-3: Create an objective function that measures how well the deep learning models performed after being trained with the provided hyperparameters. This function gauges the model's effectiveness on a validation set.

  • Step-4: Using the wolves' fitness levels, the dominance and hierarchy are determined.

  • Step-5: Update the alpha, beta, delta and omega wolf's position using the Eq. (11)

  • Step-6: Restrict the wolf’s new positions to remain inside each hyperparameter's stated ranges. A location is modified if it exceeds the limits.

  • Step-7: Verify that the termination condition-such as completing the required number of iterations or obtaining the target fitness value-is met. The optimization procedure ends if the condition is satisfied; otherwise, return to step 5.

  • Step-8: Take the optimal collection of hyperparameters for the deep learning models, which corresponds to the solution that is best and represents the wolf with the highest fitness value.

Results and discussion

Prior to discussing the outcome, the fundamental performance evaluation criteria that are frequently used to evaluate different machine learning models while they are still in the training phase as well as in the testing phase are discussed in this section.

Performance evaluation with confusion matrix

The confusion matrix [39] which is a two-dimensional table is used to determine performance metrics. It displays the actual and predicted class values which are represented by its elements as true positive (T + ve), true negative (T-ve), false positive (F + ve), and false negative (F-ve). The classification performance can be quantified using these four elements. Based on the confusion matrix, the five score metrics used in this study are as follows -

$$\mathrm{Accuracy}\;(\mathrm{Acc})\;=\frac{TP+TN}{TP+TN+FP+FN}$$
(12)
$$\mathrm{Sensitivity}\;(\mathrm{Se})\;=\;\frac{TP}{TP+FN}$$
(13)
$$\mathrm{Specificity}\;(\mathrm{Spe})\;=\frac{TN}{TN+FP}$$
(14)
$$\mathrm{Precision}\;(\mathrm P)\;=\;\frac{TP}{TP+FP}$$
(15)
$$\mathrm F1-\mathrm{score}=2\times\left(\frac{Precision\times Recall}{Precision+Recall}\right)$$
(16)

Along with the true + ve rate and false + ve rate, the ROC curve is shown on a graph which is known as the receiver operator characteristic. An area under the curve (AUC) score, or area under the curve, is also obtained.

Experimental results analysis of all the proposed models

Experimental results are described by the following subsections that display the training results by plotting the accuracy and loss curves for each model used. For each model, the confusion matrix is also created and displayed.

Training results of all the proposed deep learning models

The experiments are carried out in Python by using various packages such as keras, opencv, tensorflow 2.1, scikitlearn [40] using the system configuration of intel Core i5 processor, 8th Generation, with 16 GB RAM, and NVIDIA GEFORCE graphics combined with 8 GB memory. The standard T1, T2-weighted and SPECT DaTscan datasets are used for the study. The datasets are split into two sets i.e. train and test using an 80:20 ratio. Again, the training set is then split into train and validation sets. To train all the proposed deep learning models with GWO using the algorithm given in “Image pre-processing” section, various hyperparameters used are shown in Table 5 and the optimized hyperparameters by GWO are shown in Table 6.

Table 5 Hyperparameters used in all the proposed models
Table 6 Optimized hyperparameters using GWO

All the proposed deep learning models GWO-VGG16, GWO-DenseNet, GWO-DenseNet-LSTM, GWO-InceptionV3 and hybrid model GWO-VGG16 + InceptionV3 are pre-trained using the above hyperparameters. All the models comprise of an input layer, two hidden layers and an output layer. Every model has its own layers, such as convolutional, maxpooling, stem, global average pooling etc. Each layer consists of 256, and 128 neurons respectively. Every hidden layer ends with a dropout layer with 20 percent of neurons dropping out to overcome the overfitting problem. ReLu activation function is employed to all the hidden layers. To train all the models ‘adam’ optimizer and loss function ‘binary_crossentropy’ is used. The GWO algorithm is used for hyperparameter optimization with all the proposed models to obtain better performance. The ranges of parameters are given manually and optimized hyperparameters are shown in Table 6.

Testing results of all the proposed deep learning models

A fully separate data subset that is previously prepared, is used to test and evaluate the effectiveness of the proposed models. Figures 7a-e and 8a-e show the confusion matrix for all the proposed models for both the datasets i.e. T1,T2-weighted MRI and SPECT DaTscan respectively.

Fig. 7
figure 7

Confusion matrix for all the proposed models (a) GWO-VGG16, (b) GWO-DenseNet, (c) GWO-DenseNet-LSTM, (d) GWO-InceptionV3 and (e) Hybrid model (GWO-VGG16 + InceptionV3) using T1,T2-weighted MRI dataset

Fig. 8
figure 8

Confusion matrix of all the proposed models (a) GWO-VGG16, (b) GWO-DenseNet, (c) GWO-DenseNet-LSTM, (d) GWO-InceptionV3 and (e) Hybrid model (GWO-VGG16 + InceptionV3) using SPECT DaTscan dataset

Experimental results and discussions

Before obtaining the results of the proposed models, various preprocessing and hyperparameter optimization techniques are applied to obtain better accuracy and other performance measure results. Four deep learning models VGG16, DenseNet, DenseNet + LSTM, InceptionV3 and a hybrid model VGG16 + InceptionV3 are trained using the two standard datasets T1,T2-weighted and SPECT DaTscan with 9070 and 20,096 images. The results are briefly explained below for both datasets.

Results using T1,T2-weighted MRI dataset

The results evaluation and comparison of all the proposed models are presented in Table 7.

Table 7 Results of the proposed models using T1,T2-weighted dataset

The results of Table 7 demonstrate that all the proposed models achieved more than 99% of testing accuracy except GWO-DenseNet + LSTM which resulted in 98.29% accuracy and the hybrid model (GWO-VGG16 + InceptionV3) obtained highest accuracy 99.94% with the training loss of 0.0272 which is minimum among all models.

Results using SPECT DaTscan MRI dataset

The result evaluations and comparison of all the proposed models are given in Table 8.

Table 8 Results of the proposed models using SPECT DaTscan dataset

The aforementioned table demonstrates that all the proposed models achieved more than 99% testing accuracy. The hybrid model GWO-VGG16 + InceptionV3 obtained exactly 100% testing accuracy with the training loss values 0.0153 in comparison to other models.

The description of the validation of proposed models using independent data for both datasets is given in Table 9.

Table 9 MRI data sample distribution for independent dataset

Results using SPECT DaTscan MRI dataset

The results of Table 10 demonstrate that all the proposed models performed well and the hybrid model (GWO-VGG16 + InceptionV3) obtained highest accuracy of 99.54% with the training loss of 0.0027 which is minimum among all models.

Table 10 Validation of the proposed models using independent data of T1,T2-weighted dataset

Results using SPECT DaTscan MRI dataset

Similarly, the above Table 11, results exhibit that all the proposed models performed well and the hybrid model (GWO-VGG16 + InceptionV3) resulted 99.11% accuracy during testing with the training loss values 0.0041 in comparison to other models.

Table 11 Validation of the proposed models using independent data of SPECT DaTscan dataset

Comparison with the existing models

The proposed models' outcomes are presented in Tables 12 and 13 along with comparisons to other previously reported models. The comparison exhibits that for both datasets, the proposed deep learning models beat all other existing models in terms of performance metrics like accuracy, sensitivity, specificity, precision, f1-score and AUC score.

Table 12 Proposed models comparison with existing models using T1,T2-weighted dataset
Table 13 Proposed models’ comparison with the existing models using SPECT DaTscan dataset

The above table shows that, in terms of accuracy, from all the proposed deep learning models, hybrid model GWO-VGG16 + InceptionV3 outperforms the other eleven existing models and obtained 99.92% accuracy which is nearly similar to the model proposed by [22] with 100% accuracy.

Table 13 demonstrates that, in terms of accuracy, from all the proposed models, hybrid model GWO-VGG16 + InceptionV3 outperforms the other existing models with 100% accuracy.

When doing comparison on the basis of accuracy the reported machine learning models give the values as 99.01(Men), 96.97% (Women) [16], 85% [17], 96.4 [41] which is less than the result of the proposed deep learning models (99.94%) in T1,T2 weighted datasets. The papers [18] and [39] do comparison on the basis of AUC rather than accuracy and the values are 94.2% and 98% respectively. In the present investigation, the proposed deep learning models demonstrated an improvement of 1.9% to 4.8% in the AUC values in comparison to the existing models given in [18] and [39]. Similarly for the SPECT DaTscan dataset the proposed deep learning models exhibit better performance with 100% accuracy and 99.92% AUC in comparison to the reported literature [42] and [11] that give AUC of 97% and 99% respectively. Hence, in overall comparison, the proposed deep learning based models outperform the existing models in terms of accuracy and AUC values. The deep learning classification models presented in this study show the potential of such computational tools as future assistive diagnostic solution for doctors.

Limitations of the study

The study has certain research limitations which are listed below :

  • 1.To implement it needs high memory space more than 16GB RAM, a high-end GPU system in this case.

  • 2.Because of the high-dimension data, it takes more time to execute and also increases the time and space complexity.

  • 3.Only binary class classification problem is used for early prediction of Parkinson’s disease. Multiclass classification can also be done.

Conclusion and future work

The detection of Parkinson's disease is becoming more and more crucial today. Because PD is a tremor illness, it is increasingly difficult to make an accurate diagnosis of the condition, especially in the early stages. This study proposes a classification approach for Parkinson's disease (PD) detection that enables doctors to make an accurate and timely diagnosis. The novelty of the paper is development of the hybrid model containing VGG16 and part of InceptionV3 whose hyperparameters are updated using GWO for PD detection which is first of its kind. The paper proposes four deep learning models, VGG16, DenseNet, DenseNet + LSTM, InceptionV3 and a hybrid model VGG16 + InceptionV3. In order to optimize the hyperparameters of proposed deep learning models, the GWO algorithm is used which automatically fine tune the hyperparameters and enhance the performance of the models. Two datasets T1, T2-weighted and SPECT DaTscan are used for the experiment. Various preprocessing techniques are applied to the images to enhance the models' functionality. After pre-processing, the GWO optimization algorithm is fitted into all the models and efficiently encoded. The hybrid model GWO-VGG16 + InceptionV3 has obtained 99.92% of accuracy and 99.99% AUC with T1,T2 dataset. While the same hybrid model has resulted 100% of accuracy and 99.92% AUC for SPECT DaTscan dataset. The paper also validates the effectiveness of the proposed hybrid model using independent data and demonstrated its superiority with accuracy and AUC values of 99.54% and 99.56% for T1,T2 weighted dataset and 99.11% accuracy with 99.15% AUC for SPECT DaTscan dataset respectively.

The work can be expanded in the future by adding new hyperparameter tuning techniques. Feature extraction will be done using Region of Interest (ROI) of two regions caudate and putamen in SPECT DaTscan dataset. Segmentation will be applied to the MRI images to detect the Parkinson’s disease. Also, work can be done on multiclass classification problems and more no of patients with larger MRI images.

The proposed model will help the physicians to detect the PD disease before the detection of motor symptoms which help them to initiate treatment timely and make proper strategy for better treatment and hence improved quality of their lives. Secondly, the proposed algorithm can be integrated with wearable sensors and devices for real time tracking of PD disease and hence can reduce the treatment time.

Availability of data and materials

The datasets generated and/or analysed during the current study are available in the PPMI repository, https://www.ppmi-info.org/.

References

  1. Michael J. For Foundation for Parkinson Research, Parkinson’s disease causes, (Retrieved from https://www.michaeljfox.org/understanding-parkinsons/living-with-pd.html), 12 April 2023.

  2. Bhat S, Acharya UR, Hagiwara Y, Dadmehr N, Adeli H. Parkinson’s disease: cause factors, measurable indicators, and early diagnosis. Comput Biol Med. 2018;102:234–41.

    Article  PubMed  Google Scholar 

  3. Abdulhay E, Arunkumar N, Kumaravelu N, Vellaiappan E, Venkatraman V. Gait and tremor investigation using machine learning techniques for the diagnosis of Parkinson disease. Futur Gener Comput Syst. 2018;83:366–73.

    Article  Google Scholar 

  4. Lundervold AS, Lundervold A. An overview of deep learning in medical imaging focusing on MRI. Z Med Phys. 2019;29(2):102–27.

    Article  PubMed  Google Scholar 

  5. Acharya UK, Kumar S. Particle swarm optimized texture based histogram equalization (PSOTHE) for MRI brain image enhancement. Optik, Science Direct. 2020;224:165760. https://doi.org/10.1016/j.ijleo.2020.165760.

    Article  Google Scholar 

  6. Mirjalili S, Mirjalili SM, Lewis A. Grey wolf optimizer. Adv Eng Softw. 2014;69:46–61. Elsevier.

    Article  Google Scholar 

  7. Dorigo M, Birattari M, Stutzle T. Ant colony optimization. IEEE Comput Intell Magaz. 2006;1(4):28–39.

    Article  Google Scholar 

  8. Zhao F, Wang Z, Wang L, Xu T, Zhu N. A multi-agent reinforcement learning driven artificial bee colony algorithm with the central controller. Expert Syst Appl. 2023;219:119672. https://doi.org/10.1016/j.eswa.2023.119672. Elsevier.

    Article  Google Scholar 

  9. Shaikh MS, Hua C, Jatoi MA, Ansari MM, Qader AA. Application of grey wolf optimization algorithm in parameter calculation of overhead transmission line system. IET Sci Meas Technol. 2021;15(2):218–31.

    Article  Google Scholar 

  10. Magesh P, Myloth R, Tom R. An explainable machine learning model for early detection of Parkinson’s disease using LIME on DaTscan imagery. Comput Biol Med. 2020;126:104041. https://doi.org/10.1016/j.compbiomed.2020.104041.

    Article  CAS  PubMed  Google Scholar 

  11. Thakur M, Kuresan H, Dhanalakshmi S, Lai KW, Wu X. Soft attention based DenseNet model for Parkinson’s disease classification using SPECT images. Front Aging Neurosci. 2022;13(14):908143. https://doi.org/10.3389/fnagi.2022.908143.

    Article  Google Scholar 

  12. Kurmi S, Shreya S, Sen A, Sinitca D, Sarkar R. An ensemble of CNN models for Parkinson’s disease detection using DaTscan images. Diagnostics. 2022;12:1173. https://doi.org/10.3390/diagnostics12051173.

    Article  PubMed  PubMed Central  Google Scholar 

  13. Basnin N, Nahar N, Anika FA, Hossain MS, Andersson K. Deep learning approach to classify Parkinson’s disease from MRI samples, Brain Informatics. 2021:12960. https://doi.org/10.1007/978-3-030-86993-9_48. Springer.

  14. Camacho M, et al. Explainable classification of Parkinson’s disease using deep learning trained on a large multi-center database of T1-weighted MRI datasets. NeuroImage Clin. 2023;38:103405. https://doi.org/10.1016/j.nicl.2023.103405.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Baagil H. Neural correlates of impulse control behaviors in Parkinson’s disease: Analysis of multimodal imaging data. Neuroimage Clin. 2018;37:103315. https://doi.org/10.1016/j.nicl.2023.103315.

    Article  Google Scholar 

  16. Solana-Lavalle G, Rosas-Romero R. Classification of PPMI MRI scans with voxel-based morphometry and machine learning to assist in the diagnosis of Parkinson’s disease. Comput Methods Programs Biomed. 2021;198:105793. https://doi.org/10.1016/j.cmpb.2020.105793.

    Article  PubMed  Google Scholar 

  17. Talai AS, Sedlacik J, Boelmans K, Forckert ND. Utility of multi-modal MRI for differentiating of Parkinson’s disease and progressive supranuclear palsy using machine learning. Front Neurol. 2021;14(12):648548. https://doi.org/10.3389/fneur.2021.648548.

    Article  Google Scholar 

  18. Chakraborty S, Aich S, Kim HC. Detection of Parkinson’s disease from 3T T1 weighted MRI scans using 3D convolutional neural network. Diagnostics (Basel). 2020;10(6):402.

    Article  PubMed  PubMed Central  Google Scholar 

  19. Wingate J, Kollia I, Bidaut L, Kollias S. A unified deep learning approach for prediction of Parkinson’s disease. arXiv e. 2019. https://doi.org/10.48550/arXiv.1911.10653.

  20. Mostafa TA, Cheng I. Parkinson’s Disease Detection Using Ensemble Architecture from MR Images *. In: IEEE 20th International Conference on Bioinformatics and Bioengineering (BIBE). 2020. p. 987–92.

    Google Scholar 

  21. Sivaranjini S, Sujatha C. Deep learning based diagnosis of Parkinson’s disease using convolutional neural network. Multimedia Tools and Applications. 2020;79:15467–79. https://doi.org/10.1007/s11042-019-7469-8.

    Article  Google Scholar 

  22. Esmaeilzadeh S, Yao Y, Adeli E. End-to-end Parkinson disease diagnosis using brain MR-images by 3D-CNN. arXiv. 2018;1–7. https://doi.org/10.48550/arXiv.1806.05233.

  23. Shah PM, Zeb A, Shafi U, Zaidi SFA, Shah MA. Detection of Parkinson's disease in brain MRI using convolutional neural network. 2018 24th International Conference on Automation and Computing (ICAC), Newcastle Upon Tyne, UK; 2018. p.1–6. https://doi.org/10.23919/IConAC.2018.8749023.

  24. Mei J, Tremblay C, Stikov N, Desrosiers C, Frasnelli J. Differentiation of Parkinson’s disease and non-parkinsonian olfactory dysfunction with structural MRI data. Computer-Aided Diagnosis. International Society for Optics and Photonics; 2021. p.11597. 115971E. https://doi.org/10.1117/12.2581233.

  25. Pugalenthi R, Rajakumar RM, Ramya J, Rajinikanth V. Evaluation and classification of the brain tumor MRI using machine learning technique. J Control Eng Appl Inform. 2019;21(4):12–21.

    Google Scholar 

  26. Leung KH, Rowe SP, Pomper MG. A three-stage, deep learning, ensemble approach for prognosis in patients with Parkinson’s disease. EJNMMI Res. 2021;11:52. https://doi.org/10.1186/s13550-021-00795-6. SpringerOpen.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  27. Mohammed F, He X, Lin Y. An easy-to-use deep-learning model for highly accurate diagnosis of Parkinson’s disease using SPECT images. Comput Med Imaging Graph. 2021;87:101810. https://doi.org/10.1016/j.compmedimag.2020.101810.

    Article  PubMed  Google Scholar 

  28. Pianpanit T, et al. Neural network interpretation of the Parkinson’s disease diagnosis from SPECT imaging. arXiv: Image and Video Processing. 2019;1–7.

  29. Chien CY, Hsu SW, Lee TL, Sung PS, Lin CC. Using artificial neural network to discriminate Parkinson’s disease from other parkinsonism’s by focusing on putamen of dopamine transporter SPECT images: a retrospective study. Res Dev Med Med Sci. 2023;5:10–27.

    Google Scholar 

  30. Nalini TS, Anusha MU, Umarani K. Parkinson’s disease detection using spect images and artificial neural network for classification. Int J Eng Res Technol (IJERT) IETE. 2020;8(11):105–8.

  31. Kollia, Stafylopatis AG, Kollias S. Predicting Parkinson’s disease using latent information extracted from deep neural networks. In 2019 international joint conference on neural networks. IEEE; 2019. p. 1–8.

  32. Rumman M, Tasneem AN, Farzana S, Pavel MI, Alam MA. Early detection of Parkinson’s disease using image processing and artificial neural network, 2018 Joint 7th International Conference on Informatics, Electronics & Vision (ICIEV) and 2018 2nd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), Kitakyushu, Japan. 2018. p. 256–261.

  33. Martínez-Murcia F, et al. A 3D convolutional neural network approach for the diagnosis of Parkinson’s disease. In: International work conference on the interplay between natural and artificial computation, Springer;2017. p. 324–333. https://doi.org/10.1007/978-3-319-59740-9_32.

  34. MJFF. The Michael J Fox Foundation for Parkinson’s Research [WWW Document], 13 November 2022. https://www.michaeljfox.org.

  35. Marek K, et al. The Parkinson progression marker initiative (ppmi). Prog Neurobiol. 2011;95(4):629–35.

    Article  PubMed Central  Google Scholar 

  36. Srinivas K, Sri R, Pravallika K, Nishitha K, Polamuri D. COVID-19 prediction based on hybrid Inception V3 with VGG16 using chest X-ray images, Multimedia Tools and Application. 2023. p. 1–18. https://doi.org/10.1007/s11042-023-15903-y.

  37. Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer vision. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR). 2016. p. 2818–2826. https://doi.org/10.1109/CVPR.2016.308.

  38. Mohakud R, Dash R. Designing a grey wolf optimization based hyper-parameter optimized convolutional neural network classifier for skin cancer detection. J King Saud Univ Comput Inf Sci. 2021;34(8):6280–91.

    Google Scholar 

  39. Han J, Pei J, Kamber M. Data mining : concepts and techniques. Elsevier: Morgan Kaufmann Publishers; 2011.

    Google Scholar 

  40. Pedregosa F, Weiss R, Brucher M. Scikit-learn: machine learning in python. J Mach Learn Res. 2011;12:2825–30.

    Google Scholar 

  41. Lei H, et al. Sparse feature learning for multi-class Parkinson’s disease classification. Technol Health Care. 2018;26:1–11.

    Article  Google Scholar 

  42. Ramamurthy P, Rajakumar MP, Ramya J, Venkatesan R. Evaluation and classification of the brain tumor MRI using machine learning technique. Control Eng Appl Inform. 2019;21:12–21.

    Google Scholar 

  43. Siddiqi MH, et al. A precise medical imaging approach for brain MRI image classification. Comput Intell Neurosci. 2022;2022(6447769):1–15. https://doi.org/10.1155/2022/6447769.

  44. Ortiz J, et al. Parkinson’s disease detection using isosurfaces-based features and convolutional neural networks. Front Neuroinform. 2019;13:48.

    Article  PubMed  PubMed Central  Google Scholar 

Download references

Funding

Z.Z. was partially supported by the Professorship Fund. The funder had no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. A.L. was partially supported by Natural Science Basic Research Program of Shaanxi (Program No. 2024JC-YBMS-484).

Author information

Authors and Affiliations

Authors

Contributions

B.M., A.K., S.S.M. developed the method, wrote the main manuscript text, and S.D., S.M. prepared the figures and validated the result, A.L., Z.Z. edited the manuscript.  All authors have read and agreed to the published version of the manuscript.

Corresponding authors

Correspondence to Saurav Mallik or Zhongming Zhao.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Majhi, B., Kashyap, A., Mohanty, S.S. et al. An improved method for diagnosis of Parkinson’s disease using deep learning models enhanced with metaheuristic algorithm. BMC Med Imaging 24, 156 (2024). https://doi.org/10.1186/s12880-024-01335-z

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12880-024-01335-z

Keywords