0% found this document useful (0 votes)
21 views11 pages

Paper 3

Uploaded by

arefin.aktar02
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views11 pages

Paper 3

Uploaded by

arefin.aktar02
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 11

Proceeding Paper

Early Detection of Alzheimer’s Disease: An Extensive Review


of Advancements in Machine Learning Mechanisms Using
an Ensemble and Deep Learning Technique †
Renjith Prabhavathi Neelakandan 1, * , Ramesh Kandasamy 2 , Balasubramani Subbiyan 3
and Mariya Anto Bennet 4

1 School of Computer Science and Engineering, Vellore Institute of Technology, Chennai Campus,
Chennai 603103, Tamilnadu, India
2 Department of Computer Science and Engineering, Sri Krishna College of Engineering and Technology,
Coimbatore 641008, Tamilnadu, India; rameshk@skcet.ac.in
3 Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation,
Vaddeswaram 522502, Andrapradesh, India; bala08.ap@gmail.com
4 Department of Electronics and Communication Engineering, Vel Tech Rangaranjan Dr. Sagunthala R & D
Institute of Science and Technology, Chennai 600062, Tamilnadu, India; drmantobenet@veltech.edu.in
* Correspondence: renjith.pn@vit.ac.in
† Presented at the International Conference on Recent Advances on Science and Engineering, Dubai,
United Arab Emirates, 4–5 October 2023.

Abstract: Alzheimer’s disease (AD) is the most common form of dementia in senior individuals. It is
a progressive neurological ailment that predominantly affects memory, cognition, and behavior. An
early AD diagnosis is essential for effective disease management and timely intervention. Due to its
complexity and heterogeneity, AD is, however, difficult to diagnose precisely. This paper investigates
the integration of disparate machine learning algorithms to improve AD diagnostic accuracy. The
used dataset includes instances with missing values, which are effectively managed by employing
appropriate imputation techniques. Several feature selection algorithms are applied to the dataset
to determine the most relevant characteristics. Moreover, the Synthetic Minority Oversampling
Citation: Neelakandan, R.P.;
Technique (SMOTE) is employed to address class imbalance issues. The proposed system employs an
Kandasamy, R.; Subbiyan, B.; Bennet,
Ensemble Classification algorithm, which integrates the outcomes of multiple predictive models to
M.A. Early Detection of Alzheimer’s
enhance diagnostic accuracy. The proposed method has superior disease prediction capabilities in
Disease: An Extensive Review of
Advancements in Machine Learning
comparison to existing methods. The experiment employs a robust AD dataset from the UCI machine
Mechanisms Using an Ensemble and learning repository. The findings of this study contribute significantly to the field of AD diagnoses
Deep Learning Technique. Eng. Proc. and pave the way for more precise and efficient early detection strategies.
2023, 59, 10. https://doi.org/
10.3390/engproc2023059010 Keywords: Alzheimer’s disease (AD); machine learning; early detection; diagnosis; Ensemble
Classification
Academic Editors: Nithesh Naik,
Rajiv Selvam, Pavan Hiremath,
Suhas Kowshik CS and Ritesh
Ramakrishna Bhat
1. Introduction
Published: 11 December 2023 Alzheimer’s disease is a progressive neurological illness that affects memory, cognition,
and behavior in senior individuals and is the most common cause of dementia [1–3]. An AD
diagnosis should be accurate and timely to ensure effective disease management and timely
Copyright: © 2023 by the authors.
intervention, resulting in improved patient care and potential therapeutic interventions.
Licensee MDPI, Basel, Switzerland. For clinicians, the precise diagnosis of AD is a challenging task due to its complexity and
This article is an open access article heterogeneity. In recent years, machine learning has emerged as a powerful tool in medical
distributed under the terms and diagnoses, offering the potential to augment traditional diagnostic approaches and improve
conditions of the Creative Commons accuracy [4,5]. Figure 1 represents the IoT-based patient monitoring system. Through the
Attribution (CC BY) license (https:// integration of disparate machine learning algorithms, this study aims to improve AD
creativecommons.org/licenses/by/ diagnosis accuracy by leveraging the capabilities of machine learning. Multi-algorithm
4.0/). approaches strive to overcome the limitations of individual models by harnessing the

Eng. Proc. 2023, 59, 10. https://doi.org/10.3390/engproc2023059010 https://www.mdpi.com/journal/engproc


Eng. Proc. 2023, 59, 10 FOR PEER REVIEW 2 of 11
Eng. Proc. 2023, 59, 10 2 of 11

models by harnessing
predictive capabilitiesthe
of predictive capabilities This
multiple algorithms. of multiple algorithms.
study aims This Alzheimer’s
to enhance study aims
todisease
enhance Alzheimer’s disease (AD) diagnoses using
(AD) diagnoses using machine learning techniques.machine learning techniques.

Analyzing

Gateway Network & storage

Patient fixed with sensors

Figure 1. IoT-based patient monitoring system.


Figure 1. IoT-based patient monitoring system.
The method involves working with a dataset containing missing values. To address
this,The method involves
imputation techniques working with a dataset
are employed, containing
ensuring datasetmissing
integrityvalues. To address
and analysis qual-
this, imputation techniques are employed, ensuring dataset integrity
ity [6–8]. Additionally, a feature selection algorithm identifies key dataset characteristics and analysis quality
[6–8].
crucialAdditionally,
for accurateaAD feature selection
prediction. algorithm
Class identifies
imbalance, common key dataset
in medicalcharacteristics
datasets, iscru-tack-
cial
ledfor accurate
using AD prediction.
the Synthetic Minority Class imbalance, common
Oversampling Techniquein(SMOTE)medical datasets,
[9,10]. This is tackled
ensures
using the Synthetic
unbiased predictive Minority
modelsOversampling
and more reliable Technique
disease (SMOTE)
prediction.[9,10].ToThis ensures
further un-
improve
biased predictive models and more reliable disease prediction.
AD prediction accuracy, an automated method using machine learning is proposed. An To further improve AD
prediction
Ensemble accuracy, an automated
Classification algorithm method
is applied, using machinemultiple
combining learningpredictive
is proposed. An En-
models [11].
semble Classification
This approach enhancesalgorithm is applied,
AD detection combining
reliability multiple predictive
by amalgamating results models [11].
from various
This approachFeature
algorithms. enhances AD detection
extraction extractsreliability
fresh data byfeatures,
amalgamating results selection
while feature from various picks
algorithms.
significant Feature extraction
characteristics. extracts fresh
Appropriate featuredata features,
selection while feature
algorithms selection
are vital picks
for accurate
significant
predictioncharacteristics.
[12]. Algorithms Appropriate
like Mutualfeature selection
Information algorithms
Scores, are vital
Relief, and for accurate
Recursive Feature
prediction
Elimination [12]. Algorithms
efficiently likeessential
extract Mutual features.
Information Scores, Relief,
A Univariate and assigns
Analysis Recursive Feature
significance
scores to each
Elimination feature.extract
efficiently Class distribution is balanced
essential features. with SMOTE
A Univariate Analysis [13],assigns
addressing the
signifi-
unbalanced
cance scores to dataset issue byClass
each feature. increasing minority
distribution class rows,
is balanced withthereby
SMOTEboosting minority
[13], addressing
class
the classifier accuracy
unbalanced [14].by
dataset issue Other studiesminority
increasing have used advanced
class machine
rows, thereby learningminority
boosting methods
suchclassifier
class as deep learning
accuracyand convolutional
[14]. Other studies neural
have networks
used advancedfor ADmachine
diagnoses [15,16].meth-
learning Novel
biomarkers
ods and multimodal
such as deep learning and data integration [17,18]
convolutional neuralhave shownfor
networks promise in predicting
AD diagnoses AD.
[15,16].
This study utilizes a robust AD dataset from UCI’s machine
Novel biomarkers and multimodal data integration [17,18] have shown promise in pre- learning repository, ensuring
findings’
dicting AD.reliability
This study and generalizability.
utilizes a robust ADThe proposed
dataset approach
from UCI’s outperforms
machine existing
learning reposi-
methods in disease prediction after rigorous evaluation. Beyond
tory, ensuring findings’ reliability and generalizability. The proposed approach outper- an AD diagnosis, this
studyexisting
forms holds broader
methods implications.
in disease It offers insights
prediction into machine
after rigorous learningBeyond
evaluation. applications
an ADin
medical diagnoses,
diagnosis, this study facilitating
holds broader precise early detection
implications. It offersacross
insightsmedical domains.
into machine As AD
learning
and neurodegenerative disorders rise, accurate diagnostic methods
applications in medical diagnoses, facilitating precise early detection across medical do- are vital for patient
outcomes
mains. As AD and healthcare
and management.
neurodegenerative Moreover,
disorders this research
rise, accurate advances
diagnostic machine
methods learn-
are vital
for patient outcomes and healthcare management. Moreover, this research advances ma-is
ing’s role in medical research, beyond an AD diagnosis. The remainder of the paper
arranged
chine as follows:
learning’s role inSection
medical 2 explains
research,the materials
beyond an ADanddiagnosis.
methodology, Section 3 presents
The remainder of the
the experimentation and performance assessment, Section
paper is arranged as follows: Section 2 explains the materials and methodology, 4 demonstrates Experimental
Section 3
Configuration,
presents Results, andand
the experimentation Discussion,
performance andassessment,
Section 5 concludes with a discussion
Section 4 demonstrates Exper-of
future work.
imental Configuration, Results, and Discussion, and Section 5 concludes with a discussion
of2.future work.
Materials and Methods/Methodology
2.1. Literature Review
2. Materials and Methods/Methodology
Ensemble learning enhances an AD diagnosis through 3D convolutional neural net-
2.1. Literature Review
works and MRI. Such networks can distinguish between healthy individuals, those with
mildEnsemble
cognitivelearning enhances
impairment, an AD
and AD diagnosis
patients [1,2]. through
Transfer3D convolutional
learning neural net-
aids in detecting AD
works and MRI. Such networks can distinguish between healthy individuals,
with these networks [3]. Techniques like deep aggregation learning and stacking-based those with
en-
mild
semblecognitive impairment,
learning and
with genetic AD patients [1,2].
hyperparameter Transfer
tweaking learning
improve aids in detecting
diagnostic accuracy AD
[4,5].
with
Using these
the networks [3]. Techniques
ADNI dataset, MRI-basedlike deep aggregation
ensemble learning95.2%
learning achieves and stacking-based
accuracy distin-
ensemble
guishing learning
AD from with
NC andgenetic
77.8%hyperparameter tweakingsMCI
accuracy distinguishing improve
fromdiagnostic
pMCI [11].accuracy
However,
[4,5]. Using the ADNI dataset, MRI-based ensemble learning achieves 95.2%
this dataset’s limitations include sample size and lack of method comparison with other accuracy
Eng. Proc. 2023, 59, 10 3 of 11

state-of-the-art techniques. Additionally, the research does not assess the method’s inter-
pretability, potentially limiting its application [11]. Another study introduces an ensemble
learning architecture using 2D CNNs for an AD diagnosis [12]. This method trains on grey
matter density maps and uses ensemble models to improve prediction accuracy. How-
ever, its limitations include reliance on 2D MRI images and the need for testing on larger
datasets [12]. Research on machine learning for AD diagnoses using neuroimaging data
explored techniques like Support Vector Machines and CNNs [13,14]. While some methods
achieve significant accuracy, they often face challenges with real-world healthcare data or
require further testing on more extensive datasets [15].
A study using a stacking-genetic algorithm ensemble learning model reached a high
accuracy, precision, recall, and F1-score in early AD diagnoses [15]. Nevertheless, issues like
variable dataset validation and clinical interpretability remain. Combining MRI classifiers
offers reliable AD detection, but its applicability requires further exploration [16]. On the
other hand, Random Forest achieves high accuracy predicting AD using limited features
from MRI scans [17]. Deep learning has shown potential in AD diagnoses, especially when
studying complex disease pathways [18]. Still, its reliability in predicting AD progression
needs rigorous testing across various imaging modalities and larger datasets. The use of
a deep CNN for a stage-based AD diagnosis shows promise, but a comprehensive method-
ology comparison and general applicability assessment are essential [19]. Other methods,
such as high-pressure liquid chromatography with AI algorithms, offer insights into pre-
dicting Alzheimer’s medication properties [20]. Deep learning techniques integrating
expert knowledge and multi-source data have outperformed many ensemble methods [21].
However, the system might need substantial computational resources and could vary across
datasets. Research on ensemble learning with Conformal Predictors indicates improved
categorization, but a broader dataset is essential for validation [22]. Hierarchical ensem-
ble learning addresses some deep learning challenges, providing enhanced classification
accuracy with pre-trained neural networks [23]. However, this may require substantial
training datasets and high-quality MRI scans. Lastly, ensemble learning for regression
problems shows potential in predicting medication effects, but needs expansion for broader
applications [14,24]. Ensemble learning and advanced algorithms demonstrate significant
promise in AD diagnoses [25,26]. However, broader dataset validations, methodology
comparisons, and evaluations of real-world applicability are crucial.

2.2. Proposed Work


In the initial phase of the system, categorical attributes are converted into numeric
attributes (0 s and 1 s). The absent values in the dataset are then handled using the median
value. Feature extraction creates new data features. Next, feature selection is used to
find disease–diagnosis-relevant traits. Accurate prediction requires this step. Several
feature selection techniques are researched to choose the most useful characteristics for
an AD diagnosis. After declaring a set number of features, Recursive Feature Elimination
(RFE) removes them. A Univariate Analysis evaluates each attribute numerically. PCA
reduces dimensionality while maintaining useful data. Mutual Information Scores and
Relief automatically choose relevant features to accelerate a diagnosis. SMOTE is used to
oversample the minority class in AD datasets to address class imbalance. Fair categorization
datasets result. Ensemble classification mixes model predictions to efficiently handle textual
characteristics. Aggregating label forecasts and forecasting the majority vote improve
classification accuracy. A comprehensive and sophisticated ensemble-based model aims
to improve AD diagnoses. The model extracts critical characteristics and uses a balanced
dataset by merging several feature selection techniques and SMOTE for class imbalance,
boosting illness prediction accuracy. The ensemble classification strengthens the model’s
textual feature management. The planned study will enhance AD detection and diagnoses,
improving patient outcomes and healthcare management.
Eng. Proc. 2023, 59, 10 4 of 11

A. Pre-processing
The pre-processing phase prepares the raw AD dataset for an analysis. Categorical
attributes are transformed to numeric for compatibility with machine learning. Median
imputation addresses missing values, ensuring data completeness. Feature extraction
enriches the dataset, while feature selection pinpoints the most informative attributes.
Recursive Feature Elimination (RFE) removes less vital features iteratively. A Univariate
Analysis ranks features based on their importance. A Principal Component Analysis (PCA)
compresses data without losing critical information. This rigorous preparation creates a
solid foundation for the ensemble-based AD diagnosis model.
B. Extraction and Selection of Features
In the ensemble-based model for an AD diagnosis, feature extraction transforms the
raw AD dataset to capture essential patterns, enhancing its richness for better prediction.
Feature selection then identifies the most critical characteristics within this dataset. Several
algorithms assess which features most influence diagnostic accuracy. Recursive Feature
Elimination (RFE) methodically removes less important features to streamline the model,
while a Univariate Analysis ranks each feature’s significance in classification. A Principal
Component Analysis (PCA) compresses data, retaining essential variance for a concise rep-
resentation. By using these feature extraction and selection methods, the model highlights
the AD dataset’s key aspects, improving prediction accuracy and supporting early disease
detection for improved patient results.
Given a dataset X of size n × m (n samples, m features),

X ∈ R( n × m ) (1)

Compute the mean of each feature and subtract it from the corresponding feature in X,
resulting in a zero-mean dataset Xcentered .

1
Xcentered = s = X − × ( ∑ ( i = 1 ) n × n · Xi ) (2)
n
Calculate the covariance matrix.
Xcentered
C( Xcentered ) = Xcentered T × (3)
( n − 1)

Compute the eigen values (λ) and eigenvectors (v) of the covariance matrix C.

Xcentered
C = Xcentered T × (4)
( n − 1)
Z m
C × vi = λi × v i (5)
1

V of size m × k : V = [v1 , v2 , . . . vk ] (6)

Xnew of size n × k : Xnew = Xcentered × V (7)


C. Synthetic Minority Oversampling Technique (SMOTE)
The proposed ensemble-based model for AD diagnoses uses the Synthetic Minor-
ity Oversampling Technique (SMOTE) to tackle class imbalance often found in medical
datasets, including AD. Class imbalance can lead to biased learning, favoring the larger
class and reducing accuracy. SMOTE addresses this by creating synthetic samples for the
underrepresented class, enhancing its presence in the dataset. By adding these samples, the
model better understands minority class patterns, leading to better AD diagnosis accuracy.
Eng. Proc. 2023, 59, 10 5 of 11

Using SMOTE ensures a balanced dataset, enhancing the model’s prediction accuracy for
both classes.
Algorithm: synthetic samples depending on the minority-majority class ratio.
Input
i. Minority class samples: M
ii. k (number of nearest neighbors to consider)
Output
Synthetic samples: S
1. Create an empty synthetic sample list: S = []
2. Calculate the number of synthetic samples (n_synthetic) depending on the minority-
majority class ratio.
3. Each minority class sample m in M:
a. Find k closest neighbors of m from minority class samples, omitting m.
b. Randomly choose one of the k neighbours (nn).
c. Difference vector diff = nn − m.
d. Add a random proportion of diff to ‘m’ to create n_synthetic samples.
4. Add all newly synthesized samples to S.
5. Return synthetic sample list S.
6. The method identifies AD efficiently and correctly. Healthy and AD patients are first
separated. SMOTE fakes minority class samples for dataset balance. Representing
both groups promotes learning. Splitting the balanced dataset into training and testing
sets preserves class distribution.
7. Logistic Regression, Random Forest, or SVM predict AD. The chosen model learns
from training set characteristics and labels. The testing set evaluates the model’s
accuracy, precision, recall, and F1-score.
8. Successful classification models can discover AD in new data. Fresh instance features
suggest AD.
SMOTE builds synthetic samples along line segments linking a minority class sample
and its k nearest neighbors, extending the minority class in feature space. Logistic Regres-
sion, Random Forest, or a Support Vector Machine are used to predict AD. The selected
model learns features and annotations from the training set. To measure the model’s
efficacy, the accuracy, precision, recall, and F1-score are used on the assessment set. The
trained classification model is able to detect AD in new, unlabeled data if its performance is
adequate. By feeding the model the characteristics of new instances, it can precisely predict
the presence of AD. With careful consideration of dataset quality, feature selection, and
model selection, this algorithm provides a promising strategy for early and accurate AD de-
tection. Utilizing SMOTE to resolve class imbalance and advanced classification techniques,
the algorithm improves patient outcomes by facilitating timely diagnoses and intervention.
D. AD Prediction Using SMOTE
The proposed method efficiently classifies AD. The dataset, initially divided into
healthy and AD patients, is balanced using SMOTE. This enhances learning by representing
both classes equally. The data are then split for training and testing with equal class
distribution. The model, using Logistic Regression, Random Forest, or a Support Vector
Machine, learns from the training set and is evaluated based on the accuracy, precision,
recall, and F1-score. Once trained, the model can predict AD in new data. By addressing
class imbalances with SMOTE and using advanced techniques, this approach promises
early and accurate AD detection, improving patient outcomes.
E. Classification Procedure
A Support Vector Machine (SVM) is a key classification tool with significant potential
for an AD diagnosis. It is a versatile supervised learning algorithm suited for both linear
and nonlinear tasks. Especially useful for complex medical datasets like AD, SVM identifies
Eng. Proc. 2023, 59, 10 6 of 11

the best hyperplane to separate classes. After refining features, SVM can discern complex
patterns and relationships in the dataset. Its ability to handle nonlinear relationships
through various kernel functions and resist outliers ensures reliable predictions. When
trained on a balanced dataset from SMOTE, SVM offers high sensitivity and specificity,
vital for early AD detection.

3. Experimentation and Performance Assessment


This study evaluates the ensemble-based model’s efficacy in an AD diagnosis. Using
advanced feature extraction and selection, it trains on a dataset balanced via SMOTE,
ensuring equal representation of healthy and AD subjects. Through cross-validation,
the model’s predictive accuracy is assessed against metrics like the precision, recall, and
F1-score. This ensemble-based model excels in comparison to SVM and Logistic Regression.
This study further examines the impact of SMOTE samples and optimal hyperparameters
on performance. The results indicate superior diagnostic precision, highlighting the model’s
potential for early AD detection and improved healthcare management. The experimental
design begins with collecting comprehensive AD data, which undergoes pre-processing
and feature augmentation. SMOTE ensures dataset balance, which is then divided for
training and testing. Model efficacy is evaluated using metrics like the accuracy, precision,
recall, F1-score, and AUC-ROC, alongside a confusion matrix.

Accuracy = (True Positives + True Negatives)/Total Instances (8)

Precision = (True Positives)/((True Positives + False Positives)) (9)

Recall = (True Positives)/((True Positives + False Negatives)) (10)

F1 Score = (2× (Precision×Recall))/((Precision + Recall)) (11)


TP (True Positive) refers to instances correctly predicted as positive (having the dis-
ease), whereas FP (False Positive) refers to instances incorrectly predicted as positive,
despite not having the disease. TN (True Negative) indicates instances that were accurately
predicted as negative (not having the disease), whereas FN (False Negative) indicates
instances that were incorrectly predicted as negative but actually included the disease.

4. Results and Discussion


Feature selection and data resampling are critical for enhancing machine learning
model performance, especially with imbalanced datasets. Feature selection chooses relevant
features from the initial set, eliminating unimportant or redundant ones. This enhances
model efficiency and interpretability, and reduces overfitting. Methods like Recursive
Feature Elimination (RFE), Univariate Feature Selection (UFS), and a Principal Component
Analysis (PCA) help identify key features for accurate predictions. Data resampling adjusts
the dataset’s distribution, particularly when class imbalances exist. Oversampling, like the
SMOTE method, creates synthetic samples for the minority class, while under-sampling
removes instances from the dominant class. However, under-sampling can lead to informa-
tion loss. By integrating feature selection and resampling, models are trained on balanced
and pertinent datasets, improving accuracy and real-world applicability. These techniques
effectively address challenges like class imbalances and high-dimensional feature spaces.
A. Tuning hyperparameters
Hyperparameter optimization is essential for enhancing the ensemble-based AD
diagnosis model. This process finds the best values for key model parameters like learning
rate, depth of decision trees, or number of neighbors in K-Nearest Neighbors (KNNs).
Methods like Grid Search or Random Search are used, combined with cross-validation
Eng. Proc. 2023, 59, 10 7 of 11
Eng. Proc. 2023, 59, 10 FOR PEER REVIEW 7 of 11

to avoid overfitting. The model’s performance is tested with various hyperparameter


optimal hyperparameters,
combinations the model
using metrics like is validated
the accuracy, on unseen
F1-score, data to confirm
or AUC-ROC. its reliability.
After identifying the
This thorough
optimal tuning ensures
hyperparameters, the model’s
the model peak on
is validated accuracy
unseenindata
an AD diagnosis,
to confirm benefiting
its reliability.
patient
This care and
thorough disease
tuning management.
ensures the model’s peak accuracy in an AD diagnosis, benefiting
patient care and disease
B. Grid‐Search tuning management.
B. Grid-Search
During thetuningmodel’s hyperparameter tuning, various parameter combinations were
explored to optimize
During the model’s performance. We examined
hyperparameter the ‘Bootstrap’
tuning, various parameterparameter
combinationsusingwere
both
‘True’ and ‘False’. Handling missing data, the maximum model depth
explored to optimize performance. We examined the ‘Bootstrap’ parameter using both was tested with
values
‘True’ andof 5‘False’.
and 7, Handling
and the maximum features
missing data, the were assessed
maximum withdepth
model options
wasoftested
3 and with
4. We
also evaluated
values of 5 andthe impact
7, and theofmaximum
minimumfeatures
sample leaf
werevalues of 3 with
assessed and 4.options
The decision
of 3 andtrees’
4.
minimum sample split values were tried at 3, 5, and 7, while the number of estimators
We also evaluated the impact of minimum sample leaf values of 3 and 4. The decision trees’ was
tested with
minimum 200, 400,
sample splitand 600.were
values By assessing
tried at 3,these
5, andcombinations, we identified
7, while the number the optimal
of estimators was
configuration for the best model performance. This rigorous tuning improved
tested with 200, 400, and 600. By assessing these combinations, we identified the optimal the model’s
predictive capabilities.
configuration for the best model performance. This rigorous tuning improved the model’s
predictive capabilities.
C. Optimal RF hyperparameters
During RF
C. Optimal hyperparameter
hyperparameterstuning, we adjusted several parameters to enhance the
model’s
Duringperformance. We set tuning,
hyperparameter the “Bootstrap” to “False”,
we adjusted theparameters
several “Maximumto depth” to seven
enhance the
model’s performance. We set the “Bootstrap” to “False”, the “Maximum depth” toatseven
layers, and “Maximum features” to four. The “Minimum samples leaf” was fixed three,
while and
layers, “Minimum
“Maximumsamples split”torequired
features” four. Theseven samples.
“Minimum The model
samples leaf” employed 200
was fixed at esti-
three,
mators as indicated by the “n_estimators” value. These adjustments
while “Minimum samples split” required seven samples. The model employed 200 estima-optimized the
model’s performance, ensuring more accurate predictions. Proper hyperparameter
tors as indicated by the “n_estimators” value. These adjustments optimized the model’s tuning
is vital for improved
performance, ensuringmodel
more results
accurate and capabilities.
predictions. Proper hyperparameter tuning is vital
D.improved
for Effectiveness
modelEvaluation
results and capabilities.
For the study
D. Effectiveness article on AD diagnoses, the ensemble-based model must be evaluated
Evaluation
and compared to different classification methods at several significance levels. To correct
For the study article on AD diagnoses, the ensemble-based model must be evaluated
class imbalance, the dataset is prepared, preprocessed, and balanced using SMOTE.
and compared to different classification methods at several significance levels. To correct
Cross-validation divides the balanced dataset into training and testing sets for proper
class imbalance, the dataset is prepared, preprocessed, and balanced using SMOTE. Cross-
evaluation. Figure 2 represents the relationship with various models. The ensemble-based
validation divides the balanced dataset into training and testing sets for proper evaluation.
model is trained using optimized hyperparameters and SVM and Logistic Regression clas-
Figure 2 represents the relationship with various models. The ensemble-based model is
sifiers.
trained using optimized hyperparameters and SVM and Logistic Regression classifiers.

Figure 2. A representation of the relationship between various models.


Figure 2. A representation of the relationship between various models.
E. Precision
E. Precision
It quantifies the model’s True Positive predictions. A high accuracy score means the
model reliably predicts positive situations, whereas a low score means it produces many
erroneous positive predictions. Figure 3 represents the precision score for different mod-
It quantifies the model’s True Positive predictions. A high accuracy score means the
els.
model reliably predicts positive situations, whereas a low score means it produces many
Eng. Proc. 2023, 59, 10 Precision
erroneous positive = (True Figure
predictions. Positives)/((True Positives
3 represents + Falsescore
the precision Positives)) 8 of 11
for different mod-
els.
The precision graph represented in Figure 3 clearly illustrates the varying precision
scores of predictive algorithms.
Precision = (TrueSVM stands out with
Positives)/((True an impressive
Positives 96%, indicating accu-
+ False Positives))
It quantifies
rate positive the model’s
predictions. ExtraTrue
TreePositive
shows apredictions.
lower 76%, A highthe
while accuracy
decisionscore
tree,means the
Logistic
model The precision
reliably graphpositive
predicts represented in Figure
situations, 3 clearly
whereas a lowillustrates
score meansthe varying precision
it produces many
Regression, and XG Boost perform moderately at 81%. SVM’s dominance is evident.
scores
erroneousof predictive algorithms.Figure
positive predictions. SVM stands out with
3 represents thean impressive
precision score96%, indicating
for different accu-
models.
rate positive predictions. Extra Tree shows a lower 76%, while the decision tree, Logistic
Regression, and Precision = (True
XG Boost Positives)/((True
perform moderately atPositives + False
81%. SVM’s Positives))
dominance is evident.

Figure 3. A representation of precision scores for different models.

F. Recall
Figure 3. A representation
Recall of precision
is a performance statisticscores for different
for binary models.
classification
Figure 3. A representation of precision scores for different models. models. It tests the model’s
ability to identify all positive occurrences from the dataset’s total positive instances. Recall
F. Recall
The precision
is sensitivity and thegraph
True represented
Positive ratein Figure
(TPR). 3 clearly
Figure illustrates
4 represents the
the varying
recall scoreprecision
for dif-
scoresRecall
of is a performance
predictive algorithms.statistic
SVM for stands
binary classification
out with an models. It tests
impressive 96%, the model’s
indicating
ferent models. The ratio of True Positive predictions (properly recognized positive cases)
ability
accuratetopositive
identifypredictions.
all positive occurrences
Extra Tree fromathe
shows dataset’s
lower 76%, total positive
while the instances.
decision tree, Recall
Logistic
to the total of True Positive and False Negative predictions (positive instances mistakenly
is sensitivity
Regression, and the True perform moderately at 81%. SVM’s dominance is evident. dif-
Positive rate (TPR). Figure 4 represents the recall score for
forecasted asand XG Boost
negative) is used.
ferent models. The ratio of True Positive predictions (properly recognized positive cases)
F. Recall
to the total of TrueRecall = (True
Positive andPositives)/((True Positives + False
False Negative predictions Negatives))
(positive instances mistakenly
Recallas
forecasted is negative)
a performance statistic for binary classification models. It tests the model’s
is used.
A high
ability recall all
to identify score suggests
positive that the from
occurrences modelthecan properly
dataset’s totalidentify
positivea instances.
significantRecall
pro-
portion of positive
is sensitivity cases,
Recall
and the =True meaning
(True few
rateFalse
Positives)/((True
Positive Negatives.
(TPR). Figure 4+A
Positives lowNegatives))
False
representsrecallthescore
recallmeans
scorethefor
model misses
different many
models. Thepositive
ratio ofexamples,
True resulting
Positive in more
predictions False
(properly Negatives.
recognized Recall is
positive critical
cases)
A high recall score suggests that the model can properly identify a significant pro-
in
tomedical
the total diagnoses (to detect
of True Positive andillnesses)
Falsefew and fraud
Negative detection
predictions (to detect
(positive fraudulent
instances trans-
portion of positive cases, meaning False Negatives. A low recall score mistakenly
means the
actions)
forecastedto accurately
as negative) identify
is used.positive instances. However, optimizing one statistic might
model misses many positive examples, resulting in more False Negatives. Recall is critical
affect other metrics in a classification assignment; therefore, it is important to balance re-
in medical diagnoses (to detect illnesses) and fraud detection (to detect fraudulent trans-
call and other metrics
Recall =like accuracy.
(True Positives)/((True Positives + False Negatives))
actions) to accurately identify positive instances. However, optimizing one statistic might
affect other metrics in a classification assignment; therefore, it is important to balance re-
call and other metrics like accuracy.

Figure4.4.AArepresentation
Figure representationof
ofrecall
recallscore
score for
for different
different model.
model.

A high recall score suggests that the model can properly identify a significant propor-
tion of4.positive
Figure cases, meaning
A representation of recall few
scoreFalse Negatives.
for different A low recall score means the model
model.
misses many positive examples, resulting in more False Negatives. Recall is critical in med-
ical diagnoses (to detect illnesses) and fraud detection (to detect fraudulent transactions) to
accurately identify positive instances. However, optimizing one statistic might affect other
metrics in a classification assignment; therefore, it is important to balance recall and other
metrics like accuracy.
Eng. Proc. 2023, 59, 10 FOR PEER REVIEW 9 of 11
Eng. Proc. 2023, 59, 10 9 of 11

The recall scores for various models were evaluated to quantify their ability to accu-
The recall scores for various models were evaluated to quantify their ability to accu-
rately identify positive cases in the dataset. The SVM algorithm exhibited an outstanding
rately identify positive cases in the dataset. The SVM algorithm exhibited an outstanding
recall score of 97%, correctly identifying 97% of the positive cases. Surprisingly, the KNN
recall score of 97%, correctly identifying 97% of the positive cases. Surprisingly, the KNN
algorithm surpassed even the SVM, obtaining a recall score of 95%, demonstrating its ef-
algorithm surpassed even the SVM, obtaining a recall score of 95%, demonstrating its
fectiveness
effectiveness inincorrectly
correctlyidentifying
identifyingpositive
positivecases.
cases.InIncontrast,
contrast,thethedecision
decision tree
tree algorithm
algorithm
achieved a lower recall score of 84%, indicating that it missed a considerable
achieved a lower recall score of 84%, indicating that it missed a considerable portion portion of
of the
the positive cases. The Naive Bayes model achieved a recall score of
positive cases. The Naive Bayes model achieved a recall score of 75%, while the Logistic75%, while the Lo-
gistic Regression
Regression model model performed
performed relatively
relatively better
better with with a recall
a recall score
score ofof81%.
81%. Overall,
Overall, the
the
results
results highlight the superior performance of the KNN model in identifying positive cases
highlight the superior performance of the KNN model in identifying positive cases
compared
compared to tothe
theother
otherfour
fouralgorithms.
algorithms.The Themodels’
models’ diagnostic
diagnosticskills on the
skills testing
on the set are
testing set
assessed using the accuracy, precision, recall, F1-score, and AUC-ROC.
are assessed using the accuracy, precision, recall, F1-score, and AUC-ROC. The confusion The confusion ma-
trix alsoalso
matrix assesses
assessestrue, false,
true, positive,
false, and
positive, andnegative
negativepredictions.
predictions.External
Externalvalidation
validation onon aa
different dataset assesses
different dataset assessesthe themodel’s
model’scapacity
capacitytotogeneralize
generalize toto unobserved
unobserved datadata
byby com-
compar-
paring
ing the the proposed
proposed model’s
model’s performance
performance to baseline
to the the baseline classifiers
classifiers and using
and using statistical
statistical tests
tests to discover performance
to discover performance differences. differences.
G.
G. F1‐Score
F1-Score
A model with a high F1-score is one that effectively balances precision and recall.
A model with a high F1-score is one that effectively balances precision and recall.
Evaluating the F1-scores of the aforementioned models would provide a more thorough
Evaluating the F1-scores of the aforementioned models would provide a more thorough
comprehension
comprehension of oftheir
theiroverall
overalleffectiveness
effectiveness and
and potential
potential trade-offs
trade-offs between
between precision
precision and
and recall.
recall. Figure
Figure 5 represents
5 represents a confusion
a confusion matrix
matrix on prediction
on prediction of Alzheimer’s.
of Alzheimer’s.

Confusion matrix
Figure 5. Confusion matrix on
on prediction
prediction of Alzheimer’s.
Alzheimer’s.

Visualization methods
Visualization methods like
like ROC
ROC curves
curves andand precision–recall
precision–recall curves
curves show
show the
the model’s
model’s
discrimination performance, whereas a feature significance analysis shows feature
discrimination performance, whereas a feature significance analysis shows feature contri- contribu-
tions. Figure
butions. 5 illustrates
Figure an AD
5 illustrates prediction
an AD confusion
prediction matrix.
confusion TrueTrue
matrix. Positive (TP) occurrences
Positive (TP) occur-
are accurately predicted as including the disease, while False Positive
rences are accurately predicted as including the disease, while False Positive (FP) instances
(FP) are
in-
wrongly forecasted as positive. True Negative (TN) occurrences were accurately
stances are wrongly forecasted as positive. True Negative (TN) occurrences were accu- predicted
as negative
rately (not having
predicted the disease),
as negative while
(not having theFalse Negative
disease), while(FN) examples
False Negativewere
(FN)mistakenly
examples
forecasted as negative but included the disease. This research study evaluates the
were mistakenly forecasted as negative but included the disease. This research study eval- suggested
ensemble-based
uates the suggestedmodel for an AD diagnosis
ensemble-based model forto improve medical data
an AD diagnosis analytics
to improve and patient
medical data
care by revealing its accuracy and efficacy.
analytics and patient care by revealing its accuracy and efficacy.
5. Conclusions
5. Conclusions
This research offers an in-depth study of an AD diagnosis through machine learn-
ing. This
Using research
featureoffers an in-depth
selection and datastudy of an ADour
resampling, diagnosis
proposed through machine learning.
ensemble-based model
Using feature selection and data resampling, our proposed ensemble-based
effectively differentiates between healthy individuals and AD patients. It outperforms model effec-
tively differentiates
baseline classifiers between
like SVMhealthy individuals
and Logical and ADinpatients.
Regression accuracy, It outperforms
precision, andbaseline
other
classifiers like SVM and Logical Regression in accuracy, precision, and
metrics. Relevant features enhance the model’s clarity and effectiveness, while SMOTEother metrics. Rel-
evant features enhance the model’s clarity and effectiveness, while SMOTE
balancing addresses class imbalance. This work contributes significantly to AD diagnoses, balancing ad-
dresses class imbalance. This work contributes significantly to AD diagnoses,
promoting early detection and better patient outcomes. Future studies could explore deep promoting
early detection
learning and better
techniques, such aspatient
CNNsoutcomes.
and RNNs, Future studies could
for improved explore deep
brain imaging learning
pattern recog-
techniques, such as varied
nition. Combining CNNs data
and RNNs,
sources,for
likeimproved
genetics brain imaging
and clinical pattern
data, mightrecognition.
refine the
Combining
diagnosis. A varied data sources,
longitudinal likedata
patient genetics and clinical
analysis data,
can track mightprogression
disease refine the diagnosis.
and risk
A longitudinal patient data analysis can track disease progression
prediction. Collaborating with medical experts for real-world validation, improvingand risk prediction.
model
Eng. Proc. 2023, 59, 10 10 of 11

interpretability, and integrating it into clinical systems will further its potential in AD
diagnoses and treatment.

Author Contributions: Experiment Design and Data Pre-processing, R.P.N. and B.S.; Design, R.K.;
Review and Interpretation, M.A.B.; Data Analysis and Interpreted Result, R.P.N.; Writing—Review
and Editing, R.P.N. All authors have read and agreed to the published version of the manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Data will be provided on request.
Conflicts of Interest: The authors declare no conflict of interest.

References
1. Grueso, S.; Viejo-Sobera, R. Machine learning methods for predicting progression from mild cognitive impairment to AD dementia:
A systematic review. Alzheimer’s Res. Ther. 2021, 13, 1–29.
2. El-Sappagh, S.; Saleh, H.; Ali, F.; Amer, E.; Abuhmed, T. Two-stage deep learning model for AD detection and prediction of the
mild cognitive impairment time. Neural Comp. Appl. 2022, 34, 14487–14509. [CrossRef]
3. Iddi, S.; Li, D.; Aisen, P.S.; Rafii, M.S.; Thompson, W.K.; Donohue, M.C. AD Neuroimaging Initiative. Predicting the course of
Alzheimer’s progression. Brain Inform. 2019, 6, 1–18. [CrossRef]
4. Yuen, S.C.; Liang, X.; Zhu, H.; Jia, Y.; Leung, S.W. Prediction of differentially expressed microRNAs in blood as potential
biomarkers for AD by meta-analysis and adaptive boosting ensemble learning. Alzheimer’s Res. Ther. 2021, 13, 1–30.
5. Naz, S.; Ashraf, A.; Zaib, A. Transfer learning using freeze features for Alzheimer neurological disorder detection using ADNI
dataset. Multi. Syst. 2022, 28, 85–94. [CrossRef]
6. Wang, S.; Du, Z.; Ding, M.; Rodriguez-Paton, A.; Song, T. KG-DTI: A knowledge graph based deep learning method for
drug-target interaction predictions and AD drug repositions. Appl. Intell. 2022, 52, 846–857. [CrossRef]
7. Bermudez, C.; Graff-Radford, J.; Syrjanen, J.A.; Stricker, N.H.; Algeciras-Schimnich, A.; Kouri, N.; Vemuri, P. Plasma biomarkers
for prediction of AD neuropathologic change. Acta Neuropathol. 2023, 146, 13–29. [CrossRef]
8. Diogo, V.S.; Ferreira, H.A.; Prata, D. AD Neuroimaging Initiative. Early diagnosis of AD using machine learning: A multi-
diagnostic, generalizable approach. Alzheimer’s Res. Ther. 2022, 14, 107. [CrossRef]
9. Venkataramana, L.Y.; Jacob, S.G.; Prasad, V.; Athilakshmi, R.; Priyanka, V.; Yeshwanthraa, K.; Vigneswaran, S. Geometric
SMOTE-Based Approach to Improve the Prediction of Alzheimer’s and Parkinson’s Diseases for Highly Class-Imbalanced Data.
In AI, IoT, and Blockchain Breakthroughs in E-Governance; IGI Global: Hershey, PA, USA, 2023; pp. 114–137.
10. Zhang, P.; Lin, S.; Qiao, J.; Tu, Y. Diagnosis of AD with ensemble learning classifier and 3D convolutional neural network. Sensors
2021, 21, 7634. [CrossRef]
11. Rao, K.N.; Gandhi, B.R.; Rao, M.V.; Javvadi, S.; Vellela, S.S.; Basha, S.K. Prediction and Classification of AD using Machine
Learning Techniques in 3D MR Images. In Proceedings of the 2023 International Conference on Sustainable Computing and
Smart Systems (ICSCSS), Coimbatore, India, 14–16 June 2023; pp. 85–90.
12. Khoei, T.T.; Labuhn, M.C.; Caleb, T.D.; Hu, W.C.; Kaabouch, N. A stacking-based ensemble learning model with genetic algorithm
for detecting early stages of AD. In Proceedings of the 2021 IEEE International Conference on Electro Information Technology
(EIT), Mt. Pleasant, MI, USA, 14–15 May 2021; pp. 215–222.
13. Tambe, P.; Saigaonkar, R.; Devadiga, N.; Chitte, P.H. Deep Learning techniques for effective diagnosis of AD using MRI images.
ITM Web Conf. 2021, 40, 03021. [CrossRef]
14. Ghali, U.M.; Usman, A.G.; Chellube, Z.M.; Degm, M.A.A.; Hoti, K.; Umar, H.; Abba, S.I. Advanced chromatographic technique
for performance simulation of anti-Alzheimer agent: An ensemble machine learning approach. SN Appl. Sci. 2020, 2, 1–12.
[CrossRef]
15. An, N.; Ding, H.; Yang, J.; Au, R.; Ang, T.F. Deep ensemble learning for AD classification. J. Biomed. Inform. 2020, 105, 103411.
16. Pereira, T.; Cardoso, S.; Silva, D.; Guerreiro, M.; de Mendonça, A.; Madeira, S.C. Ensemble learning with Conformal Predictors:
Targeting credible predictions of conversion from Mild Cognitive Impairment to AD. arXiv 2018, arXiv:1807.01619.
17. Wang, R.; Li, H.; Lan, R.; Luo, S.; Luo, X. Hierarchical Ensemble Learning for AD Classification. In Proceedings of the 2018 7th
International Conference on Digital Home (ICDH), Guilin, China, 30 November–1 December 2018; pp. 224–229.
18. Orhobor, O.I.; Soldatova, L.N.; King, R.D. Federated ensemble regression using classification. In Proceedings of the Discovery
Science: 23rd International Conference, DS 2020, Thessaloniki, Greece, 19–21 October 2020; Springer International Publishing:
Berlin/Heidelberg, Germany; Volume 23, pp. 325–339.
19. Kang, W.; Lin, L.; Zhang, B.; Shen, X.; Wu, S. AD Neuroimaging Initiative. Multi-model and multi-slice ensemble learning
architecture based on 2D convolutional neural networks for AD diagnosis. Comput. Biol. Med. 2021, 136, 104678.
Eng. Proc. 2023, 59, 10 11 of 11

20. Mirzaei, G.; Adeli, H. Machine learning techniques for diagnosis of Alzheimer disease, mild cognitive disorder, and other types
of dementia. Biomed. Sign. Process. Control 2022, 72, 103293. [CrossRef]
21. Nguyen, D.K.; Lan, C.H.; Chan, C.L. Deep ensemble learning approaches in healthcare to enhance the prediction and diagnosing
performance: The workflows, deployments, and surveys on the statistical, image-based, and sequential datasets. Int. J. Environ.
Res. Public Health 2021, 18, 10811. [CrossRef]
22. Shaikh, T.A.; Ali, R. Enhanced computerised diagnosis of AD from brain MRI images using a classifier merger strategy. Int. J.
Inform. Technol. 2021, 14, 1–13.
23. Song, M.; Jung, H.; Lee, S.; Kim, D.; Ahn, M. Diagnostic classification and biomarker identification of AD with random forest
algorithm. Brain Sci. 2021, 11, 453. [CrossRef]
24. Hemalatha, B.; Renukadevi, M. Analysis of Alzheimer disease prediction using machine learning techniques. Inf. Technol. Ind.
2021, 9, 519–525.
25. Alamro, H.; Thafar, M.A.; Albaradei, S.; Gojobori, T.; Essack, M.; Gao, X. Exploiting machine learning models to identify novel
AD biomarkers and potential targets. Sci. Rep. 2023, 13, 4979. [CrossRef]
26. Albahri, A.S.; Alwan, K.J.; Taha, Z.K.; Ismail, S.F.; Hamid, R.A.; Zaidan, A.A.; Albahri, O.S.; Zaidan, B.B.; Alamoodi, A.H.;
Alsalem, M.A. IoT-based telemedicine for disease prevention and health promotion: State-of-the-Art. J. Netw. Comput. Appl. 2021,
173, 102873. [CrossRef]

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.

You might also like