<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">32752</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2023.032752</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A Deep Learning for Alzheimer&#x2019;s Stages Detection Using Brain Images</article-title>
<alt-title alt-title-type="left-running-head">A Deep Learning for Alzheimer&#x2019;s Stages Detection Using Brain Images</alt-title>
<alt-title alt-title-type="right-running-head">A Deep Learning for Alzheimer&#x2019;s Stages Detection Using Brain Images</alt-title>
</title-group>
<contrib-group content-type="authors">
<contrib id="author-1" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Ullah</surname><given-names>Zahid</given-names></name><xref ref-type="aff" rid="aff-1">1</xref><email>zasultan@kau.edu.sa</email></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Jamjoom</surname><given-names>Mona</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University</institution>, <addr-line>Jeddah</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-2"><label>2</label><institution>Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University</institution>, <addr-line>Riyadh</addr-line>, <country>Saudi Arabia</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Zahid Ullah. Email: <email>zasultan@kau.edu.sa</email></corresp>
</author-notes>
<pub-date pub-type="epub" date-type="pub" iso-8601-date="2022-08-16"><day>16</day>
<month>08</month>
<year>2022</year></pub-date>
<volume>74</volume>
<issue>1</issue>
<fpage>1457</fpage>
<lpage>1473</lpage>
<history>
<date date-type="received"><day>28</day><month>5</month><year>2022</year></date>
<date date-type="accepted"><day>01</day><month>7</month><year>2022</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2023 Ullah and Jamjoom</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Ullah and Jamjoom</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_32752.pdf"></self-uri>
<abstract>
<p>Alzheimer&#x2019;s disease (AD) is a chronic and common form of dementia that mainly affects elderly individuals. The disease is dangerous because it causes damage to brain cells and tissues before the symptoms appear, and there is no medicinal or surgical treatment available yet for AD. AD causes loss of memory and functionality control in multiple degrees according to AD&#x2019;s progression level. However, early diagnosis of AD can hinder its progression. Brain imaging tools such as magnetic resonance imaging (MRI), computed tomography (CT) scans, positron emission tomography (PET), etc. can help in medical diagnosis of AD. Recently, computer-aided diagnosis (CAD) such as deep learning applied to brain images obtained with these tools, has been an established strategic methodology that is widely used for clinical assistance in prognosis of AD. In this study, we proposed an intelligent methodology for building a convolutional neural network (CNN) from scratch to detect AD stages from the brain MRI images dataset and to improve patient care. It is worth mentioning that training a deep-learning model requires a large amount of data to produce accurate results and prevent the model from overfitting problems. Therefore, for better understanding of classifiers and to overcome the model overfitting problem, we applied data augmentation to the minority classes in order to increase the number of MRI images in the dataset. All experiments were conducted using Alzheimer&#x2019;s MRI dataset consisting of brain MRI scanned images. The performance of the proposed model determines detection of the four stages of AD. Experimental results show high performance of the proposed model in that the model achieved a 99.38&#x0025; accuracy rate, which is the highest so far. Moreover, the proposed model performance in terms of accuracy, precision, sensitivity, specificity, and f-measures is promising when compared to the very recent state-of-the-art domain-specific models existing in the literature.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Dementia</kwd>
<kwd>MRI images</kwd> 
<kwd>CNN model</kwd> 
<kwd>detection</kwd> 
<kwd>data augmentation</kwd> 
<kwd>decision support system</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1"><label>1</label><title>Introduction</title>
<p>AD is one of the chronic neurodegenerative diseases and mainly causes dementia among the middle-aged and elderly [<xref ref-type="bibr" rid="ref-1">1</xref>]. This is an irreversible disease with high chances of failure in medical trials for AD drugs [<xref ref-type="bibr" rid="ref-2">2</xref>]. AD is pathologically classified by steady and gradual deterioration in memory and other cognitive functions [<xref ref-type="bibr" rid="ref-3">3</xref>]. The present clinical prognosis of this disease is waste of time without success [<xref ref-type="bibr" rid="ref-4">4</xref>]. In 2006, the number of AD patients reached more than 26 million worldwide [<xref ref-type="bibr" rid="ref-5">5</xref>]. While an estimate presented by [<xref ref-type="bibr" rid="ref-6">6</xref>] shows that around 50 million people are living with dementia, and this number is predicted to increase by 2050 to 152 million, especially in poor and middle-income countries, which is a significant challenge of the twenty-first century. As discussed by [<xref ref-type="bibr" rid="ref-7">7</xref>], AD is the 6th leading cause of death in the US. The mental health ability of patients deteriorates such that patients are unable to live normal lives without their family members&#x2019; assistance [<xref ref-type="bibr" rid="ref-8">8</xref>]. But early prognosis and treatment can hinder the progression of AD [<xref ref-type="bibr" rid="ref-9">9</xref>].</p>
<p>The cause of AD is still not fully known [<xref ref-type="bibr" rid="ref-10">10</xref>]; however, some potential causes are environmental and genetic factors, head injuries, and chemicals [<xref ref-type="bibr" rid="ref-8">8</xref>] that lead to damaged brain cells and tissue and thus decreasing its functionalities [<xref ref-type="bibr" rid="ref-11">11</xref>]. AD patients have common symptoms like loss of memory, communication and recognition issues, cognition issues, and disruptions in mood and behavior [<xref ref-type="bibr" rid="ref-8">8</xref>]. Primarily, AD affects portions of the brain that regulate memory and language, resulting in AD subjects undergoing memory loss, confusion, and difficulties reading, writing, and speaking [<xref ref-type="bibr" rid="ref-12">12</xref>]. There is no remedial treatment for AD because the brain cells die, which ultimately causes loss of memory and cognitive power [<xref ref-type="bibr" rid="ref-8">8</xref>]. But on the other hand, early prognosis and treatment [<xref ref-type="bibr" rid="ref-9">9</xref>], exercise, and memory coaching [<xref ref-type="bibr" rid="ref-1">1</xref>] can slow down the progression of AD.</p>
<p>Over the past couple of decades, researchers have presented new imaging modalities to medical doctors such as computer-aided diagnosis (CAD) to be applied to medical images that assist clinicians to make sense of large amounts of non-invasive medical data and assist them in diagnostic tasks [<xref ref-type="bibr" rid="ref-13">13</xref>]. CAD systems make use of state-of-the-art technologies such as X-rays, MRI, CT scans, PET, and other revolutionary studies such as DNA sequencing, gene expression [<xref ref-type="bibr" rid="ref-9">9</xref>], and ultrasound are nowadays indispensable for diagnosis and prognosis of different types of cancer, bone and skin diseases, and AD [<xref ref-type="bibr" rid="ref-14">14</xref>]. Moreover, the American Academy of Neurology advocated that the most appropriate primary evaluation for patients with dementia is the structural imaging of the brain (either CT scan or MRI), which shows diffuse cortical and intellectual weakness in AD patients [<xref ref-type="bibr" rid="ref-8">8</xref>]. So, CT scan and MRI provide deep insight into the study of the brain and also into the features of the brain that can thoroughly assess the disease and ultimately discover the best mechanisms for diagnosing AD. Thus, they can be utilized with CAD systems [<xref ref-type="bibr" rid="ref-13">13</xref>] that help clinicians to avoid misdiagnosis due to less experience or exhaustion. In addition, it may provide earlier and more accurate diagnosis, which in turn reduces the costs of caring for AD patients [<xref ref-type="bibr" rid="ref-14">14</xref>].</p>
<p>Advances in state-of-the-art technologies and the current growth of deep learning (DL) have impacted every field of life, and their applications in (for example) medical imaging have shown tremendous potential for clinical decision support that helps physicians diagnose various stages of disease. The strength of DL compared to other shallow learning models is that DL can learn the most predictive features from the raw data [<xref ref-type="bibr" rid="ref-15">15</xref>]. Unlike artificial neural networks (ANN), DL models use a data-driven approach by utilizing deep neural networks (DNN) with multiple layers in a hierarchical structure to capture high-level features of the raw data [<xref ref-type="bibr" rid="ref-16">16</xref>].</p>
<p>The application of DNN for extracting features such as a stack autoencoder or Deep Boltzmann machine (DBM) outperformed conventional ML methods [<xref ref-type="bibr" rid="ref-17">17</xref>]. Due to several limitations of the traditional machine-learning (ML) methods with respect to image data and the increased success of the DL model, which requires less effort when preprocessing images and can automatically extract salient features from the raw data without feature selection in advance, leading DL to a more objective and less biased model [<xref ref-type="bibr" rid="ref-1">1</xref>]. Moreover, deep-learning models remained progressive in performance and seem to hold potential for diagnostic classification of AD using multimodal neuroimaging data [<xref ref-type="bibr" rid="ref-18">18</xref>]. The motivation of this study is due to these properties of DL models and their success with imaging data in detecting disease. In this study, we have proposed a DL model for detecting different stages of AD based on MRI images. Our proposed model is using a data-driven strategy with a convolutional neural network (CNN), which consists of multiple convolutional layers. The benefits of using CNN include extracting the most salient features from the images, detecting different stages of AD and, on the basis of that information, providing an effective decision support system for physicians to help them in their diagnostic work. We have tried both approaches (i.e., transfer learning and CNN from scratch) for building the model. In transfer learning, we used VGG-16 and ResNet-50, but unfortunately, the results of both transfer learning models were not promising, both in terms of validation and of testing data. The results of CNN from scratch achieved the highest results in terms of accuracy and other classification measures, tested and validated on a publicly MRI dataset available on Kaggle. Therefore, this study has reported the results of CNN from scratch.</p>
</sec>
<sec id="s2"><label>2</label><title>Previous Works</title>
<p>Research is an ongoing process whereby researchers analyze raw data from different perspectives using different methodologies and mechanisms and reaching conclusions by providing in-depth insights drawn from raw data and building decision support models that can help decision-makers in their jobs across different fields. Similarly, the detection of AD has attracted the recent attention of researchers, and various methodologies were proposed with ML and DL as common approaches for automatic detection. Since this study is more concerned with the DL methodology, we will therefore limit the scope of our discussion to only DL models in the literature.</p>
<p>A study conducted by [<xref ref-type="bibr" rid="ref-1">1</xref>] used three MRI and one fMRI image datasets for classification into binary and multi-classes using a deep ensemble approach with three trained networks of fine-tuned AlexNet, ResNet-101, and Inception-ResNet-v2. Moreover, the study has increased the number of images using data augmentation before training the model. The results reported show the higher accuracy of their proposed model achieving 98.51&#x0025; on binary classification and 98.67&#x0025; on multi-class classification. Similarly, the study of [<xref ref-type="bibr" rid="ref-8">8</xref>] has applied an VGG model to an Alzheimer&#x2019;s disease national initiative (ADNI) dataset. Due to the insufficient number of 379 patients&#x2019; images in the ADNI dataset, data augmentation using flipping, rotation, illumination, and zooming techniques was applied in order to increase the number of images that are ideal for training the CNN model. The dataset was augmented from 3925 to 37,590 images. The results of the proposed VGG model achieved the test accuracies of binary classification&#x2014;i.e., 99.27&#x0025; (MCI <italic>vs.</italic> AD), 97.06&#x0025; (MCI <italic>vs.</italic> CN), and 98.89&#x0025; (AD <italic>vs.</italic> CN).</p>
<p>Reference [<xref ref-type="bibr" rid="ref-9">9</xref>] used a DL framework for early detection of AD with an embedded feature selection method. In this approach, the feature selection was performed using conventional ML methods with k-folds cross-validation for model evaluation. They found that AdaBoost (AB) with 5-fold cross-validation yielded optimal results in terms of feature selection, which were used for training the models&#x2014;i.e., recurrent neural network (RNN), CNN, deep RNN, and enhanced deep RNN. The results showed that EDRNN outperformed the other three models with a classification rate of 88.7&#x0025;.</p>
<p>In [<xref ref-type="bibr" rid="ref-19">19</xref>] the researchers used different models with CNN architecture and applied them to an ADNI dataset for classifying different stages of AD. According to them, 29 different pre-trained models were used on MRI images and evaluated the learning models using accuracy, precision, sensitivity, and specificity. Among these models, EfficientNetB0 achieved high accuracy of 92.98&#x0025;, while precision of the EfficientNetB3 model was 89.78&#x0025;, sensitivity of the EfficientNetB2 model was 94.42&#x0025;, and specificity of the EfficientNetB3 model was 97.28&#x0025;.</p>
<p>An attempt made by [<xref ref-type="bibr" rid="ref-20">20</xref>] used a CNN-based DEMNET (dementia network) model for detecting four different stages of AD. The dataset used in the study was collected from Kaggle containing 6400 MRI images. The data augmentation was performed using the SMOTE method. The results of the DEMNET model achieved a testing accuracy of 95.23&#x0025;. Similarly, [<xref ref-type="bibr" rid="ref-21">21</xref>] applied a hybrid classical quantum NN to a Kaggle dataset containing 6400 MRI images for automated detection of AD, and ResnNet34 was used for feature extraction. The proposed hybrid classical quantum network achieved the highest testing accuracy of 97.2&#x0025;.</p>
<p>Reference [<xref ref-type="bibr" rid="ref-22">22</xref>] used DL and hybrid DL methods for analyzing MRI images for early diagnosis of AD and dementia. They have used OASIS and Kaggle MRI datasets for training AlexNet and ResNet-50 models, as well as AlexNet&#x002B;SVM and ResNet-50&#x002B;SVM hybrid models. For data augmentation, the SMOTE method was used for an OASIS dataset to balance the classes, while data augmentation was used for the Kaggle dataset. The AlexNet&#x002B;SVM outperformed other methods applied to the Kaggle MRI dataset after augmentation and attained the high classification rate of 94.8&#x0025;.</p>
<p>The detection of AD using DL was performed by [<xref ref-type="bibr" rid="ref-23">23</xref>], who used a dataset of digital subtracted angiogram scans of AD subjects. The feature extraction was made using pre-trained models&#x2014;i.e., InceptionV3 and DenseNet201. Data augmentation was used to increase the number of scans in the dataset, while principal component analysis (PCA) was used for feature selection with a criterion of explained variance ratio of 0.99. The model was trained and concluded with an accuracy score of 99&#x0025;. Likewise, [<xref ref-type="bibr" rid="ref-24">24</xref>] used three CNN-based models for early diagnosis of AD. For that, the models were applied to MRI images from an ADNI dataset. The reported results showed model performance with a classification rate of 97.65&#x0025; for AD <italic>vs.</italic> MCI and 88.37&#x0025; for MCI <italic>vs.</italic> CN.</p>
<p>A volumetric feature-based AD diagnosis method using CNN and DNN was proposed by [<xref ref-type="bibr" rid="ref-25">25</xref>]. The proposed AD diagnosis method was based on volumetric feature extraction from the left and right hippocampi of structured MRI data. The dataset (GARD) containing MRI scans were collected in South Korea. The proposed models were trained using the preprocessed GARD dataset and achieved an average accuracy of 94.82&#x0025; and 94.02&#x0025;, respectively. Reference [<xref ref-type="bibr" rid="ref-26">26</xref>] proposed GoogleNet, ResNet-18, and ResNet-152 to detect the four stages of AD. The experimental results show ResNet-18 achieved an accuracy of 98.88&#x0025;, and ResNet-152 98.14&#x0025;. Similarly, [<xref ref-type="bibr" rid="ref-27">27</xref>] has used CNN and MMSE methods to identify MCI <italic>vs.</italic> CN from MRI images and achieved an accuracy rate of 90.9&#x0025;.</p>
</sec>
<sec id="s3"><label>3</label><title>Methodology</title>
<sec id="s3_1"><label>3.1</label><title>Dataset Descriptions</title>
<p>The Alzheimer&#x2019;s brain MRI dataset of 6400 images was collected from Kaggle [<xref ref-type="bibr" rid="ref-28">28</xref>]. The dataset was divided into four different classes: mildly demented, moderately demented, non-demented, and very mildly demented. The number of images per class were of variable length in that mildly demented contained 896, moderately demented contained 64, non-demented contained 3200, and very mildly demented contained 2240 images, which comes to a total of 6400 images, as shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. The dataset was processed with all images resized into (128&#x2009;&#x00D7;&#x2009;128) pixels. <xref ref-type="fig" rid="fig-2">Fig. 2</xref> shows sample images of the Alzheimer&#x2019;s dataset displayed from the training set.</p>
<fig id="fig-1"><label>Figure 1</label><caption><title>Alzheimer dataset</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-1.png"/></fig>
<fig id="fig-2"><label>Figure 2</label><caption><title>Sample images in training set</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-2.png"/></fig>
</sec>
<sec id="s3_2"><label>3.2</label><title>Data Augmentation</title>
<p>Recently, the application of DNN has attracted the attention of researchers for detecting and diagnosing AD. Neural networks (NN) are the fundamental building blocks of the deep-learning model, where DNN contains millions of parameters to learn; therefore, numerous iterations are required to find the optimal values for classification [<xref ref-type="bibr" rid="ref-29">29</xref>]. The data size is an essential factor because a large amount of data can help better learn the model parameters and generalize better, since it captures the intrinsic data effectively [<xref ref-type="bibr" rid="ref-30">30</xref>]. One of the major issues with DNN applications in AD detection and diagnosis is the unavailability of a large amount of training data for learning discriminative patterns in large dimensional feature spaces [<xref ref-type="bibr" rid="ref-17">17</xref>] because large data size prevents deep-learning models from experiencing overfitting problems [<xref ref-type="bibr" rid="ref-22">22</xref>].</p>
<p>Since the collected dataset was hugely imbalanced, as shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>, class balancing was necessary to increase the number of images in the minority classes in order to reach a better understanding of the learning model&#x2019;s classifier. Although it is challenging to produce a large, benchmarked, medically tested set of images [<xref ref-type="bibr" rid="ref-31">31</xref>], this challenge can be overcome using data augmentation containing numerous methods that bring random variations to existing training images while maintaining class labels [<xref ref-type="bibr" rid="ref-32">32</xref>,<xref ref-type="bibr" rid="ref-33">33</xref>]. Moreover, although a contingency simulation can generate a large amount of data, the process is burdensome and error-prone, while data augmentation artificially increases representative and diversified training data with labels preserving transformation in a low-cost and effectual manner&#x00A0;[<xref ref-type="bibr" rid="ref-34">34</xref>].</p>
<p>In this study, data augmentation was performed using several techniques, including rotation, shearing, zooming, and horizontal and vertical flipping. The number of images increased after data augmentation to a total of 10,074 images, as shown in <xref ref-type="fig" rid="fig-3">Fig. 3</xref>.</p>
<fig id="fig-3"><label>Figure 3</label><caption><title>Alzheimer&#x2019;s dataset after data augmentation</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-3.png"/></fig>
</sec>
<sec id="s3_3"><label>3.3</label><title>CNN Architecture</title>
<p>In this study, we proposed a deep-learning model based on CNN architecture built from scratch. CNN is one of the most commonly used types of ANN [<xref ref-type="bibr" rid="ref-35">35</xref>] that has revealed amazing solutions to image data problems [<xref ref-type="bibr" rid="ref-36">36</xref>]. The architecture of CNN normally comprises 3 main layers: convolution, pooling, and fully connected layers [<xref ref-type="bibr" rid="ref-37">37</xref>] with an additional output layer. In convolution layers, the input images are convoluted with filters and generate feature maps [<xref ref-type="bibr" rid="ref-38">38</xref>]. A pooling layer is normally used after the convolution layer for reducing the dimensionality of an image while preserving the main features of the feature maps [<xref ref-type="bibr" rid="ref-39">39</xref>]. The fully connected layers contain various layers in which each node of a previous layer is connected to the nodes of the next layer [<xref ref-type="bibr" rid="ref-36">36</xref>]. A flattened layer is also part of the CNN architecture that normally occurs between the last convolution and the fully connected layers and is used to vectorize the feature maps and pass it on as input to the fully connected layers.</p>
</sec>
<sec id="s3_4"><label>3.4</label><title>Proposed Model</title>
<p>The proposed model comprises 3 convolutional layers, 3 pooling layers, 2 fully connected layers, and one output layer, as shown in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>. The pixel size of all images in the proposed method represented size of (128&#x2009;&#x00D7;&#x2009;128). The input layer takes the images in three dimensions, which are height, width, and dimension (i.e., H&#x2009;&#x00D7;&#x2009;W&#x2009;&#x00D7;&#x2009;D), where D represents whether the image is RGB or in a greyscale. In the proposed method, the image was provided in RGB form. Therefore, the input image was provided at size (128&#x2009;&#x00D7;&#x2009;128&#x2009;&#x00D7;&#x2009;3). Moreover, all convolution layers are convoluted according to the kernel size of (3&#x2009;&#x00D7;&#x2009;3), which is normally appropriate for classification purposes [<xref ref-type="bibr" rid="ref-40">40</xref>]. Similarly, ReLu is used as an activation function, as shown in <xref ref-type="disp-formula" rid="eqn-1">Eq. (1)</xref>, while same padding is used to avoid image shrinkage.
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>L</mml:mi><mml:mi>u</mml:mi><mml:mspace width="thickmathspace" /><mml:mo stretchy="false">(</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mi>z</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></disp-formula></p>
<fig id="fig-4"><label>Figure 4</label><caption><title>Architecture of the proposed model</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-4.png"/></fig>
<p>Likewise, the dropout operation was also performed in order to prevent the model from overfitting. As in the deep-learning models, the input layer is used to provide the image in a specific shape, while the actual learning of the model started at the convolution layers. Therefore, the input layers have zero trainable parameters, as shown in <xref ref-type="table" rid="table-2">Tab. 2</xref>.</p>
<p>The first convolutional layer in our proposed model used 16 filters with respect to the kernel size of (3&#x2009;&#x00D7;&#x2009;3) and same padding. The output shape of the first convolution layer per formula (filters &#x002B; 2&#x2009;&#x00D7;&#x2009;padding &#x2013; (kernel_size &#x2013; 1)) is (128 &#x00D7;&#x2009;128&#x2009;&#x00D7;&#x2009;16). Moreover, the activation shape is the product of the three dimensions (i.e., 128 &#x00D7;&#x2009;128&#x2009;&#x00D7;&#x2009;16), which is 262,144 elements, as shown in <xref ref-type="table" rid="table-1">Tab. 1</xref>. Similarly, the number of trainable parameters in each layer is calculated using <xref ref-type="disp-formula" rid="eqn-2">Eq. (2)</xref>.
<disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:msub><mml:mi>Z</mml:mi><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x2211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2217;</mml:mo><mml:mspace width="thickmathspace" /><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>n</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:math></disp-formula>where f is the current kernel size, x is the previous layer output, b is a bias constant, and z is the output vector.</p>
<table-wrap id="table-1"><label>Table 1</label><caption><title>Output shape and activation size of layers</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">No.</th>
<th align="left">Layer</th>
<th align="left">Output shape</th>
<th align="left">Size</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">1.</td>
<td align="left">Input layer</td>
<td align="left">128 &#x00D7;&#x2009;128&#x2009;&#x00D7;&#x2009;3</td>
<td align="left">49152</td>
</tr>
<tr>
<td align="left">2.</td>
<td align="left">Convolution 1</td>
<td align="left">128 &#x00D7;&#x2009;128&#x2009;&#x00D7;&#x2009;16</td>
<td align="left">262144</td>
</tr>
<tr>
<td align="left">3.</td>
<td align="left">Max-pooling1</td>
<td align="left">64&#x2009;&#x00D7;&#x2009;64&#x2009;&#x00D7;&#x2009;16</td>
<td align="left">65536</td>
</tr>
<tr>
<td align="left">4</td>
<td align="left">Convolution 2</td>
<td align="left">64&#x2009;&#x00D7;&#x2009;64&#x2009;&#x00D7;&#x2009;32</td>
<td align="left">131072</td>
</tr>
<tr>
<td align="left">4.</td>
<td align="left">Max-pooling 2</td>
<td align="left">32&#x2009;&#x00D7;&#x2009;32&#x2009;&#x00D7;&#x2009;32</td>
<td align="left">32768</td>
</tr>
<tr>
<td align="left">5.</td>
<td align="left">Convolution 3</td>
<td align="left">32&#x2009;&#x00D7;&#x2009;32&#x2009;&#x00D7;&#x2009;64</td>
<td align="left">65536</td>
</tr>
<tr>
<td align="left">6.</td>
<td align="left">Max-pooling 3</td>
<td align="left">16&#x2009;&#x00D7;&#x2009;16&#x2009;&#x00D7;&#x2009;64</td>
<td align="left">16384</td>
</tr>
<tr>
<td align="left">7.</td>
<td align="left">Fully connected layer 1</td>
<td align="left">128</td>
<td align="left">---</td>
</tr>
<tr>
<td align="left">8.</td>
<td align="left">Fully connected layer 2</td>
<td align="left">64</td>
<td align="left">---</td>
</tr>
<tr>
<td align="left">9.</td>
<td align="left">Fully connected layer 3</td>
<td align="left">4</td>
<td align="left">---</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="table-2"><label>Table 2</label><caption><title>Proposed model layers and trainable parameters</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">No.</th>
<th align="left">Layer</th>
<th align="left">Trainable parameters</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">1.</td>
<td align="left">Input layer</td>
<td align="left">0</td>
</tr>
<tr>
<td align="left">2.</td>
<td align="left">Convolution 1</td>
<td align="left">448</td>
</tr>
<tr>
<td align="left">3.</td>
<td align="left">Max-pooling1</td>
<td align="left">0</td>
</tr>
<tr>
<td align="left">4</td>
<td align="left">Convolution 2</td>
<td align="left">4640</td>
</tr>
<tr>
<td align="left">4.</td>
<td align="left">Max-pooling 2</td>
<td align="left">0</td>
</tr>
<tr>
<td align="left">5.</td>
<td align="left">Convolution 3</td>
<td align="left">18496</td>
</tr>
<tr>
<td align="left">6.</td>
<td align="left">Max-pooling 3</td>
<td align="left">0</td>
</tr>
<tr>
<td align="left">7.</td>
<td align="left">Fully connected layer 1</td>
<td align="left">2097280</td>
</tr>
<tr>
<td align="left">8.</td>
<td align="left">Fully connected layer 2</td>
<td align="left">8256</td>
</tr>
<tr>
<td align="left">9.</td>
<td align="left">Fully connected layer 3</td>
<td align="left">260</td>
</tr>
<tr>
<td align="left" colspan="2">Total trainable parameters</td>
<td align="left">2,129,380</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The total number of trainable parameters in the first convolution layer is [((3&#x2009;&#x00D7;&#x2009;3&#x2009;&#x00D7;&#x2009;3) &#x002B; 1)&#x2009;&#x00D7;&#x2009;16] &#x003D; 448 as shown in <xref ref-type="table" rid="table-2">Tab. 2</xref>. After the first convolution layer, a max-pooling layer was used to reduce the dimensionality of the images. As described by [<xref ref-type="bibr" rid="ref-41">41</xref>], this layer is normally utilized for capturing the important features of the feature maps. In our proposed model, a max-pooling of size (2&#x2009;&#x00D7;&#x2009;2) was used in order to reduce the dimensionality of images, thus making a layer of size (64&#x2009;&#x00D7;&#x2009;64&#x2009;&#x00D7;&#x2009;16), as shown in <xref ref-type="table" rid="table-1">Tab. 1</xref>. A dropout operation was also performed after the max-pooling layer in order to avoid overfitting.</p>
<p>The second convolution layer was used with 32 filters of the Alzheimer&#x2019;s images, kernel size (3&#x2009;&#x00D7;&#x2009;3), and same padding. The output shape of the second layer is (64&#x2009;&#x00D7;&#x2009;64&#x2009;&#x00D7;&#x2009;32), and the activation shape has 131,072 elements, as shown in <xref ref-type="table" rid="table-1">Tab. 1</xref>. The total number of trainable parameters in the second layer is 4,640, as shown in <xref ref-type="table" rid="table-2">Tab. 2</xref>. A max-pooling layer was added after the second convolution layer of size (2&#x2009;&#x00D7;&#x2009;2) for dimensionality reduction, followed by a dropout layer to avoid overfitting.</p>
<p>A third convolution layer with 64 filters, kernel size of (3&#x2009;&#x00D7;&#x2009;3), and same padding was used. The output shape of this layer was (32&#x2009;&#x00D7;&#x2009;32&#x2009;&#x00D7;&#x2009;64), and the activation size was 6,5536 elements. The total number of trainable parameters in this layer was 18,496. Like the first two layers, a max-pooling and dropout layers were also added after the third convolution layer for dimensionality reduction and preventing model overfitting problems, respectively. After the analysis of the convolution layers, the features were vectorized using the flattened layer and passed to the fully connected layers. The fully connected layers typically accept the feature maps in a vector shape, which is the responsibility of the flattened layer.</p>
<p>The last step in building the proposed model was the fully connected layers, comprised of various layers depending on the nature of the problem, in which each node (neuron) of a previous layer is linked to the nodes of the next layer [<xref ref-type="bibr" rid="ref-40">40</xref>]. In the proposed model, two fully connected layers were built of 128 and 64 neurons, respectively. The last fully connected layer of 4 neurons was an output layer, which was predicting the sorting of Alzheimer&#x2019;s images into their corresponding classes. In the fully connected layers, ReLu was used as an activation function except for the last layer, where softmax was used instead, as shown in <xref ref-type="disp-formula" rid="eqn-4">Eq. (4)</xref>. The softmax layer consisted of 4 neurons and was used to classify each image with respect to their class based on the predicted value.
<disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:mi>&#x03C3;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Z</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msup><mml:mrow><mml:msubsup><mml:mrow><mml:mo>&#x2211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msubsup><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>A hyperparameter tuning was performed to finalize parameters for the proposed model. According to the [<xref ref-type="bibr" rid="ref-42">42</xref>,<xref ref-type="bibr" rid="ref-43">43</xref>] setting of proper hyperparameters is the key to attain optimal outcomes in classification problems. In this tuning, the Adam optimizer provided promising results and was ultimately chosen for the compilation of our proposed model. Moreover, sparse categorical cross-entropy was utilized for measuring loss of the proposed model in classifying the Alzheimer&#x2019;s MRI images based on the predicted values.</p>
</sec>
</sec>
<sec id="s4"><label>4</label><title>Experimental Setup</title>
<p>First of all, the dataset was divided into three different sets: training set, validation set, and testing set with ratios of 80&#x0025;, 10&#x0025;, and 10&#x0025;, respectively. The model was compiled using the Adam optimizer, while loss was computed using sparse categorical cross-entropy. The proposed model was executed using 100 epochs in a Google Colab notebook in a GPU environment. Moreover, the model was also executed with different numbers of iterations, but 100 iterations achieved promising results. The average execution time for each epoch was around 6 s. The batch size was set to 64, which has achieved better results compared to the alternate numbers. After the execution of the proposed model for all epochs, the model was saved to the directory for future use. Furthermore, the validation set was used to validate our proposed model in terms of training and validation losses and accuracies. <xref ref-type="fig" rid="fig-5">Fig. 5</xref> shows training and validation accuracy with respect to epochs. Generally, the training accuracy is higher, but in this case both curves coincide as the epochs progress.</p>
<p>Similarly, training and validation loss was exhibited in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. As the epochs progress, both curves come closer to each other in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. At the start, both curves proceed with a very minor gap between them, but after epoch 25, the curves are very close to each other, and as the epochs continue, the curves coincide. This overlap of training and validation loss curves demonstrates the lack of model overfitting. Moreover, model overfitting was also prevented using dropout operations.</p>
<fig id="fig-5"><label>Figure 5</label><caption><title>Training and validation accuracy with respect to epochs</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-5.png"/></fig>
<fig id="fig-6"><label>Figure 6</label><caption><title>Training and validation loss with respect to epochs</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-6.png"/></fig>
</sec>
<sec id="s5"><label>5</label><title>Results and Discussion</title>
<p>The experiment was repeated several times to validate the results of the proposed model, and the best classifier according to the accuracy metric was utilized to report the results. The model was evaluated using the test set and achieved the highest testing accuracy of 99.38&#x0025;, which gives a classification error of only 0.62&#x0025;. This shows the reliability of the proposed model to be used as an intelligent model for automatic detection of Alzheimer&#x2019;s patients based on unseen MRI images. The accuracy of the proposed model was calculated using <xref ref-type="disp-formula" rid="eqn-4">Eq. (4)</xref>.
<disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:mrow><mml:mtext mathvariant="italic">Accuracy</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>Moreover, in the implementation phase of testing the proposed model, a confusion matrix was computed as shown in <xref ref-type="fig" rid="fig-7">Fig. 7</xref>, which provides some significant values that are calculated based on accurately and inaccurately classified MRI images. These values are commonly referred to as True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN).</p>
<fig id="fig-7"><label>Figure 7</label><caption><title>Confusion matrix</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-7.png"/></fig>
<p>The proposed model was further evaluated using precision, sensitivity, and F-measure, which are calculated based on the values of the confusion matrix.
<disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:mrow><mml:mtext mathvariant="italic">Precision</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula>
<disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:mrow><mml:mtext mathvariant="italic">Recall</mml:mtext></mml:mrow><mml:mspace width="thickmathspace" /><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mspace width="thickmathspace" /><mml:mrow><mml:mtext mathvariant="italic">Sensitivity</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula>
<disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:mrow><mml:mtext>F</mml:mtext></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mtext mathvariant="italic">measure</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo>(</mml:mo><mml:mn>2</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mrow><mml:mtext mathvariant="italic">Precision</mml:mtext></mml:mrow><mml:mo>&#x2217;</mml:mo><mml:mrow><mml:mtext mathvariant="italic">Recall</mml:mtext></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mtext mathvariant="italic">Precision</mml:mtext></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mtext mathvariant="italic">Recall</mml:mtext></mml:mrow></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p>Based on the above equation, the classification report is exhibited in <xref ref-type="table" rid="table-3">Tab. 3</xref>.</p>
<table-wrap id="table-3"><label>Table 3</label><caption><title>Classification report</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Class</th>
<th align="left">Precision</th>
<th align="left">Recall/Sensitivity</th>
<th align="left">F-measure</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Mild demented</td>
<td align="left">1.00</td>
<td align="left">0.98</td>
<td align="left">0.99</td>
</tr>
<tr>
<td align="left">Moderate demented</td>
<td align="left">1.00</td>
<td align="left">1.00</td>
<td align="left">1.00</td>
</tr>
<tr>
<td align="left">Non demented</td>
<td align="left">1.00</td>
<td align="left">0.99</td>
<td align="left">1.00</td>
</tr>
<tr>
<td align="left">Very mild demented</td>
<td align="left">0.99</td>
<td align="left">1.00</td>
<td align="left">0.99</td>
</tr>
<tr>
<td align="left">Accuracy</td>
<td align="left" colspan="3"><bold>0.99</bold></td>
</tr>
<tr>
<td align="left">Macro avg</td>
<td align="left">1.00</td>
<td align="left">0.99</td>
<td align="left">0.99</td>
</tr>
<tr>
<td align="left">Weighted avg</td>
<td align="left">0.99</td>
<td align="left">0.99</td>
<td align="left">0.99</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The values in <xref ref-type="table" rid="table-3">Tab. 3</xref> show the reliability of the proposed model in terms of precision, sensitivity, and F-measure to be used as a decision support system for detecting the stages of Alzheimer&#x2019;s disease and can help medical practitioners with its diagnosis.</p>
<p>We compared our results using the proposed model with those of existing state-of-the-art studies using the same dataset. Our proposed method outperformed existing studies in terms of accuracy, precision, sensitivity, specificity, and F-measure, as shown in <xref ref-type="table" rid="table-4">Tab. 4</xref>.</p>
<table-wrap id="table-4"><label>Table 4</label><caption><title>Comparison of proposed method with existing studies</title></caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th align="left">Study</th>
<th align="left">Method/ Model</th>
<th align="left">Task</th>
<th align="left">Accuracy</th>
<th align="left">Prec.</th>
<th align="left">Sens.</th>
<th align="left">Spec.</th>
<th align="left">F-measure</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left">Murugan et&#x00A0;al.&#x00A0;[<xref ref-type="bibr" rid="ref-20">20</xref>]</td>
<td align="left">CNN based DEMNET</td>
<td align="left">Multi-class</td>
<td align="left">95.23&#x0025;</td>
<td align="left">96&#x0025;</td>
<td align="left">95&#x0025;</td>
<td align="left">---</td>
<td align="left">95.27&#x0025;</td>
</tr>
<tr>
<td align="left">Shahwar et&#x00A0;al.&#x00A0;[<xref ref-type="bibr" rid="ref-21">21</xref>]</td>
<td align="left">Hybrid classical&#x2013;quantum CNN based on ResNet34</td>
<td align="left">Binary class</td>
<td align="left">97&#x0025;</td>
<td align="left">---</td>
<td align="left">---</td>
<td align="left">---</td>
<td align="left">---</td>
</tr>
<tr>
<td align="left">Mohammed et&#x00A0;al.&#x00A0;[<xref ref-type="bibr" rid="ref-22">22</xref>]</td>
<td align="left">AlexNet&#x002B;<break/>SVM (hybrid model)</td>
<td align="left">Multi-class</td>
<td align="left">94.8&#x0025;,</td>
<td align="left">---</td>
<td align="left">93&#x0025;,</td>
<td align="left">97.75&#x0025;</td>
<td align="left">---</td>
</tr>
<tr>
<td align="left"><bold>Proposed Model</bold></td>
<td align="left"><bold>CNN from scratch</bold></td>
<td align="left">Multi-class</td>
<td align="left"><bold>99.38&#x0025;</bold></td>
<td align="left"><bold>99&#x0025;</bold></td>
<td align="left"><bold>99&#x0025;</bold></td>
<td align="left"><bold>99&#x0025;</bold></td>
<td align="left"><bold>99&#x0025;</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Murugan&#x00A0;et&#x00A0;al.&#x00A0;[<xref ref-type="bibr" rid="ref-20">20</xref>] used a CNN-based DEMNET model for classifying the stages of AD. The dataset was collected from Kaggle containing a total of 6400 MRI images, which was initially imbalanced. The class balancing was conducted using a conventional SMOTE ML method, resulting in increasing the number of images to 12800. Moreover, the dataset was split into 80&#x0025; training, 10&#x0025; validation, and 10&#x0025; testing sets. The input image was resized into (172 &#x00D7; 172). The proposed DEMNET architecture contained 2 convolution layers followed by a pooling layer. The DEMNET block comprises the stacked layer of two convolution layers, batch normalization, and a pooling layer. The blocks consisted of distinctive filters (i.e., 32, 64, 128, and 256). Dropout operation was also performed after the third and fourth blocks to reduce the model overfitting problem. Moreover, three dense layers and a softmax layer was used to output the classification results. The model was compiled using root mean square propagation optimizer, and the loss was measured using categorical cross entropy. In contrast to the study of [<xref ref-type="bibr" rid="ref-20">20</xref>], we proposed a CNN from scratch model and applied it to the same dataset. The main differences include that, for example, data augmentation is more powerful than the conventional SMOTE method for increasing the number of images. We used input image of size (128&#x2009;&#x00D7;&#x2009;128) and dropout operation after each pooling layer. Moreover, for model compilation, we used an Adam optimizer and sparse categorical cross-entropy for measuring loss, which was optimal in our proposed model. These are the main reasons for achieving higher results.</p>
<p>Similarly, [<xref ref-type="bibr" rid="ref-21">21</xref>] used a hybrid classical&#x2013;quantum NN model in which the feature extraction was based on CNN-based ResNet34 that converted the given input into 512 feature vectors, and the dimensionality reduction was performed using a quantum variational circuit. Data augmentation was not performed. The model output was a binary classification of MRI images, either demented or non-demented, with an accuracy rate of 97&#x0025;. Likewise, in Mohammed&#x00A0;et&#x00A0;al.&#x00A0;[<xref ref-type="bibr" rid="ref-22">22</xref>], data augmentation was performed on the whole MRI dataset using cropping, rotation, brightness, flipping, and contrast methods and increased the images to 29,090. In contrast, we used data augmentation in only minority classes that initially had fewer images and achieved high performance of the model in terms of accuracy, sensitivity, specificity, and F-measure. The high performance of our proposed model shows the reliability of the proposed method to be used as a reliable decision support system for detecting AD that can help clinicians diagnose the disease.</p>
<p>Furthermore, our proposed model outperformed other state-of-the-art models that used different datasets, as discussed in Section 2. We have collected and discussed the most recent literature in the same domain of our study. The trend of using pre-trained models is obvious in the previous studies, most of which have used transfer learning methods for binary or multi-class classification of AD MRI images. However, our proposed model shows higher performance in terms of accuracy and other classification measures compared to existing state-of-the-art studies that used the same or different datasets, as discussed in Section 2 of this study.</p>
<p>Moreover, the proposed model was tested on the unseen MRI images of the test set and achieved promising results, as shown in <xref ref-type="fig" rid="fig-8">Fig. 8</xref>. Our proposed model predicted the same brain MRI images of Alzheimer&#x2019;s patients with respect to classes that actually belong to those classes. This also leads us to conclude that our proposed model is reliable for detecting AD stages and can be used as a medical decision support system that helps physicians in their diagnosing tasks.</p>
<fig id="fig-8"><label>Figure 8</label><caption><title>Model-predicted results</title></caption><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-8a.png"/><graphic mimetype="image" mime-subtype="png" xlink:href="CMC_32752-fig-8b.png"/></fig>
</sec>
<sec id="s6"><label>6</label><title>Conclusion</title>
<p>In this study, we present a CNN model built from scratch to detect the four stages of AD. The aim of this study was the detection of AD and providing a decision-based model that can help clinicians to diagnose AD. For this, we also attempted to use some transfer learning models of CNN-based architectures such as VGG-16 and ResNet50. The transfer learning models were employed in the same dataset, but unfortunately the results were not promising. Therefore, in this scenario, we concluded that the CNN model built from scratch provides better detection results in than the pre-learned models. Moreover, the proposed CNN from scratch model was also compared to recent existing state-of-the-art models. The performance of the proposed model showed a high testing accuracy rate of 99.38&#x0025;, which is the highest so far. Moreover, the performance of the model in terms of precision, sensitivity, specificity, and F-measure was 99&#x0025;, 99&#x0025;, and 99&#x0025; respectively, outperforming the existing models in the literature. This shows that the proposed model is reliable for detecting the four stages of AD and can be used as an intelligent decision support model for clinical diagnosis of AD. In future, the dataset can be enriched with real MRI images collected from different hospitals, clinics, and research centers to validate the results. New models, especially transfer learning models and models with different parameters, can also be applied to the new dataset, and the results will be evaluated and compared to the existing models.</p>
</sec>
</body>
<back>
<ack><p>Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R104), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.</p></ack>
<fn-group>
<fn fn-type="other"><p><bold>Funding Statement:</bold> Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R104), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.</p></fn>
<fn fn-type="conflict"><p><bold>Conflicts of Interest:</bold> The authors declare that they have no conflicts of interest to report regarding the present study.</p></fn>
</fn-group>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Loddo</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Buttau</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Di Ruberto</surname></string-name></person-group>, &#x201C;<article-title>Deep learning based pipelines for Alzheimer&#x2019;s disease diagnosis: A comparative study and a novel deep-ensemble method</article-title>,&#x201D; <source>Computers in Biology and Medicine</source>, vol. <volume>141</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>15</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Mirzaei</surname></string-name> and <string-name><given-names>H.</given-names> <surname>Adeli</surname></string-name></person-group>, &#x201C;<article-title>Machine learning techniques for diagnosis of Alzheimer disease, mild cognitive disorder, and other types of dementia</article-title>,&#x201D; <source>Biomedical Signal Processing and Control</source>, vol. <volume>72</volume>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Dadar</surname></string-name>, <string-name><given-names>T. A.</given-names> <surname>Pascoal</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Manitsirikul</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Misquitta</surname></string-name>, <string-name><given-names>V. S.</given-names> <surname>Fonov</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Validation of a regression technique for segmentation of white matter hyperintensities in Alzheimer&#x2019;s disease</article-title>,&#x201D; <source>IEEE Transactions on Medical Imaging</source>, vol. <volume>36</volume>, no. <issue>8</issue>, pp. <fpage>1758</fpage>&#x2013;<lpage>1768</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Lei</surname></string-name>, <string-name><given-names>E.</given-names> <surname>Liang</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Yang</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Yang</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Zhou</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Predicting clinical scores for Alzheimer&#x2019;s disease based on joint and deep learning</article-title>,&#x201D; <source>Expert Systems with Applications</source>, vol. <volume>187</volume>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Brookmeyer</surname></string-name>, <string-name><given-names>E.</given-names> <surname>Johnson</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Ziegler-Graham</surname></string-name> and <string-name><given-names>H. M.</given-names> <surname>Arrighi</surname></string-name></person-group>, &#x201C;<article-title>Forecasting the global burden of Alzheimer&#x2019;s disease</article-title>,&#x201D; <source>Alzheimer&#x2019;s and Dementia</source>, vol. <volume>3</volume>, no. <issue>3</issue>, pp. <fpage>186</fpage>&#x2013;<lpage>191</lpage>, <year>2007</year>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Livingston</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Huntley</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Sommerlad</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Ames</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Ballard</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Dementia prevention, intervention, and care: 2020 report of the lancet commission</article-title>,&#x201D; <source>The Lancet</source>, vol. <volume>396</volume>, no. <issue>10248</issue>, pp. <fpage>413</fpage>&#x2013;<lpage>446</lpage>, <fpage>2020</fpage>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Kumar</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Oh</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Schindler</surname></string-name>, <string-name><given-names>A. M.</given-names> <surname>Lai</surname></string-name>, <string-name><given-names>P. R. O.</given-names> <surname>Payne</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Machine learning for modeling the progression of Alzheimer disease dementia using clinical data: A systematic literature review</article-title>,&#x201D; <source>JAMIA Open</source>, vol. <volume>4</volume>, no. <issue>3</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>10</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Naz</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Ashraf</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Zaib</surname></string-name></person-group>, &#x201C;<article-title>Transfer learning using freeze features for Alzheimer neurological disorder detection using ADNI dataset</article-title>,&#x201D; <source>Multimedia Systems</source>, vol. <volume>28</volume>, no. <issue>1</issue>, pp. <fpage>85</fpage>&#x2013;<lpage>94</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Mahendran</surname></string-name> and <string-name><given-names>D. R. V.</given-names> <surname>PM</surname></string-name></person-group>, &#x201C;<article-title>A deep learning framework with an embedded-based feature selection approach for the early detection of the Alzheimer&#x2019;s disease</article-title>,&#x201D; <source>Computers in Biology and Medicine</source>, vol. <volume>141</volume>, no. <issue>September 2021</issue>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Afzal</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Maqsood</surname></string-name>, <string-name><given-names>U.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Mehmood</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Nawaz</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Alzheimer disease detection techniques and methods: A review</article-title>,&#x201D; <source>International Journal of Interactive Multimedia and Artificial Intelligence</source>, vol. <volume>6</volume>, no. <issue>7</issue>, pp. <fpage>26</fpage>&#x2013;<lpage>38</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Bi</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Xiao</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Wang</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Computer aided Alzheimer&#x2019;s disease diagnosis by an unsupervised deep learning technology</article-title>,&#x201D; <source>Neurocomputing</source>, vol. <volume>392</volume>, pp. <fpage>296</fpage>&#x2013;<lpage>304</lpage>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Islam</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>Early diagnosis of Alzheimer&#x2019;s disease: A neuroimaging study with deep learning architectures</article-title>,&#x201D; in <conf-name>IEEE Conf. on Computer Vision and Pattern Recognition Workshops</conf-name>, <conf-loc>Salt Lake City, USA</conf-loc>, pp. <fpage>1994</fpage>&#x2013;<lpage>1996</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>P.</given-names> <surname>Padilla</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Lopez</surname></string-name>, <string-name><given-names>J. M.</given-names> <surname>Gorriz</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Ramirez</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Salas-Gonzalez</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>NMF-SVM based CAD tool applied to functional brain images for the diagnosis of Alzheimer&#x2019;s disease</article-title>,&#x201D; <source>IEEE Transactions on Medical Imaging</source>, vol. <volume>31</volume>, no. <issue>2</issue>, pp. <fpage>207</fpage>&#x2013;<lpage>216</lpage>, <year>2012</year>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Gao</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Jiang</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Zhou</surname></string-name> and <string-name><given-names>D.</given-names> <surname>Chen</surname></string-name></person-group>, &#x201C;<article-title>Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview</article-title>,&#x201D; <source>Mathematical Biosciences and Engineering</source>, vol. <volume>16</volume>, no. <issue>6</issue>, pp. <fpage>6536</fpage>&#x2013;<lpage>6561</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Venugopalan</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Tong</surname></string-name>, <string-name><given-names>H. R.</given-names> <surname>Hassanzadeh</surname></string-name> and <string-name><given-names>M. D.</given-names> <surname>Wang</surname></string-name></person-group>, &#x201C;<article-title>Multimodal deep learning models for early detection of Alzheimer&#x2019;s disease stage</article-title>,&#x201D; <source>Scientific Reports</source>, vol. <volume>11</volume>, no. <issue>1</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>13</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>E.</given-names> <surname>Lin</surname></string-name>, <string-name><given-names>C. H.</given-names> <surname>Lin</surname></string-name> and <string-name><given-names>H. Y.</given-names> <surname>Lane</surname></string-name></person-group>, &#x201C;<article-title>Deep learning with neuroimaging and genomics in Alzheimer&#x2019;s disease</article-title>,&#x201D; <source>International Journal of Molecular Sciences</source>, vol. <volume>22</volume>, no. <issue>15</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>20</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Lu</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Popuri</surname></string-name>, <string-name><given-names>G. W.</given-names> <surname>Ding</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Balachandar</surname></string-name> and <string-name><given-names>M. F.</given-names> <surname>Beg</surname></string-name></person-group>, &#x201C;<article-title>Multimodal and multiscale deep neural networks for the early diagnosis of Alzheimer&#x2019;s disease using structural MR and FDG-PET images</article-title>,&#x201D; <source>Scientific Reports</source>, vol. <volume>8</volume>, no. <issue>1</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>13</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Jo</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Nho</surname></string-name> and <string-name><given-names>A. J.</given-names> <surname>Saykin</surname></string-name></person-group>, &#x201C;<article-title>Deep learning in Alzheimer&#x2019;s disease: Diagnostic classification and prognostic prediction using neuroimaging data</article-title>,&#x201D; <source>Frontiers in Aging Neuroscience</source>, vol. <volume>11</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>14</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Sava&#x015F;</surname></string-name></person-group>, &#x201C;<article-title>Detecting the stages of Alzheimer&#x2019;s disease with pre-trained deep learning architectures</article-title>,&#x201D; <source>Arabian Journal for Science and Engineering</source>, vol. <volume>47</volume>, no. <issue>2</issue>, pp. <fpage>2201</fpage>&#x2013;<lpage>2218</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Murugan</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Venkatesan</surname></string-name>, <string-name><given-names>M. G.</given-names> <surname>Sumithra</surname></string-name>, <string-name><given-names>X. Z.</given-names> <surname>Gao</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Elakkiya</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>DEMNET: A deep learning model for early diagnosis of Alzheimer diseases and dementia from mr images</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>9</volume>, pp. <fpage>90319</fpage>&#x2013;<lpage>90329</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Shahwar</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Zafar</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Almogren</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Zafar</surname></string-name>, <string-name><given-names>A. U.</given-names> <surname>Rehman</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Automated detection of Alzheimer&#x2019;s via hybrid classical quantum neural networks</article-title>,&#x201D; <source>Electronics</source>, vol. <volume>11</volume>, no. <issue>5</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>19</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>B. A.</given-names> <surname>Mohammed</surname></string-name>, <string-name><given-names>E. M.</given-names> <surname>Senan</surname></string-name>, <string-name><given-names>T. H.</given-names> <surname>Rassem</surname></string-name>, <string-name><given-names>N. M.</given-names> <surname>Makbol</surname></string-name>, <string-name><given-names>A. A.</given-names> <surname>Alanazi</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Multi-method analysis of medical records and mri images for early diagnosis of dementia and Alzheimer&#x2019;s disease based on deep learning and hybrid methods</article-title>,&#x201D; <source>Electronics</source>, vol. <volume>10</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>20</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Gharaibeh</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Almahmoud</surname></string-name>, <string-name><given-names>M. Z.</given-names> <surname>Ali</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Al-Badarneh</surname></string-name>, <string-name><given-names>M.</given-names> <surname>El-Heis</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Early diagnosis of Alzheimer&#x2019;s disease using cerebral catheter angiogram neuroimaging: A novel model based on deep learning approaches</article-title>,&#x201D; <source>Big Data and Cognitive Computing</source>, vol. <volume>6</volume>, no. <issue>2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>23</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Ji</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>W. Q.</given-names> <surname>Yan</surname></string-name> and <string-name><given-names>R.</given-names> <surname>Klette</surname></string-name></person-group>, &#x201C;<article-title>Early diagnosis of Alzheimer&#x2019;s disease using deep learning</article-title>,&#x201D; in <conf-name>Proc. of the 2nd Int. Conf. on Control and Computer Vision</conf-name>, <conf-loc>Korea</conf-loc>, pp. <fpage>87</fpage>&#x2013;<lpage>91</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Basher</surname></string-name>, <string-name><given-names>B. C.</given-names> <surname>Kim</surname></string-name>, <string-name><given-names>K. H.</given-names> <surname>Lee</surname></string-name> and <string-name><given-names>H. Y.</given-names> <surname>Jung</surname></string-name></person-group>, &#x201C;<article-title>Volumetric feature-based Alzheimer&#x2019;s disease diagnosis from smri data using a convolutional neural network and a deep neural network</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>9</volume>, pp. <fpage>29870</fpage>&#x2013;<lpage>29882</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Farooq</surname></string-name>, <string-name><given-names>S. M.</given-names> <surname>Anwar</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Awais</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Rehman</surname></string-name></person-group>, &#x201C;<article-title>A deep CNN based multi-class classification of Alzheimer&#x2019;s disease using MRI</article-title>,&#x201D; in <conf-name>2017 IEEE Int. Conf. on Imaging Systems and Techniques (IST)</conf-name>, <conf-loc>Beijing, China</conf-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>7</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Qiu</surname></string-name>, <string-name><given-names>G. H.</given-names> <surname>Chang</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Panagia</surname></string-name>, <string-name><given-names>D. M.</given-names> <surname>Gopal</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Au</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Fusion of deep learning models of MRI scans, mini&#x2013;mental state examination, and logical memory test enhances diagnosis of mild cognitive impairment,&#x201D;<italic>Alzheimer&#x2019;s &#x0026; Dementia: Diagnosis</italic></article-title>, <source>Assessment &#x0026; Disease Monitoring</source>, vol. <volume>10</volume>, pp. <fpage>737</fpage>&#x2013;<lpage>749</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="web"><person-group person-group-type="author"><collab>Kaggle</collab></person-group>, &#x201C;<article-title>Alzheimer MRI preprocessed dataset</article-title>,&#x201D; <uri xlink:href="https://www.kaggle.com/datasets/sachinkumar413/alzheimer-mri-dataset">https://www.kaggle.com/datasets/sachinkumar413/alzheimer-mri-dataset</uri>, <year>2022</year> (accessed Apr. 16, 2022).</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="web"><person-group person-group-type="author"><string-name><given-names>J. P.</given-names> <surname>Maheswari</surname></string-name></person-group>, &#x201C;<article-title>Breaking the curse of small data sets in machine learning: Part 2</article-title>,&#x201D; <uri xlink:href="https://towardsdatascience.com/breaking-the-curse-of-small-data-sets-in-machine-learning-part-2-894aa45277f4#:~:text=Neural%20networks%20are%20the%20basic,iteration%20can%20result%20in%20overfitting">https://towardsdatascience.com/breaking-the-curse-of-small-data-sets-in-machine-learning-part-2-894aa45277f4#:~:text=Neural%20networks%20are%20the%20basic,iteration%20can%20result%20in%20overfitting</uri>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="web"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Ding</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Tong</surname></string-name> and <string-name><given-names>H.</given-names> <surname>Liu</surname></string-name></person-group>, &#x201C;<article-title>Data augmentation for deep graph learning: A survey</article-title>,&#x201D; arXiv preprint arXiv, pp. <fpage>08235</fpage>, <year>2022</year>, [Online]. Available: <uri xlink:href="http://arxiv.org/abs/2202.08235">http://arxiv.org/abs/2202.08235</uri>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><given-names>P. P.</given-names> <surname>Kumar</surname></string-name>, <string-name><given-names>V. M. A.</given-names> <surname>Bai</surname></string-name> and <string-name><given-names>G. G.</given-names> <surname>Nair</surname></string-name></person-group>, &#x201C;<chapter-title>Augmentation techniques on mammogram images for cnn based breast cancer classification</chapter-title>,&#x201D; in <source>Emergent Converging Technologies and Biomedical Systems</source>, Singapore, Springer, pp. <fpage>635</fpage>&#x2013;<lpage>649</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>P.</given-names> <surname>Oza</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Sharma</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Patel</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Adedoyin</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Bruno</surname></string-name></person-group>, &#x201C;<article-title>Image augmentation techniques for mammogram analysis</article-title>,&#x201D; <source>Journal of Imaging</source>, vol. <volume>8</volume>, no. <issue>5</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>22</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Ahmad</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Ahmad</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Asif</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Rehman</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Alharbi</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>Evolution-based performance prediction of star cricketers</article-title>,&#x201D; <source>Computers, Materials and Continua</source>, vol. <volume>69</volume>, no. <issue>1</issue>, pp. <fpage>1215</fpage>&#x2013;<lpage>1232</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Zhang</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Chen</surname></string-name></person-group>, &#x201C;<article-title>A deep-learning intelligent system incorporating data augmentation for short-term voltage stability assessment of power systems</article-title>,&#x201D; <source>Applied Energy</source>, vol. <volume>308</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>17</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>U. R.</given-names> <surname>Acharya</surname></string-name>, <string-name><given-names>S. L.</given-names> <surname>Oh</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Hagiwara</surname></string-name>, <string-name><given-names>J. H.</given-names> <surname>Tan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Adam</surname></string-name> <etal>et al.,</etal></person-group> &#x201C;<article-title>A deep convolutional neural network model to classify heartbeats</article-title>,&#x201D; <source>Computers in Biology and Medicine</source>, vol. <volume>89</volume>, pp. <fpage>389</fpage>&#x2013;<lpage>396</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Altwaijry</surname></string-name> and <string-name><given-names>I.</given-names> <surname>Al-Turaiki</surname></string-name></person-group>, &#x201C;<article-title>Arabic handwriting recognition system using convolutional neural network</article-title>,&#x201D; <source>Neural Computing and Applications</source>, vol. <volume>33</volume>, pp. <fpage>2249</fpage>&#x2013;<lpage>2261</lpage>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Sultana</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Sufian</surname></string-name> and <string-name><given-names>P.</given-names> <surname>Dutta</surname></string-name></person-group>, &#x201C;<article-title>Advancements in image classification using convolutional neural network</article-title>,&#x201D; in <conf-name>4th IEEE Int. Conf. on Research in Computational Intelligence and Communication Networks</conf-name>, <conf-loc>Kolakata, India</conf-loc>, pp. <fpage>122</fpage>&#x2013;<lpage>129</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Yu</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Jia</surname></string-name> and <string-name><given-names>C.</given-names> <surname>Xu</surname></string-name></person-group>, &#x201C;<article-title>Convolutional neural networks for hyperspectral image classification</article-title>,&#x201D; <source>Neurocomputing</source>, vol. <volume>219</volume>, pp. <fpage>88</fpage>&#x2013;<lpage>98</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Abiwinanda</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Hanif</surname></string-name>, <string-name><given-names>S. T.</given-names> <surname>Hesaputra</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Handayani</surname></string-name> and <string-name><given-names>T. R.</given-names> <surname>Mengko</surname></string-name></person-group>, &#x201C;<chapter-title>Brain tumor classification using convolutional neural network</chapter-title>,&#x201D; in <source>World Congress on Medical Physics and Biomedical Engineering</source>, Singapore, Springer, pp. <fpage>183</fpage>&#x2013;<lpage>189</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Ullah</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Jamjoom</surname></string-name></person-group>, &#x201C;<article-title>An intelligent approach for arabic handwritten letter recognition using convolutional neural network</article-title>,&#x201D; <source>PeeJ Computer Science</source>, vol. <volume>8</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>22</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Gamb&#x00E4;ck</surname></string-name> and <string-name><given-names>U. K.</given-names> <surname>Sikdar</surname></string-name></person-group>, &#x201C;<article-title>Using convolutional neural networks to classify hate-speech</article-title>,&#x201D; in <conf-name>Proc. of the First Workshop on Abusive Language Online</conf-name>, <conf-loc>Vancouver, Canada, no. 7491</conf-loc>, pp. <fpage>85</fpage>&#x2013;<lpage>90</lpage>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X. R.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Zhou</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Sun</surname></string-name> and <string-name><given-names>S. K.</given-names> <surname>Jha</surname></string-name></person-group>, &#x201C;<article-title>A lightweight CNN based on transfer learning for COVID-19 diagnosis</article-title>,&#x201D; <source>Computers, Materials &#x0026; Continua</source>, vol. <volume>72</volume>, no. <issue>1</issue>, pp. <fpage>1123</fpage>&#x2013;<lpage>1137</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>W.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>G. C.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>X. R.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Zhang</surname></string-name> and <string-name><given-names>N. N.</given-names> <surname>Ge</surname></string-name></person-group>, &#x201C;<article-title>Fine-grained vehicle type classification using lightweight convolutional neural network with feature optimization and joint learning strategy</article-title>,&#x201D; <source>Multimedia Tools and Applications</source>, vol. <volume>80</volume>, no. <issue>20</issue>, pp. <fpage>30803</fpage>&#x2013;<lpage>30816</lpage>, <year>2021</year>.</mixed-citation></ref>
</ref-list>
</back>
</article>