Abstract
Magnetic resonance imaging (MRI), a non-ionizing radiation imaging method, is widely utilized in diagnosing Alzheimer’s disease (AD) due to its excellent imaging ability for brain soft tissues and high-resolution slice imaging. However, the enormous size of 3D MRI makes it difficult to process and analyze it. Therefore, a challenge is to accurately mine encephalatrophy patches from 3D MRI and build a patch-based diagnostic model. The existing patch-based methods mainly extract and fuse features of each patch in isolation, ignoring mutual information between patches, which makes the diagnosis performance unsatisfactory. We propose a novel dual-branch AD diagnostic model based on distinguishing atrophic patch localization. (1) We propose a Distinguishing Atrophic Patch Localization (DAPL) algorithm based on Distinguishing Index (DI) and Spatial Contact Ratio (SCR) to extract lesion areas that have significant impacts on diagnosis from 3D MRI. Meanwhile, we proposed a Discontinuity-voxel-based Dynamic Voxel Wrapping algorithm (DV\(^2\)W) to calculate DI for each patch. (2) A dual-branch diagnostic network (DBDN) is constructed to obtain intra-patch and inter-patch features synchronously. The intra-feature extraction branch extracts feature from each patch through a parallel multi-channel network and fuse them. The Inter-feature extraction branch defines the Spatial Context Mixing matrix (SCM) and performs feature extraction on SCM to obtain mutual information. The performance evaluation demonstrates that our DBDN model has adequate diagnostic performance compared to state-of-the-art methods. In addition, the distinguishing pathological locations identified by our DAPL algorithm can effectively guide inexperienced clinical doctors to identify lesion areas and guide doctors in diagnosis quickly.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
AD is a kind of irreversible dementia and a prevalent neurodegenerative disorder that mainly affects older adults [5]. According to the World Health Organization (WHO), over 50 million people globally are living with dementia, and Alzheimer’s disease is the most common form, accounting for 60-70% of cases. According to statistics released by the International Alzheimer’s Association, the global number of AD patients is projected to reach 76 million by 2030 and 135 million by 2050 [24]. AD typically impairs an individual’s cognitive, behavioral, and emotional abilities. With the increasing life expectancy of modern society, the aging population has intensified, leading to a rise in the occurrence of AD and placing a considerable burden on communities [47]. The precise etiology of AD remains uncertain, and no efficacious medications exist for its cure, underscoring the significance of early diagnosis and medical intervention to decelerate the progression of the disease.
The brain’s anatomical structure in AD patients presents diffuse encephalatrophy, with the primary morphological changes characterized by narrowed gyrus and widened sulcus [43]. The most prominent atrophy locations are concentrated in the hippocampus, temporal, parietal, and prefrontal areas. Magnetic Resonance Imaging (MRI) can clearly find the lesions of encephalatrophy. MRI is a neuroimaging method that utilizes nuclear magnetic resonance phenomena to capture electromagnetic signals emitted by the human body in order to reconstruct its internal structure. Compared to other imaging approaches such as Computed Tomography (CT) and Positron Emission Tomography (PET), MRI stands as the most widely used noninvasive neuroimaging technology as it poses no radioactive tracers or harmful gamma rays [48]. Magnetization Prepared Rapid Gradient Echo (MP-RAGE), a form of T1-weighted structural MRI (sMRI), effectively portrays brain structures such as gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF) and is proficient in detecting encephalatrophy in clinical settings. Figure 1 displays the MP-RAGE scans of AD patients with encephalatrophy.
Advancements in computer vision technology have led researchers to explore feature extraction from neuroimaging to achieve computer-aided diagnosis (CAD) [1, 54, 63]. Due to its wide application in clinically diagnosing AD, MRI is a neuroimaging modality upon which many methods rely to construct CAD models. Besides MRI, several methods utilize different modalities of neuroimaging to build CAD models, such as fMRI (functional MRI) [49], PET (Positron Emission Tomography) [55], DTI (Diffusion Tensor Imaging) [39], EEG (Electroencephalogram) [40], and others. These diverse neuroimaging modalities present information from various perspectives of patients, aiding in a multifaceted understanding of disease diagnosis. The use of MRI for AD diagnosis is driven by its high spatial resolution, non-invasive nature, and capability to provide detailed brain images. MRI enables early detection of structural changes, quantitative analysis of brain regions, and identification of comorbid conditions without using ionizing radiation. These advantages make MRI a crucial tool for accurate and comprehensive assessment of AD, thereby facilitating timely intervention and effective monitoring of disease progression. Hence, our study proposed an MRI-based CAD model.
Early AD diagnosis through CAD can assist clinicians in identifying potential lesions and providing diagnostic recommendations, thus improving diagnostic efficiency and reducing misdiagnoses. Depending on the scale of input data for feature extraction, current computer-aided AD diagnosis methods can be separated into four classifications: subject-based methods [3, 8, 64], voxel-based methods [44, 62], ROI (region of interest)-based methods [23, 52], and patch-based methods [28, 42]. Subject-based methods typically use 2D slices [3, 64] or 3D images [8] of patients to extract features. The 2D slice methods often focus on core cerebral regions for feature extraction while neglecting other parts and the contextual correlation features between brain tissues. On the other hand, the 3D methods heavily rely on model design quality for diagnostic performance, and their interpretability is insufficient. Voxel-based methods usually extract voxel values from registered images as features through statistical analysis [4], directly measuring the density and volume of GM, WM, and CSF components [41]. However, voxel-based methods indiscriminately treat each voxel value, resulting in high feature dimensionality and complexity in extraction and calculations. ROI-based methods utilize the cerebral anatomical structure to extract features from specific brain regions related to diseases for diagnosis [56]. These methods often require prior knowledge or experienced clinicians’ input and lack consideration for cross-regional areas, frequently leading to suboptimal results. In the early stage of AD, structural changes often appear in local areas, such as the hippocampus and amygdala [47], rather than the whole brain or isolated voxels. Patch-based methods extract patches related to AD diagnosis from the entire image and extract features from the patches to capture cerebral disease-related patterns. These patch-based methods construct patch-level feature extraction models, rendering auxiliary diagnostic tasks more straightforward and efficient [34]. Simultaneously, patch-based methods overcome the limitations of anatomical regions and can identify features of cross-tissue lesion areas. However, the existing patch-based methods often depend on specific prior knowledge and lack adaptability to different patients. For different diagnostic tasks (e.g., AD vs. CN and sMCI vs. pMCI), existing methods frequently extract patches from the same locations [32] and build diagnostic networks to achieve CAD. Meanwhile, when constructing diagnostic networks, current studies often extract features from each patch separately [29, 42], failing to mine the contextual relationships between patches and resulting in the subpar diagnostic performance of patch-based methods. Consequently, patch-based methods are promising but face two main challenges:
-
(1)
One challenge confronting patch-based feature extraction methods is how to precisely locate distinguishing atrophic patches for various diagnostic tasks while pruning diagnostically irrelevant patches.
-
(2)
Developing a more accurate diagnostic model that simultaneously extracts and fuses inter- and intra-features of patches is another challenge in constructing a patch-based computer-aided AD diagnostic model.
To address these challenges, we propose a novel dual-branch Alzheimer’s disease diagnostic model based on distinguishing atrophic patch localization. The schematic of our proposed method is illustrated in Fig. 2. Initially, we introduce a data-driven method for distinguishing atrophic patch localization. This involves defining a distinguishing index for the patches and subsequently proposing a discontinuity voxel-based dynamic voxel wrapping algorithm to calculate the distinguishing index of various patches in the brain for disease diagnosis. Furthermore, we propose a selection and pruning algorithm to achieve more precise localization of distinguishing atrophic patches. To simultaneously extract inter- and intra-features from these distinguishing atrophic patches, we present a dual-branch diagnostic network structure. The intra-feature extraction branch employs a multi-layer parallel network structure designed to extract features from each patch individually and fuse the feature vectors. The inter-feature extraction branch integrates the multiple distinguishing atrophic patches by defining a spatial context mixing matrix, which is then input into a Convolutional Neural Network (CNN) feature extraction network to obtain mutual information between patches. Finally, the features from both branches are fused to produce the final diagnostic result. This approach ensures that both local and global context information is effectively utilized, enhancing the accuracy and reliability of Alzheimer’s disease diagnosis. The main contributions of our proposed method are as follows:
-
A data-driven patch localization method is proposed to locate atrophic patches with high discrimination for disease diagnosis. We present a Discontinuity Voxel-based Dynamic Voxel Wrapping (DV\(^2\)W) algorithm to calculate the various patches’ Distinguishing Index (DI). And we propose a patch selection and pruning algorithm DAPL (Distinguishing Atrophic Patch Localization) based on DI and Spatial Contact Ratio (SCR) to achieve distinguishing atrophic patch localization.
-
A dual-branch diagnostic network that integrates inter-features and intra-features of patches is proposed to diagnose diseases accurately. The intra-feature extraction branch constructs a parallel network structure to synchronously extract and fuse features from each patch. The inter-feature extraction branch integrates the patches and creates a Spatial Context Mixing (SCM) matrix, which is then utilized to construct a CNN network. The fusion of two components yields the final diagnostic result.
-
Experiments on real patient datasets demonstrate our diagnostic model’s superiority in two diagnostic tasks compared to existing state-of-the-art methods. The visualization of distinguishing atrophic patch locations can intuitively display areas that require attention for disease diagnosis and assist clinicians in quickly discovering and identifying potential lesion locations to enhance diagnostic efficiency.
The remaining sections are structured as follows: Section 2 provides a summary of current related works across four feature extraction strategies. Section 3 describes our proposed method and the dataset used. In Section 4, we present the experimental settings and results of our approach. Section 5 discusses the advancements and implications of our proposed method. Section 6 offers the conclusion.
2 Related work
In this section, we introduce feature extraction methods for four categories involved in computer-aided AD diagnosis: subject-based methods, voxel-based methods, ROI-based methods, and patch-based methods.
Subject-based methods usually utilize 3D neuroimages or 2D slices as input to construct a model for feature extraction and disease diagnosis. Cobbinah et al. [8] developed a diagnostic model with a soft attention mechanism and residual blocks, aligning different protocols of 3D MRI scans with the proposed adversarial autoencoder as input. Feng et al. [11] built a 3D CNN architecture for extracting feature representations from neuroimages. Subsequently, a fully stacked bidirectional long short-term memory (LSTM) network was suggested to capture spatially relevant deep-level information in the feature map, enhancing diagnostic performance. Wang et al. [57] input 3D MRI into an ensemble of DenseNet with five branches. Although the 3D methods can use all the information of neuroimages, the diagnostic network constructed usually has many parameters and high computational complexity. Meanwhile, the diagnostic performance is often unsatisfactory due to unavoidable image noise. Zhang et al. [64] input the 2D coronal gray matter slices into a feature extraction model with channel attention. Liu et al. [33] assembled a diagnostic network cascading CNN and recurrent neural network (RNN). CNN extracts feature from slices, while a bidirectional gate recurrent unit is employed to extract spatial information between different slices. The methods based on 2D slices usually select partial brain images as input data, often leading to information loss. In addition, using 2D slices often makes it challenging to mine spatial information, resulting in poor diagnostic performance.
Subject-based methods eliminate the need for prior knowledge or predefined lesion areas and data preprocessing. Deep learning advancements reduce manual feature design dependence, enhancing diagnostic accuracy through global and contextual information capture. However, performance relies heavily on network design and training optimization, facing challenges with complex medical images for clinical understanding.
Voxel-based methods typically require the registration of neuroimages into a public space. The processed images extract meaningful voxel values as features through statistical analysis methods. This method also measures the volume and density of GM, WM, and CSF. Zhang et al. [62] use MRI data and a voxel-based morphometry method (VBM) [4] to diagnose AD. Richhariya et al. [44] proposed a voxel-based feature selection and classification method - universum SVM-RFE. This method can achieve better accuracy by reducing the number of features. In addition, Takao et al. [53] also utilized VBM to resolve the dependability of longitudinal changes in GM volume. Voxel-based feature extraction methods usually deal with very high voxel dimensions, which makes feature extraction and selection more complex and carries the risk of overfitting. At the same time, neglecting the consideration of the correlation between voxels and the importance of voxels can make the diagnostic model unable to achieve satisfactory performance.
Voxel-based methods simplify MRI analysis but struggle with high-dimensional features, particularly with complex imaging modalities like fMRI and large datasets. They may overlook regional information crucial for disease discrimination, impacting diagnostic performance.
ROI-based methods need to utilize the prior cerebral morphological features to divide the anatomical structure of the human brain into different regions [56]. These approaches usually focus on localized tissue regions and specific parts of the brain rather than the whole brain. Suk et al. [52] constructed feature sequences based on the extracted features from different ROIs, then input them into the deep autoencoder model to obtain potential features and combined them with a hidden Markov model to build a state transition matrix to predict the state of the disease. Wang et al. [58] used the brain tissue extraction tool to obtain hippocampus slices and constructed a CNN model for feature extraction. Li et al. [26] extracted 93 ROI-based volume features from multimodal neuroimages and utilized principal component analysis to diagnose disease. Cui et al. [9] diagnosed the disease by analyzing ROIs’ shape and volume features, such as the hippocampus. Although the ROI-based methods have good interpretability, they require much prior knowledge. In addition, morphological abnormalities caused by neurological diseases do not constantly occur in pre-defined ROIs. They may also exist in locations across ROIs, which can make the performance of ROI-based methods unstable and affect diagnostic performance.
ROI-based methods in CAD offer interpretability for clinicians, reducing computational complexity compared to voxel-based methods. However, they rely on prior knowledge for ROI selection, impacting diagnostic performance. Neurological disorders may affect multiple ROIs, potentially destabilizing CAD model performance.
The early changes in the brain structure of AD often occur in parts or multiple regions of the brain. Patch-based methods divide the brain into patches with a specified size and extract features from these patches. These methods avoid the enormous computational burden of processing large 3D images and can more accurately obtain disease-related features [50, 66]. Lian et al. [28] proposed a one-stage patch selection and disease diagnostic model. A hierarchical fully convolutional neural network is proposed for feature extraction of multiple patches, and patches that impact diagnosis are selected through network pruning algorithms. Pan et al. [42] divided MRI into patches of the same size and isolated each patch into a constructed CNN feature extraction network to obtain features. The features were fused and input into the classifier for disease diagnosis. Suk et al. [51] cropped the images into multiple patches and input them into a Deep Boltzmann Machine for feature fusion. Finally, the Support Vector Machine was used for diagnosis. Liu et al. [32] constructed a multimodal cascaded network using MRI and PET(Positron Emission Computed Tomography) data. They input the patches corresponding to the positions of the two modalities into the network for feature extraction and disease diagnosis. In addition, [34] extracted features from the patches determined using prior knowledge, fused the subjects’ demographic information, and jointly input the fused features into the classifier for disease diagnosis.
Although patch-based methods have many advantages, such as being more targeted in processing objects and having lower computational complexity compared to subject-based and voxel-based methods, and having the same patch size compared to ROI-based methods, making it easier to construct feature extraction networks for different patches, there are still some challenges. The first challenge is that almost all patch-based methods do not select different patches for different diagnostic tasks in AD, and some models do not even perform feature selection. Instead, they directly divide the images into patches and input them into feature extraction networks, which affects the diagnosis performance. We propose a data-driven distinguishing atrophic patch localization method to select different patches with distinguishing for different diagnostic tasks. The second challenge is that existing patch-based methods often extract features independently of each patch for fusion diagnosis, lacking in mining the contextual relationships between patches. To this end, we propose a dual-branch diagnostic network that simultaneously mines inter-features and intra-features and fuses them into the classifier for diagnosis.
3 Materials and methods
3.1 Dataset and data pre-processing
In this study, we investigate T1-weighted structural MR images of 1461 subjects, including 218 AD, 328 pMCI, 448 sMCI, and 467 CN. Mild cognitive impairment (MCI) is a condition situated between cognitive normal (CN) and AD and serves as a significant risk factor for dementia [16]. Clinically, MCI patients are classified into two subtypes based on whether they will progress into AD within 36 months of their MCI diagnosis: stable MCI (sMCI) and progressive MCI (pMCI). The demographic characteristics of these subjects are illustrated in Fig. 3. All these neuroimages are acquired from Alzheimer’s Disease Neuroimaging Initiative [36]. The slice thickness of all MRIs equals 1.2 mm. The sagittal scanning images are selected because they can clearly display the structure of the sulcus and gyrus. Some necessary preprocessing for these images is done. We utilize tool [17] for intensity correction, FreeSurfer [12] for skull stripping, and image registration to the MNI152 template [21].
3.2 Methodology
In clinical, physicians usually focus on multiple local locations of the brain when diagnosing Alzheimer’s disease through neuroimaging and combine these regions to determine the diagnostic results comprehensively. In order to design a more accurate AD computer-aided diagnostic model, we simulate the process of disease diagnosis by clinicians. Firstly, we propose a data-driven distinguishing atrophic patch localization method, which can locate patches with a high impact on disease diagnosis. Furthermore, we propose a dual-branch disease diagnostic model that extracts intra-features and inter-features from the selected patches and fuses the extracted features to obtain diagnostic results. The workflow of our proposed AD diagnostic model is shown in Fig. 4. The following two subsections provide a detailed introduction to the two main parts of our model.
3.2.1 Distinguishing atrophic patch localization
When diagnosing a specific disease, clinicians usually prioritize observing the pathological changes at specific locations in the neuroimaging based on experience [14]. We propose a method for CAD to identify the locations of these lesions on where they need to be focused. First, we divide the registered 3D MRI into N patches of the same size, then define Distinguishing Index (DI) for each patch. Let \(I_{X\times Y\times Z}\) represent a 3D MRI; \(l\left( i\right) =\left( l_x\left( i\right) ,l_y\left( i\right) ,l_z\left( i\right) \right) \) denotes a voxel in I, and \(i\in [1,N]\). \(v\left( l\left( i\right) \right) \) is the voxel value mapping function. \(P\left( l\left( i\right) \right) \) represents a patch with \(l\left( i\right) \) as the first voxel and \(l\times l\times l\) as the size. DI represents the ability of different patches to distinguish between different diagnostic tasks. We propose a Discontinuity Voxel-based Dynamic Voxel Wrapping (DV\(^2\)W) algorithm to calculate the DI of each patch \(P\left( l\left( i\right) \right) \) with specified diagnostic tasks. Meanwhile, we defined Spatial Contact Ratio (SCR) between different patches. Finally, based on DI and SCR, we propose a Distinguishing Atrophic Patch Localization (DAPL) algorithm to filter out M patches with influential distinguishing significance for diagnostic tasks. To calculate DI, we need to calculate the difference between two patches obtained from two classifications. Unlike calculating the differences in 3D matrices, it is easier to calculate the differences (or distances) between sequence data. Therefore, we consider transforming the patches of three-dimensional matrix structures into sequential data. Here we define the h-order outstretching image sequence.
Definition 1
(h-Order Outstretching Image Sequence) Given image \(I_{a\times b\times c}\), cut the image into multiple slices along the h direction, each slice being a two-dimensional matrix. Expand each matrix into a sequence and concatenate the sequences of each slice expansion. The flattened sequence is called the h-order outstretching sequence of the image I, represented as
where \(k=a\times b\times c\) is the number of voxels in the image I. For a 3D MRI image, \(h\in H=\left\{ axial,\ sagittal,\ coronal\right\} \).
For two image sequences, calculating the degree of difference can be seen as calculating similarity. The standard calculation of similarity between sequences is based on Euclidean distance measurement and requires that the points in the two sequences correspond one-to-one. However, for images from two different types of diseases, voxel points cannot achieve bijection due to differences in the location and grade of brain atrophy, as well as errors in the registration process. Using traditional similarity calculation methods will not be able to effectively calculate the similarity (difference) between two image sequences. To achieve this, we propose a DV\(^2\)W algorithm based on discontinuity voxels to calculate the difference between image sequences. Before introducing the algorithm, we will provide some relevant definitions.
Definition 2
(Matching Function) Given two image sequences \(Seq\left( I_1,h\right) \) and \(Seq\left( I_2,h\right) \), The matching function \(\mathcal {F}_{Seq\left( I_1,h\right) ,Seq\left( I_2,h\right) }\left( f,g\right) \) contains two mappings, with the mapping rules being:
where \(\Lambda \) represents null.
Generally, the matching function satisfies monotonicity, that is:
-
(1)
If \(f\left( seq\left( I_1,h\right) _i\right) \rightarrow \ seq\left( I_2,h\right) _j\), then for all \(i^\prime >i\), \(f\left( seq\left( I_1,h\right) _{i^\prime }\right) \rightarrow seq\left( I_2,h\right) _{j^\prime }/\Lambda \), where \(j^\prime \) satisfies \(j^\prime >j\).
-
(2)
If \(g\left( seq\left( I_2,h\right) _i\right) \rightarrow \ seq\left( I_1,h\right) _j\), then for all \(i^\prime >i\), \(g\left( seq\left( I_2,h\right) _{i^\prime }\right) \rightarrow seq\left( I_1,h\right) _{j^\prime }/\Lambda \), where \(j^\prime \) satisfies \(j^\prime >j\).
Definition 3
(Discontinuity Voxel) Given two sequences \(Seq\left( I_1,h\right) \) and \(Seq\left( I_2,h\right) \), and the matching function \(\mathcal {F}_{Seq\left( I_1,h\right) ,Seq\left( I_2,h\right) }\left( f,g\right) \). If we have \(f\left( seq\left( I_1,h\right) _i\right) \rightarrow \Lambda \) or \(g\left( seq\left( I_2,h\right) _j\right) \rightarrow \Lambda \), then the voxel points \(seq\left( I_1,h\right) _i\) and \(seq\left( I_2,h\right) _j\) are called discontinuity voxels, abbreviated as DV.
Definition 4
(Voxel Distance) Let \(seq\left( I_1,h\right) _i\) and \(seq\left( I_2,h\right. \) \(\left. \right) _j\) be two voxel points of two image sequences, the voxel distance between \(seq\left( I_1,h\right) _i\) and \(seq\left( I_2,h\right) _j\) are defined as
where \(\delta \) is the critical threshold.
Definition 5
(Distinguishing Distance) Given two sequences \(Seq\left( I_1,h\right) \) and \(Seq\left( I_2,h\right) \), find an optimal matching function \(\mathcal {F}_{Seq\left( I_1,h\right) ,Seq\left( I_2,h\right) }\left( f,g\right) \) such that
then \(DS\left( Seq\left( I_1,h\right) ,Seq\left( I_2,h\right) \right) \) is called the distinguishing distance between images \(I_1\) and \(I_2\). \(\tau \) represents the maximum of voxel distance.
Definition 6
(Distinguishing Index) Given the image data of two classifications \(CL_1\) and \(CL_2\), the distinguishing index of patch \(P\left( l\left( i\right) \right) \) in the task of diagnosing \(CL_1\) and \(CL_2\) is defined as:
where \(S_\alpha \) and \(S_\beta \) are the number of images in \(CL_1\) and \(CL_2\), respectively.
\(DI\left( l\left( i\right) \right) \) describes the diagnostic ability of a patch at a specified location \(l\left( i\right) \) in diagnosing \(CL_1\) and \(CL_2\) tasks, which is the degree of encephalatrophy at that location. The key to calculating DI is to calculate distinguishing distance, meaning we need to know the distinguishing distance between any two patches clearly.
Let’s consider the outstretching image sequences \(Seq\left( A,\right. \) \(\left. h\right) =\left\{ \rho _1,\rho _2,\ldots ,\rho _m\right\} \) and \(Seq\left( B,h\right) =\left\{ \theta _1,\theta _2,\ldots ,\theta _m\right\} \) of two image patches. When calculating distinguishing distance, each voxel in the two sequences has two possible states, continuity voxel, and discontinuity voxel. A voxel is defined as a continuity voxel state, abbreviated as CV when it can discover a corresponding voxel in another sequence. On the contrary, if the current voxel cannot find a corresponding voxel in another sequence, it is in a DV state. We use \(DS\left( \rho _i,\theta _j\right) \) to represent the distinguishing distance of the two sequences \(\left\{ \rho _1,\rho _2,\ldots ,\rho _i\right\} \) and \(\left\{ \theta _1,\theta _2,\ldots ,\theta _j\right\} \). The two voxels \(\rho _i\) and \(\theta _j\) have four possible states, that is (1) \(\rho _i\) is DV, and \(\theta _j\) is DV; (2) \(\rho _i\) is DV, and \(\theta _j\) is CV; (3) \(\rho _i\) is CV, and \(\theta _j\) is DV; (4) \(\rho _i\) is CV, and \(\theta _j\) is CV. So we can calculate \(DS\left( \rho _i,\theta _j\right) \) by the following equation:
where \({DS}_{DD},{DS}_{DC},{DS}_{CD},{DS}_{CC}\) are the four state matrices of the current voxel pair \(\left( \rho _i,\theta _j\right) \). Below, we will provide specific calculation methods for these four matrices.
In order to better match voxels between two sequences, we propose a dynamic strategy. The cumulative distance of the current voxel pair \(\left( \rho _i,\theta _j\right) \) comes from \(\left( \rho _{i-1},\theta _j\right) \), \(\left( \rho _i,\theta _{j-1}\right) \), and \(\left( \rho _{i-1},\theta _{j-1}\right) \); similarly, the state matrices also come from these three parts. Let’s take \({DS}_{DD}\left( \rho _i,\theta _j\right) \) as an example to illustrate the calculation process. If the cumulative distance comes from \(\left( \rho _{i-1},\theta _j\right) \), since \(\theta _j\) is DV, only the state of \(\rho _{i-1}\) needs to be concerned, which can be separated into two situations: \({DS}_{DD}\left( \rho _{i-1},\theta _j\right) \) and \({DS}_{CD}\left( \rho _{i-1},\theta _j\right) \). If the cumulative distance comes from \(\left( \rho _i,\theta _{j-1}\right) \), since \(\rho _i\) is DV, only the state of \(\theta _{j-1}\) needs to be concerned, which can be separated into two situations: \({DS}_{DD}\left( \rho _i,\theta _{j-1}\right) \) and \({DS}_{DC}\left( \rho _i,\theta _{j-1}\right) \). If the cumulative distance comes from \(\left( \rho _{i-1},\theta _{j-1}\right) \), we only need to consider \(DS\left( \rho _{i-1},\theta _{j-1}\right) \). For the current step \(\left( \rho _i,\theta _j\right) \), the cumulative distance varies depending on the state to which it belongs. For the state (1), (2) and (3), we set the cumulative distance to \(\tau \). When the state is (4), the cumulative distance equals to \(DIS\left( \rho _i,\theta _j\right) \). The solving procedure of \({DS}_{DC}\left( \rho _i,\theta _j\right) \),\({DS}_{CD}\left( \rho _i,\theta _j\right) \), and \({DS}_{CC}\left( \rho _i,\theta _j\right) \) can be received similarly.
We calculate distinguishing distance biunique for the patches at location \(l\left( i\right) \) in two categories \(CL_1\) and \(CL_2\) of data samples and calculate the weighted average to obtain DI. The calculation methodology of DI is systematized into Algorithm 1.
Algorithm 1 provides the calculation procedure of DI. Lines 5-17 describe the calculation method of distinguishing distance. By calculating distinguishing distance pairwise between patches at location \(l\left( i\right) \) from two categories \(CL_1\) and \(CL_2\) and performing weighted averaging, DI can be obtained.
We require the selection of patches to have a strong distinguishing ability on the one hand and to have smaller overlapping parts between patches on the other hand. An enormous overlap can increase redundancy by extracting similar features by the diagnostic network. Therefore, we propose the concept of spatial contact ratio to measure the degree of overlap between two patches.
Definition 7
(Spatial Contact Ratio) Let \(l\left( i\right) =\left( l_x\left( i\right) ,l_y\left( i\right) ,l_z\left( i\right) \right) \) and \(l\left( j\right) =\left( l_x\left( j\right) ,l_y\left( j\right) ,l_z\left( j\right) \right) \) be two voxel locations of MRI, SCR is defined as the proportion of voxels that overlap two patches \(P\left( l\left( i\right) \right) \) and \(P\left( l\left( j\right) \right) \) with the size \(l\times l\times l\), and the calculation equation is as follows:
where ReLU [37] is the linear rectification function.
By calculating SCR, we can obtain the capacity of voxel overlap between any two patches. When selecting atrophic patches, to avoid neural networks learning too many repetitive features, we need to control the repetition of the selected patches.
The DI allows different patches to distinguish diagnostic tasks, and the SCR can effectively measure voxel overlap. Based on DI and SCR, we propose a distinguishing atrophic patch localization (DAPL) algorithm to filter and prune N partitioned patches. We first arrange these N patches in DI order from high to low. Then we start filtering from the first patch, and the patch that enters the selection queue must ensure that the SCR value of all patches in the current queue is more petite than \(\varepsilon \). Otherwise, the patch will not be retained in the selection queue. Traverse all patches, and the top-M patches remaining in the selection queue are the distinguishing atrophic patches we have located. The above calculation process is summarized as Algorithm 2.
The PS set output by Algorithm 2 is the distinguishing atrophic patches selected in the classification task \(CL_1\) vs. \(CL_2\). The DAPL algorithm achieves precise localization and screening of areas that significantly impact diagnosis, which helps to better extract features from neural images and perform high-precision disease diagnosis. We provide a data-driven approach to select different patches personalized for different diagnostic tasks. Furthermore, it is also convenient to adjust the number and size of the selected patches, showing our proposed method’s progressiveness.
3.2.2 Dual-branch AD diagnostic model
We selected the M patches with the most discriminative ability through the above distinguishing atrophic patch localization algorithm. Next, we will construct a neural network to extract features from these patches. Existing patch-based AD diagnostic models often extract features from each patch in isolation and fuse the extracted features. Considering the mutual information between different patches, we propose a dual-branch patch feature extraction and disease diagnosis network. The network consists of two branches, the intra-feature extraction branch (DFB) and the inter-feature extraction branch (IFB). DFB is a typical network branch that extracts the internal features of each patch and fuses them. IFB integrates different patches into an input matrix at the voxel level through the defined Spatial Context Mixing matrix (SCM), and constructs a feature extraction network to extract features from the SCM, obtaining the interactive features between patches. Two branches perform feature-level fusion and connect classifiers to diagnose different diseases. The following will introduce these two branches separately.
(1) The intra-feature extraction branch DFB
In the DFB branch, we built a shallow convolutional neural network for each patch to extract features. The network structure is shown in Fig. 5. The network architecture details of the single a patch feature extraction network in DFB are shown in Table 1. Each subnetwork comprises six convolutional layers, two max-pooling layers, and one global average pooling (GAP) layer. We connected the batch normalization (BN) [19] layer behind each convolutional layer and added the ReLU [37] function. Adding the BN layer can accelerate our model’s learning speed, improve network learning stability, and avoid gradient disappearance. The final GAP layer of each subnetwork can significantly reduce the number of neurons, thereby reducing the number of parameters and avoiding overfitting. The features extracted from each patch are concatenated and implicitly represented through a fully connected (FC) layer, resulting in intra-patch features. M subnetworks are distributed in parallel, and during the training process, rapid training is achieved through parameter sharing.
(2) The inter-feature extraction branch IFB
The DFB branch can effectively extract features within the selected patches, but lacks the extraction of mutual features between various patches. In order to better implement mutual feature extraction, we propose a voxel-wise network structure that extracts mutual features between patches by extracting features from the constructed SCM. Firstly, we give the definition of SCM.
Definition 8
(Spatial Context Mixing matrix) Let \(PS=\left\{ P_1,P_2,\ldots ,P_M\right\} \) be a selected patch set, and the size of each patch is \(l\times l\times l\). We flatten each patch into a vector of \(l^3\), and stack M expanded one-dimensional vectors into a matrix of size \(\sqrt{M}\times \sqrt{M}\times l^3\). This three-dimensional matrix is called the spatial context mixing matrix, abbreviated as SCM.
SCM can be regarded as screening the original MR images, selecting locations with significant differentiation significance for the disease, and reassembling them into an input matrix. The calculation process of SCM is illustrated in Fig. 6. The different patches are flattened and arranged one by one to form a new matrix, so that the feature representation of the potential Mutual information between patches can be well obtained when the matrix is feature extracted, which makes up the defect of the single feature extraction method for each patch.
Next, we perform feature extraction on SCM to obtain mutual features between patches. In the IFB branch, SCM serves as the input of the network. Due to the size of the input matrix, we constructed a deep-level feature extraction network based on CNN. The network contains eight convolutional layers, five max-pooling layers, and two fully connected layers. Similarly, the BN layer and ReLU function are connected after each convolutional layer. The network structure of the IFB branch is shown in Fig. 7. The network architecture details of the IFB are shown in Table 2. We can obtain more profound features correlated to the mutual information between patches through feature extraction and compression of SCM layer by layer.
(3) Dual branch feature fusion
Our proposed dual-branch AD diagnostic network combines the DFB and IFB branches. Specifically, we first concatenate the features extracted from two branches, then design an artificial neural network composed of two fully connected layers, and finally obtain the classification results through softmax. The design of ANN can be used to learn the joint representation of more complex intra-patch and inter-patch features. By continuously optimizing the model through training, it is possible to achieve deep extraction of disease diagnostic features and provide accurate diagnosis of diseases.
4 Experiments and results
4.1 Experimental settings
We conducted experiments on a computer equipped with an NVIDIA GeForce RTX 2080Ti 11G GPU to implemented our proposed model. Our deep learning framework of choice was Pytorch, and we used SGD [45] as an optimizer. Our method was implemented in two diagnostic tasks: AD vs. CN task and sMCI vs. pMCI task. To evaluate the performance of our proposed model, we selected four indicators: accuracy (ACC), sensitivity (SEN), specificity (SPE), and area under the curve (AUC).
4.2 parameter selection
Our proposed DAPL algorithm offers flexibility by allowing for adjustable numbers and sizes of selected patches to adapt to various diagnostic tasks and data types. We conducted experiments to determine the optimal hyperparameters and selected a suitable range based on relevant research and our algorithm. The results, displayed in Fig. 8, show the impact of patch size and number on diagnostic performance. Figure 8(a) demonstrates that a patch size of \(25\times 25\times 25\) resulted in the highest diagnostic accuracy for two tasks, while Fig. 8(b) shows that selecting 36 patches achieved optimal performance due to the construction of the SCM matrix. Too many patches introduce unnecessary data, while too few patches result in incomplete information extraction. Therefore, we recommend a patch size of \(25\times 25\times 25\) and 36 patches for our method. Different hyperparameters can be set for various tasks, allowing for model reuse and good generalization ability.
4.3 Performance of dual-branch AD diagnostic model
This section provides the experimental results of our proposed dual-branch AD diagnostic model. Firstly, we conducted ablation experiments on each branch in the model, and the results are displayed in Table 3 (Note: The bold value indicates the best result in that column.). The table demonstrates that the DBDN model combining the intra-feature extraction branch and inter-feature extraction branch has significantly improved diagnostic performance compared to DFB and IFB alone for diagnosis. Comparing IFB with DFB, we can find that the IFB method, which contains mutual information between patches, achieves better performance, which confirms the advantages of SCM and the necessity of IFB. Further, to more intuitively show the performance of different models in different diagnostic tasks, we give the confusion matrix of each model in two diagnostic tasks, as shown in Fig. 9. The confusion matrix indicates that the DBDN model integrating two branches can divide the test samples into the correct classification as far as possible. It has a lower error than using a single branch alone. The results of the ablation experiment verified that fusing the two branches can improve diagnostic performance on the one hand, and on the other hand, constructing SCM to extract mutual features between patches has better performance than fusing each patch after feature extraction in isolation.
We evaluated our proposed DBDN model against several patch-based approaches, and the results are presented in Table 4 (Note: The bold value indicates the best result in that column.). All methods in the table use MRI data from ADNI. As seen from the table, our accuracy has improved by 0.05 for AD diagnostic tasks compared with the best method in baselines. For MCI diagnostic tasks, our accuracy has improved by 0.109. The two diagnostic tasks are heavy. Almost all the evaluation indicators of our DBDN method are higher than the existing patch-based computer-aided diagnostic methods, indicating that our patch selection, dual-branch feature extraction, and diagnostic model are compelling.
To guarantee the success of our proposed DBDN model, we have contained a graph (Fig. 10) that displays the changes in training and validation loss curves for two tasks. Upon analyzing the chart, it is evident that the model started to converge after about thirty epochs and reached a plateau at around 50 epochs. This suggests that our training process was appropriately executed, without any overfitting or underfitting phenomenon, and the convergence process was accurate. As a result, the saved model is the most ideal diagnostic model.
We provide the ROC (Receiver Operating Characteristic) curves to further illustrate the performance of our model, shown as Fig. 11. The ROC curve is an essential tool for evaluating the effectiveness of classification models, especially in medical diagnostics. It displays the trade-off between the true positive rate and the false positive rate across various threshold settings. Our ROC curves demonstrate the high diagnostic accuracy of our DBDN model in the two diagnostic tasks, with the curves approaching the top-left corner, indicating excellent discrimination between the different classes. The high AUC values associated with the ROC curves underscore the model’s robustness and reliability in accurately diagnosing conditions based on neural imaging data.
To validate the robustness and generalization ability of our model, we conducted 10-fold cross-validation. Specifically, the dataset was divided into ten equal parts. Each part was used as a validation set while the remaining nine parts were used as training sets. This process was repeated ten times to ensure that each data point was utilized for both training and validation. During the 10-fold cross-validation, we recorded ACC, SEN, SPE, and AUC values for each fold. These findings indicate that our model maintains stable performance across various data splits, showing no significant signs of overfitting. we plotted the distribution of these metrics across different folds, as shown in Fig. 12. The box plots illustrate that the performance metrics are closely clustered, with small standard deviations, which further confirms the robustness of the model. In addition, we recorded the training and inference times for the DBDN model. For each diagnostic task, training the model takes approximately 8 hours. Once the model is trained, the average inference time for predicting the condition of each patient is approximately 84 milliseconds.
4.4 Results of distinguishing atrophic patch localization
Here, we showcase the results of our proposed method for identifying atrophic patches. We divided the data into two diagnostic tasks: AD vs. CN and sMCI vs. pMCI. For each task, we selected 36 patches that contained the most atrophic tissues. To better show the locations of the patches identified by our DAPL algorithm, we visualized the top ten cerebral regions corresponding to the selected patches for both tasks. These visualizations are depicted in Figs. 13-14. Typically, the majority of atrophic areas were found in the Limbic system and temporal lobe. The Limbic system plays a crucial role in regulating mood, motivation, decision-making, and learning [14, 61]. If this system malfunctions, it can result in various psychiatric and neurological disorders. The temporal lobe is particularly important for encoding and retrieving memories [6], so damage to this region is a common symptom of early Alzheimer’s disease.
We analyzed the brain tissue information in 36 selected patches, as shown in Fig. 15. Our findings indicate that these patches are concentrated in various brain parts, including the limbic system [14], basal nuclei [59], temporal lobe [6], parietal lobe [14], frontal lobe [13], and occipital lobe [18]. The Hippocampus and surrounding Limbic system play a significant role in diagnosing cognitive diseases. Tissues in the temporal lobe also show good differentiation ability. This suggests that patients with cognitive impairment may experience atrophy in the Hippocampus, temporal lobe, and surrounding tissues. Furthermore, compared to patients with mild cognitive impairment (MCI), those with Alzheimer’s disease (AD) exhibit more extensive encephalatrophy coverage, with atrophic lesions in the basal nuclei, parietal lobe, frontal lobe, and occipital lobe.
5 Discussion
In this section, we initially compare our diagnostic model DBDN with several state-of-the-art previous AD diagnostic models. Following that, we quantitatively analyze the lesion area using the proposed DAPL algorithm and offer diagnostic recommendations for specific locations that require special attention.
5.1 Comparison with previous AD diagnostic model
Some computer-aided diagnostic models are already constructed for AD diagnosis in existing research. We have compiled the diagnostic results of state-of-the-art reported in the existing literature on AD and MCI diagnostic tasks, as shown in Table 5 (Note: The bold value indicates the best result in that column.). These methods include traditional feature engineering methods [20, 31, 60] and deep learning methods. All references use MRI neuroimages from the ADNI dataset. The table demonstrates that the method proposed in this paper performs more satisfactorily than most methods, whether in AD diagnostic tasks or MCI diagnostic tasks. Our model has achieved optimal performance owing to two factors. Our DAPL algorithm is developed to accurately identify disease-related locations for each diagnostic task, making it more precise. Additionally, our dual-branch feature extraction and diagnosis model effectively captures deep representations of features at various patch scales. It is worth noticing that although our method uses different amounts of data compared to other methods in the list, it can achieve optimal diagnostic performance regardless of whether it uses more data for training or models with fewer data. Therefore, it can be verified that our method is effective and advanced.
5.2 Qualitative analysis of pathological changes in patients
In Section 4.3, we used the DAPL algorithm to identify distinguishing patches and visualize their locations. Calculating the DI for different patches, we provide a list of potential lesion areas that should receive special attention in clinical practice, as shown in Table 6. The organizations listed in the table are named according to the AAL3V1 template [46], with the code column indicating the corresponding numbers in the template. The list is sorted based on the locations’ differentiation ability, with most being concentrated in the Limbic system, including the Hippocampus, Parahippocampal gyrus, and Amygdala. Other regions, such as the temporal lobe, basal nuclei, parietal lobe, frontal lobe, and occipital lobe, are also involved. For the two diagnostic tasks, the Hippocampus and surrounding areas are the tissues with the most substantial differentiation ability, consistent with the conclusion recorded in the literature [6]. For the temporal lobe, the Superior temporal gyrus has good discriminative ability in both diagnostic tasks. Compared to the Middle temporal gyrus and Inferior temporal gyrus, some frontal lobe areas exhibit more extraordinary discriminative ability in AD diagnostic tasks.
Meanwhile, the results show that when diagnosing AD, doctors should not only consider the Limbic system and temporal lobe, but also pay attention to the pathological changes in the patient’s occipital lobe. The table can also reflect the differences in the impact of different diseases on the brain. MCI patients often have lesions mainly concentrated in the limbic system and temporal lobe, affecting memory, cognition, thinking, and other functions. In contrast, AD significantly impacts the patient’s brain and a more comprehensive range of design.
In addition to facilitating computer-aided disease diagnosis, our results can provide clinical doctors with potential lesion areas to assist inexperienced doctors in identifying and treating lesions quickly. This impact can be profoundly positive, given the potential implications for early diagnosis and treatment for patients with neurodegenerative diseases.
5.3 Limitations and future Work
The proposed method, while demonstrating significant advancements in the automatic diagnosis of Alzheimer’s disease using neuroimaging data, is not without its limitations. One limitation is the method’s adaptability to varying input sizes of neuroimaging data. The current model is optimized for inputs of uniform dimensions. This lack of flexibility can potentially reduce the model’s effectiveness in diverse clinical settings. Another limitation concerns the separation of the patch selection and feature extraction processes. In our approach, patch selection is conducted independently of feature extraction. The quality and relevance of the selected patches directly impact the subsequent feature extraction and the overall accuracy of the disease diagnosis. If the patch selection algorithm does not perform well, it can hinder the model’s ability to extract meaningful features and make accurate predictions.
Looking ahead, our future work will address these limitations by developing a more integrated and adaptable approach. One primary direction is to design a unified model that combines patch selection and feature extraction into a single, cohesive process. This integrated approach aims to enhance the efficiency and accuracy of the diagnostic process by ensuring that the patches selected are inherently suitable for effective feature extraction, thereby improving the overall performance of the model. Additionally, we plan to develop a model capable of handling neuroimaging inputs of various scales. This model would be designed to process images acquired using different protocols, enhancing its adaptability and robustness in diverse clinical environments. Furthermore, we intend to explore the use of multimodal neuroimaging data to create a comprehensive diagnostic tool. Leveraging information from different imaging modalities can provide a more holistic view of patients’ condition, potentially leading to more accurate and insightful diagnosis.
6 Conclusion
In this study, a novel dual-branch Alzheimer’s disease diagnostic model based on distinguishing atrophic patch localization is developed. First, we propose a distinguishing atrophic patch localization algorithm by defining and calculating the distinguishing index and spatial contact ratio. When calculating DI, considering the difference in intensity between voxels with atrophic lesions and regular voxels, we propose a discontinuity voxel-based dynamic voxel wrapping algorithm to calculate the DI of a patch at a specified location. The DAPL algorithm can accurately select patches that significantly impact disease diagnosis, improving the efficiency of feature extraction in diagnostic networks and providing a reference for clinical doctors in locating patches. Through experiments, we selected 36 patches, which significantly reduced the computational complexity of the model and achieved efficient processing and analysis of image data. Then, we propose a dual-branch AD diagnostic model that includes an intra-feature extraction branch and an inter-feature extraction branch. The DFB branch obtains features within different patches and fuses them. The IFB branch extracts mutual information between patches through the defined spatial context mixing matrix and the constructed network structure. The fusion of the two parts improves the accuracy of diagnosis. Our method demonstrated optimal performance by comparing patch-based methods with various feature extraction strategies. At the same time, based on the analysis of the pathological site of the brain where the patch is located, we provide recommendations for diagnosing potential lesion areas for different diseases.
Structural MRI can display the structural lesions of the brain. In the future, we will consider more types of neuroimaging, such as functional MRI, Positron Emission Computed Tomography, and Diffusion Tensor Imaging, to explore more in-depth lesions related to brain function. At the same time, we are also considering constructing an imaging omics fusion diagnostic model that combines multiple neuroimages.
Data Availability
The dataset used in this study was obtained from Alzheimer’s Disease Neuroimaging Initiative (ADNI). More information regarding ADNI can be obtained from the following link: http://adni.loni.usc.edu/ (accessed on 16 July 2023).
References
Abbas SQ, Chi L, Chen YPP (2023) Transformed domain convolutional neural network for alzheimer’s disease diagnosis using structural mri. Pattern Recognit 133:109031
Alinsaif S, Lang J, Initiative ADN et al (2021) 3d shearlet-based descriptors combined with deep features for the classification of alzheimer’s disease based on mri data. Comput Biol Med 138:104879
Arafa DA, Moustafa HED, Ali HA et al (2024) A deep learning framework for early diagnosis of alzheimer’s disease on mri images. Multimed Tools Appl 83(2):3767–3799
Ashburner J, Friston KJ (2000) Voxel-based morphometry-the methods. Neuroimage 11(6):805–821
Blennow K, de Leon MJ, Zetterberg H (2006) Alzheimer’s disease. The Lancet 368(9533):387–403
Chen J, Duan X, Shu H et al (2016) Differential contributions of subregions of medial temporal lobe to memory system in amnestic mild cognitive impairment: insights from fmri study. Sci Rep 6(1):1–14
Chen Y, Xia Y (2021) Iterative sparse and deep learning for accurate diagnosis of alzheimer’s disease. Pattern Recognit 116:107944
Cobbinah BM, Sorg C, Yang Q et al (2022) Reducing variations in multi-center alzheimer’s disease classification with convolutional adversarial autoencoder. Med Image Anal 82:102585
Cui R, Liu M (2018) Hippocampus analysis by combination of 3-d densenet and shapes for alzheimer’s disease diagnosis. IEEE J Biomed Health Inf 23(5):2099–2107
Cui R, Liu M, Initiative ADN et al (2019) Rnn-based longitudinal analysis for diagnosis of alzheimer’s disease. Comput Med Imaging Graph 73:1–10
Feng C, Elazab A, Yang P et al (2019) Deep learning framework for alzheimer’s disease diagnosis via 3d-cnn and fsbi-lstm. IEEE Access 7:63605–63618
Fischl B (2012) Freesurfer. Neuroimage 62(2):774–781
Fonseca LM, Yokomizo JE, Bottino CM et al (2016) Frontal lobe degeneration in adults with down syndrome and alzheimer’s disease: a review. Dement Geriatr Cogn Disord 41(3–4):123–136
Foundas AL, Leonard CM, Mahoney SM et al (1997) Atrophy of the hippocampus, parietal cortex, and insula in alzheimer’s disease: a volumetric magnetic resonance imaging study. Cogn Behav Neurol 10(2):81–89
Gao X, Shi F, Shen D et al (2021) Task-induced pyramid and attention gan for multimodal brain image imputation and classification in alzheimer’s disease. IEEE J Biomed Health Inf 26(1):36–43
Gauthier S, Reisberg B, Zaudig M et al (2006) Mild cognitive impairment. The Lancet 367(9518):1262–1270
Gorgolewski K, Burns CD, Madison C, et al (2011) Nipype: a flexible, lightweight and extensible neuroimaging data processing framework in python. Frontiers in Neuroinformatics p 13
Holroyd S, Shepherd ML, Downs JH III (2000) Occipital atrophy is associated with visual hallucinations in alzheimer’s disease. J Neuropsychiatry Clin Neurosci 12(1):25–28
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, pmlr, pp 448–456
Janoušová E, Vounou M, Wolz R, et al (2012) Biomarker discovery for sparse classification of brain images in alzheimer’s disease. Annals of the BMVA (2)
Jenkinson M, Bannister P, Brady M et al (2002) Improved optimization for the robust and accurate linear registration and motion correction of brain images. Neuroimage 17(2):825–841
Jie B, Zhang D, Cheng B et al (2015) Manifold regularized multitask feature learning for multimodality disease classification. Hum Brain Mapp 36(2):489–507
Karwath A, Hubrich M, Kramer S, et al (2017) Convolutional neural networks for the identification of regions of interest in pet scans: a study of representation learning for diagnosing alzheimer’s disease. In: Artificial intelligence in medicine: 16th conference on artificial intelligence in medicine, AIME 2017, Vienna, Austria, June 21-24, 2017, Proceedings 16, Springer, pp 316–321
Kerwin D, Abdelnour C, Caramelli P et al (2022) Alzheimer’s disease diagnosis and management: perspectives from around the world. Alzheimer’s & Dementia: Diagnosis, Assessment & Disease Monitoring 14(1):e12334
Korolev S, Safiullin A, Belyaev M, et al (2017) Residual and plain convolutional neural networks for 3d brain mri classification. In: 2017 IEEE 14th international symposium on biomedical imaging (ISBI 2017), IEEE, pp 835–838
Li F, Tran L, Thung KH et al (2015) A robust deep model for improved classification of ad/mci patients. IEEE J Biomed Health Inf 19(5):1610–1616
Li H, Tan Y, Miao J et al (2023) Attention-based and micro designed efficientnetb2 for diagnosis of alzheimer’s disease. Biomed Signal Process Control 82:104571
Lian C, Liu M, Zhang J et al (2018) Hierarchical fully convolutional network for joint atrophy localization and alzheimer’s disease diagnosis using structural mri. IEEE Trans Pattern Anal Mach Intell 42(4):880–893
Lin W, Tong T, Gao Q et al (2018) Convolutional neural networks-based mri image analysis for the alzheimer’s disease prediction from mild cognitive impairment. Front Neurosci 12:777
Lin W, Lin W, Chen G et al (2021) Bidirectional mapping of brain mri and pet with 3d reversible gan for the diagnosis of alzheimer’s disease. Front Neurosci 15:646013
Liu M, Zhang D, Shen D (2016) Relationship induced multi-template learning for diagnosis of alzheimer’s disease and mild cognitive impairment. IEEE Trans Med Imaging 35(6):1463–1474
Liu M, Cheng D, Wang K et al (2018a) Multi-modality cascaded convolutional neural networks for alzheimer’s disease diagnosis. Neuroinformatics 16:295–308
Liu M, Cheng D, Yan W et al (2018b) Classification of alzheimer’s disease by combination of convolutional and recurrent neural networks using fdg-pet images. Front Neuroinform 12:35
Liu M, Zhang J, Adeli E et al (2018c) Joint classification and regression via deep multi-task multi-channel learning for alzheimer’s disease diagnosis. IEEE Trans Biomed Eng 66(5):1195–1206
Liu M, Li F, Yan H et al (2020) A multi-model deep convolutional neural network for automatic hippocampus segmentation and classification in alzheimer’s disease. Neuroimage 208:116459
Mueller SG, Weiner MW, Thal LJ et al (2005) The alzheimer’s disease neuroimaging initiative. Neuroimaging Clin 15(4):869–877
Nair V, Hinton GE (2010) Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th international conference on machine learning (ICML-10), pp 807–814
Ning Z, Xiao Q, Feng Q et al (2021) Relation-induced multi-modal shared representation learning for alzheimer’s disease diagnosis. IEEE Trans Med Imaging 40(6):1632–1645
O’Dwyer L, Lamberton F, Bokde AL et al (2012) Using support vector machines with multiple indices of diffusion for automated classification of mild cognitive impairment. PloS one 7(2):e32441
Oltu B, Akşahin MF, Kibaroğlu S (2021) A novel electroencephalography based approach for alzheimer’s disease and mild cognitive impairment detection. Biomed Signal Process Control 63:102223
Ortiz A, Munilla J, Gorriz JM et al (2016) Ensembles of deep learning architectures for the early diagnosis of the alzheimer’s disease. Int J Neural Syst 26(07):1650025
Pan Y, Liu M, Lian C et al (2018) Synthesizing missing pet from mri with cycle-consistent generative adversarial networks for alzheimer’s disease diagnosis. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part III 11, Springer, pp 455–463
Pini L, Pievani M, Bocchetta M et al (2016) Brain atrophy in alzheimer’s disease and aging. Ageing Res Rev 30:25–48
Richhariya B, Tanveer M, Rashid AH et al (2020) Diagnosis of alzheimer’s disease using universum support vector machine based recursive feature elimination (usvm-rfe). Biomed Signal Process Control 59:101903
Robbins H, Monro S (1951) A stochastic approximation method. The Annals of Mathematical Statistics 22(3)
Rolls ET, Huang CC, Lin CP et al (2020) Automated anatomical labelling atlas 3. Neuroimage 206:116189
Scheltens P, De Strooper B, Kivipelto M et al (2021) Alzheimer’s disease. The Lancet 397(10284):1577–1590
Sharma R, Goel T, Tanveer M, et al (2023) Deep learning based diagnosis and prognosis of alzheimer’s disease: a comprehensive review. IEEE Transactions on Cognitive and Developmental Systems
Sheng J, Wang B, Zhang Q et al (2019) A novel joint hcpmmp method for automatically classifying alzheimer’s and different stage mci patients. Behav Brain Res 365:210–221
Suk HI, Lee SW, Shen D et al (2014a) Hierarchical feature representation and multimodal fusion with deep learning for ad/mci diagnosis. NeuroImage 101:569–582
Suk HI, Lee SW, Shen D et al (2014b) Hierarchical feature representation and multimodal fusion with deep learning for ad/mci diagnosis. NeuroImage 101:569–582
Suk HI, Wee CY, Lee SW et al (2016) State-space model with deep learning for functional dynamics estimation in resting-state fmri. NeuroImage 129:292–307
Takao H, Amemiya S, Abe O et al (2021) Reliability of changes in brain volume determined by longitudinal voxel-based morphometry. J Magn Reson Imaging 54(2):609–616
Tu Y, Lin S, Qiao J et al (2022) Alzheimer’s disease diagnosis via multimodal feature fusion. Comput Biol Med 148:105901
Tu Y, Lin S, Qiao J et al (2024) Multimodal fusion diagnosis of alzheimer’s disease based on fdg-pet generation. Biomed Signal Process Control 89:105709
Tzourio-Mazoyer N, Landeau B, Papathanassiou D et al (2002) Automated anatomical labeling of activations in spm using a macroscopic anatomical parcellation of the mni mri single-subject brain. Neuroimage 15(1):273–289
Wang H, Shen Y, Wang S et al (2019) Ensemble of 3d densely connected convolutional network for diagnosis of mild cognitive impairment and alzheimer’s disease. Neurocomputing 333:145–156
Wang SH, Phillips P, Sui Y et al (2018) Classification of alzheimer’s disease based on eight-layer convolutional neural network with leaky rectified linear unit and max pooling. J Med Syst 42:1–11
Xiong Y, Ye C, Chen Y et al (2022) Altered functional connectivity of basal ganglia in mild cognitive impairment and alzheimer’s disease. Brain Sci 12(11):1555
Xu L, Yao Z, Li J et al (2019) Sparse feature learning with label information for alzheimer’s disease classification based on magnetic resonance imaging. IEEE Access 7:26157–26167
Yang H, Xu H, Li Q, et al (2019) Study of brain morphology change in alzheimer’s disease and amnestic mild cognitive impairment compared with normal controls. Gen Psychiatr 32(2)
Zhang F, Tian S, Chen S et al (2019) Voxel-based morphometry: improving the diagnosis of alzheimer’s disease based on an extreme learning machine method from the adni cohort. Neuroscience 414:273–279
Zhang P, Lin S, Qiao J et al (2021) Diagnosis of alzheimer’s disease with ensemble learning classifier and 3d convolutional neural network. Sensors 21(22):7634
Zhang Y, Teng Q, Liu Y et al (2022) Diagnosis of alzheimer’s disease based on regional attention with smri gray matter slices. J Neurosci Methods 365:109376
Zhu T, Cao C, Wang Z et al (2020) Anatomical landmarks and dag network learning for alzheimer’s disease diagnosis. IEEE Access 8:206063–206073
Zhu W, Sun L, Huang J et al (2021) Dual attention multi-instance deep learning for alzheimer’s disease diagnosis with structural mri. IEEE Trans Med Imaging 40(9):2354–2366
Acknowledgements
This research was funded by Science and Technology Development Project of Liaoning Province of China under Grant number 2021JH6/10500127 and Science Research Project of Liaoning Department of Education of China under Grant number LJKZ0008.
Author information
Authors and Affiliations
Contributions
Yue Tu: Conceptualization, Methodology, Software, Validation, Formal Analysis, Writing - Original Draft, Visualization. Shukuan Lin: Resources, Supervision, Writing - Review & Editing. Jianzhong Qiao: Resources, Supervision. Kuankuan Hao: Writing - Review & Editing, Investigation, Validation, Response to Reviewers. Yilin Zhuang: Data Curation, Software. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Competing of Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Ethical and Informed Consent for Data Used
A benchmark dataset, Alzheimer’s Disease Neuroimaging Initiative(ADNI), which was used in our work, has obtained the informed consent from the participants. We have obtained permission to use this dataset. More information can be found in the following link: http://adni.loni.usc.edu/study-design/ (accessed on 16 July 2023).
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Tu, Y., Lin, S., Qiao, J. et al. A novel dual-branch Alzheimer’s disease diagnostic model based on distinguishing atrophic patch localization. Appl Intell 54, 9067–9087 (2024). https://doi.org/10.1007/s10489-024-05663-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-024-05663-z