Machine learning methods for predicting progression from mild cognitive impairment to Alzheimer’s disease dementia: a systematic review

Background An increase in lifespan in our society is a double-edged sword that entails a growing number of patients with neurocognitive disorders, Alzheimer’s disease being the most prevalent. Advances in medical imaging and computational power enable new methods for the early detection of neurocognitive disorders with the goal of preventing or reducing cognitive decline. Computer-aided image analysis and early detection of changes in cognition is a promising approach for patients with mild cognitive impairment, sometimes a prodromal stage of Alzheimer’s disease dementia. Methods We conducted a systematic review following PRISMA guidelines of studies where machine learning was applied to neuroimaging data in order to predict whether patients with mild cognitive impairment might develop Alzheimer’s disease dementia or remain stable. After removing duplicates, we screened 452 studies and selected 116 for qualitative analysis. Results Most studies used magnetic resonance image (MRI) and positron emission tomography (PET) data but also magnetoencephalography. The datasets were mainly extracted from the Alzheimer’s disease neuroimaging initiative (ADNI) database with some exceptions. Regarding the algorithms used, the most common was support vector machine with a mean accuracy of 75.4%, but convolutional neural networks achieved a higher mean accuracy of 78.5%. Studies combining MRI and PET achieved overall better classification accuracy than studies that only used one neuroimaging technique. In general, the more complex models such as those based on deep learning, combined with multimodal and multidimensional data (neuroimaging, clinical, cognitive, genetic, and behavioral) achieved the best performance. Conclusions Although the performance of the different methods still has room for improvement, the results are promising and this methodology has a great potential as a support tool for clinicians and healthcare professionals.


Background
The increase in lifespan experienced in Western societies has largely been driven by medical and technological advances [1]; however, this improvement has resulted in an increasing number of people diagnosed with neurocognitive disorders. In 2010, dementia was associated with $604 billion of healthcare expenses in the USA [2]. Every year, ten million new cases of dementia are registered, and by 2050, it is estimated that 135 million people will have some degree of dementia [3]. Age is the main risk factor for dementia; the prevalence is 1-2% at the age of 65 but reaches 30% at the age of 85. From all neurodegenerative disorders, about 60-90% are characterized as Alzheimer's disease (AD) dementia subtype (depending on the diagnostic criteria used) [4], for which there is yet no cure.
Patients are typically diagnosed with AD when the symptoms of a cognitive decline have already manifested, i.e., when dementia has appeared. In such cases, the diagnosis is determined too late, failing to implement preventive protocols to reduce cognitive decline. Pharmacological and non-pharmacological treatments have proven to be effective in reducing cognitive and behavioral symptoms in the early stages of the disease [5]. In light of these treatments, recent studies have focused on detecting patients with cognitive impairment that have not reached dementia in order to delay or prevent its development. The last edition of the Diagnostic and Statistical Manual of Mental Disorders (DSM-5) includes a specific category for this type of patients called a mild neurocognitive disorder, analogous to the mild cognitive impairment (MCI) whose main characteristic is having minor memory impairment [4] (throughout this review, MCI will be used instead of mild neurocognitive disorder as it is more frequent in the scientific literature). MCI can, in some cases, be a prodromal stage of dementia, especially for AD [6]. It is worth mentioning at this point that AD should be considered as a continuum, where patients with MCI that will eventually progress to AD dementia already have AD, but the cognitive symptoms have not yet fully manifested. For this reason, it is important to differentiate between those MCI patients that will progress to AD dementia and those who will remain stable.
In late stages, when dementia symptoms have already appeared, AD is easier to confirm with neuroimaging techniques and cerebrospinal fluid evaluations for the presence of neurofibrillary clews, beta-amyloid and tau proteins [7], and temporal cortex atrophy [4]. Nevertheless, in the early stages, although biomarkers may be present in magnetic resonance image (MRI) and/or positron emission tomography (PET) results, the detection of MCI to AD dementia progression remains challenging in the clinical practice [8,9]. To overcome this challenge, the scientific community now has access to thousands of neuroimaging longitudinal datasets from healthy, MCI, and AD subjects along with other variables (i.e., demographic, genetic, and cognitive measurements, etc.) stored in public databases such as the Alzheimer disease neuroimaging initiative (ADNI) (http://adni.loni.usc.edu). These datasets can be compared and analyzed to perform classification and automatic detection of AD and MCI progression [10,11] using newly developed computeraided techniques like machine learning (ML) algorithms. Then, these new tools could be transferred to the clinic to assist in the early diagnosis and prognosis.
The ML paradigm consists of training an algorithm with a dataset; in this case, neuroimaging results together with other clinical variables, to extract common factors that help classify subjects according to a variable of interest. In the case of an early diagnosis of AD and distinction from a stable MCI condition, for example, the algorithm learns to classify the data according to the specific diagnosis and extracts which factors have been the most relevant for the differentiation between the groups. Subsequently, the trained algorithm can be used to classify a specific individual for which we do not know the diagnosis and thus manage to assist in the therapeutic approach [12][13][14]. This technique can be applied to any disease that occurs with morphological changes or with characteristic neural patterns. See Arbabshirani, Plis, Sui, & Calhoun [15] for a review of the same objective and methodology but applied to autism, attention deficit disorder, and schizophrenia.
Recent work has demonstrated that ML algorithms are able to classify images from AD, MCI, and healthy participants with very high accuracy levels [16,17]. Although such classification has provided valuable information about AD biomarkers, for this technology to have a more substantial clinical impact by empowering a clinician to administer a customized treatment protocol, it is necessary to determine and predict whether a MCI patient will progress to AD dementia or remain stable. The goal of this systematic review is to analyze the existing classification methods based in ML algorithms applied to neuroimaging data in combination with other variables for predicting MCI to AD dementia progression.

Methods
To perform this systematic review, we followed the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) guidelines [18,19]. A systematic search was done to find studies that included ML methods to predict MCI to AD dementia progression using neuroimaging techniques. Progression to AD dementia from MCI is established when, during a follow-up period (3 years for ADNI and 1 year for AddNeuroMed databases), a patient that was initially classified as MCI, is diagnosed with Alzheimer (that is a "progressive MCI" or pMCI) based on clinical criteria (MMSE and CDR scales, and NINCDS/ADRDA criteria for probable AD dementia [20,21]). Patients are considered "stable MCI" (sMCI), when they were diagnosed as MCI at baseline and the diagnosis remained as MCI during the follow-up.
Only articles written in English and published between January 2010 and May 2021 (included) were selected. Articles published before 2010 were not included because the technological (e.g., computational power, graphical processing units) and methodological (e.g., ML and deep learning algorithm development) gap between those studies and the current standards make them hardly comparable. In fact, even the comparison between articles published in the early vs. the late 2010s presents methodological gaps. These differences are not only due to better methods but mainly to technological advances that were not possible before and early 2010s, and the growth of the ADNI database.
After removing duplicates, the eligibility criteria were applied by two independent reviewers (SG and RVS) to select only the articles that included (1) prediction of MCI to AD dementia conversion, (2) use of neuroimaging data, (3) classification methods based on ML algorithms, and (4) accuracy results.
Once the selection of studies was concluded, the following data was extracted for each study: (1) first author and year of publication, (2) groups, (3) sample size and mean age, (4) database, (5) neuroimaging technique used and variables selected, (6) classification method, (7) validation method, (8) accuracy achieved, and (9) area under the ROC curve.
We also analyzed the risk of bias of the selected studies. The aspects considered in the analysis of bias were based on the Cochrane guidelines for systematic reviews [22], but the exact criteria were adapted by taking into account the particular methodology and goals of the studies, focused on creating and validating a classification model in large datasets. The criteria used are detailed in Table 1.
We also performed an interpretability analysis based on the framework proposed by Kohoutová et al. [23]. These authors have developed three levels of assessment for the interpretability of ML models in neuroimaging based on the model itself, the feature selection and characteristics, and biological factors; also, each level has several sublevels. Model-level assessment consists of evaluating the model as a whole and testing it in different contexts and conditions. The sublevels include sensitivity and specificity, generalizability, behavioral analysis, representational analysis, and analysis of confounds. Feature-level assessment consists of evaluating the significance of individual features used in prediction, including stability, feature importance, and visualization. Finally, the biology-level assessment is a validation of the model based on its neurobiological plausibility and it has two sublevels: literature (relationship of the model with previous literature) and invasive studies (the possibility of using more invasive methods).
We assessed whether the studies included in the review complied with each of the sublevels, but we did not include behavioral analysis, representational analysis, and invasive studies sub-levels. Behavior analysis sublevel was not considered because the only "behavior" of the model is to classify subjects, and the behavior is measured as accuracy, which is included in the sensitivity and specificity sublevels. Representational analysis compares the model with other models, other brain regions, or other experimental settings; in our review, the main goal of almost all studies was to find neural patterns that predict AD dementia, and therefore, it is common to use the whole brain as a feature. Also, there is only one experimental setting aimed to find maximum classification accuracy so it cannot be compared to similar experiments in the same study, only with similar literature (which represents another sub-level). Finally, the invasive studies sub-level is not applicable because the long-term objective of these investigations is to find a non-invasive method of predicting AD dementia as soon as possible.

Results
As shown in Fig. 1, the workflow followed for the article selection included the four phases (identification, screening, eligibility, and inclusion) proposed by the PRISMA guidelines [18,19]. The 452 articles remaining after eliminating duplicates were screened, and after applying the exclusion criteria, 117 articles were selected for the review. The risk of bias analysis is shown in Fig. 2 and Table 2. The overall risk of bias of all the studies was considered low. From the 117 articles selected at the eligibility stage, only one study [65] was not included in the qualitative analysis because of the high risk of bias. The sample size in this study was seven subjects, and it did not include any validation method. Therefore, the final number of studies included in the qualitative analysis was 116.
The studies selected for the qualitative analysis are presented in Table 3  MRI was the most common kind of neuroimaging used (in 76 out of 116 studies), followed by PET (11 studies), 26 studies included data from both techniques (MRI and PET), two studies used magnetoencephalography (MEG) data, and one study used MRI and MEG data.
Regarding the source of the datasets, 107 out of 116 studies used the ADNI database in any of its versions (ADNI-1, 2, 3, or GO) to obtain samples of healthy, MCI, and AD subjects. Of the remaining eight studies, three used data from AddNeuroMed (h t t p s : / / consortiapedia.fastercures.org/consortia/anm/) database, one used the Australian Imaging, Biomarker & Lifestyle Flagship Study of Ageing (AIBL) database, and four collected their own data. Li et al. [108] used both ADNI and AIBL.
Although almost all studies used the same database, the cohorts selected varied across them. Most articles (66 out of 116 studies) divided their participants into four groups: healthy controls, stable MCI patients (sMCI), progressive MCI patients (pMCI), and AD patients. 21 articles selected three cohorts of MCI, AD, and healthy subjects, although in order to predict the progression to AD dementia, they had to distinguish between pMCI and sMCI patients. The remaining 29 studies used different groups of participants: 21 only sMCI and pMCI, six had healthy controls, and MCI with two of them separating sMCI and pMCI. Wee et al. [113] differentiated between early and late MCI, and Li et al. [106] did not specify the cohorts selected.
The sample size also varied across studies. Wee et al. [37] has the smallest sample with 27 subjects, and Bae et al. [130] has the largest sample with 3940 subjects; the mean sample size was 546 participants. The sample size follows an ascendant trend across years, which may be attributed to the increased data availability in the ADNI database. The mean age ranged from 56 to 79 years old.
Although 30 studies did not include the mean age of the sample, they used an ADNI database, and therefore, the age range might be similar to the rest of the studies. The variations in age between studies may be due to differences in participant selection and the moment when the study was conducted (since the ADNI database has been incorporating more data over the years).
As for feature selection, the most common were whole-brain volumes, selected in 70 articles, and intensity measurements of glucose metabolism, selected in 31 PET studies, also 16 studies included genetic features (APOE4 genotype). Other selected features were neuropsychological test results (18 out of 116 studies) and demographic variables such as age (15 out of 116 studies). 42 studies only used one type of features such as 3D MRI data or whole-brain gray matter volumes, and 74 studies selected two or more different types of features. As for the algorithm results, the most useful areas to discriminate between AD patients and healthy subjects or sMCI, were mainly located in the temporal, parietal, and frontal lobes. In particular, the most relevant regions were the hippocampus, amygdala, entorhinal cortex, precuneus, cingulate gyrus, and rostral and caudal areas of the medial frontal lobe.
Regarding the ML methods used to classify the patients and detect probable MCI to AD dementia progression, the most popular were those based on support vector machine (SVM). SVM was used in 60 out of the 116 studies; this method is a supervised ML algorithm that has demonstrated its utility in neuroimaging-based applications, especially in the classification of future clinical outcomes [139]. SVM takes every measurement from every subject as a single point in a multidimensional space, with the number of dimensions being the total number of features of that particular dataset (for example, 93 gray matter volumes from regions of interest). The algorithm then finds the maximal margin     separating hyperplane that optimally differentiates groups of data points representing different classes (e.g., pMCI vs. sMCI, or AD vs. HC). The data instances closest to the group boundaries are the support vectors and are, by definition, the ones that determine the position of the hyperplane. The mapping into a higher dimensional space is done by a kernel function, usually polynomial or Gaussian [26]. The SVM algorithm is trained with labeled data (indicating whether the data belongs to a healthy person, sMCI, pMCI, or AD dementia patient, for example) to generate this multidimensional space.
Once the model has been trained, we can introduce a new subject with MCI and it will be classified in the multidimensional space into the boundaries of one of the previously defined groups (i.e., sMCI, AD dementia).
For example, if the new MCI patient is classified as belonging to the AD or pMCI group, we can infer that this subject is more likely to develop a future AD dementia due to being more similar to subjects in that group. The different groups for classification will depend on the specific methodology of each study.
The combination of SVM with other methods allows to improve feature selection and to avoid overfitting of data, and this will facilitate the generalization of the model (i.e., achieving high accuracy when applied to different datasets). For example, Thung et al. [71] used SVM with multiple kernels (linear, Gaussian, and polynomial) after feature selection with least squares and logistic elastic net regressions and also matrix completion with label-guided low-rank matrix completion method.  Note. This table shows the results of the bias analysis performed based on Higgins et al. [22] with the punctuations specified in Table 1           On the other hand, Toussaint et al. [36] used non-linear SVM with Gaussian Radial Basis Function kernel but only after a two-sample t test and a spatial independent component analysis, performed for the detection of glucose metabolism and characteristic region patterns of AD patients. Other classification methods used were random forest or neural networks that can have different architectures, but the most commonly used for image classification tasks where convolutional neural networks, applied in 16 studies.
As for validation methods, cross-validation was selected in 72 studies, with different numbers of folds and/ or iterations. Cross-validation consists in dividing the sample into two sets, one to train the algorithm (training set) and another one for validation (testing set). This partition can be done several times by changing the train/test split of the data, and the accuracy of each iteration can be averaged to obtain a more robust quantification of the model performance instead of just validating the model on one test sample. Another validation method is the leave-one-out cross-validation, selected in 13 studies. In this case, the model is trained with all the data except for one data point, then it tries to classify the data point left out and does the same with the rest of the sample in subsequent iterations. The train/test method was selected in 27 studies with different percentages of data partitions. Finally, three studies validated the model on an independent test set and Ardekani et al. [64] used "out of bag" as a validation method.
The results of ML classification algorithms can be assessed based on their sensitivity (percentage of correctly detected pMCI patients or true positive ratio) and specificity (percentage of healthy or sMCI subjects correctly identified or true negative ratio), or accuracy (percentage of correctly classified subjects). By changing the decision threshold of the classifier, we can compensate the ratio between true positive/true negative and generate a graphic representation of that ratio, or what is known as the receiver operating characteristic (ROC) curve [140]. The calculation of the area under the ROC curve (AUC ROC) represents a good quantitative index to compare the classification models, since it indicates the ability of the model to predict both the presence and non-presence of disease, or in this case, the progression or lack of progression from MCI to AD dementia [141]. An AUC ROC of one implies a perfect classification of every subject in the sample. The maximum accuracies achieved by every study in the prediction of AD dementia progression from MCI patients or the accuracy of the method in discriminating between a progressive/stable MCI are shown in the "Results" column of Table 3; the AUC coefficient is presented when available.
The mean accuracy of studies that used MRI as neuroimaging technique was 74.5% and 76.9% for studies that selected PET scans. The combination of both techniques achieved even better results with a mean accuracy of 77.5%. AUC ROC results follow a similar pattern with a mean AUC of 0.79, 0.80, and 0.80 for MRI, PET, and MRI+PET, respectively. López et al. [69] used MRI and MEG and achieved an accuracy of 100 and an AUC of 0.97 but with a sample of 33 subjects and finally Pusil et al. [110] (with 54 subjects) and Xu et al. [126] (with 129 subjects) used MEG with accuracies of 100% and 87%, respectively.
Considering the two most popular classification methods (SVM and CNN), the mean accuracies were 75.4% for SVM and 78.5% for CNN. The best results with the SVM algorithm were obtained again by Pusil et al. [110] with 100% accuracy, but using a small sample of 54 that makes the model hardly generalizable. In studies with bigger samples, Guerrero et al. [49] had the highest accuracy results (97.2% with 511 subjects) followed by Lin et al. [122] (97.3% with 164 subjects). Finally, although 43 studies did not report AUC ROC values, the highest reported was 0.98 in Lin et al. [122] followed by 0.98 in Hojjati et al. [85] with 80 subjects. The lowest accuracy reported was found in the study by Plant et al. [24] (50%), who included a small sample size [63] and is the oldest study of all the analyzed, making the comparison between studies difficult. The second-lowest accuracy was achieved by Sorensen et al. [93] (55% with 400 subjects); in this case, the study was part of a data science competition (Kaggle) where data was already preprocessed for all participants and could not be manipulated by the authors. The lowest AUC ROC reported was found in Liu et al. [45] (0.53 and 414 subjects) using SVM and in Pan et al. [123] (0.59 with 787 subjects) with CNN.
Finally, the interpretability analysis is shown in Table 4. Most of the studies presented results of specificity and sensitivity (113 out of 116), about half of them (60 out of 116) made an analysis of confounds, but only Lebedev et al. [50], Li et al. [106], and Syaifullah et al. [135] complied with the generalizability sublevel, testing their model in different datasets. All of the studies performed a stability measurement of their model, and only 25 did not specify which features were the most important for the classification task. On the other hand, less than half of the studies (51 out of 116) presented their results along with some kind of visualization of the most relevant brain areas for the prediction of MCI conversion. Finally, the comparison of the results with the existing literature was done in all of the studies except for four of them.

Discussion
In this systematic review, we analyzed 116 studies, conducted over the last 10 years, which used neuroimaging data to predict conversion to AD dementia from MCI using ML algorithms. The complexity of neuroimaging results and the amplitude of the deterioration and symptoms present in multiple areas and functions in AD, make its detection very complex in patients with MCI by simply visualizing a single patient neuroimaging data. Nevertheless, using the publicly available data collected over the last decades, together with the newly developed ML algorithms, researchers can not only distinguish the brains of AD patients and healthy people with high accuracy, but also predict MCI patient's disease progression (i.e., whether a MCI patient will progress to AD dementia or remain stable in the future). This information is highly valuable for clinicians in order to achieve a more accurate prognosis and therefore set treatment plans that can slow down the development of the disease and prevent higher degrees of cognitive impairment. The 116 studies analyzed reached different levels of accuracy using classification methods based on ML algorithms. Only 24 studies focused exclusively on predicting MCI progression, most studies also tried to find the main differences between healthy controls and AD patients. The specific search for AD biomarkers is much more abundant in the literature than predicting progression from a MCI or even from healthy subjects [142]. In any case, in the studies that carried out both tasks and in studies that focused on the prediction of AD progression, the distinction between controls and AD was always more accurate than the distinction between pMCI vs. sMCI, showing the difficulty of finding biomarkers before functional impairment/dementia appear.
One of the main challenges of this review was to compare studies with highly variable methodologies including different samples, preprocessing techniques, types of neuroimaging data, and also different classification and validation methods. Still, studies that achieved higher levels of accuracy have in common the use of multimodal and multidimensional data combined with increasingly complex classification methods. Easy-toimplement algorithms, such as those based on SVM, are leaving room for more complex algorithms based on deep learning paradigms such as neural networks, capable of identifying dementia-associated subtle changes of brain morphology in a way able to increase the number of correctly classified subjects. All methods seemed to benefit from the inclusion of demographic variables and cognitive measurements, and even genetic variables if these were available. Nevertheless, in order for these techniques to be able to help clinicians in their everyday practice, a balance is needed between the most advanced data and algorithms that achieve the higher performance, and the data and methods that might be available in the clinical practice. In this sense, future studies might need to focus more on achieving high performance using large datasets with more essential (and easily obtainable) data such as structural MRI, demographic, and cognitive results.
Regarding the sample, most studies use the publicly available data from the ADNI database. This database is still incorporating new data and the most recent studies even use ADNI-2 and ADNI-3 [113,117]. The main problem of the studies performed 10 years ago is their smaller sample size. Furthermore, even if two studies report similar accuracies, a study with a bigger sample size will have results that are more generalizable. For example, Plant et al. [24] and Popuri et al. [92] obtained similar accuracies of 75% and 79% of correct classifications, respectively, but Popuri et al. [92] used a sample of about 30 times larger. Upon review of the recent developments in the field, it is apparent that as these patient populations continue to age and their disease progression matures and is clinically diagnosed, the inclusion of their ongoing neuroimaging data will enrich these public databases, thereby enabling improved validation of classification methods. Nevertheless, it should be noted that a follow-up of 2 or 3 years might not be long enough to detect progression to AD dementia. Therefore, subjects considered as stable MCI or even as healthy subjects might, in fact, progress to AD dementia in the long term. This problem will always be present with the inclusion of new cases in the ADNI database, but the follow-ups recorded will be increasingly longlasting, thus being more useful. On the other hand, in order for these methods to be clinically useful, the models have to be tested, not only in big samples, but also in more variable and diverse groups of people, other than the ADNI sample. This approach is highlighted by Lebedev et al. [50], for example, who applied the model to the ADNI and then to the AddNeuroMed cohort, achieving similar performance and accuracy results in both, making it more robust for future clinical implementation.
Another interesting result from the inclusion of neuroimaging data in ML algorithms is the possibility of finding out which brain regions are more relevant to predict the conversion from MCI to AD dementia. In this case, the highlighted regions as informed by the algorithms (hippocampus, amygdala, entorhinal cortex, precuneus, cingulate gyrus, and medial frontal lobe) have been widely validated by the scientific literature as relevant in the progression to AD [8,33,34,[143][144][145][146][147]. This coincidence between the literature and the algorithm results supports the notion that the classification methods can detect differences between groups based on relevant neuroimaging features.
In terms of accuracy, although the algorithms are useful and able to discriminate the brain characteristics of AD, the performance of the algorithms are far from being specific enough to leave complete diagnosis in the hands of automated methods, so the judgment of a clinical professional will remain crucial in the near future. Also, clinical criteria can achieve sometimes similar results in terms of predicting the conversion from MCI to AD dementia 1 year before its onset [148]. Nevertheless, computer-aided diagnosis, when implemented in the clinical practice, will offer a faster, easier to perform, and earlier detection way of predicting the potential progression to AD dementia in MCI patients. Therefore, the automated methods discussed above present a low-cost approach that can be useful as a first approximation, a method to discriminate ambiguous cases, and a support tool for large datasets.
Clinical research is moving towards a broader and more open context where professionals from very different disciplines might be interested in these types of studies. As such, it is important to present the results from complex neuroimaging classification studies as clearly as possible. The framework to interpret ML models provided by Kohoutová et al. [23] is a helpful starting point for this purpose. Most or all of the studies reviewed here included information about the specificity and sensitivity (model level), the stability of the models, and the most important features selected (feature level), along with a comparison with the previous literature (biology level). However, there are some important issues that should be addressed in future studies such as the inclusion of visualizations of the most relevant brain areas to predict MCI conversion, an adequate analysis of confounds, and generalization methods. These specific improvements would provide more comprehensive and comparable studies.

Limitations
Regarding the limitations of the review, it is worth mentioning that we did not include methodological details such as the preprocessing methods to obtain the neuroimaging results or the mathematical development of the algorithms. This information could have provided a better understanding of each model's performance and how the data is classified to differentiate between groups, but these deep methodological analyses were out of the scope of this review given its more clinically oriented focus.

Conclusions
The recent trend in research to find diagnostic automation methods presents great potential in the early detection of neurodegenerative diseases. Since structural changes appear before the clinical symptoms manifest, there is a valuable period of time in which the morphological and functional changes in the brain can be detected and, therefore, used to predict and provide clinical treatment to slow down the future development of a neurological disease.
Research in this field is still rapidly advancing, new increasingly complex algorithms continue to be developed, and access to higher levels of computational capacity is also increasing, as well as the precision and resolution of neuroimaging techniques. In the future, we can expect faster, more precise, and more efficient classification methods that may be directly incorporated into the neuroimaging techniques themselves that enable the generation of a diagnostic hypothesis with a simple scan of a patient's brain. However, the challenge to translate this knowledge into daily practice remains. This challenge will be overcome on the one hand by increasing the generalizability of the classification methods as they are applied to more diverse samples and, on the other hand, by finding the trade-off between the higher precision achieved when including complex information and a sufficient performance using only the clinical data commonly available for the clinicians. Thus, future studies should focus on obtaining good results using data easily available in the clinical practice (structural MRI, demographic data, and cognitive results, for example) and making their models as much generalizable as possible using more diverse and inclusive samples.