Published on in Vol 13 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/79841, first published .
Reducing Educational Bias in Cognitive Assessment via Dynamic Support Vector Machine Weighting: Validation Study on an Education-Stratified Dataset

Reducing Educational Bias in Cognitive Assessment via Dynamic Support Vector Machine Weighting: Validation Study on an Education-Stratified Dataset

Reducing Educational Bias in Cognitive Assessment via Dynamic Support Vector Machine Weighting: Validation Study on an Education-Stratified Dataset

1School of Humanities and Social Sciences, University of Science and Technology of China, Hefei, China

2School of the Gifted Young, University of Science and Technology of China, Hefei, China

3Department of Rehabilitation Medicine, The First Affiliated Hospital of USTC, Division of Life Sciences and Medicine, University of Science and Technology of China, Tian'e Hu No.1, Hefei, China

4Department of Rehabilitation Medicine, The Second People's Hospital, Wuhu, China

5Shuguang Hospital Anhui Branch Affiliated to Shanghai University of Traditional Chinese Medicine, Hefei, China

*these authors contributed equally

Corresponding Author:

Ming Wu, BSc


Background: The Mini-Mental State Examination (MMSE) remains widely used for cognitive screening, yet its performance varies substantially across educational backgrounds. Linear education corrections fail to capture the nonlinear interference patterns among subitems.

Objective: This study aimed to analyze how educational level shapes MMSE subitem contributions and to develop an education-adaptive optimization strategy using support vector machine–based weighting.

Methods: MMSE data from 812 participants were stratified into 4 education groups. Subitem deletion experiments quantified each subitem’s contribution (Δ). Education-specific support vector machine models were then constructed to derive dynamic weighting coefficients. Performance improvements were assessed before and after weighting.

Results: The illiterate group relied heavily on spatial orientation and memory, whereas university-educated individuals depended more on executive and calculation functions. Several education-dependent interference items were identified (eg, visuospatial construction in the primary group and basic orientation tasks in the university group). Dynamic weighting improved accuracy in all cohorts, most notably among illiterate individuals (Δ=7.25%; P=.06), followed by the primary school group (Δ=3.12%; P=.03).

Conclusions: Education-stratified weighting enhances the fairness and interpretability of MMSE-based screening. External validation confirmed generalizability, although multicenter studies are needed.

JMIR Rehabil Assist Technol 2026;13:e79841

doi:10.2196/79841

Keywords



The Mini-Mental State Examination (MMSE) is one of the most widely used cognitive screening tools in clinical and community settings, designed for the rapid detection of cognitive impairment [1]. Since its development by Folstein et al in 1975 [2], the MMSE has become a core instrument for assessing five cognitive domains—orientation, immediate recall, attention or calculation, language, and visuospatial ability—and can typically be completed within 10 to 15 minutes [3]. Its broad adoption reflects its operational simplicity and solid psychometric performance, with reported sensitivity of 80% to 85% and specificity of 75% to 80% for dementia screening [4].

Alzheimer disease is characterized by progressive neurodegeneration, including amyloid-β deposition and structural decline. As Alzheimer disease progresses in a relatively predictable neuroanatomical sequence, MMSE subdomains provide clinically meaningful stage markers: orientation and memory deficits often correspond to early hippocampal involvement, while language and visuospatial impairments reflect later temporoparietal degeneration [5,6]. This multidimensional structure allows the MMSE to map aspects of the disease trajectory beyond a single total score.

Nevertheless, the diagnostic performance of the MMSE varies considerably across educational levels [7,8]. Individuals with higher education often exhibit false-negative results due to two well-recognized mechanisms: compensatory neuroplasticity that delays the manifestation of cognitive symptoms [9] and the use of test-taking strategies that allow them to maintain near-normal scores despite underlying pathology [10,11]. In contrast, individuals with limited formal education show substantially higher false-positive rates, particularly on education-dependent items, such as literacy tasks and object naming, with error rates increasing by approximately 30% to 35% compared with education-neutral instruments [12-15]. As a result, the MMSE demonstrates reduced sensitivity (68%‐72%) and specificity (65%‐70%) when a conventional cutoff score of 24 points is applied across heterogeneous educational backgrounds [16-18]. Current clinical guidelines therefore emphasize the need to incorporate educational history into cognitive assessment to minimize diagnostic inaccuracies [19,20].

Machine learning (ML) has emerged as a transformative approach in medical diagnostics, with particular relevance to cognitive assessment optimization. Support Vector Machines (SVMs) offer several advantages for psychometric refinement, including (1) robust performance in limited-sample settings (typically n<1000) [21,22], (2) effective handling and interpretability of high-dimensional feature spaces [23,24], and (3) strong compatibility with multimodal data integration frameworks that combine neuropsychological assessments with imaging- or biomarker-derived measures [25-28]. Earlier ML classifiers trained solely on MMSE total scores or subitems—most commonly logistic regression, SVM, or random forest—have demonstrated moderate diagnostic performance, with reported accuracies of 72% to 85% and area under the curve (AUC) values of 0.75 to 0.85 [29]. In contrast, multimodal models integrating magnetic resonance imaging (MRI) radiomics, positron emission tomography (PET) signatures, or speech biomarkers consistently achieve accuracies exceeding 90% to 95% [30,31]. This performance gap highlights the need for methodological innovations that enhance unimodal MMSE-based classifiers while preserving their scalability and low implementation cost.

This study introduces an education-sensitive adaptation of the MMSE based on an SVM-guided dynamic weighting framework. We analyzed cognitive screening data (n=812) collected from Chinese tertiary hospitals and community health centers and stratified participants into four educational cohorts: illiterate (0 y), primary (≤6 y), secondary (7‐12 y), and tertiary (≥13 y). The model uses systematic item-response analysis combined with SVM-guided weighting to adjust the relative contribution of each MMSE subitem in an education-specific manner.

This work provides 2 methodological innovations. First, it moves beyond conventional linear education corrections by using nonlinear SVM modeling to identify “cognitive interference” patterns unique to each education level. Second, it introduces a dynamic weighting strategy specifically designed to mitigate the disproportionate false-positive burden experienced by low-literacy populations. Together, these innovations aim to deliver a more equitable and clinically practical adaptation of the MMSE, particularly relevant for resource-limited settings.

Although unimodal or condensed classifiers derived solely from MMSE subitems are highly accessible, they are inherently constrained by limited feature richness and therefore tend to exhibit moderate predictive performance. In contrast, multimodal diagnostic systems—such as those leveraging MRI, PET, radiomics, or speech biomarkers—achieve substantially higher accuracies but remain impractical for widespread screening due to cost, technical requirements, and limited availability in primary care. Thus, methodological advances that enhance the diagnostic utility of unimodal MMSE-based approaches, while maintaining their affordability and scalability, are urgently needed. This study addresses this gap by proposing an education-stratified dynamic weighting method designed to improve predictive performance within the intrinsic limitations of unimodal cognitive classifiers.


Ethical Considerations

This study followed the principles of the Declaration of Helsinki and was approved by the Ethics Committee of the First Affiliated Hospital of the University of Science and Technology of China (approval number 2024-RE-431). All clinical and community datasets were obtained with authorization from the original data custodians. During data processing, only anonymized records were used, and no personally identifiable information (eg, names, addresses) was accessed, ensuring compliance with the Declaration of Helsinki.

Sample

Cognitive assessment data were obtained from two sources: the Rehabilitation Medicine Department of a tertiary hospital and a community-based cognitive screening program. A total of 812 valid records were included. Among these, 404 samples were collected from outpatients and inpatients with clinically diagnosed cognitive impairment. The inclusion and exclusion criteria are summarized in Table 1.

Table 1. Data inclusion and exclusion criteria.
SourceNumber (n)Inclusion criteriaExclusion criteriaData assessment requirements
Clinical dataset404
  • Outpatients and inpatients with clinically confirmed cognitive impairment
  • Complete clinical and assessment records
  • Patients who underwent surgery or died during hospitalization
  • Missing MMSEa assessment data
  • For clinical data: All MMSE assessors completed a standardized 12-h training program and passed certification.
  • Postcollection, automated missing value detection and manual verification were performed.
Community dataset408
  • Community residents with no medical consultation for cognitive complaints within the past 12 mo
  • Intact consciousness and ability to complete cognitive assessments
  • Individuals with diagnosed mild cognitive impairment or suspected cognitive decline
  • Individuals with incomplete assessment procedures or severe organic neurological conditions
  • For community data: Assessments followed standardized community screening procedures. Assessments were administered following standardized community screening procedures with verification of data completeness and adherence to protocol.

aMMSE: Mini-Mental State Examination.

Scale Selection and Scoring Criteria

The MMSE was used as the primary cognitive assessment tool. The scale comprises five cognitive domains: orientation, memory, attention and calculation, language, and visuospatial construction. Education-adjusted cutoff scores were based on Chinese normative standards [3]: illiterate group: ≤17, primary school group: ≤20, secondary school group: ≤22, and university group: ≤23.

Data Processing and Grouping

Overview

A total of 1000 assessments were collected initially. After quality control, 188 (18.8%) records were excluded due to missing data (n=107, 10.7%), nonstandardized test administration (n=35, 3.5%), or unqualified assessors (n=46, 4.6%). Before merging the hospital and community datasets, baseline demographic comparability was examined using an independent-samples t test (age) and chi-square test (gender) to minimize potential selection bias.

The final dataset (N=812) was stratified by educational attainment as follows: illiterate (n=108, 13.3%), primary school (n=105, 12.9%), secondary school (n=364, 44.8%), and university (n=235, 28.9%). Within each group, patient and control counts were recorded as illiterate (patients: n=60, 7.4%; and controls: n=48, 5.9%), primary (patients: n=53, 6.5%; and controls: n=52, 6.4%), secondary (patients: n=185, 22.8%; and controls: n=179, 22.0%), and university (patients: n=106, 13.1%; and controls: n=129, 15.9%). This preprocessing ensured a standardized data foundation for subsequent ML procedures.

Model Selection

A SVM classifier with a Radial Basis Function kernel was selected due to its suitability for medium-sized datasets and its capacity to model nonlinear decision boundaries. A stratified 5-fold cross-validation scheme was used to maximize data utilization while preserving the original case-control ratio within each educational group.

Hyperparameter Optimization

Hyperparameters were optimized through a grid search embedded within the cross-validation framework. The search grid included penalty parameter C: {0.1, 1, 5, 10, 100} and kernel coefficient γ: {0.001, 0.01, 0.1, 1}.

For each parameter pair, the model was trained on 4 folds and evaluated on the remaining fold, and the average accuracy across all 5 folds was used to identify the optimal configuration. Education-specific models were then retrained on the full dataset using the selected hyperparameters to ensure maximal predictive performance while avoiding information leakage.

Experimental Design

The overall experimental workflow is presented in Figure 1.

Figure 1. Contribution of individual Mini-Mental State Examination (MMSE)j items across educational strata. Feature importance was quantified using the percentage change in cross-validated accuracy (Δi) after deleting each subitem. We operationally defined critical factors as items whose removal caused a performance decrease of Δ<−1.0 and interference factors as items whose removal increased accuracy by Δ>+0.5. SVM: support vector machine.

The process began with data fusion and initial grouping: clinical and community data were merged to construct the initial dataset (n=812), and participants were stratified into 4 educational groups (ie, illiterate, primary school, secondary school, and university) to ensure subsequent analyses reflected education-specific cognitive characteristics. Next, a Support Vector Machine classifier with a radial basis function kernel was trained using a stratified 5-fold cross-validation scheme, preserving the original case-control ratio within each group. To quantify the contribution of each MMSE subitem, a systematic ablation procedure was performed. For each educational group, each subitem was removed in turn, and an SVM model was retrained. The feature contribution was quantified using the cross-validated change in prediction accuracy (Δ), calculated as:

Δ=AccuracypostdeletionAccuracybaseline

A negative Δ indicates a positive contribution (performance dropped when the item was removed), while a positive Δ suggests the item acted as noise. Finally, based on the subitem contribution profiles, an education-specific dynamic weighting scheme was constructed to adjust the relative importance of MMSE subitems before classification.

Other Statistical Techniques

Routine statistical analyses were performed using SPSS version 24.0 (IBM Corp), and ML model construction and evaluation were conducted using Python (version 3.10.10; Python Software Foundation)

For categorical variables, chi-square or Fisher exact tests were applied depending on expected frequencies. Continuous variables with a normal distribution were expressed as mean (SD) and compared using independent-samples t tests. Variables not following a normal distribution were summarized using median (IQR) and analyzed using nonparametric tests.

A 2-tailed P<.05 was considered statistically significant.


Demographic Comparison

Following data cleansing, demographic characteristics of the clinical and community datasets were compared to verify baseline homogeneity before integration. As presented in Table 2, differences in age and gender distributions were not statistically significant (P>.05).

Table 2. Demographic characteristics and data composition stratified by educational background.
CharacteristicIlliterate (n=108)Primary school (n=105)Secondary school (n=364)University (n=235)P value
Age, mean (SD)59.02 (8.5)58.55 (7.8)58.62 (8.1)56.7 (7.4).05
Gender, n (%).37
Male67 (62.0)60 (57.1)219 (60.2)140 (59.6)
Female41 (38.0)45 (42.9)145 (39.8)95 (40.4)
Data source, n (%).12
Clinicala (patient)60 (55.6)53 (50.5)185 (50.8)106 (45.1)
Communityb (control)48 (44.4)52 (49.5)179 (49.2)129 (54.9)

aHospital dataset: patients diagnosed or evaluated in the tertiary hospital.

bCommunity dataset: participants recruited from community-based cognitive screening programs.

Item-Wise Contribution Analysis

To assess the contribution of individual MMSE items across educational groups, we conducted an item deletion experiment. The change in cross-validated accuracy (Δi) after removing each subitem was used to quantify its importance. A negative Δi (< -1.0) indicated a critical item whose removal impaired performance, while a positive Δi (> +0.5) indicated an interference item whose removal improved accuracy. The results of this analysis are visualized in Figure 2.

Figure 2. Heatmap visualization of feature contribution analysis. The color spectrum reflects prediction accuracy after item deletion. Cooler colors (blue and green) denote "critical features" (accuracy drop), while warmer colors (yellow and red) denote "neutral/interfering features" (accuracy stable/increase). The y-axis represents educational groups with the following sample sizes (N_patient/N_control): illiterate (60/48), primary school (53/52), secondary school (185/179), and university (106/129). DM: delayed memory; EA: execution ability; IM: immediate memory; PP: penmanship and praxis; SS: spatial structure.

Illiterate Group

In the illiterate cohort (baseline accuracy=78.83%), all subitems demonstrated positive contributions. The most influential items were spatial orientation (Δ=−6.58% postremoval), immediate memory (Δ=−3.85%), and writing (Δ=−2.86%). No interference items were identified.

Primary School Group

In the primary school group (baseline accuracy=85.71%), spatial construction functioned as an interference factor (Δ =+0.95%), while calculation showed a negligible effect (Δ=0). Removal of immediate memory, delayed recall, reading, or writing reduced accuracy to 83.81% (Δ=−1.90%).

Secondary School Group

In the secondary school group (baseline accuracy=69.78%), multiple interference factors were observed: spatial orientation (Δ=+3.57%), delayed recall (Δ=+3.02%), executive function (Δ=+2.19%), reading (Δ=+1.64%), immediate memory (Δ=+2.47%), writing (Δ=+0.20%), and spatial construction (Δ=+1.10%). Only temporal orientation (Δ=−0.83%) and calculation ability (Δ=−0.80%) demonstrated positive contributions.

University Group

In the university group (baseline accuracy=71.49%), five interference items were identified: spatial orientation (Δ=+2.98%), calculation (Δ=+2.98%), immediate memory (Δ=+1.70%), naming (Δ=+1.70%), and delayed recall (Δ=+1.42%). Repetition (Δ=−0.57%) and reading (Δ=−0.30%) showed small positive contributions.

Item Weight Analysis

On the basis of these contribution patterns, education-specific weighted coefficients were calculated for all MMSE subitems. The weighted diagnostic score was defined as follows:

S=t111 (vi×w1)

where S is the final weighted diagnostic score, vi is the raw score of the i-th MMSE subitem, and wi is the dynamic weighting coefficient assigned to the i-th subitem.

Classification was determined by comparing S to a group-specific classification threshold (T): S<T indicated cognitive impairment, and S≥T indicated normal cognition.

After normalization, the maximum possible weighted score was 60. Education-specific diagnostic thresholds were defined as follows: illiterate (t=30), primary (t=31), secondary (t=32), and university (t=33). As presented in Tables 3 and 4, accuracy improved across all educational strata after weight calibration.

Table 3. Model performance and weighting coefficients for core cognitive domains by educational group.
GroupPrediction accuracy before weightingPrediction accuracy after weightingOrientation (time)Orientation (space)Memory (immediate)Memory (delayed)Attention and calculation
Illiterate (%)79.8386.0813221
Primary school (%)85.7188.1623221
Secondary school (%)69.7869.8323212
University (%)71.4973.6213222
Table 4. Weighting coefficients for language and visuospatial domains by educational group.
GroupNamingRepetitionExecution abilityReadingWritingSpatial structure
Illiterate (%)332323
Primary school (%)123221
Secondary school (%)222111
University (%)231122

Table 5 summarizes the diagnostic performance before and after dynamic weighting, including accuracy, sensitivity, specificity, AUC, and the statistical significance of performance changes. The largest improvement occurred in the illiterate group (Δ=+7.25%), whereas the secondary school group showed the smallest improvement (Δ=+0.05%).

Weight distributions differed across education levels. For basic cognitive domains, spatial orientation weights were uniformly set to 3 across all groups, and immediate memory weights remained at 2. In the illiterate group, naming, repetition, and spatial construction received the highest weights (all=3). In the primary school group, temporal orientation was weighted at 2 and spatial construction was weighted at 1. In the secondary school group, executive function and calculation were weighted at 2, while spatial construction remained at 1. In the university group, executive function received the highest weight (3), calculation was weighted at 2, and spatial construction, reading, and writing each received a weight of 1.

Table 5. Comparison of diagnostic performance metrics and statistical significance before and after dynamic weighting.
Education group and metricBefore weightingAfter weightingImprovement (Δ)P value
Illiterate7.25%.06
Accuracy (%)78.8386.08
Sensitivity (%)81.4584.92
Specificity (%)76.2187.24
AUCa0.8350.912
Primary school2.45%.03
Accuracy (%)85.7188.16
Sensitivity (%)86.1088.54
Specificity (%)85.3287.78
AUC0.8940.928
Secondary school0.05%.46
Accuracy (%)69.7869.83
Sensitivity (%)70.2570.31
Specificity (%)69.3169.35
AUC0.7520.753
University2.13%.18
Accuracy (%)71.4973.62
Sensitivity (%)72.1574.88
Specificity (%)70.8372.36
AUC0.7680.795

aAUC: area under the curve.


Principal Findings

This study developed an education-sensitive optimization framework for the MMSE using a dynamic SVM-based weighting strategy. Across 812 education-stratified participants, the model demonstrated that cognitive subitems contribute differently to diagnostic prediction depending on years of formal education. Dynamic weighting enhanced the performance of unimodal MMSE classifiers, particularly in low-literacy groups, and reduced education-related diagnostic bias. These findings refine the diagnostic utility of condensed cognitive screening tools and support their broader applicability in resource-limited settings.

Comparison With Prior Work

ML has been widely applied to cognitive assessment [32-35], yet existing studies typically use MMSE scores as features or labels without addressing education-driven measurement bias [36,37]. Prior condensed MMSE-only classifiers generally achieve accuracies of 72% to 85%, whereas multimodal MRI-, PET-, and biomarker-based models often exceed 90% [32,33]. Although our model remains unimodal, the dynamic weighting mechanism improved prediction by amplifying high-value signals and attenuating education-dependent interference, partially narrowing the performance gap with multimodal systems while sustaining low cost and operational simplicity.

Traditional MMSE optimization approaches—such as linear education corrections or standard ML models—lack the capacity to capture nonlinear, education-specific feature interactions [38]. In contrast, our framework integrates item deletion experiments with SVM-informed dynamic weighting to generate transparent, subgroup-adaptive coefficients. This approach offers an interpretable alternative to black box ensemble models and provides an effective mechanism for addressing item-level educational bias in cognitive assessment. A comprehensive comparison between the proposed dynamic weighted SVM model and existing cognitive assessment optimization strategies is presented in Table 6.

Table 6. Comparison of the proposed dynamic weighted Support Vector Machine (SVM) model with existing cognitive assessment optimization strategies.
Optimization strategyKey methodologyData requirementsAdvantagesLimitations
Traditional linear correction (eg, MMSEa-E)Linear regression and fixed point additionMMSE scores+demographicsSimple calculation and clinically familiarIgnored nonlinear “cognitive interference” and low precision
Standard machine learning (eg, RFb, ANNc)Black box classification using raw scoresMMSE subitems or total scoresHigh classification accuracyLow interpretability (“black box”) and hard to explain to clinicians
Multimodal fusion modelsDeep learning (CNNd/RNNe) integrationMRIf/PETg imaging+biomarkers+scaleHighest accuracy (>90%) and comprehensive pathology mappingHigh cost, low accessibility in primary care, and complex deployment
This study (dynamic weighted SVM)Nonlinear dynamic weighting via SVMMMSE subitems only (low cost)High interpretability (visible weights), education adaptive, and high accessibilityAccuracy is lower than multimodal models (feature limitation)

aMMSE: Mini-Mental State Examination.

bRF: random forest.

cANN: artificial neural network.

dCNN: convolutional neural network.

eRNN: recurrent neural network.

fMRI: magnetic resonance imaging.

gPET: positron emission tomography.

Interpretation of Education-Specific Patterns

Education-stratified analyses revealed distinct, cognitive reliance patterns. The illiterate group depended heavily on orientation and immediate memory, consistent with basic functional domains commonly preserved in low-literacy populations. The primary school group demonstrated changes in weighting for temporal orientation, whereas secondary school participants exhibited varied interference across multiple domains and minimal model improvement (Δ=0.05), suggesting high within-group heterogeneity. University-educated participants showed stronger reliance on executive functioning and calculation. These patterns align with prior evidence that cognitive performance becomes increasingly distributed and strategy dependent with higher educational attainment [39,40].

The dynamic weighting strategy provided incremental accuracy improvements across all educational cohorts while reducing false-positive rates in lower-education groups. Subitem deletion further allowed visualization of feature importance (Figure 2), conceptually analogous to Shapley Additive Explanations–based explainability, thereby enhancing the transparency of SVM decision processes. Clinical validation by senior rehabilitation physicians confirmed that weight distributions reflected plausible neurocognitive patterns.

External validation using an independent community dataset (n=314) demonstrated stable performance (overall accuracy: 82.48% and AUC: 0.88), with the highest accuracy observed in the illiterate group (88.89%). These results support the generalizability and robustness of the proposed dynamic weighting model and mitigate concerns about overfitting.

Limitations

This study has several limitations. First, the sample was derived from a single geographic region, leading to educational imbalance across subgroups and potentially limiting generalizability. Broader multicenter sampling is needed. Second, demographic confounders beyond age and gender—such as socioeconomic status, occupational complexity, and urban-rural residence—were not available in the dataset, restricting the ability to fully disentangle education effects from related social determinants. Third, although the MMSE provides valuable screening utility, ceiling effects in highly educated individuals limit sensitivity to early cognitive decline. Integrating additional modalities, such as the Montreal Cognitive Assessment or imaging markers, may complement the proposed model. Finally, the SVM model relied on grid-search hyperparameter tuning; future research may explore AutoML-based optimization to improve efficiency and predictive performance.

Conclusions

This study proposes a dynamic SVM-based weighting framework that enhances the diagnostic fairness of MMSE-based cognitive screening across diverse educational backgrounds. By quantifying item-level contributions and adapting subitem weights for each education group, the method addresses a longstanding source of measurement bias in cognitive assessment. The approach retains the accessibility and scalability of condensed cognitive screening tools while improving prediction accuracy and interpretability. These findings provide a practical foundation for developing equitable cognitive assessment strategies, particularly in resource-limited regions.

Acknowledgments

The authors sincerely acknowledge all individuals and institutions that contributed to this work. This research was jointly supported by the Health Research Program of Anhui Province (project number AHWJ2023A10123) and the Anhui Provincial Teaching Research Project (project number 2022jyxm1851) titled "Research and Implementation of a Standardized Training Curriculum System for Rehabilitation Therapists Based on Competency-Based Education." Special thanks are extended to the funding bodies for their support.

Qi Zhang is co-corresponding author (email: ustczhangqi@163.com; phone +86 17333070277).

Funding

This work was supported by the Anhui Provincial Teaching Research Project (grant 2022jyxm1851), the Health Research Program of Anhui (grant AHWJ2023A10123), and the Anhui Provincial Graduate Academic Innovation Project “Optimization and Prediction of Alzheimer’s Disease-Related Cognitive Scales Based on Machine Learning” (grant 2024xscx009). This research received no commercial sponsorship or financial support that could compromise its objectivity. All authors declare no financial or proprietary interests relevant to this study. The funding agencies provided only financial support and had no involvement in study design, data collection and analysis, manuscript preparation, or submission for publication. The research team assumes full responsibility for the study's independence and scientific integrity.

Data Availability

The clinical dataset used in this study was obtained under ethical approval (2024-RE-431) and contains sensitive patient information; therefore, it cannot be made publicly available. Deidentified clinical records (n=404) may be shared upon reasonable request to the corresponding author under a data sharing agreement. Community screening data (n=408) are available with permission from the participating community health centers.

The complete source code for data preprocessing, Support Vector Machine modeling, dynamic weighting, and statistical analysis is openly available on GitHub [41].

Authors' Contributions

QL: Conceptualization, writing – original draft, formal analysis, data curation, methodology

CM: Conceptualization, writing – original draft, formal analysis, data curation, methodology

ML: Investigation, resources

SC: Data curation (clinical)

MY: Data curation (clinical and community)

LX: Data curation (clinical and community)

QZ: Supervision, project administration

MW: Supervision, project administration

All authors have reviewed and approved the final manuscript, agreeing to assume public responsibility for its content.

QZ is the co-corresponding author.

Conflicts of Interest

None declared.

  1. Chun CT, Seward K, Patterson A, Melton A, MacDonald-Wicks L. Evaluation of available cognitive tools used to measure mild cognitive decline: a scoping review. Nutrients. Nov 8, 2021;13(11):3974. [CrossRef] [Medline]
  2. Folstein MF, Folstein SE, McHugh PR. “Mini-mental state”. A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res. Nov 1975;12(3):189-198. [CrossRef] [Medline]
  3. Jia X, Wang Z, Huang F, et al. A comparison of the Mini‑Mental State Examination (MMSE) with the Montreal Cognitive Assessment (MoCA) for mild cognitive impairment screening in Chinese middle‑aged and older population: a cross‑sectional study. BMC Psychiatry. Oct 4, 2021;21(1):485. [CrossRef] [Medline]
  4. Patnode CD, Perdue LA, Rossom RC, et al. Screening for cognitive impairment in older adults: updated evidence report and systematic review for the US Preventive Services Task Force. JAMA. Feb 25, 2020;323(8):764-785. [CrossRef] [Medline]
  5. Scheltens P, De Strooper B, Kivipelto M, et al. Alzheimer’s disease. Lancet. Apr 24, 2021;397(10284):1577-1590. [CrossRef] [Medline]
  6. Yin TT, Cao MH, Yu JC, et al. T1‑weighted imaging‑based hippocampal radiomics in the diagnosis of Alzheimer’s disease. Acad Radiol. Dec 2024;31(12):5183-5192. [CrossRef] [Medline]
  7. Cardoso S, Barros R, Marôco J, de Mendonça A, Guerreiro M. Different MMSE domains are associated to cognitive decline and education. Appl Neuropsychol Adult. 2024;31(4):533-539. [CrossRef] [Medline]
  8. Bird HR, Canino G, Stipec MR, Shrout P. Use of the Mini‑Mental State Examination in a probability sample of a Hispanic population. J Nerv Ment Dis. Dec 1987;175(12):731-737. [CrossRef] [Medline]
  9. Koepsell TD, Kurland BF, Harel O, Johnson EA, Zhou XH, Kukull WA. Education, cognitive function, and severity of neuropathology in Alzheimer disease. Neurology. May 6, 2008;70(19 Pt 2):1732-1739. [CrossRef] [Medline]
  10. Baker DW, Gazmararian JA, Sudano J, Patterson M. The association between age and health literacy among elderly persons. J Gerontol B Psychol Sci Soc Sci. Nov 2000;55(6):S368-S374. [CrossRef] [Medline]
  11. Ouvrard C, Berr C, Meillon C, et al. Norms for standard neuropsychological tests from the French CONSTANCES cohort. Eur J Neurol. May 2019;26(5):786-793. [CrossRef] [Medline]
  12. Aprahamian I, Martinelli JE, Cecato J, Yassuda MS. Screening for Alzheimer’s disease among illiterate elderly: accuracy analysis for multiple instruments. J Alzheimers Dis. 2011;26(2):221-229. [CrossRef] [Medline]
  13. Rezende GP, Cecato J, Martinelli JE. Cognitive abilities screening instrument-short form, Mini-Mental State Examination and functional activities questionnaire in the illiterate elderly. Dement Neuropsychol. 2013;7(4):410-415. [CrossRef] [Medline]
  14. Katzman R, Zhang MY, et al. A Chinese version of the Mini-Mental State Examination; impact of illiteracy in a Shanghai dementia survey. J Clin Epidemiol. 1988;41(10):971-978. [CrossRef] [Medline]
  15. Weiss BD, Reed R, Kligman EW, Abyad A. Literacy and performance on the Mini-Mental State Examination. J Am Geriatr Soc. Jul 1995;43(7):807-810. [CrossRef] [Medline]
  16. Paddick SM, Gray WK, McGuire J, Richardson J, Dotchin C, Walker RW. Cognitive screening tools for identification of dementia in illiterate and low-educated older adults, a systematic review and meta-analysis. Int Psychogeriatr. Jun 2017;29(6):897-929. [CrossRef] [Medline]
  17. Pellicer-Espinosa I, Díaz-Orueta U. Cognitive screening instruments for older adults with low educational and literacy levels: a systematic review. J Appl Gerontol. Apr 2022;41(4):1222-1231. [CrossRef] [Medline]
  18. Brucki SMD, Mansur LL, Carthery-Goulart MT, Nitrini R. Formal education, health literacy and Mini-Mental State Examination. Dement Neuropsychol. 2011;5(1):26-30. [CrossRef] [Medline]
  19. Carnero-Pardo C. Should the Mini-Mental State Examination be retired? Neurologia. Oct 2014;29(8):473-481. [CrossRef] [Medline]
  20. Cai H, Shao Y, Liu XY, et al. Interpretable prognostic modeling for long-term survival of Type A aortic dissection patients using Support Vector Machine algorithm. Eur J Med Res. Apr 15, 2025;30(1):277. [CrossRef] [Medline]
  21. Luo W. Predicting cervical cancer outcomes: statistics, images, and machine learning. Front Artif Intell. 2021;4:627369. [CrossRef] [Medline]
  22. Cai F, Cherkassky V. Generalized SMO algorithm for SVM-based multitask learning. IEEE Trans Neural Netw Learn Syst. Jun 2012;23(6):997-1003. [CrossRef] [Medline]
  23. Huang C, Chung FL, Wang S. Multi-view L2-SVM and its multi-view core vector machine. Neural Netw. Mar 2016;75:110-125. [CrossRef] [Medline]
  24. Mastropietro A, Feldmann C, Bajorath J. Calculation of exact Shapley values for explaining Support Vector Machine models using the radial basis function kernel. Sci Rep. Nov 10, 2023;13(1):19561. [CrossRef] [Medline]
  25. Jasodanand VH, Bellitti M, Kolachalama VB. An AI-first framework for multimodal data in Alzheimer’s disease and related dementias. Alzheimers Dement. Sep 2025;21(9):e70719. [CrossRef] [Medline]
  26. Rathore S, Habes M, Iftikhar MA, Shacklett A, Davatzikos C. A review on neuroimaging-based classification studies and associated feature extraction methods for Alzheimer’s disease and its prodromal stages. Neuroimage. Jul 15, 2017;155:530-548. [CrossRef] [Medline]
  27. Zhang D, Wang Y, Zhou L, Yuan H, Shen D. Multimodal classification of Alzheimer’s disease and mild cognitive impairment. Neuroimage. Apr 1, 2011;55(3):856-867. [CrossRef] [Medline]
  28. Jasodanand VH, Kowshik SS, Puducheri S, et al. AI-driven fusion of multimodal data for Alzheimer’s disease biomarker assessment. Nat Commun. Aug 11, 2025;16(1):7407. [CrossRef] [Medline]
  29. Dincer A, Gordon BA, Hari-Raj A, et al. Comparing cortical signatures of atrophy between late-onset and autosomal dominant Alzheimer disease. Neuroimage Clin. 2020;28:102491. [CrossRef] [Medline]
  30. Qiu S, Miller MI, Joshi PS, et al. Multimodal deep learning for Alzheimer’s disease dementia assessment. Nat Commun. Jun 20, 2022;13(1):3404. [CrossRef] [Medline]
  31. Castellano G, Esposito A, Lella E, Montanaro G, Vessio G. Automated detection of Alzheimer’s disease: a multi-modal approach with 3D MRI and amyloid PET. Sci Rep. Mar 3, 2024;14(1):5210. [CrossRef] [Medline]
  32. Vyas A, Aisopos F, Vidal ME, Garrard P, Paliouras G. Identifying the presence and severity of dementia by applying interpretable machine learning techniques on structured clinical records. BMC Med Inform Decis Mak. Oct 17, 2022;22(1):271. [CrossRef] [Medline]
  33. Feng F, Wang P, Zhao K, et al. Radiomic features of hippocampal subregions in Alzheimer’s disease and amnestic mild cognitive impairment. Front Aging Neurosci. 2018;10:290. [CrossRef] [Medline]
  34. Battista P, Salvatore C, Castiglioni I. Optimizing neuropsychological assessments for cognitive, behavioral, and functional impairment classification: a machine learning study. Behav Neurol. 2017;2017:1850909. [CrossRef] [Medline]
  35. Khatun S, Morshed BI, Bidelman GM. A single‑channel EEG‑based approach to detect mild cognitive impairment via speech‑evoked brain responses. IEEE Trans Neural Syst Rehabil Eng. May 2019;27(5):1063-1070. [CrossRef] [Medline]
  36. Wu Y, Jia M, Xiang C, Lin S, Jiang Z, Fang Y. Predicting the long-term cognitive trajectories using machine learning approaches: a Chinese nationwide longitudinal database. Psychiatry Res. Apr 2022;310:114434. [CrossRef] [Medline]
  37. Martin SA, Townend FJ, Barkhof F, Cole JH. Interpretable machine learning for dementia: a systematic review. Alzheimers Dement. May 2023;19(5):2135-2149. [CrossRef] [Medline]
  38. Kantayeva G, Lima J, Pereira AI. Application of machine learning in dementia diagnosis: a systematic literature review. Heliyon. Nov 2023;9(11):e21626. [CrossRef] [Medline]
  39. Livingston G, Huntley J, Sommerlad A, et al. Dementia prevention, intervention, and care: 2020 report of the Lancet Commission. Lancet. Aug 8, 2020;396(10248):413-446. [CrossRef] [Medline]
  40. Stern Y. Cognitive reserve in ageing and Alzheimer’s disease. Lancet Neurol. Nov 2012;11(11):1006-1012. [CrossRef] [Medline]
  41. Dynamic-SVM-weighting-for-MMSE-optimization. GitHub. URL: https://github.com/masC1962/Dynamic-SVM-Weighting-for-MMSE-Optimization [Accessed 2026-01-05]


AUC: area under the curve
ML: machine learning
MMSE: Mini-Mental State Examination
MRI: magnetic resonance imaging
PET: positron emission tomography
SVM: support vector machine


Edited by Sarah Munce; submitted 30.Jun.2025; peer-reviewed by Ali M Hasan, Hongxing Kan, Susanne Neufang, Zhe-Yuan Jiang; accepted 23.Dec.2025; published 25.Feb.2026.

Copyright

© Qing Liu, Chi Ma, Mengyuan Liu, Suhui Chen, Mengting Yu, Lijuan Xia, Qi Zhang, Ming Wu. Originally published in JMIR Rehabilitation and Assistive Technology (https://rehab.jmir.org), 25.Feb.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Rehabilitation and Assistive Technology, is properly cited. The complete bibliographic information, a link to the original publication on https://rehab.jmir.org/, as well as this copyright and license information must be included.