Abstract
Background: Artificial intelligence (AI) has demonstrated superior diagnostic accuracy compared with medical practitioners, highlighting its growing importance in health care. SMART-Pred (Shiny Multi-Algorithm R Tool for Predictive Modeling) is an innovative AI-based application for Alzheimer disease (AD) prediction using handwriting analysis.
Objective: This study aimed to develop and evaluate a noninvasive, cost-effective AI tool for early AD detection, addressing the need for accessible and accurate screening methods.
Methods: The study used principal component analysis for dimensionality reduction of handwriting data, followed by training and evaluation of 10 diverse AI models, including logistic regression, naïve Bayes, random forest, adaptive boosting, support vector machine, and neural network. Model performance was assessed using accuracy, sensitivity, precision, specificity, F1-score, and area under the curve (AUC) metrics. The DARWIN (Diagnosis Alzheimer With Handwriting) dataset, comprising handwriting samples from 174 participants (89 patients with AD and 85 healthy controls), was used for validation and testing.
Results: The neural network classifier achieved an accuracy of 91% (95% CI 0.79‐0.97) and an AUC of 94% on the test set after identifying the most significant features for AD prediction. These performance results surpass those of current clinical diagnostic tools, which typically achieve around 81% accuracy. SMART-Pred’s performance aligns with recent AI advancements in AD prediction, such as Cambridge scientists’ AI tool achieving 82% accuracy in identifying AD progression within 3 years, using cognitive tests and magnetic resonance imaging scans. The variables “air_time” and “paper_time” consistently emerged as critical predictors for AD across all 10 AI models, highlighting their potential importance in early detection and risk assessment. To augment transparency and interpretability, we incorporated the principles of explainable AI, specifically using Shapley Additive Explanations, a state-of-the-art method to emphasize the features responsible for our model’s efficacy.
Conclusions: SMART-Pred offers noninvasive, cost-effective, and efficient AD prediction, demonstrating the transformative potential of AI in health care. While clinical validation is necessary to confirm the practical applicability of the identified key variables, the findings of this study contribute to the growing body of research on AI-assisted AD diagnosis and may lead to improved patient outcomes through early detection and intervention.
doi:10.2196/70272
Keywords
Introduction
Artificial intelligence (AI) is revolutionizing medical diagnosis, offering unprecedented accuracy and efficiency in health care []. AI algorithms can analyze vast amounts of complex medical data, including images, patient records, and genetic information, to detect diseases earlier and more precisely than traditional methods [-]. In some cases, AI outperforms human practitioners in diagnostic accuracy, particularly in fields like radiology and pathology [-]. For instance, AI systems can identify subtle patterns in electrocardiograms to detect genetic heart conditions and analyze retinal images to diagnose diabetic retinopathy in its early stages [-]. By augmenting human expertise, AI not only improves diagnostic speed and accuracy but also has the potential to extend quality health care to remote areas, ultimately leading to better patient outcomes and more personalized treatment plans [,,,-].
Alzheimer disease (AD) is a progressive brain disorder that gradually impairs memory and cognitive abilities, ultimately affecting the capacity to perform even the most basic tasks [-]. In most cases, especially those involving late-onset AD, symptoms usually appear in individuals during their mid-60s. Alzheimer disease and related dementias (ADRD) represent the primary cause of dementia among older adults [-]. In contrast, early-onset ADRD, which is relatively uncommon, can develop in people aged 30-65 years []. Aging is common to all cells and organelles in living organisms [-]. In recent times, there has been a disproportionate increase in the occurrence of ADRD, Huntington disease, Parkinson disease, amyotrophic lateral sclerosis, and other neurodegenerative diseases [-]. It is estimated that about 50 million people across the globe are living with ADRD [,]. ADRD are associated with a large family and caregiver burden. In 2018, the United States spent as much as US $277 billion on care for patients with ADRD. It is estimated that over 100 million people will be living with ADRD by the year 2050, for which the associated cost of treatment will be US $1200 billion [-]. Even though the prevalences of most diseases, such as breast cancer, prostate cancer, heart disease, stroke, and HIV, have decreased from 2000 to 2015, the prevalence of ADRD has increased by 123% within the same period [].
Although many interventions, including novel nonsteroidal anti-inflammatory drugs and small molecules, have been explored for the treatment of ADRD [-], there is currently no cure for ADRD. Consequently, early diagnosis and the identification of ADRD biomarkers are critical to tailoring appropriate interventions for ADRD, reinforcing the need to design optimal methods for the prediction of ADRD and identify individuals at higher risk of ADRD in a timely manner [,-].
The rapid growth of big data has necessitated the development of robust AI-driven algorithms for predictive modeling [-]. These algorithms are essential for efficiently handling large datasets and extracting relevant information through machine learning (ML) techniques []. Illustratively, ML algorithms, specifically hierarchical clustering, have been applied to visualize and examine the effects of point mutations that are associated with different phenotypes of diseases, like COVID-19, cystic fibrosis, and sickle cell anemia, resulting in an improved understanding of how point mutations lead to various variants of diseases [-]. AI and ML models have also found their place in disease prediction and risk analysis [,,] for various diseases (neurodegenerative disorders, cardiovascular diseases, renal diseases, infectious diseases, diabetes, chronic diseases, neurological and psychiatric disorders, obesity, COVID-19, skin diseases, brain tumors, thyroid diseases, and liver diseases) [-]. However, no single algorithm can optimally solve every problem, highlighting the need for diverse ML approaches in predictive modeling and analytics. R, a widely adopted programming language for statistical data analysis, has become a powerful tool in data science []. The introduction of the Shiny package by RStudio revolutionized web development within the R ecosystem, making it accessible to both experienced data scientists and those with limited programming experience [,].
Methods
Development and Framework of the Shiny Multi-Algorithm R Tool for Predictive Modeling Application
SMART-Pred (Shiny Multi-Algorithm R Tool for Predictive Modeling) is an innovative AI-based R Shiny application developed to address the increasing demand for accessible data science tools and to simplify the complexity of ML workflows [,,]. This user-friendly platform automates essential steps, such as data preprocessing, model training, hyperparameter tuning, model validation, and performance evaluation. It features a web interface that allows users to upload and preview datasets, configure variables, and compute descriptive statistics. SMART-Pred incorporates 10 classic ML algorithms, including logistic regression, random forest, adaptive boosting (AdaBoost), support vector machine (SVM), and neural network, for robust classification tasks. The tool also offers interactive data frames for comparing model evaluation metrics, alongside visualization capabilities like multimodel receiver operating characteristic (ROC) curve comparisons, feature importance ranking (by each selected algorithm saved as a CSV file in the same working directory in addition to the graphical display), and the Shapley Additive Explanations (SHAP) framework, which can provide the direction of influence of each feature. With one-click server deployment, SMART-Pred ensures accessibility anytime and anywhere, creating a convenient and intuitive environment for ML modeling that bridges the gap between complex AI algorithms and practical applications, potentially accelerating research and innovation across data-driven fields.
The DARWIN (Diagnosis Alzheimer With Handwriting) dataset [] was used to validate SMART-Pred’s classification capabilities and benchmark its performance against prior studies. Specifically curated for early AD detection, the DARWIN dataset comprises handwriting data from 174 participants (89 patients with AD and 85 healthy controls) across 452 columns, including unique identifiers, classification labels (“P” for patient and “H” for healthy), and 451 features extracted from 25 structured tasks (graphic, copy, memory, and dictation exercises) with 18 features per task. Collected using standardized protocols (eg, Wacom tablets at 200 Hz), this dataset captures fine motor and cognitive variations linked to neurodegeneration, offering a balanced sample to minimize algorithmic bias and richer multidimensional data compared to single-task datasets. Unlike neuroimaging-focused alternatives like Alzheimer’s Disease Neuroimaging Initiative (ADNI), the DARWIN dataset supports noninvasive, cost-effective, and immediately deployable screening in clinical settings, aligning with SMART-Pred’s goal of accessible early AD detection [,]. Its feature richness and task diversity mitigate overfitting risks and provide direct insights into cognitive-motor integration, a hallmark of AD progression, while prioritizing actionable, low-resource biomarkers over complex biological data, thus enabling it to offer pragmatic solutions for early diagnosis in diverse health care environments. The SMART-Pred application and data are provided in .
The ultimate objective of this study is to leverage handwriting characteristics from the DARWIN dataset to distinguish between patients with AD and healthy individuals, contributing to earlier, more accessible diagnostic methods that can inform tailored interventions. We hypothesized that the integration of user-configurable and automated ML methods within the SMART-Pred workflow will enable the reliable classification of patients with AD versus healthy individuals based on handwriting features from the DARWIN dataset, demonstrating the potential of handwriting-based digital biomarkers for early, scalable AD detection. SMART-Pred’s end-to-end ML workflow for classification tasks is structured across 4 core components: data input and preview, data preprocessing, model training and validation, and model evaluation. This framework integrates user-configurable options with automated processing to ensure flexibility and methodological rigor. The process starts with data upload and preprocessing, advances through model training with multiple algorithmic choices, and culminates in comprehensive performance evaluation and visualization, as detailed in the flowchart presented in . By democratizing access to advanced ML techniques through an intuitive interface, SMART-Pred aims to enhance research capabilities and clinical applicability in AD detection and beyond.

SHAP-Based Interpretability Method
SHAP is a model-agnostic interpretability technique grounded in cooperative game theory, which quantifies how much each feature contributes to a model’s prediction by computing Shapley values, ensuring consistent and locally accurate explanations for individual outcomes. In its general, model-agnostic form, SHAP approximates these values using a weighted linear regression approach (Kernel SHAP), though specialized algorithms exist for specific model types. SHAP is applied to interpret predictions from all 10 ML classifiers trained on AD data, revealing the relative impact of each input feature on classification outcomes [,]. The explanation model approximates the original complex model by representing the prediction as a linear combination of feature coalitions given by:
where denotes a binary vector indicating the presence or absence of features in the coalition, is the total number of AD features, and corresponds to the Shapley value, reflecting the individual feature’s contribution [].
Through optimization, the Shapley value provides a theoretically grounded and consistent measure of feature importance, enhancing the interpretability of the ML classifier’s predictions.
Data Input and Preview
SMART-Pred allows users to upload local data files through a web interface. Upon uploading, users can preview the dataset in an interactive table format () and access basic descriptive statistics (). For descriptive statistics, the interface provides options for variable and measure configuration, where users can choose specific statistical measures: mean, SD, minimum, median, and maximum for numerical variables, and mode and unique count for character variables. After data upload, users proceed to configure the analysis parameters through a structured variable selection process (). First, users specify the analysis type by selecting either “Classification” or “Regression” from the task type dropdown menu. For the target variable selection, users choose the dependent variable that represents the outcome to be predicted. The interface then presents a comprehensive list of available independent variables in a scrollable selection box, including various measurement features and other kinematic parameters. Users can either select individual variables or use the “Select all independent variables” checkbox for comprehensive analysis. This structured approach ensures proper specification of the prediction task and relevant features for subsequent analysis. The interface also provides checkboxes for selecting from 10 ML models ().




Data Preprocessing
The preprocessing pipeline () implements a systematic approach to data preparation. The preprocessing pipeline in SMART-Pred begins with a stratified 70‐30 split to preserve class balance between patients with AD and controls, ensuring representative distributions in the training (70%) and test (30%) sets. For missing values, mean/mode imputation is systematically applied: numerical features use subgroup-specific means (eg, imputing based on AD stage or age cohort) to retain biological context, while categorical variables use mode replacement or an “Unknown” category to avoid information loss. This approach balances simplicity with clinical relevance, as arbitrary deletion would disproportionately affect smaller AD subgroups.
Principal component analysis (PCA) was strategically implemented to address the DARWIN dataset’s high dimensionality (451 features), focusing on spatiotemporal handwriting variables like air_time and paper_time. PCA condensed these into orthogonal components that captured 95% cumulative variance, eliminating redundant kinematic features (eg, overlapping stroke velocity metrics) while preserving discriminative power for AD-related motor-cognitive patterns. The threshold of 95% variance retention was chosen to minimize noise from minor handwriting variations unrelated to neurodegeneration. Prior to PCA, feature filtering removed near-zero variance metrics (<5% unique values) and highly correlated variables (Pearson r>0.9), reducing multicollinearity that could distort component interpretation.
For non-DARWIN datasets with missing values, SMART-Pred uses context-aware imputation: numerical features use median values stratified by diagnostic status to avoid dilution of AD-specific patterns, while categorical variables adopt a hybrid approach-mode replacement for features with <10% missingness and “Missing” flagging for higher rates. This dual strategy ensures robustness across heterogeneous data sources while maintaining compatibility with the tool’s multi-algorithm framework. The preprocessing workflow optimizes computational efficiency without sacrificing clinical interpretability, enabling focused analysis of neurodegeneration-sensitive handwriting biomarkers.
Model Training and Validation
The model training process uses a robust validation strategy involving 10-fold cross-validation repeated 5 times. The system has 10 classic ML algorithms (). The models range from traditional approaches (eg, logistic regression and naïve Bayes) to advanced ensemble methods (eg, random forest and gradient boosting) and deep learning (eg, neural network), and the selection has been driven by the need to balance algorithmic diversity, clinical utility, and technical robustness. These models have been chosen to cover a broader spectrum of approaches: linear models (eg, logistic regression) provide interpretability for clinical stakeholders, ensemble methods (eg, random forest and AdaBoost) address nonlinear relationships while reducing overfitting, and deep learning models (eg, neural network) capture complex interactions. Criteria, such as accuracy, computational efficiency, and explainability, have guided the selection. Domain-specific validation ensures that the models excel at analyzing temporal handwriting dynamics and mitigates class imbalance. The inclusion of both high-precision neural networks and interpretable models optimizes the tradeoff between complexity and clinical transparency, while cross-validation confirms consistent performance. This multi-algorithm strategy ensures adaptability to diverse patient profiles and aligns with SMART-Pred’s goal of delivering a robust, deployable tool for early AD detection. Detailed information about each algorithm’s implementation and theoretical foundation has been provided previously [-]. Hyperparameter optimization is performed through a grid search process to identify the best model configuration.
| Algorithm | R function | Reference |
| Logistic regression | glm | [] |
| Naïve Bayes | Naïve_bayes | [] |
| K-nearest neighbor | knn | [] |
| Decision tree | rpart | [] |
| Random forest | rf | [] |
| Gradient boosting | gbm | [] |
| Adaptive boosting | ada | [] |
| Extreme gradient boosting | xgbTree | [] |
| Support vector machine | svmRadial | [] |
| Neural network | nnet | [] |
aSMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling.
Evaluation Output
The evaluation module provides a comprehensive assessment through multiple metrics and visualizations. The system generates validation model performance metrics ( and ), including accuracy [], sensitivity [], specificity [], and F1-score [], for model selection; test model performance metrics (), including accuracy [], sensitivity [], specificity [], F1-score [], and ROC [], for final evaluation; and ROC curves () for comparing model performance. Additionally, variable importance analysis () is conducted to identify key predictive features across different models. Lastly, SHAP analysis is conducted to interpret the predictions made by the 10 ML classifiers. All evaluation results are presented through interactive visualizations and data frames, allowing the examination of model performance from multiple perspectives.





Ethical Considerations
This study used secondary, deidentified data and therefore did not require Institutional Review Board approval. All data used in this study have been anonymized to protect privacy.
Results
Preliminary Findings
In SMART-Pred, the validation results () demonstrated the diverse array of models analyzed, revealing significant variations in performance metrics, such as accuracy, sensitivity, specificity, F1-score, and area under the curve (AUC), for the DARWIN dataset among the models. The neural network and AdaBoost classifiers emerged as top performers, exhibiting high accuracy scores of 81% and 80%, respectively. Moreover, AdaBoost and neural network achieved commendable AUC scores (92% and 91%, respectively), suggesting a robust capability to distinguish between positive and negative AD cases. Conversely, models, such as k-nearest neighbors (KNN) and naïve Bayes, were noted for lower overall performance, revealing potential limitations in handling the DARWIN dataset’s intricacies.
| Model name | Accuracy, value (95% CI) | Recall | Precision | Specificity | F1-score | AUC | |
| Logistic regression | 0.75 (0.71‐0.78) | 0.77 | 0.73 | 0.73 | 0.75 | 0.82 | |
| Naïve Bayes | 0.67 (0.64‐0.69) | 0.94 | 0.60 | 0.41 | 0.73 | 0.87 | |
| K-nearest neighbors | 0.64 (0.61‐0.66) | 0.97 | 0.58 | 0.32 | 0.72 | 0.84 | |
| Decision tree | 0.69 (0.67‐0.71) | 0.58 | 0.73 | 0.79 | 0.65 | 0.80 | |
| Random forest | 0.78 (0.76‐0.80) | 0.83 | 0.75 | 0.74 | 0.79 | 0.90 | |
| Gradient boosting | 0.80 (0.79‐0.82) | 0.81 | 0.80 | 0.80 | 0.80 | 0.91 | |
| Adaptive boosting | 0.80 (0.79‐0.81) | 0.83 | 0.78 | 0.77 | 0.80 | 0.92 | |
| Extreme gradient boosting | 0.80 (0.79‐0.80) | 0.80 | 0.79 | 0.79 | 0.79 | 0.91 | |
| Support vector machine | 0.72 (0.70‐0.74) | 0.85 | 0.66 | 0.59 | 0.74 | 0.86 | |
| Neural network | 0.81 (0.80‐0.82) | 0.82 | 0.80 | 0.80 | 0.81 | 0.91 | |
aSMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling.
bAUC: area under the curve.
Full Model With All Features
When the model was used for prediction in the test set (), the AdaBoost classifier stood out with a remarkable accuracy of 85%, a balanced sensitivity of 88%, an F1-score of 85%, and an AUC of 92%, illustrating its ability to generalize well beyond the validation dataset. Such consistency between the validation and test phases highlights the robustness of the AdaBoost model for practical applications. The extreme gradient boosting (XGBoost) classifier also demonstrated reliable performance across the dataset, maintaining stable accuracy and specificity, further reinforcing its candidacy for real-world deployment. Many other algorithms in our pipeline, such as the gradient boosting and random forest classifiers, achieved commendable evaluation scores across the evaluation metrics, reaffirming the robustness of the pipeline for classification-based predictive modeling.
| Model name | Accuracy, value (95% CI) | Recall | Precision | Specificity | F1-score | AUC | |
| Logistic regression | 0.74 (0.60‐0.85) | 0.65 | 0.77 | 0.81 | 0.71 | 0.78 | |
| Naïve bayes | 0.64 (0.50‐0.77) | 0.85 | 0.59 | 0.44 | 0.70 | 0.82 | |
| K-nearest neighbors | 0.62 (0.48‐0.75) | 0.96 | 0.57 | 0.30 | 0.71 | 0.80 | |
| Decision tree | 0.72 (0.58‐0.83) | 0.77 | 0.69 | 0.67 | 0.73 | 0.73 | |
| Random forest | 0.79 (0.66‐0.89) | 0.77 | 0.80 | 0.81 | 0.78 | 0.89 | |
| Gradient boosting | 0.85 (0.72‐0.93) | 0.81 | 0.88 | 0.89 | 0.84 | 0.89 | |
| Adaptive boosting | 0.85 (0.72‐0.93) | 0.88 | 0.82 | 0.81 | 0.85 | 0.92 | |
| Extreme gradient boosting | 0.81 (0.68‐0.91) | 0.73 | 0.86 | 0.89 | 0.79 | 0.88 | |
| Support vector machine | 0.83 (0.70‐0.92) | 0.81 | 0.84 | 0.85 | 0.82 | 0.83 | |
| Neural network | 0.79 (0.66‐0.89) | 0.69 | 0.86 | 0.89 | 0.77 | 0.88 | |
aSMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling.
bAUC: area under the curve.
On integrating findings from both evaluation phases, it becomes evident that ensemble models (random forest, gradient boosting, AdaBoost, and XGBoost) hold significant promise due to their balanced performance metrics and strong generalization capabilities [-]. These models provide a compelling combination of sensitivity and specificity, which is crucial for the diagnostics required in AD detection. The observed fluctuations in the performance of the neural network classifier from validation to testing (especially for recall) suggest that while it possesses inherent strengths, there is room for enhancement.
Our comparative analysis underscores ensemble models (random forest, gradient boosting, AdaBoost, and XGBoost) as the most robust candidates for AD diagnostics in the scope of this work, given their superior performance across both the validation and test datasets [-]. These ensemble models have good, balanced accuracy in this regard, making them particularly suited for clinical settings where precise prediction and identification are critical. The insights garnered from this study pave the way for continued exploration into model optimization, potentially enhancing the applicability and accuracy of ML techniques in terms of advancing AD research and the early diagnosis of other diseases, which can be considered crucial in tailoring interventions to target diseases.
Based on the ROC curves (), the ensemble models (AdaBoost: AUC=0.92; random forest: AUC=0.89; gradient boosting: AUC=0.89; XGBoost: AUC=0.88) had the strongest performance among the 10 ML models considered in this study. The findings suggest that the AdaBoost ensemble model is highly effective in correctly identifying positive and negative instances for the given classification task. AdaBoost was closely followed by random forest and gradient boosting, both of which demonstrated impressive AUC values of 0.89. The XGBoost and neural network classifiers both achieved an impressive AUC value of 0.88.

In contrast, the decision tree model exhibited the weakest performance, with an AUC of 0.73. While this is still better than random guessing, it suggests that the decision tree algorithm may not be the most suitable choice. The remaining models, including logistic regression, naïve Bayes, SVM, and KNN, had AUC values ranging from 0.78 to 0.83, indicating reasonably good predictive capabilities. The intersecting nature of the ROC curves highlights the better tradeoffs between different models [-]. Certain algorithms may excel at specific levels of the false positive rate, which can be an important consideration when selecting the most appropriate model for the task at hand [].
Building upon our variable importance analysis of SMART-Pred (), the partial variable set was constructed by selecting the most influential features among all the algorithms identified through our importance ranking analysis, including “air_time” and “paper_time” variables. Results for the reduced feature model analysis via SMART-Pred are shown in and .
| Model name | Accuracy, value (95% CI) | Recall | Precision | Specificity | F1-score | AUC | |
| Logistic regression | 0.85 (0.82‐0.88) | 0.85 | 0.85 | 0.86 | 0.85 | 0.89 | |
| Naïve Bayes | 0.85 (0.83‐0.87) | 0.87 | 0.83 | 0.84 | 0.85 | 0.93 | |
| K-nearest neighbors | 0.83 (0.81‐0.84) | 0.95 | 0.76 | 0.71 | 0.84 | 0.93 | |
| Decision tree | 0.82 (0.80‐0.84) | 0.77 | 0.85 | 0.87 | 0.81 | 0.89 | |
| Random forest | 0.88 (0.87‐0.90) | 0.88 | 0.89 | 0.89 | 0.88 | 0.95 | |
| Gradient boosting | 0.88 (0.87‐0.89) | 0.88 | 0.88 | 0.88 | 0.88 | 0.94 | |
| Adaptive boosting | 0.89 (0.88‐0.90) | 0.89 | 0.88 | 0.88 | 0.88 | 0.96 | |
| Extreme gradient boosting | 0.87 (0.87‐0.88) | 0.87 | 0.87 | 0.88 | 0.87 | 0.94 | |
| Support vector machine | 0.86 (0.85‐0.88) | 0.85 | 0.87 | 0.87 | 0.86 | 0.93 | |
| Neural network | 0.86 (0.85‐0.87) | 0.87 | 0.85 | 0.85 | 0.86 | 0.95 | |
aSMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling.
bAUC: area under the curve.
| Model name | Accuracy, value (95% CI) | Recall | Precision | Specificity | F1-score | AUC | |
| Logistic regression | 0.87 (0.74‐0.95) | 0.88 | 0.85 | 0.85 | 0.87 | 0.88 | |
| Naïve Bayes | 0.81 (0.68‐0.91) | 0.85 | 0.79 | 0.78 | 0.81 | 0.84 | |
| K-nearest neighbors | 0.81 (0.68‐0.91) | 0.92 | 0.75 | 0.70 | 0.83 | 0.90 | |
| Decision tree | 0.81 (0.68‐0.91) | 0.88 | 0.77 | 0.74 | 0.82 | 0.81 | |
| Random forest | 0.87 (0.75‐0.95) | 0.85 | 0.88 | 0.89 | 0.86 | 0.91 | |
| Gradient boosting | 0.87 (0.75‐0.95) | 0.85 | 0.88 | 0.89 | 0.86 | 0.90 | |
| Adaptive boosting | 0.85 (0.72‐0.93) | 0.81 | 0.88 | 0.89 | 0.84 | 0.91 | |
| Extreme gradient boosting | 0.85 (0.72‐0.93) | 0.81 | 0.88 | 0.89 | 0.84 | 0.87 | |
| Support vector machine | 0.89 (0.77‐0.96) | 0.85 | 0.92 | 0.93 | 0.88 | 0.92 | |
| Neural network | 0.91 (0.79‐0.97) | 0.88 | 0.92 | 0.93 | 0.90 | 0.94 | |
aSMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling.
bAUC: area under the curve.

Reduced Model With Important Features
The comparative analysis of the reduced feature model and full feature model showed that models trained on the selected partial variables (reduced feature model) consistently demonstrated equal or superior performance compared to those trained on all variables (full feature model), as seen in . The comparison between the full and partial variable sets (reduced feature model) revealed patterns in both validation () and test () metrics.


The pattern was particularly profound for models, such as SVM, naïve Bayes, KNN, and neural network, which showed substantial improvements across all metrics when using the selected important variables (most notable was the neural network’s approximately 20% increment in the reduced feature model’s test recall, from 69% to 88%). Ensemble methods like random forest, gradient boosting, AdaBoost, and XGBoost demonstrated slight improvements in performance across both variable sets, highlighting their inherent robustness regardless of feature selection [,]. Notably, the improvements observed in the validation metrics were largely mirrored in the test metrics, indicating that the benefits of using carefully selected important variables generalize well to unseen data.
The consistent enhancement across various performance indicators, including accuracy, recall, precision, F1-score, and ROC values, in both the validation and test scenarios reinforces the effectiveness of our variable importance–based feature selection approach. These findings strongly validate our strategy of using variable importance ranking for feature selection, demonstrating that focusing on the most influential variables not only maintains or improves model performance but also potentially reduces computational complexity and simplifies the model optimization process [-]. In the reduced feature model, the SVM classifier achieved an accuracy of 89% (95% CI 0.77‐0.96) with an AUC of 92%, while the neural network algorithm reached an accuracy of 91% (95% CI 0.79‐0.97) with an AUC of 94%, after identifying the most significant features for AD prediction in the test set. The comparison of ROC curves is presented in .

SHAP Analysis on the DARWIN Dataset
Interpreting the results and understanding the decision-making process of ML models has become paramount, especially in critical fields like medical diagnostics. In the context of AD detection, understanding which features or biological markers most influence a model’s prediction (specifically direction of influence) can provide invaluable insights for medical professionals. In lieu of this, this study performed SHAP analysis. However, for brevity and ease of presentation, the SHAP analysis of the 2 best models (neural network and SVM classifiers) is presented in this section, while the SHAP analysis of the remaining 8 models has been presented in .
The SHAP analysis of the neural network classifier indicated that air_time24 and air_time5 (mean SHAP values of −0.0029 and −0.0008, respectively) had the greatest negative influence on AD prediction, while air_time2 and paper_time8 (mean SHAP values of +0.0018 and +0.0012, respectively) had the greatest positive influence on AD prediction ().

For the support vector classifier, paper_time23 and air_time17 (mean SHAP values of −0.0013 and −0.0004, respectively) had the greatest negative influence on AD prediction, while air_time6 and paper_time8 (mean SHAP values of +0.0013 and +0.0009, respectively) had the greatest positive influence on AD prediction ().
Discussion
Robustness of the SMART-Pred Application
AI is transforming medical practice and research by enhancing disease diagnostics, accelerating biomarker discovery, and revolutionizing disease surveillance [-]. In diagnostics, AI algorithms can examine complex medical data, such as imaging findings, genetic information, and clinical records, to identify subtle patterns that might be missed by human observers [-]. This capability leads to earlier and more accurate disease detection, potentially improving patient outcomes. In biomarker identification, AI’s ability to process vast amounts of multiomic data is uncovering novel indicators of disease presence, progression, and treatment response [,]. These AI-discovered biomarkers are paving the way for personalized medicine approaches and more targeted therapeutic interventions. For disease surveillance, AI systems are proving to be invaluable in monitoring and predicting disease outbreaks on a global scale []. By analyzing diverse data sources, including social media, environmental factors, and health records, AI can identify emerging health threats in real-time, enabling rapid response and containment strategies [,]. Moreover, AI’s impact extends to drug discovery, treatment optimization, and health care resource allocation, promising a future of more efficient, effective, and accessible health care delivery [,].
The SMART-Pred application demonstrates a robust, user-friendly, and customizable approach to variable importance analysis across multiple ML models. This innovative tool offers valuable insights into the predictive power of various features in disease diagnostics, biomarker identification, and disease surveillance, particularly in the context of AD, as demonstrated in this study. By using the “varImp()” function and SHP analysis in R in SMART-Pred, our study identified the top-most influential variables for each model, providing a comprehensive view of feature relevance in AD prediction.
Discussion of the Findings
A thorough analysis of variable importance across all models revealed consistent patterns and model-specific characteristics. The “air_time” features, particularly air_time23, air_time2, and air_time6, emerged as consistently important variables across multiple models, including naïve Bayes, KNN, AdaBoost, and SVM. Additionally, “paper_time” features (paper_time17 and paper_time23) showed significant importance across several models, indicating the crucial role of temporal features in prediction. The selection of air_time (pen movement duration above paper) and paper_time (pen contact duration) in SMART-Pred’s reduced feature model is rooted in their direct biological ties to AD pathology. Air_time reflects dysfunction of the prefrontal cortex, a region critical for motor planning and task initiation [,]. AD-related neurodegeneration is reflected in handwriting performance, with patients showing longer overall writing times and substantially increased air times between pen strokes across everyday copying tasks, indicating hesitation and disrupted sequencing of movements as they struggle to translate cognitive intentions into fluent motor output [-]. Paper_time, conversely, correlates with motor execution deficits caused by degeneration of the basal ganglia and cerebellar region, which impairs fine motor control. Quantitative tablet-based studies have reported that kinematic and pressure features, such as time in air, speed, and complexity, can correctly classify healthy controls, patients with mild cognitive impairment, and patients with AD, with roughly 70%-80% accuracy in 3-group models and above 90% accuracy in several pairwise contrasts. However, it was observed that the accuracy of classification was significantly influenced by the characteristics of the groups being categorized and the particular task, with rates varying from 63.5% to 100%. Moreover, the diagnostic accuracy derived from kinematic measurements exhibited higher specificity in differentiating normal cognition from impaired cognition (mild cognitive impairment and AD), whereas enhanced sensitivity was seen in discriminating between the degrees of impaired cognition (mild cognitive impairment vs AD) [-]. More recent digital pen work has shown that handwriting characteristics distinguish patients with AD from controls, with the task accuracy of the XGBoost model achieving a maximum value of 96.55% and the AUC reaching up to 99.10%, proving that handwriting characteristic analysis is promising in auxiliary AD screening or AD diagnosis []. On the other hand, the spatiotemporal dynamics of simple loop handwriting separate patients with early-stage AD from healthy older adults, with about 74% classification accuracy []. These findings support total duration, time in air, and related temporal variability as sensitive quantitative markers that differentiate healthy aging, mild cognitive impairment, and AD and that can contribute to early staging and screening.
The validity of the variables is further supported by multimodal evidence. Neuroimaging and digital clock drawing studies have linked temporal handwriting features, including “time in air” and “time on paper,” to global brain and gray matter volumes, while time-based handwriting metrics have shown significant negative correlations with global cognitive scores, such as the MoCA (Montreal Cognitive Assessment), supporting the interpretation of prolonged air_time and paper_time variability as markers of cognitive decline [-]. Compared with invasive biomarkers, such as cerebrospinal fluid Aβ42, which require lumbar puncture and typically achieve diagnostic sensitivities of around 80% at a cost of approximately US $1000 per test [,], and FDG-PET (fluorodeoxyglucose positron emission tomography) scans, which show similar median sensitivity but are substantially more expensive (often around US $3000 per scan) [,], digital clock drawing metrics, such as air_time and paper_time, provide noninvasive, low-cost screening with high diagnostic accuracy for early cognitive impairment and AD [], with our SMART-Pred protocol achieving 88% sensitivity at a cost of approximately US $5 per test (our study). Their ability to capture both executive dysfunction (via planning delays) and visuospatial deficits (via erratic stroke execution) provides a holistic view of AD’s cognitive-motor integration breakdown, aligning with National Institutes of Health priorities for functional digital biomarkers [-]. This dual focus on mechanistic relevance and clinical practicality makes them ideal for scalable, early AD detection in routine care settings.
Consistent with the literature, our findings showed that random forest had a more balanced feature utilization [,], with paper_time6 showing maximum importance (100.00), followed by air_time23 (98.27) and total_time9 (94.82). The model also highlighted the significance of diverse features, such as paper_time23 (90.20), mean_speed_in_air25 (89.03), and mean_jerk_in_air17 (84.03), suggesting a more comprehensive feature integration approach [-]. Gradient boosting and XGBoost showed similar patterns but with distinct emphases []. Gradient boosting prioritized air_time16 (100.00), while XGBoost highlighted air_time15 (100.00) as the most important feature. The gradient boosting model incorporated additional variables like pressure_mean8 and mean_speed_on_paper2 in the predictions, while the XGBoost model incorporated additional variables like gmrt_in_air1 and total_time3 in the predictions. The neural network model exhibited a unique feature importance pattern, emphasizing variables that were less prominent in other models. While including air_time features, it notably prioritized gmrt_on_paper2 (100.00), mean_acc_in_air13 (98.61), and num_of_pendown19 (98.58), suggesting its ability to capture better patterns in the data, particularly those related to acceleration and movement characteristics. These findings regarding variable importance support the notion that interpretability is crucial in AI systems, especially in medical applications [-]. The ability to identify and understand underlying diagnostic biomarkers is not only beneficial for enhancing the diagnostic process but also plays a vital role in tailoring treatment plans for patients [-].
ROC curve analysis provided a comparative evaluation of various ML models when assessed with all available variables (full feature model) versus a carefully selected subset of the most important variables (“air_time” and “paper_time” variables; reduced feature model). In the full feature model, AdaBoost demonstrated exceptional predictive capability with an AUC of 0.92, while decision tree showed relatively lower performance with an AUC of 0.73. Strong performance was also noted for random forest and gradient boosting (both with an AUC of 0.89) and XGBoost and neural network (both with an AUC of 0.89). In the reduced feature model, neural network and SVM classifiers outperformed all other algorithms (AUC values of 0.94 and 0.92, respectively), revealing their effectiveness even with fewer inputs. Ensemble models, including random forest, gradient boosting, XGBoost, and AdaBoost, continued to perform well, achieving AUC values ranging from 0.87 to 0.91, which reflect their robustness [-]. Interestingly, the decision tree’s AUC improved to 0.81, indicating that it benefited significantly from targeted variable selection in the reduced feature model. Consistent with the literature, our results from the SMART-Pred analysis highlight the consistent strength of ensemble models, such as random forest, gradient boosting, XGBoost, and AdaBoost, across varying variable sets and support the strategic approach of leveraging variable importance ranking to enhance model performance [-]. By focusing on critical features, models can maintain or even improve accuracy, as demonstrated in this work. However, it should be emphasized that nothing in this work demonstrated causality between these variables and AD.
Comparison of SMART-Pred With Other AI Tools for AD Diagnostics
SMART-Pred offers a unique approach to AD diagnostics by employing handwriting analysis as a noninvasive biomarker, distinguishing itself from other AI tools in the field. Unlike most existing tools that use cognitive tests, magnetic resonance imaging scans, or blood-based markers, SMART-Pred leverages the DARWIN dataset, which includes handwriting samples from 174 participants (89 patients with AD and 85 healthy controls) across 25 tasks, identifying critical predictors like air_time and paper_time. This focus provides a novel, cost-effective method for early AD detection. The tool integrates 10 diverse ML models, including random forest, AdaBoost, and neural network, within a user-friendly R Shiny application, democratizing access to advanced AI for nonspecialists and enabling robust performance comparison and customization. With neural network achieving 91% accuracy and SVM achieving 89% accuracy on the test set of the reduced feature model, SMART-Pred surpasses many clinical diagnostic tools (typically having around 81% accuracy) using noninvasive data, and its emphasis on interpretability through variable importance ranking (eg, air_time and paper_time) via the “varImp()” function in R aids physician acceptance and hypothesis formulation.
SMART-Pred stands out due to its reliance on accessible handwriting data compared to resource-intensive methods like magnetic resonance imaging scans or genetic testing used by tools, such as the USC model (approximately 85% accuracy) [] and the Cambridge AI tool (82% accuracy for progression prediction) []. While a quantum ML approach has reported an exceptional 99.89% accuracy, its practical application is limited by computational complexity []. On the other hand, SMART-Pred prioritizes ease of use and scalability. Unlike the UCSF model, which predicts AD risk up to 7 years in advance using clinical records (72% accuracy), SMART-Pred focuses on immediate diagnostic classification for direct clinical application. Its ability to bridge complex AI algorithms with user-friendly interfaces addresses a critical gap in clinical adoption, and this is further enhanced by its focus on interpretability, a challenge for many AI models. By pinpointing specific handwriting features as predictors, SMART-Pred not only facilitates early detection but also contributes to understanding potential risk factors, paving the way for tailored interventions.
The interplay between AI interpretability and physician acceptance is a key factor in integrating AI systems into health care. SMART-Pred, a multifaceted R Shiny AI-based predictive modeling application, demonstrated robust predictive power in AD diagnosis using the DARWIN dataset. Incorporating 10 ML algorithms with tunable parameters, it achieved optimal results, with ensemble models like random forest and AdaBoost excelling across evaluation metrics. This provides an efficient, customizable approach to AI modeling, particularly for complex datasets and disease diagnosis, which is extendable to other classification-based scenarios. Its ability to identify potential risk factors makes it valuable for guiding hypothesis formulation and experimentation across health care domains. Research indicates that physicians are more likely to adopt AI technologies when the reasoning behind insights is clear, especially in complex fields like AD diagnosis and management [-].
Future studies could focus on clinical assessments to evaluate whether top variables like air_time and paper_time are sufficient for AD diagnosis, strengthening the synergy between AI outcomes and experimental findings. Additionally, SMART-Pred could be integrated into electronic health record systems to identify biomarkers of successful aging, enhance disease diagnosis and surveillance, and analyze medical records, addressing public health concerns through data-driven insights and personalized medicine strategies. The use of smart home data for predicting AD symptoms has also shown promise, highlighting the potential of diverse data sources in AD research, with SMART-Pred’s handwriting analysis offering a complementary approach.
SMART-Pred demonstrated competitive performance compared to recent AI models for AD prediction. Its unique use of handwriting analysis and accessibility as an R Shiny application contribute significantly to the field. Future work could also explore integrating multiple data modalities to enhance its predictive accuracy and robustness. Additional clinical research is needed to evaluate the practical applicability of the top-ranked variables (air_time and paper_time) in real-world AD risk assessment and diagnosis.
A comprehensive and comparative analysis of the performance of AI tools for AD detection and prediction is provided in .
| AI tool | Primary data source | Accuracy | Key features/differentiators |
| SMART-Pred (our study) | Handwriting analysis (DARWIN dataset) | 91.00% (neural network) with 95% CI: 0.79‐0.97), AUC=94% | Multi-algorithm R Shiny application (10 ML models) with automated feature selection, explainable AI (SHAP), and cross-validation; noninvasive, cost-effective, user-friendly, generalizable framework applicable to diverse biomarkers and deployable beyond handwriting analysis |
| Cambridge AI tool [] | Cognitive tests, MRI scans | 82.00% (progression prediction) | Predicts AD progression within 3 years; 3 times more accurate than standard clinical markers |
| USC model [] | MRI scans, genetic data, biomarkers | 85.00% (CNN) with control and MCI/AD groups | Uses deep learning for high accuracy in brain scan analysis; integrates diverse data types |
| Quantum ML approach [] | Unspecified (advanced computational methods) | 99.89% | Exceptional accuracy but limited by complexity and resource needs; less practical for immediate use |
| Deep learning CNN [] | Handwriting analysis (DARWIN dataset) | 90.40% | Proposes a novel CNN as a cheap, fast, and accurate solution with a high accuracy, but with no XAI |
| Particle swarm optimization (PSO) algorithm [] | Handwriting analysis (DARWIN dataset) | 90.57% (LR, AdaBoost, and RF, with RF dominating the accuracy battle with the best feature dimension of 20) | Uses a hybrid ML model for predicting AD using PSO-based feature selection, but with no XAI. |
| ML [] | Handwriting analysis (DARWIN dataset) | 85.29% (RF) | Proposes a single classifier merging the features from the best tasks, but with no XAI. |
| ML [] | DARWIN dataset analysis | 82.43% (RF) | Uses several ML techniques with selected tasks of different types and complexities to evaluate the healthy status of patients and a combination of different tasks, but with no XAI |
| ML [] | Handwriting analysis (DARWIN dataset) | 85% (XGBoost) | Uses several ensemble techniques by implementing a 5-fold cross-validation that ensured robust and unbiased results, but with no XAI |
| ML [] | Handwriting analysis (DARWIN dataset) | 85.86% (LR) | Uses several ML techniques involving extensive hyperparameter tuning, but with no XAI |
aAI: artificial intelligence.
bAD: Alzheimer disease.
cSMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling.
dDARWIN: Diagnosis Alzheimer WIth Handwriting.
eMRI: magnetic resonance imaging.
fCNN: convolutional neural network.
gMCI: mild cognitive impairment.
hML: machine learning.
iXAI: explainable artificial intelligence.
jLR: linear regression.
kAdaBoost: adaptive boosting.
lRF: random forest.
mXGBoost: extreme gradient boosting.
Practical Design Implications of SMART-Pred
The implementation of SMART-Pred, an AI-based tool for early AD detection via handwriting analysis, holds transformative practical design implications across clinical diagnostics, health care technology, and neurodegenerative research. In clinical settings, SMART-Pred provides a noninvasive, cost-effective screening option, achieving up to 91% accuracy with the neural network model (with 95% CI: 0.79‐0.97 and AUC=94%), which can be seamlessly integrated into primary care where access to advanced neuroimaging or cerebrospinal fluid testing is limited. Using simple handwriting tasks from the DARWIN dataset, the application enables early identification of motor-cognitive decline, potentially reducing diagnostic delays, especially in underserved populations. This supports value-based care by enhancing efficiency and patient outcomes, aligning with broader trends of AI adoption in health care, and prioritizing patients for further evaluation or early interventions.
Beyond clinical use, SMART-Pred influences health care technology by showcasing a user-friendly R Shiny interface that automates complex ML workflows (data preprocessing, model training, and evaluation) for nonspecialist clinical researchers and providers while addressing challenges like infrastructure compatibility and algorithmic bias. Its focus on explainability through feature importance ranking (air_time and paper_time) promotes transparency, building trust among clinicians who are wary of black-box systems. In neurodegenerative research, SMART-Pred inspires exploration of noninvasive biomarkers for other disorders like Parkinson disease and Huntington disease, and future validation on datasets like ADNI could lead to hybrid models combining handwriting with imaging or genetic data for enhanced precision. By prioritizing clinical utility, SMART-Pred supports future AI development in radiology and beyond, encouraging interdisciplinary collaboration and clinician involvement in design to address real-world health care needs and drive innovation in diagnostics.
Limitations and Future Work
The SMART-Pred tool has demonstrated promising capabilities for AD prediction through handwriting analysis, but it faces a few critical limitations in its basic version. Computationally, its exclusive reliance on CPU processing within the R environment restricts parallel computing capabilities, leading to inefficiencies when handling large datasets or complex ML algorithms like neural network []. This architectural constraint directly impacts scalability, particularly for resource-intensive tasks such as hyperparameter optimization and multimodal data integration. A fundamental reproducibility challenge stems from the inability to save key outputs (model metrics, ROC curves, and trained model states), hindering longitudinal studies and comparative analyses [-]. These limitations collectively restrict the tool’s utility in real-world clinical workflows.
To address these issues, the authors propose an advanced version integrating Python-based frameworks like TensorFlow for GPU-accelerated neural network models and enhanced computational efficiency. Planned upgrades include flexible hyperparameter optimization, automated parameter distribution analysis, and comprehensive result-saving functionalities. Future validation will expand beyond the DARWIN dataset to incorporate multimodal data from the ADNI, encompassing neuroimaging, genetic, and biochemical biomarkers. This broader validation aims to assess generalizability across diverse populations and clinical settings while enabling cross-modal biomarker comparisons, a critical step toward establishing SMART-Pred as a clinically actionable tool for early AD detection.
Comparison With Prior Work
Our study on SMART-Pred aligns with the growing trend of using AI for the early detection and prediction of AD. The performance of our model (91% accuracy with neural network, 95% CI: 0.79‐0.97; AUC=94%) compares favorably with recent advancements in the field. presents a concise comparison of studies that used ML, explainable AI, and the DARWIN dataset for AD detection.
| Dataset | ML techniques used | XAI methods applied | Key insights and features identified | Performance metrics |
| DARWIN dataset analysis [] | LR, RF, SVM, and EBM | SHAP and EBM interpretability | “air_time” (in-air movement duration) identified as the most decisive feature group | 89% (SD 5%) accuracy with EBM on the reduced dimensionality dataset |
| DARWIN dataset analysis [] | ML and stacking | SHAP XAI features importance and direction of influence | “air_time” and “paper_time” (pen contact duration) features dominate the feature importance analysis using SHAP | 88.57% accuracy with the stacking ensemble model |
| DARWIN dataset analysis [] | LR, AdaBoost, XGBoost, KNN, RF, SVM, etc | LIME and ELI5 | “air_time” and “paper_time” features were the most important features using ELI5 | 91.00% accuracy with RF |
| SMART-Pred (current study) | LR, AdaBoost, XGBoost, KNN, RF, SVM, neural network, etc | SHAP features importance and direction of influence | “air_time” and “paper_time” as primary biomarkers using SHAP XAI. | 91.00% accuracy with neural network; 95% CI 0.79‐0.97; AUC=94% |
aML: machine learning.
bXAI: explainable artificial intelligence.
cDARWIN: Diagnosis Alzheimer WIth Handwriting.
dLR: logistic regression.
eRF: random forest.
fSVM: support vector machine.
gEBM: explainable boosting machine.
hSHAP: Shapley Additive Explanations.
iAdaBoost: adaptive boosting.
jXGBoost: extreme gradient boosting.
kKNN: k-nearest neighbors.
lLIME: Local Interpretable Model-Agnostic Explanations.
mELI5: Explain Like I’m 5.
nSMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling.
The comparative analysis revealed 3 critical insights. First, the studies demonstrate feature consensus, identifying temporal handwriting features like “air_time” as pivotal predictors, thereby validating their biological relevance to AD motor-cognitive decline through their association with prefrontal cortex dysfunction and motor planning deficits. Second, SMART-Pred represents a methodological evolution from prior work by expanding model diversity (10 algorithms vs up to a maximum of 7 algorithms in earlier studies), incorporating neural networks to capture complex nonlinear patterns in handwriting dynamics, and advancing biomarker identification through dual focus on “air_time” (in-air movement duration) and “paper_time” (pen contact duration) rather than single-feature analysis. Third, despite architectural differences, SMART-Pred classifiers performed very well (with neural network achieving an accuracy of 91% [95% CI 0.79-0.97]), highlighting the DARWIN dataset’s robustness for handwriting-based diagnosis while indicating that model selection impacts biomarker discovery more than absolute accuracy in this domain. This progression from single-feature interpretation to multi-algorithm, multibiomarker analysis reflects an intentional shift toward clinical translatability, balancing explainability with predictive power.
Conclusion of the Study
The SMART-Pred R Shiny application represents a significant advancement in AI-based predictive modeling for medical diagnostics, offering a user-friendly interface and comprehensive algorithm selection. This tool holds value for researchers and clinicians in personalized medicine and public health surveillance, demonstrating potential to enhance early detection and screening strategies for conditions such as AD and Parkinson disease. By bridging the gap between computational modeling and clinical application, SMART-Pred exemplifies the growing trend of leveraging interactive data visualization and AI-powered analysis in health care settings.
Future development of SMART-Pred will focus on clinical validation and address the challenges associated with imbalanced datasets. The roadmap includes expanding algorithmic support to encompass advanced ML techniques, image recognition architectures, and analysis capabilities for diverse data types such as text, time series, and graph data. Additionally, plans to improve visualization capabilities and user interface design aim to increase accessibility and power. These ongoing enhancements are designed to create a more flexible and comprehensive platform for data analysis across various scientific domains, solidifying SMART-Pred’s role in advancing predictive modeling in health care and beyond.
Acknowledgments
This work was supported by start-up funds provided to SK by Celia Scott Weatherhead School of Public Health and Tropical Medicine at Tulane University. The article processing charges for this manuscript were supported by the Weatherhead Foundation through a gift to the Tulane University Celia Scott Weatherhead School of Public Health & Tropical Medicine in support of the School’s AI and Data Science Strategic Initiative. This generous support has enabled us to advance work that has strong potential to contribute to the development and application of AI in public health, in alignment with the Dean’s strategic AI initiative.
Funding
This work was supported by start-up funds provided to SK by Celia Scott Weatherhead School of Public Health and Tropical Medicine at Tulane University.
Data Availability
The DARWIN (Diagnosis Alzheimer With Handwriting) dataset [,] was used to validate the classification capabilities of the SMART-Pred (Shiny Multi-Algorithm R Tool for Predictive Modeling) application. A fully reproducible workflow using the provided R codes, including all necessary packages and code snippets for SMART-Pred, is available in our GitHub repository []. These scripts are designed to facilitate reproducibility and allow users to replicate the machine learning classification tasks outlined in our study.
Authors' Contributions
Conceptualization: SK
Data curation: HW, EFA, SK
Formal analysis: HW, EFA, SK, SKS
Funding acquisition: SK
Investigation: HW, EFA, SK, SKS
Methodology: HW, EFA, SK
Project administration: SK
Resources: SK
Software: HW, EFA, SK
Supervision: SK, SKS, JGS
Validation: HW, EFA, SK, SKS, JGS
Visualization: HW, EFA, SK
Writing – original draft: SK, HW
Writing – review & editing: SK, EFA, JGS, SKS
Conflicts of Interest
None declared.
Multimedia Appendix 1
SMART-Pred (Shiny Multi-Algorithm R Tool for Predictive Modeling) application and data.
ZIP File, 356 KBMultimedia Appendix 2
Shapley Additive Explanations analysis results for the additional machine learning classifiers.
DOCX File, 125 KBReferences
- Alowais SA, Alghamdi SS, Alsuhebany N, et al. Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med Educ. Sep 22, 2023;23(1):689. [CrossRef] [Medline]
- Choubey A, Choubey SB, K P, Daulatabad VS, John N. Healthcare transformation: artificial intelligence is the dire imperative of the day. Cureus. Jun 2024;16(6):e62652. [CrossRef] [Medline]
- Bajwa J, Munir U, Nori A, Williams B. Artificial intelligence in healthcare: transforming the practice of medicine. Future Healthc J. Jul 2021;8(2):e188-e194. [CrossRef] [Medline]
- Khalifa M, Albadawy M. AI in diagnostic imaging: revolutionising accuracy and efficiency. Computer Methods and Programs in Biomedicine Update. 2024;5:100146. [CrossRef]
- Hosny A, Parmar C, Quackenbush J, Schwartz LH, Aerts H. Artificial intelligence in radiology. Nat Rev Cancer. Aug 2018;18(8):500-510. [CrossRef] [Medline]
- Yu F, Moehring A, Banerjee O, Salz T, Agarwal N, Rajpurkar P. Heterogeneity and predictors of the effects of AI assistance on radiologists. Nat Med. Mar 2024;30(3):837-849. [CrossRef] [Medline]
- Shafi S, Parwani AV. Artificial intelligence in diagnostic pathology. Diagn Pathol. Oct 3, 2023;18(1):109. [CrossRef] [Medline]
- Sarwar S, Dent A, Faust K, et al. Physician perspectives on integration of artificial intelligence into diagnostic pathology. NPJ Digit Med. 2019;2(1):28. [CrossRef] [Medline]
- Ghenciu LA, Dima M, Stoicescu ER, Iacob R, Boru C, Hațegan OA. Retinal imaging-based oculomics: artificial intelligence as a tool in the diagnosis of cardiovascular and metabolic diseases. Biomedicines. Sep 23, 2024;12(9):2150. [CrossRef] [Medline]
- Kong M, Song SJ. Artificial intelligence applications in diabetic retinopathy: what we have now and what to expect in the future. Endocrinol Metab (Seoul). Jun 2024;39(3):416-424. [CrossRef] [Medline]
- Lopez-Jimenez F, Attia Z, Arruda-Olson AM, et al. Artificial intelligence in cardiology: present and future. Mayo Clin Proc. May 2020;95(5):1015-1039. [CrossRef] [Medline]
- Li Z, Yin S, Wang S, Wang Y, Qiang W, Jiang J. Transformative applications of oculomics-based AI approaches in the management of systemic diseases: a systematic review. J Adv Res. Sep 2025;75:679-696. [CrossRef]
- Barry T, Farina JM, Chao CJ, et al. The role of artificial intelligence in echocardiography. J Imaging. Feb 20, 2023;9(2):9. [CrossRef] [Medline]
- Maleki Varnosfaderani S, Forouzanfar M. The role of AI in hospitals and clinics: transforming healthcare in the 21st century. Bioengineering (Basel). Mar 29, 2024;11(4):337. [CrossRef] [Medline]
- Olawade DB, Wada OZ, Odetayo A, David-Olawade AC, Asaolu F, Eberhardt J. Enhancing mental health with artificial intelligence: current trends and future prospects. Journal of Medicine, Surgery, and Public Health. Aug 2024;3:100099. [CrossRef]
- Bekbolatova M, Mayer J, Ong CW, Toma M. Transformative potential of AI in healthcare: definitions, applications, and navigating the ethical landscape and public perspectives. Healthcare (Basel). Jan 5, 2024;12(2):125. [CrossRef] [Medline]
- Quer G, Arnaout R, Henne M, Arnaout R. Machine learning and the future of cardiovascular care: JACC state-of-the-art review. J Am Coll Cardiol. Jan 26, 2021;77(3):300-313. [CrossRef] [Medline]
- Drachman DA. Aging of the brain, entropy, and Alzheimer disease. Neurology (ECronicon). Oct 24, 2006;67(8):1340-1352. [CrossRef] [Medline]
- Alzheimer’s Association. 2018 Alzheimer’s disease facts and figures. Alzheimer’s & Dementia. Mar 2018;14(3):367-429. [CrossRef]
- Breijyeh Z, Karaman R. Comprehensive review on Alzheimer’s disease: causes and treatment. Molecules. Dec 8, 2020;25(24):5789. [CrossRef] [Medline]
- Kumar A, Sidhu J, Lui F, Tsao JW. Alzheimer Disease. StatPearls Publishing; 2023. URL: https://www.ncbi.nlm.nih.gov/books/NBK499922/ [Accessed 2025-11-29]
- Jahn H. Memory loss in Alzheimer’s disease. Dialogues Clin Neurosci. Dec 2013;15(4):445-454. [CrossRef] [Medline]
- Crecelius C. Diagnosis and treatment of non-Alzheimer’s dementias. J Am Med Dir Assoc. 2003;4(4 Suppl):H25-H29. [CrossRef] [Medline]
- Giau VV, Senanarong V, Bagyinszky E, An SSA, Kim S. Analysis of 50 neurodegenerative genes in clinically diagnosed early-onset Alzheimer’s disease. Int J Mol Sci. Mar 26, 2019;20(6):1514. [CrossRef] [Medline]
- Kennison M, Long E. The long journey of Alzheimer’s disease. J Christ Nurs. 2018;35(4):218-227. [CrossRef] [Medline]
- Cummings JL. Current perspectives in Alzheimer’s disease. Neurology (ECronicon). Jul 1998;51(1_suppl_1):S2-S17. [CrossRef]
- Mendez MF. Early-onset Alzheimer disease. Neurol Clin. May 2017;35(2):263-281. [CrossRef] [Medline]
- DiLoreto R, Murphy CT. The cell biology of aging. Mol Biol Cell. Dec 15, 2015;26(25):4524-4531. [CrossRef] [Medline]
- Bouska M, Huang K, Kang P, Bai H. Organelle aging: lessons from model organisms. J Genet Genomics. Apr 20, 2019;46(4):171-185. [CrossRef] [Medline]
- Niccoli T, Partridge L. Ageing as a risk factor for disease. Curr Biol. Sep 11, 2012;22(17):R741-R752. [CrossRef] [Medline]
- Morawe T, Hiebel C, Kern A, Behl C. Protein homeostasis, aging and Alzheimer’s disease. Mol Neurobiol. Aug 2012;46(1):41-54. [CrossRef] [Medline]
- Cummings JL, Gonzalez MI, Pritchard MC, May PC, Toledo-Sherman LM, Harris GA. The therapeutic landscape of tauopathies: challenges and prospects. Alz Res Therapy. 2023;15(1):168. [CrossRef]
- Pais M, Martinez L, Ribeiro O, et al. Early diagnosis and treatment of Alzheimer’s disease: new definitions and challenges. Braz J Psychiatry. Aug 2020;42(4):431-441. [CrossRef] [Medline]
- Nayab DE, Din FU, Ali H, et al. Nano biomaterials based strategies for enhanced brain targeting in the treatment of neurodegenerative diseases: an up-to-date perspective. J Nanobiotechnology. Dec 13, 2023;21(1):477. [CrossRef] [Medline]
- Singh A, Dawson TM, Kulkarni S. Neurodegenerative disorders and gut-brain interactions. J Clin Invest. Jul 1, 2021;131(13):13. [CrossRef] [Medline]
- GBD 2016 Dementia Collaborators. Global, regional, and national burden of Alzheimer’s disease and other dementias, 1990-2016: a systematic analysis for the Global Burden of Disease Study 2016. Lancet Neurol. Jan 2019;18(1):88-106. [CrossRef] [Medline]
- Maresova P, Hruska J, Klimova B, Barakovic S, Krejcar O. Activities of daily living and associated costs in the most widespread neurodegenerative diseases: a systematic review. Clin Interv Aging. 2020;15:1841-1862. [CrossRef] [Medline]
- Alzheimer’s Association. 2016 Alzheimer’s disease facts and figures. Alzheimer’s & Dementia. Apr 2016;12(4):459-509. [CrossRef]
- Rizzi L, Rosset I, Roriz-Cruz M. Global epidemiology of dementia: Alzheimer’s and vascular types. Biomed Res Int. 2014;2014:908915. [CrossRef] [Medline]
- Aisen PS, Cummings J, Jack CR Jr, et al. On the path to 2025: understanding the Alzheimer’s disease continuum. Alzheimers Res Ther. Aug 9, 2017;9(1):60. [CrossRef] [Medline]
- Bowroju SK, Mainali N, Ayyadevara S, et al. Design and synthesis of novel hybrid 8-hydroxy quinoline-indole derivatives as inhibitors of Aβ self-aggregation and metal chelation-induced Aβ aggregation. Molecules. Aug 8, 2020;25(16):16. [CrossRef] [Medline]
- Kakraba S, Ayyadevara S, Penthala NR, et al. A novel microtubule-binding drug attenuates and reverses protein aggregation in animal models of Alzheimer’s disease. Front Mol Neurosci. 2019;12:310. [CrossRef] [Medline]
- Kakraba S, Ayyadevara S, Mainali N, et al. Thiadiazolidinone (TDZD) analogs inhibit aggregation-mediated pathology in diverse neurodegeneration models, and extend C. elegans life- and healthspan. Pharmaceuticals (Basel). Oct 20, 2023;16(10):10. [CrossRef] [Medline]
- Ayyadevara S, Balasubramaniam M, Kakraba S, Alla R, Mehta JL, Shmookler Reis RJ. Aspirin-mediated acetylation protects against multiple neurodegenerative pathologies by impeding protein aggregation. Antioxid Redox Signal. Dec 10, 2017;27(17):1383-1396. [CrossRef] [Medline]
- Lee LY, Vaghari D, Burkhart MC, et al. Robust and interpretable AI-guided marker for early dementia prediction in real-world clinical settings. EClinicalMedicine. Aug 2024;74:102725. [CrossRef] [Medline]
- Schöll M, Verberk IMW, Del Campo M, et al. Challenges in the practical implementation of blood biomarkers for Alzheimer’s disease. Lancet Healthy Longev. Oct 2024;5(10):100630. [CrossRef] [Medline]
- Agostinho D, Simões M, Castelo-Branco M. Predicting conversion from mild cognitive impairment to Alzheimer’s disease: a multimodal approach. Brain Commun. 2024;6(4):fcae208. [CrossRef] [Medline]
- Chudzik A, Śledzianowski A, Przybyszewski AW. Machine learning and digital biomarkers can detect early stages of neurodegenerative diseases. Sensors (Basel). Feb 29, 2024;24(5):1572. [CrossRef] [Medline]
- Olawade DB, David-Olawade AC, Wada OZ, Asaolu AJ, Adereni T, Ling J. Artificial intelligence in healthcare delivery: prospects and pitfalls. Journal of Medicine, Surgery, and Public Health. Aug 2024;3:100108. [CrossRef]
- Chakraborty C, Bhattacharya M, Pal S, Lee SS. From machine learning to deep learning: advances of the recent data-driven paradigm shift in medicine and healthcare. Current Research in Biotechnology. 2024;7:100164. [CrossRef]
- Jamarani A, Haddadi S, Sarvizadeh R, Haghi Kashani M, Akbari M, Moradi S. Big data and predictive analytics: a systematic review of applications. Artif Intell Rev. 2024;57(7):176. [CrossRef]
- Soori M, Arezoo B, Dastres R. Artificial intelligence, machine learning and deep learning in advanced robotics, a review. Cognitive Robotics. 2023;3:54-70. [CrossRef]
- Kakraba S, Yadem AC, Abraham KE. Unraveling protein secrets: machine learning unveils novel biologically significant associations among amino acids. Preprints. Preprint posted online on May 6, 2025. [CrossRef]
- Yang Z, Zhou H, Srivastav S, et al. Optimizing Parkinson’s disease prediction: a comparative analysis of data aggregation methods using multiple voice recordings via an automated artificial intelligence pipeline. Preprints. Preprint posted online on Dec 5, 2024. [CrossRef]
- Parkinson-s-disease-prediction-code-and-data-repository-main. ResearchGate. 2025. URL: https://www.researchgate.net/publication/387796710_Parkinson-s-Disease-Prediction-Code-and-Data-Repository-main [Accessed 2025-11-29]
- Netsey EK, Naandam SM, Asante Jnr J, et al. Structural and functional impacts of SARS-CoV-2 spike protein mutations: insights from predictive modeling and analytics (Preprint). JMIR Bioinform Biotechnol. Dec 8, 2025;6:e73637. [CrossRef] [Medline]
- Kakraba S, Ayyadevara S, Yadem Clement A, Abraham KE, Compadre CM, Shmookler Reis RJ. Machine Learning–Enhanced Quantitative Structure-Activity Relationship Modeling for DNA Polymerase Inhibitor Discovery: Algorithm Development and Validation. JMIR AI. Dec 3, 2025;4:e77890. [CrossRef] [Medline]
- Kakraba S, Agyemang EF, Shmookler Reis RJ. Accelerating Discovery of Leukemia Inhibitors with AI-Driven QSAR Modeling (forthcoming). JMIR AI. Dec 5, 2025;05/12/2025:81552. URL: https://preprints.jmir.org/preprint/81552 [CrossRef] [Medline]
- Mahesh B. Machine learning algorithms - a review. International Journal of Science and Research. Jan 5, 2020;9(1):381-386. [CrossRef]
- Netsey EK, Kakraba S, Naandam SM, Yadem AC. A mathematical graph-theoretic model of single point mutations associated with sickle cell anemia disease. JBT. 2021;9:1-14. [CrossRef]
- Kakraba S, Knisley D. A graph-theoretic model of single point mutations in the cystic fibrosis transmembrane conductance regulator. JBT. 2016;6(1):780-786. [CrossRef]
- Kakraba S. A hierarchical graph for nucleotide binding domain 2. East Tennessee State University; 2015. URL: https://dc.etsu.edu/etd/2517/ [Accessed 2025-11-29]
- Ahsan MM, Luna SA, Siddique Z. Machine-learning-based disease diagnosis: a comprehensive review. Healthcare (Basel). Mar 15, 2022;10(3):541. [CrossRef] [Medline]
- Shivahare BD, Singh J, Ravi V, et al. Delving into machine learning’s influence on disease diagnosis and prediction. The Open Public Health Journal. Apr 26, 2024;17(1). [CrossRef]
- Xu D, Xu Z. Machine learning applications in preventive healthcare: a systematic literature review on predictive analytics of disease comorbidity from multiple perspectives. Artif Intell Med. Oct 2024;156:102950. [CrossRef] [Medline]
- Park DJ, Park MW, Lee H, Kim YJ, Kim Y, Park YH. Development of machine learning model for diagnostic disease prediction based on laboratory tests. Sci Rep. Apr 7, 2021;11(1):7567. [CrossRef] [Medline]
- Dixon D, Sattar H, Moros N, et al. Unveiling the influence of AI predictive analytics on patient outcomes: a comprehensive narrative review. Cureus. May 2024;16(5):e59954. [CrossRef] [Medline]
- Jia L, Yao W, Jiang Y, et al. Development of interactive biological web applications with R/Shiny. Brief Bioinformatics. Jan 17, 2022;23(1). [CrossRef]
- Giorgi FM, Ceraolo C, Mercatelli D. The R language: an engine for bioinformatics and data science. Life (Basel). Apr 27, 2022;12(5):648. [CrossRef] [Medline]
- Perkins MA, Carrier JW. Using R Shiny to develop a dashboard using IPEDS, U.S. Census, and bureau of labor statistics data. PLOS ONE. 2023;18(1):e0278573. [CrossRef] [Medline]
- Bastian G, Baker GH, Limon A. Bridging the divide between data scientists and clinicians. Intelligence-Based Medicine. 2022;6:100066. [CrossRef]
- Singh P. Systematic review of data-centric approaches in artificial intelligence and machine learning. Data Science and Management. Sep 2023;6(3):144-157. [CrossRef]
- Cilia ND, De Gregorio G, De Stefano C, Fontanella F, Marcelli A, Parziale A. Diagnosing Alzheimer’s disease from on-line handwriting: a novel dataset and performance benchmarking. Eng Appl Artif Intell. May 2022;111:104822. [CrossRef]
- DARWIN. UCI Machine Learning Repository. URL: https://archive.ics.uci.edu/dataset/732/darwin [Accessed 2025-11-29]
- Nohara Y, Matsumoto K, Soejima H, Nakashima N. Explanation of machine learning models using shapley additive explanation and application for real data in hospital. Comput Methods Programs Biomed. Feb 2022;214:106584. [CrossRef] [Medline]
- Ponce‐Bobadilla AV, Schmitt V, Maier CS, Mensing S, Stodtmann S. Practical guide to SHAP analysis: explaining supervised machine learning model predictions in drug development. Clinical Translational Sci. Nov 2024;17(11):e70056. [CrossRef]
- Feng J, Liang J, Qiang Z, et al. A hybrid stacked ensemble and Kernel SHAP-based model for intelligent cardiotocography classification and interpretability. BMC Med Inform Decis Mak. Nov 28, 2023;23(1):273. [CrossRef] [Medline]
- The R Core Team. R: a language and environment for statistical computing. R Project; 2024. URL: https://cran.r-project.org/doc/manuals/r-release/fullrefman.pdf [Accessed 2025-11-29]
- Weihs C, Ligges U, Luebke K, Raabe N. KlaR analyzing German business cycles. In: Baier D, Decker R, Schmidt-Thieme L, editors. Data Analysis and Decision Support Studies in Classification, Data Analysis, and Knowledge Organization. Springer; 2005:335-343. [CrossRef]
- Kuhn M. Building predictive models in R using the caret package. J Stat Soft. 2008;28(5):1-26. [CrossRef]
- Therneau T, Atkinson B, Ripley B. rpart: recursive partitioning and regression trees. R Project; 2023. URL: https://cran.r-project.org/web/packages/rpart/rpart.pdf [Accessed 2025-11-29]
- Liaw A, Wiener M. Classification and regression by randomForest. R News. 2002;2(3):18-22. URL: https://journal.r-project.org/articles/RN-2002-022/RN-2002-022.pdf [Accessed 2025-11-29]
- Ridgeway G, Edwards D, Kriegler B, et al. gbm: generalized boosted regression models. R Project; 2024. URL: https://cran.r-project.org/web/packages/gbm/gbm.pdf [Accessed 2025-11-29]
- Culp M, Johnson K, Michailidis G. The R package ada for stochastic boosting. R Project; 2016. URL: https://cran.r-project.org/web/packages/ada/ada.pdf [Accessed 2025-11-29]
- Chen T, He T. xgboost: extreme gradient boosting. R Project; 2024. URL: https://cran.r-project.org/web/packages/xgboost/vignettes/xgboost.pdf [Accessed 2025-11-29]
- Venables WN, Ripley BD. Modern Applied Statistics with S. 4th ed. Springer; 2002. [CrossRef]
- Uddin MF. Addressing accuracy paradox using enhanched weighted performance metric in machine learning. Presented at: Sixth HCT Information Technology Trends (ITT); Nov 20-21, 2019. [CrossRef]
- Salihoğlu YS, Uslu Erdemir R, Aydur Püren B, et al. Diagnostic performance of machine learning models based on 18F-FDG PET/CT radiomic features in the classification of solitary pulmonary nodules. Mol Imaging Radionucl Ther. Jun 27, 2022;31(2):82-88. [CrossRef] [Medline]
- Ávila-Jiménez JL, Cantón-Habas V, Carrera-González MDP, Rich-Ruiz M, Ventura S. A deep learning model for Alzheimer’s disease diagnosis based on patient clinical records. Comput Biol Med. Feb 2024;169:107814. [CrossRef] [Medline]
- Sun Y, Zhang J, Zhang Y. Adaboost algorithm combined multiple random forest models (Adaboost-RF) is employed for fluid prediction using well logging data. Phys Fluids (1994). Jan 1, 2024;36(1). [CrossRef]
- Hossain S, Hasan MK, Faruk MO, Aktar N, Hossain R, Hossain K. Machine learning approach for predicting cardiovascular disease in Bangladesh: evidence from a cross-sectional study in 2023. BMC Cardiovasc Disord. Apr 18, 2024;24(1):214. [CrossRef] [Medline]
- Ghasemieh A, Lloyed A, Bahrami P, Vajar P, Kashef R. A novel machine learning model with Stacking Ensemble Learner for predicting emergency readmission of heart-disease patients. Decision Analytics Journal. Jun 2023;7:100242. [CrossRef]
- Hussain SS, Zaidi SSH. AdaBoost ensemble approach with weak classifiers for gear fault diagnosis and prognosis in DC motors. Appl Sci (Basel). 2024;14(7):3105. [CrossRef]
- Dostmohammadi M, Pedram MZ, Hoseinzadeh S, Garcia DA. A GA-stacking ensemble approach for forecasting energy consumption in a smart household: a comparative study of ensemble methods. J Environ Manage. Jul 2024;364:121264. [CrossRef] [Medline]
- Sarica A, Cerasa A, Quattrone A. Random forest algorithm for the classification of neuroimaging data in Alzheimer’s disease: a systematic review. Front Aging Neurosci. 2017;9:329. [CrossRef] [Medline]
- Moore PJ, Lyons TJ, Gallacher J, Alzheimer’s Disease Neuroimaging Initiative. Random forest prediction of Alzheimer’s disease using pairwise selection from time series data. PLoS ONE. 2019;14(2):e0211558. [CrossRef] [Medline]
- Tripoliti EE, Fotiadis DI, Argyropoulou M, Manis G. A six stage approach for the diagnosis of the Alzheimer’s disease based on fMRI data. J Biomed Inform. Apr 2010;43(2):307-320. [CrossRef] [Medline]
- Song M, Jung H, Lee S, Kim D, Ahn M. Diagnostic classification and biomarker identification of Alzheimer’s disease with random forest algorithm. Brain Sci. Apr 2, 2021;11(4):453. [CrossRef] [Medline]
- Metcalf CJE, Tate AT, Graham AL. Demographically framing trade-offs between sensitivity and specificity illuminates selection on immunity. Nat Ecol Evol. Nov 2017;1(11):1766-1772. [CrossRef] [Medline]
- Dimitriadis SI, Liparas D, Tsolaki MN. Random forest feature selection, fusion and ensemble strategy: combining multiple morphological MRI measures to discriminate among healhy elderly, MCI, cMCI and Alzheimer’s disease patients: from the Alzheimer’s disease neuroimaging initiative (ADNI) database. J Neurosci Methods. May 15, 2018;302:14-23. [CrossRef] [Medline]
- Shree HV, Mouleeswaran SK. A comprehensive study of predictive models for mortality risk assessment in medical data. Presented at: 4th International Conference on Data Engineering and Communication Systems (ICDECS); Mar 22-23, 2024. [CrossRef]
- Borkan D, Dixon L, Sorensen J, Thain N, Vasserman L. Nuanced metrics for measuring unintended bias with real data for text classification. Presented at: WWW ’19: Companion Proceedings of The 2019 World Wide Web Conference; May 13-17, 2019. [CrossRef]
- Dumitrescu E, Hué S, Hurlin C, Tokpavi S. Machine learning for credit scoring: improving logistic regression with non-linear decision-tree effects. Eur J Oper Res. Mar 2022;297(3):1178-1192. [CrossRef]
- Zhang YP, Zhang XY, Cheng YT, et al. Artificial intelligence-driven radiomics study in cancer: the role of feature engineering and modeling. Military Med Res. 2023;10(1):22. [CrossRef]
- Zhang S, Pan M. An AutoML-powered analysis framework for forest fire forecasting: adapting to climate change dynamics. Atmosphere (Basel). 2024;15(12):1481. [CrossRef]
- Sui Q, Ghosh SK. Active learning for stacking and AdaBoost-related models. Stats. 2024;7(1):110-137. [CrossRef]
- Ahmed SF, Alam M, Hassan M, et al. Deep learning modelling techniques: current progress, applications, advantages, and challenges. Artif Intell Rev. Nov 2023;56(11):13521-13617. [CrossRef] [Medline]
- Khalil M, AlSayed A, Liu Y, Vanrolleghem PA. An integrated feature selection and hyperparameter optimization algorithm for balanced machine learning models predicting N2O emissions from wastewater treatment plants. Journal of Water Process Engineering. Jun 2024;63:105512. [CrossRef]
- Xu J, Luo Y, Peng K, et al. Supplementary motor area driving changes of structural brain network in blepharospasm. Brain (Bacau). Apr 19, 2023;146(4):1542-1553. [CrossRef] [Medline]
- Wang L, Zhao J. Artificial intelligence. In: Strategic Blueprint for Enterprise Analytics Studies in Big Data, Vol 150. Springer; 2024:111-138. [CrossRef]
- Yadav S, Singh A, Singhal R, Yadav JP. Revolutionizing drug discovery: the impact of artificial intelligence on advancements in pharmacology and the pharmaceutical industry. Intelligent Pharmacy. Jun 2024;2(3):367-380. [CrossRef]
- ZainEldin H, Gamel SA, Talaat FM, et al. Silent no more: a comprehensive review of artificial intelligence, deep learning, and machine learning in facilitating deaf and mute communication. Artif Intell Rev. 2024;57(7):188. [CrossRef]
- Pinto-Coelho L. How artificial intelligence is shaping medical imaging technology: a survey of innovations and applications. Bioengineering (Basel). Dec 18, 2023;10(12):1435. [CrossRef] [Medline]
- Moore I, Magnante C, Embry E, et al. Doctor AI? A pilot study examining responses of artificial intelligence to common questions asked by geriatric patients. Front Artif Intell. 2024;7:1438012. [CrossRef] [Medline]
- Li R, Wang X, Lawler K, Garg S, Bai Q, Alty J. Applications of artificial intelligence to aid early detection of dementia: a scoping review on current capabilities and future directions. J Biomed Inform. Mar 2022;127:104030. [CrossRef] [Medline]
- Ahuja S, Zaheer S. Advancements in pathology: digital transformation, precision medicine, and beyond. J Pathol Inform. Jan 2025;16:100408. [CrossRef] [Medline]
- Bohr A, Memarzadeh K. Chapter 2 - the rise of artificial intelligence in healthcare applications. In: Bohr A, Memarzadeh K, editors. Artificial Intelligence in Healthcare. Academic Press; 2020:25-60. [CrossRef]
- Hirani R, Noruzi K, Khuram H, et al. Artificial intelligence and healthcare: a journey through history, present innovations, and future possibilities. Life (Basel). Apr 26, 2024;14(5):557. [CrossRef] [Medline]
- Iqbal J, Cortés Jaimes DC, Makineni P, et al. Reimagining healthcare: unleashing the power of artificial intelligence in medicine. Cureus. Sep 2023;15(9):e44658. [CrossRef] [Medline]
- Arya SS, Dias SB, Jelinek HF, Hadjileontiadis LJ, Pappa AM. The convergence of traditional and digital biomarkers through AI-assisted biosensing: a new era in translational diagnostics? Biosens Bioelectron. Sep 1, 2023;235:115387. [CrossRef] [Medline]
- Singh M, Kumar A, Khanna NN, et al. Artificial intelligence for cardiovascular disease risk assessment in personalised framework: a scoping review. EClinicalMedicine. Jul 2024;73:102660. [CrossRef] [Medline]
- DeGroat W, Abdelhalim H, Peker E, et al. Multimodal AI/ML for discovering novel biomarkers and predicting disease using multi-omics profiles of patients with cardiovascular diseases. Sci Rep. Nov 3, 2024;14(1):26503. [CrossRef] [Medline]
- Kapustina O, Burmakina P, Gubina N, Serov N, Vinogradov V. User-friendly and industry-integrated AI for medicinal chemists and pharmaceuticals. Artificial Intelligence Chemistry. Dec 2024;2(2):100072. [CrossRef]
- Olawade DB, Wada OZ, Ige AO, Egbewole BI, Olojo A, Oladapo BI. Artificial intelligence in environmental monitoring: advancements, challenges, and future directions. Hygiene and Environmental Health Advances. Dec 2024;12:100114. [CrossRef]
- Yelne S, Chaudhary M, Dod K, Sayyad A, Sharma R. Harnessing the power of AI: a comprehensive review of its impact and challenges in nursing science and healthcare. Cureus. Nov 2023;15(11):e49252. [CrossRef] [Medline]
- Blanco-González A, Cabezón A, Seco-González A, et al. The role of AI in drug discovery: challenges, opportunities, and strategies. Pharmaceuticals (Basel). Jun 18, 2023;16(6):891. [CrossRef] [Medline]
- Kakraba S, Ayyadevara S, Yadem Clement A, Abraham KE, Compadre CM, Reis RJS. Advanced QSAR modeling with machine learning for drug discovery: targeting DNA polymerase inhibitors (Preprint). JMIR AI. Preprint posted online on Aug 16, 2025. [CrossRef]
- Ho NTN, Gonzalez P, Gogovi GK. Writing the signs: an explainable machine learning approach for Alzheimer’s disease classification from handwriting. Healthc Technol Lett. 2025;12(1):e70006. [CrossRef] [Medline]
- Onofri E, Mercuri M, Donato G, Ricci S. Cognitive fluctuations in connection to dysgraphia: a comparison of Alzheimer’s disease with dementia Lewy bodies. Clin Interv Aging. 2015;10:625-633. [CrossRef] [Medline]
- Fernandes CP, Montalvo G, Caligiuri M, Pertsinakis M, Guimarães J. Handwriting changes in Alzheimer’s disease: a systematic review. J Alzheimers Dis. 2023;96(1):1-11. [CrossRef] [Medline]
- Schröter A, Mergl R, Bürger K, Hampel H, Möller HJ, Hegerl U. Kinematic analysis of handwriting movements in patients with Alzheimer’s disease, mild cognitive impairment, depression and healthy subjects. Dement Geriatr Cogn Disord. 2003;15(3):132-142. [CrossRef] [Medline]
- Werner P, Rosenblum S, Bar-On G, Heinik J, Korczyn A. Handwriting process variables discriminating mild Alzheimer’s disease and mild cognitive impairment. J Gerontol B Psychol Sci Soc Sci. Jul 2006;61(4):228-236. [CrossRef] [Medline]
- Lanctôt KL, Boada M, Tariot PN, et al. Association between clinical dementia rating and clinical outcomes in Alzheimer’s disease. Alzheimers Dement (Amst). 2024;16(1):e12522. [CrossRef] [Medline]
- Aggarwal NT, Wilson RS, Beck TL, Bienias JL, Bennett DA. Motor dysfunction in mild cognitive impairment and the risk of incident Alzheimer disease. Arch Neurol. Dec 2006;63(12):1763-1769. [CrossRef] [Medline]
- Garre-Olmo J, Faúndez-Zanuy M, López-de-Ipiña K, Calvó-Perxas L, Turró-Garriga O. Kinematic and pressure features of handwriting and drawing: preliminary results between patients with mild cognitive impairment, Alzheimer disease and healthy controls. Curr Alzheimer Res. 2017;14(9):960-968. [CrossRef] [Medline]
- Qi H, Zhang R, Wei Z, et al. A study of auxiliary screening for Alzheimer’s disease based on handwriting characteristics. Front Aging Neurosci. 2023;15:1117250. [CrossRef] [Medline]
- Kahindo C, El-Yacoubi MA, Garcia-Salicetti S, Rigaud AS, Cristancho-Lacroix V. Characterizing early-stage Alzheimer through spatiotemporal dynamics of handwriting. IEEE Signal Process Lett. 25(8):1136-1140. [CrossRef]
- Zheng R, Zhang Y, Yang Z, Han S, Cheng J. Reduced brain gray matter volume in patients with first-episode major depressive disorder: a quantitative meta-analysis. Front Psychiatry. 2021;12:671348. [CrossRef] [Medline]
- Peng X, Srivastava S, Sutton F, Zhang Y, Badran BW, Kautz SA. Compensatory increase in ipsilesional supplementary motor area and premotor connectivity is associated with greater gait impairments: a personalized fMRI analysis in chronic stroke. Front Hum Neurosci. 2024;18:1340374. [CrossRef] [Medline]
- Koppelmans V, Silvester B, Duff K. Neural mechanisms of motor dysfunction in mild cognitive impairment and Alzheimer’s disease: a systematic review. J Alzheimers Dis Rep. 2022;6(1):307-344. [CrossRef] [Medline]
- Yang J, Yuan C, Chai Y, et al. Dynamic handwriting features for cognitive assessment in inflammatory demyelinating diseases: a machine learning study. Appl Sci (Basel). 15(11):6257. [CrossRef]
- Yuan J, Au R, Karjadi C, et al. Associations between the digital clock drawing test and brain volume: large community-based prospective cohort (Framingham Heart Study). J Med Internet Res. Apr 15, 2022;24(4):e34513. [CrossRef] [Medline]
- Mattsson N, Zetterberg H, Hansson O, et al. CSF biomarkers and incipient Alzheimer disease in patients with mild cognitive impairment. JAMA. Jul 22, 2009;302(4):385-393. [CrossRef] [Medline]
- Papaliagkas V, Kalinderi K, Vareltzis P, Moraitou D, Papamitsou T, Chatzidimitriou M. CSF biomarkers in the early diagnosis of mild cognitive impairment and Alzheimer’s disease. Int J Mol Sci. May 19, 2023;24(10):8976. [CrossRef] [Medline]
- Morbelli S, Garibotto V, Van De Giessen E, et al. A Cochrane review on brain [¹⁸F]FDG PET in dementia: limitations and future perspectives. Eur J Nucl Med Mol Imaging. Sep 2015;42(10):1487-1491. [CrossRef] [Medline]
- Wittenberg R, Knapp M, Karagiannidou M, Dickson J, Schott JM. Economic impacts of introducing diagnostics for mild cognitive impairment Alzheimer’s disease patients. A&D Transl Res & Clin Interv. Jan 2019;5(1):382-387. [CrossRef]
- Müller S, Preische O, Heymann P, Elbing U, Laske C. Increased diagnostic accuracy of digital vs. conventional clock drawing test for discrimination of patients in the early course of Alzheimer’s disease from cognitively healthy individuals. Front Aging Neurosci. 2017;9:101. [CrossRef] [Medline]
- Palmqvist S, Zetterberg H, Mattsson N, et al. Detailed comparison of amyloid PET and CSF biomarkers for identifying early Alzheimer disease. Neurology (ECronicon). Oct 6, 2015;85(14):1240-1249. [CrossRef] [Medline]
- Andreasen N, Minthon L, Davidsson P, et al. Evaluation of CSF-tau and CSF-Abeta42 as diagnostic markers for Alzheimer disease in clinical practice. Arch Neurol. Mar 2001;58(3):373-379. [CrossRef] [Medline]
- Krishnan K, Rossetti H, Hynan LS, et al. Changes in montreal cognitive assessment scores over time. Assessment. Sep 2017;24(6):772-777. [CrossRef] [Medline]
- Speiser JL. A random forest method with feature selection for developing medical prediction models with clustered and longitudinal data. J Biomed Inform. May 2021;117:103763. [CrossRef] [Medline]
- Huang BFF, Boutros PC. The parameter sensitivity of random forests. BMC Bioinformatics. Sep 1, 2016;17(1):331. [CrossRef] [Medline]
- Islam KI, Elias E, Carroll KC, Brown C. Exploring random forest machine learning and remote sensing data for streamflow prediction: an alternative approach to a process-based hydrologic modeling in a snowmelt-driven watershed. Remote Sens (Basel). 2023;15(16):3999. [CrossRef]
- Acharjee A, Larkman J, Xu Y, Cardoso VR, Gkoutos GV. A random forest based biomarker discovery and power analysis framework for diagnostics research. BMC Med Genomics. Nov 23, 2020;13(1):178. [CrossRef] [Medline]
- Chen C, Liang J, Sun W, Yang G, Meng X. An automatically recursive feature elimination method based on threshold decision in random forest classification. Geo-spatial Information Science. Jul 4, 2025;28(4):1494-1519. [CrossRef]
- Shao Z, Ahmad MN, Javed A. Comparison of random forest and XGBoost classifiers using integrated optical and SAR features for mapping urban impervious surface. Remote Sens (Basel). 2024;16(4):665. [CrossRef]
- Ali S, Akhlaq F, Imran AS, Kastrati Z, Daudpota SM, Moosa M. The enlightening role of explainable artificial intelligence in medical & healthcare domains: a systematic literature review. Comput Biol Med. Nov 2023;166:107555. [CrossRef] [Medline]
- Sadeghi Z, Alizadehsani R, Cifci MA, et al. A review of explainable artificial intelligence in healthcare. Computers and Electrical Engineering. Aug 2024;118:109370. [CrossRef]
- Frasca M, La Torre D, Pravettoni G, Cutica I. Explainable and interpretable artificial intelligence in medicine: a systematic bibliometric review. Discov Artif Intell. 2024;4(1):15. [CrossRef]
- Farah L, Murris JM, Borget I, Guilloux A, Martelli NM, Katsahian SIM. Assessment of performance, interpretability, and explainability in artificial intelligence-based health technologies: what healthcare stakeholders need to know. Mayo Clin Proc Digit Health. Jun 2023;1(2):120-138. [CrossRef] [Medline]
- Abgrall G, Holder AL, Chelly Dagdia Z, Zeitouni K, Monnet X. Should AI models be explainable to clinicians? Crit Care. Sep 12, 2024;28(1):301. [CrossRef] [Medline]
- Ahmad A, Imran M, Ahsan H. Biomarkers as biomedical bioindicators: approaches and techniques for the detection, analysis, and validation of novel biomarkers of diseases. Pharmaceutics. May 31, 2023;15(6):1630. [CrossRef] [Medline]
- Zhou Y, Tao L, Qiu J, et al. Tumor biomarkers for diagnosis, prognosis and targeted therapy. Signal Transduct Target Ther. May 20, 2024;9(1):132. [CrossRef] [Medline]
- Khalifa M, Albadawy M. Artificial intelligence for clinical prediction: exploring key domains and essential functions. Computer Methods and Programs in Biomedicine Update. 2024;5:100148. [CrossRef]
- Lawler M, Keeling P, Kholmanskikh O, et al. Empowering effective biomarker-driven precision oncology: a call to action. Eur J Cancer. Sep 2024;209:114225. [CrossRef] [Medline]
- Passaro A, Al Bakir M, Hamilton EG, et al. Cancer biomarkers: emerging trends and clinical implications for personalized treatment. Cell. Mar 28, 2024;187(7):1617-1635. [CrossRef] [Medline]
- Llitjos JF, Carrol ED, Osuchowski MF, et al. Enhancing sepsis biomarker development: key considerations from public and private perspectives. Crit Care. Jul 13, 2024;28(1):238. [CrossRef] [Medline]
- Marques L, Costa B, Pereira M, et al. Advancing precision medicine: a review of innovative in silico approaches for drug development, clinical pharmacology and personalized healthcare. Pharmaceutics. Feb 27, 2024;16(3):332. [CrossRef] [Medline]
- Nadkarni SB, Vijay GS, Kamath RC. Comparative study of random forest and gradient boosting algorithms to predict airfoil self-noise. Presented at: RAiSE-2023 International Conference on Recent Advances in Science and Engineering; Oct 4-5, 2023. [CrossRef]
- Wei Y, Ding J, Yang S, Yang X, Wang F. Comparisons of random forest and stochastic gradient treeboost algorithms for mapping soil electrical conductivity with multiple subsets using Landsat OLI and DEM/GIS-based data at a type oasis in Xinjiang, China. European Journal of Remote Sensing. Jan 1, 2021;54(1):158-181. [CrossRef]
- Patri A, Patnaik Y. Random forest and stochastic gradient tree boosting based approach for the prediction of airfoil self-noise. Procedia Comput Sci. 2015;46:109-121. [CrossRef]
- Sahin EK. Assessing the predictive capability of ensemble tree methods for landslide susceptibility mapping using XGBoost, gradient boosting machine, and random forest. SN Appl Sci. Jul 2020;2(7):1308. [CrossRef]
- Naderi Yaghouti AR, Zamanian H, Shalbaf A. Machine learning approaches for early detection of non-alcoholic steatohepatitis based on clinical and blood parameters. Sci Rep. Jan 30, 2024;14(1):2442. [CrossRef] [Medline]
- Perera LRD, Ganegoda GU. Alzheimer’s disease prediction using clinical data approach. Presented at: 8th International Conference on Information Technology Research (ICITR); Dec 7-8, 2023. [CrossRef]
- Asadi B, Hajj R. Prediction of asphalt binder elastic recovery using tree-based ensemble bagging and boosting models. Constr Build Mater. Jan 2024;410:134154. [CrossRef]
- Nguyen KL, Trinh HT, Pham TM. Prediction of punching shear strength in flat slabs: ensemble learning models and practical implementation. Neural Comput & Applic. Mar 2024;36(8):4207-4228. [CrossRef]
- Nazat S, Abdallah M. XAI-based feature ensemble for enhanced anomaly detection in autonomous driving systems. ArXiv. Preprint posted online on Oct 20, 2024. URL: https://arxiv.org/abs/2410.15405 [Accessed 2025-11-29]
- Mahdizadeh Gharakhanlou N, Perez L. From data to harvest: leveraging ensemble machine learning for enhanced crop yield predictions across Canada amidst climate change. Sci Total Environ. Nov 15, 2024;951:175764. [CrossRef] [Medline]
- Ascher S, Wang X, Watson I, Sloan W, You S. Interpretable machine learning to model biomass and waste gasification. Bioresour Technol. Nov 2022;364:128062. [CrossRef] [Medline]
- Chattopadhyay T, Ozarkar SS, Buwa K, et al. Comparison of deep learning architectures for predicting amyloid positivity in Alzheimer’s disease, mild cognitive impairment, and healthy aging, from T1-weighted brain structural MRI. Front Neurosci. 2024;18:1387196. [CrossRef] [Medline]
- Jenber Belay A, Walle YM, Haile MB. Deep ensemble learning and quantum machine learning approach for Alzheimer’s disease detection. Sci Rep. Jun 20, 2024;14(1):14196. [CrossRef] [Medline]
- Rosenbacke R, Melhus Å, McKee M, Stuckler D. How explainable artificial intelligence can increase or decrease clinicians’ trust in AI applications in health care: systematic review. JMIR AI. Oct 30, 2024;3:e53207. [CrossRef] [Medline]
- Markus AF, Kors JA, Rijnbeek PR. The role of explainability in creating trustworthy artificial intelligence for health care: a comprehensive survey of the terminology, design choices, and evaluation strategies. J Biomed Inform. Jan 2021;113:103655. [CrossRef] [Medline]
- Gaube S, Suresh H, Raue M, et al. Non-task expert physicians benefit from correct explainable AI advice when reviewing X-rays. Sci Rep. Jan 25, 2023;13(1):1383. [CrossRef] [Medline]
- Erdogmus P, Kabakus AT. The promise of convolutional neural networks for the early diagnosis of the Alzheimer’s disease. Eng Appl Artif Intell. Aug 2023;123:106254. [CrossRef]
- Subha R, Nayana B, Selvadass M. Hybrid machine learning model using particle swarm optimization for effectual diagnosis of Alzheimer’s disease from handwriting. Presented at: 4th International Conference on Circuits, Control, Communication and Computing; Dec 21-23, 2022. [CrossRef]
- Gattulli V, Impedovo D, Pirlo G, Semeraro G. Handwriting task-selection based on the analysis of patterns in classification results on Alzheimer dataset. Presented at: IEEESDS’23: Data Science Techniques for Datasets on Mental and Neurodegenerative Disorders; Jun 22, 2023. URL: https://ceur-ws.org/Vol-3521/paper2.pdf [Accessed 2025-11-29]
- Önder M, Şentürk Ü, Polat K, Paulraj D. Diagnosis of Alzheimer’s disease using boosting classification algorithms. Presented at: 2023 International Conference on Research Methodologies in Knowledge Management, Artificial Intelligence and Telecommunication Engineering (RMKMATE); Nov 1-2, 2023. [CrossRef]
- Anacur CA, Yilmaz AG, Aykut M. Handwriting analysis for Alzheimer’s diagnosis: feature enrichment and machine learning classification. Presented at: 2024 8th International Artificial Intelligence and Data Processing Symposium (IDAP); Sep 21-22, 2024. [CrossRef]
- Humphreys D, Kupresanin A, Boyer MD, et al. Advancing fusion with machine learning research needs workshop report. J Fusion Energ. Aug 2020;39(4):123-155. [CrossRef]
- Olorunnimbe K, Viktor H. Deep learning in the stock market-a systematic survey of practice, backtesting, and applications. Artif Intell Rev. 2023;56(3):2057-2109. [CrossRef] [Medline]
- Scheidt FV, Medinová H, Ludwig N, Richter B, Staudt P, Weinhardt C. Data analytics in the electricity sector – a quantitative and qualitative literature review. Energy and AI. Aug 2020;1:100009. [CrossRef]
- Sedlakova J, Daniore P, Horn Wintsch A, et al. Challenges and best practices for digital unstructured data enrichment in health research: a systematic narrative review. PLOS Digit Health. Oct 2023;2(10):e0000347. [CrossRef] [Medline]
- Kazemi P, Ghisi A, Entezami A. Artificial intelligence-powered computational strategies in selecting and augmenting data for early design of tall buildings with outer diagrids. Buildings. 2024;14(4):1118. [CrossRef]
- Moreira A, Ferreira A, Leite N. Prediction of Alzheimer disease on the DARWIN dataset with dimensionality reduction and explainability techniques. Presented at: 1st International Conference on Explainable AI for Neural and Symbolic Methods; Nov 20-22, 2024. [CrossRef]
- Singh SK, Chaturvedi A. Leveraging handwriting dynamics, explainable AI and machine learning for Alzheimer prediction. In: Singh JP, Singh MP, Singh AK, Mukhopadhyay S, Mandal JK, Dutta P, editors. Computational Intelligence in Communications and Business Analytics CICBA 2024 Communications in Computer and Information Science, Vol 2366. Springer; 2025:338-351. [CrossRef]
- Samrat N, Islam M, Sharmi S, et al. Explainable AI-driven handwriting analysis for early Alzheimer’s disease detection using machine learning and ensemble models. Presented at: 2024 27th International Conference on Computer and Information Technology (ICCIT); Dec 20-22, 2024. [CrossRef]
- SMART-Pred. GitHub. URL: https://github.com/Agyemang1z/SMART-Pred-Shiny-Multi-Algorithm-R-Tool-for-Predictive-Modeling-Manuscript [Accessed 2025-12-10]
Abbreviations
| AD: Alzheimer disease |
| AdaBoost: adaptive boosting |
| ADNI: Alzheimer’s Disease Neuroimaging Initiative |
| ADRD: Alzheimer disease and related dementias |
| AI: artificial intelligence |
| AUC: area under the curve |
| DARWIN: Diagnosis Alzheimer With Handwriting |
| KNN: k-nearest neighbors |
| PCA: principal component analysis |
| ROC: receiver operating characteristic |
| SHAP: Shapley Additive Explanations |
| SMART-Pred: Shiny Multi-Algorithm R Tool for Predictive Modeling |
| SVM: support vector machine |
| XGBoost: extreme gradient boosting |
Edited by Jinjiao Wang; submitted 18.Dec.2024; peer-reviewed by Fereshtehossadat Shojaei, Gideon Gogovi, Sadhasivam Mohanadas, Xueting Ding, Yixuan Li; final revised version received 29.Sep.2025; accepted 05.Nov.2025; published 30.Dec.2025.
Copyright© Han Wenzheng, Edmund F Agyemang, Sudesh K Srivastav, Jeffrey G Shaffer, Samuel Kakraba. Originally published in JMIR Aging (https://aging.jmir.org), 30.Dec.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Aging, is properly cited. The complete bibliographic information, a link to the original publication on https://aging.jmir.org, as well as this copyright and license information must be included.

