- Review
- Open access
- Published:
A scoping review of machine learning models to predict risk of falls in elders, without using sensor data
Diagnostic and Prognostic Research volume 9, Article number: 11 (2025)
Abstract
Objectives
This scoping review assesses machine learning (ML) tools that predicted falls, relying on information in health records without using any sensor data. The aim was to assess the available evidence on innovative techniques to improve fall prevention management.
Methods
Studies were included if they focused on predicting fall risk with machine learning in elderly populations and were written in English. There were 13 different extracted variables, including population characteristics (community dwelling, inpatients, age range, main pathology, ethnicity/race). Furthermore, the number of variables used in the final models, as well as their type, was extracted.
Results
A total of 6331 studies were retrieved, and 19 articles met criteria for data extraction. Metric performances reported by authors were commonly high in terms of accuracy (e.g., greater than 0.70). The most represented features included cardiovascular status and mobility assessments. Common gaps identified included a lack of transparent reporting and insufficient fairness assessments.
Conclusions
This review provides evidence that falls can be predicted using ML without using sensors if the amount of data and its quality is adequate. However, further studies are needed to validate these models in diverse groups and populations.
Background
Falls in elders are common and potentially devastating events [1]. Falls are a major cause of morbidity and mortality in elderly populations with 32,000 elders dying as a result each year [2]. In addition to the human cost caused by falls, the monetary costs are staggering, amounting to US $50 billion in 2018 alone [3, 4]. The high risk of falls is highlighted by in-hospital falls being included as a “never” event — occurrences that are clearly identifiable, preventable, and serious in their consequences — by the Center for Medicare and Medicaid Services (CMS), as well as numerous authors [5,6,7].
Falls are not exclusive to older citizens, but elderly patients have an elevated risk for falls [8]. There are many reasons for increased risk of falls with aging such as reduced muscle strength, deterioration in balance control, and use of prescription drugs which can cause dizziness. To address this problem, the first step is identification of those at risk so that specific fall prevention resources, such as fall prevention equipment allocation, physical therapy, or medication reconciliation, can be efficiently allocated.
To identify patients at high risk for falls, diverse groups have developed tools through complex computational methods. Developing these tools is challenging since coded data is not always reliable and outpatient data is often missing [9]. Furthermore, assessing individual risks in any clinical setting remains complex [10]. Given the close relationship between balance, gait, and the occurrence of falls, several groups have studied using machine learning (ML) models to predict falls using sensor data and have obtained promising results [11,12,13]. However, not all clinical settings are able to provide patients with high-tech devices and sensors, nor can we expect patients to purchase such devices for themselves. Adding to the cost and availability challenges these devices pose, their efficacy is also dependent on patients’ compliance, capacity, and their technological literacy and proficiency [14,15,16,17].
For this scoping review, machine learning was defined as a collection of computational methods that automatically learn patterns and relationships from data—in this case, routinely collected clinical records—to build predictive models capable of forecasting outcomes such as the risk of falls in elderly populations.
This study aims to assess the current literature on machine learning algorithms to predict the risk of falls in elders, without using sensor data, and provide guidance for future implementation.
Methods
A scoping review was conducted following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) approach [18]. Two searches were carried out: the initial search was implemented on June 22, 2022, and included articles from January 01, 2002, to December 31, 2021; the final search was implemented on February 09, 2025, and it included articles from January 01, 2022, to December 31, 2024. The search query consisted of terms considered pertinent by the authors.
Searches
The search included peer-reviewed publications in PubMed and Scopus. Title, abstracts, and indexed fields were searched in two searches operated at two different times. The first search was launched on June 22, 2022, using the following string:
-
PubMed: “((Elde* OR old OR senil*) AND (fall OR trip) AND (Machine Learning OR"AI"OR algorithm)) AND (("2002/01/01"[Date - Publication] :"2021/12/31"[Date - Publication]))”.
-
Scopus: “((elde$ OR old OR senil$) AND (fall OR trip) AND (machine AND learning OR"AI"OR algorithm)) AND PUBYEAR > 2002 AND PUBYEAR < 2022 AND (LIMIT-TO(SRCTYPE,"j") OR EXCLUDE (SRCTYPE,"p")) AND (LIMIT-TO(DOCTYPE,"ar")) AND (LIMIT-TO(LANGUAGE,"English")) AND (LIMIT-TO(SUBJAREA,"COMP") OR LIMIT-TO(SUBJAREA,"MEDI") OR LIMIT-TO (SUBJAREA,"SOCI"))”
The second search was launched on February 09, 2025, using the following string:
-
PubMed: ((Elde* OR old OR senil*) AND (fall OR trip) AND (Machine Learning OR"AI"OR algorithm)) AND (("2021/12/31"[Date—Publication]:"2024/12/31"[Date—Publication]))
-
Scopus: ((elde$ OR old OR senil$) AND (fall OR trip) AND (machine learning OR"AI"OR algorithm)) AND PUBYEAR > 2021 AND PUBYEAR < 2025 AND (LIMIT-TO(SRCTYPE,"j") OR EXCLUDE(SRCTYPE,"p")) AND LIMIT-TO(DOCTYPE,"ar") AND LIMIT-TO(LANGUAGE,"English") AND (LIMIT-TO(SUBJAREA,"COMP") OR LIMIT-TO(SUBJAREA,"MEDI") OR LIMIT-TO(SUBJAREA,"SOCI"))
Study inclusion and exclusion criteria
Studies were included if they focused on predicting fall risk with machine learning in elderly populations and were written in English. For the purposes of this review, “elderly” was defined as individuals aged 65 years or older. A sensitive inclusion approach was adopted: articles were included if the study population included individuals aged 65 or above, even if the overall cohort also included younger participants.
Studies were excluded if they did not focus on the prediction of falls in elders or if they predicted risk by relying on sensor data. Systematic reviews, meta-analyses, opinion papers, case reports, and editorials were excluded.
Data extraction
Data was extracted by one reviewer (A. C.) and double checked by another (A. S.).
Descriptive variables extracted from each article were as follows: publication year, first author’s affiliation (country and country’s income level), population number, population characteristics (community dwelling, inpatients, age range, main pathology, ethnicity/race), number and name of key variables used in the final ML model, subsets of variables created by authors (if present), and number of models and their results.
Regarding models’ results, “accuracy” (proportion of correctly classified instances or predictions out of the total number of instances evaluated), and area under the receiving curve and receiver operating characteristic (AUC and ROC) were extracted.
Results
Figure 1 shows the PRISMA flowchart for the included studies. Table 1 shows the included studies. Out of the 19 included studies [19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37] (Fig. 1), 16 (84%) were published by authors affiliated with high-income countries as defined by the World Bank [38], with Japan and the USA tying first place with 4 studies each.
High-income countries had collectively drawn data from a total of 631,696 subjects (min: 42 patients, max: 275,940), while upper-middle-income countries, namely China, Iran, and Turkey, used data from 109,260 subjects (min: 120, max: 108,940). Twelve of 19 studies (63%) [19,20,21,22, 24, 26, 27, 29, 30, 33, 34, 37] presented data from non-hospitalized individuals (total: 425,230), and 7 studies [23, 25, 28, 31, 32, 35, 36] presented data from inpatients, totaling 315,726 patients.
In the 19 studies assessed, the population was either enrolled inside the hospital (inpatients) or through the community the subjects lived in (community dwellers). None of the community dwellers (425,230 subjects) were enrolled for a specific pathology. Thus, no disease could be differentiated as a singular confounding factor. The inpatient group consisted of 315,726 subjects, of which 1090 were enrolled after a stroke, 226 were enrolled because of their type 2 diabetes, and 42 patients were enrolled because of their Alzheimer’s disease diagnosis. The remainder was not specified by the authors of the included studies.
A total of 455 variables were included in the best-performing models. We divided these variables into 19 categories, the full extent of which can be found in Table S1. The most common variables were related to miscellaneous conditions, ranging from “handgrip strength score” to “cancer” to “visit in December”; the second most common type of variable was related to cardiovascular measurements/diseases, closely followed by mobility- and equilibrium-related variables.
More than 65 models were described in the aforementioned studies. The highest-performing models and their accuracies and/or AUCs have been summarized in Table 2.
The three most represented model categories were neural network models (14 out of 70 occurrences, 20%), logistic regression variants (13 out of 70 occurrences, 19%), and random forest approaches (11 out of 70 occurrences, 16%). The highest performing model was an SVM that achieved an AUC of 1.0 and 100% accuracy (hold-out validation method).
Some studies, such as the one published by Razmara et al. [19], had summarized variables. These variable summaries, along with three others [24, 28, 29], can be found in Supplemental Material S2. All variables used and the highest performing used in the included studies can be found in Supplemental Material S3.
The best results in terms of accuracy amidst the cross-sectional studies [19,20,21, 23,24,25, 30, 33, 35, 36] were described by Razmara et al., who for their best model reported 91.3% accuracy (using NN, with psychological and public factors).
The best results in terms of accuracy and AUC amidst the cohort studies [22, 26,27,28,29, 31, 32, 34, 37] were described by T. Deschamps et al., P. W. Patterson et al., and Soylemez et al. who reported respectively 82% accuracy (using decision tree, with an hold-out validation method) and a median AUC of 0.76 across all models (using random forest and variously penalized logistic regressions, with an hold-out validation method) and AUC = 1 and 100% accuracy (Naïve Bayes, with an hold-out validation method). They enrolled, respectively, 426, 9687, and 120 patients [26, 29, 37].
On the other hand, the best model’s poorest results were described by Suzuki et al. [28] who obtained an accuracy of 65% using a CNN and just three factors (mini-mental state + normalized knee extension strength + FIM locomotion).
However, we note that these machine learning model results are not directly comparable to each other, as they were fitted on different datasets, with potentially different distributions.
Discussion
The results highlight significant variations in the prediction of falls among patients, with diverse approaches and model performances across the included studies. Notably, most studies originated from high-income countries, drawing data from a substantial number of subjects. Variable categories ranged from cardiovascular measurements to psychological factors, reflecting the multifaceted nature of fall prediction. While the neural network model incorporating psychological and public factors achieved a remarkable 91.3% accuracy, other models showed varying degrees of success. These findings underscore the complexity of fall prediction and the influence of model selection and variables on outcomes.
Nonetheless, unbiased comparisons are not entirely possible here, since not all algorithms were directly compared over the same datasets, potentially leading to unfair comparisons.
The most studied population was people aged 60 and older, and no study included subjects younger than 40. This pattern was expected as falls are more common in elders, often resulting in serious injuries. Furthermore, given how most falls happen during common day-to-day activities [39], most included studies focused on data collected from community-dwelling adults, and just seven studies analyzed inpatients. This result underlines how prediction based on clinical data might improve community care, potentially lightening the burden on hospitals; however, falls were usually defined as “self-reported” in community-dwelling studies, therefore limiting the analyzed studies findings’ reliability.
Analyzed studies were most commonly published by high-income countries; therefore, their conclusions should be valued in their own demographics, and since there were no multicentric validations or piloted attempts at clinical implementation, we cannot generalize their results to the global population.
Complex vs. simple algorithms
In analyzing the performance of various machine learning models, it becomes evident that model complexity and the availability of training data can influence predictive accuracy. Simpler models, such as decision trees and logistic regression, have shown robust performance, particularly when trained on limited datasets. For instance, Deschamps et al. [26] employed a decision tree algorithm achieving an accuracy of 82% with a sample size of 426, while Womack et al. [22] utilized logistic regression to attain an AUC of 0.76 on a substantially larger sample of 275,940 individuals. Conversely, more complex models like neural networks seem to require substantial amounts of data to perform effectively. Suzuki et al. [28] implemented a CNN using only 42 samples (and bootstrapping them), resulting in a lower accuracy of 65%. This underscores the necessity for large training data when deploying neural networks, as insufficient data can lead to suboptimal performance. Notably, when complex models are trained on extensive datasets, they can achieve superior predictive capabilities. For example, Sharma et al. [36] applied a CatBoost algorithm to a validation dataset of 203,584 samples, resulting in an AUC of 0.70. Similarly, Thapa et al. [30] utilized eXtreme Gradient Boosting on 2785 samples, achieving an AUC of 0.85.
These observations align with existing literature, which empirically suggests that while complex models such as neural networks have the potential for higher predictive power, they necessitate larger datasets to realize this potential [40,41,42]. Selecting an appropriate machine learning model for fall prediction in the elderly should involve an assessment of model complexity relative to the available training data. Simpler models may offer sufficient performance with limited data and are easier to implement, whereas complex models may provide enhanced accuracy when supported by large, high-quality datasets.
Common and important variables
The most common homogeneous variables were mainly related to cardiovascular status and mobility assessed without sensors (see Supplemental S1). A selection of other variables, classified as “miscellaneous,” often correlated with the risk of falls and ranged from “arrived by self or with family” to “day of the week for the visit” and “drugs for peptic ulcer.”
The variety of predictors highlights the need for broad inclusions to allow for the discovery of important but unexpected variables.
Table S3 summarizes the most important variables per each study. The five most important variable categories were respectively mobility/equilibrium, miscellaneous, cardiovascular, mental health, and demographic data. We believe this last set of variables might be worthy of future research, as it underlines how social determinants of health are associated with falls.
Fairness
Fairness and minimizing bias—understood here in both its statistical and social dimensions—are increasingly recognized as critical to precision medicine. Here, we refer to statistical bias as to systematic errors in model performance, while social bias denotes the unfair or inequitable treatment that may result when such errors disproportionately harm underrepresented groups [43]. It is an increasing movement to dismantle the hard coding of sensitive attributes in AI-/ML-derived tools, with much attention focused on race and ethnicity. This transition necessitates a more deliberate and thoughtful approach to model design, underpinned by a deeper understanding of how various social determinants of health contribute to data-driven outcomes in medical decision-making [44]. This was consistent with our findings since one of the best achieving algorithm, achieved by Razmara et al. [19], used “public factors,” which included age, gender, education level, and employment among others, but not race; recognizing and addressing bias in healthcare algorithms necessitate a holistic approach that encompasses not only racial bias but also factors related to other social determinants of health. We argue that using more social determinants of health variables could considerably increase the models’ fairness and their outputs.
Common gaps
This study found the analysis of algorithms’ performance was inconsistent across studies. The extraction of accuracy, whenever possible, may not be the best measure of performance evaluation. Accuracy is a single measure and can be misleading when presented without event distribution information, potentially leading to misinterpretations. For example, if the entire population has an event, then a faulty algorithm only capable of one positive output will always have 100% accuracy. This was not the case with the included studies, but more transparent and homogeneous reporting can prevent low-quality science. Additionally, a probability threshold has to be determined to classify a fall from a non-fall, which is not necessarily trivial. Metrics like AUROC [45] or the area under the precision-recall curve (AUPRC) are threshold agnostic and might be the better choice to represent discrimination performance. Adopting transparent reporting practices and considering threshold-agnostic metrics like AUROC and AUPRC are essential steps towards ensuring robust and meaningful evaluations of algorithm performance in healthcare studies.
Limitations
This study comes with a few limitations. It is possible that our search strategy may not have included certain studies, especially those published in computer science journals not indexed in the searched databases. Furthermore, other studies may have been missed due to the keywords used in our search strategy, the incompleteness of which is often embedded in reviews. Additionally, a notable limitation of this review is the heterogeneity in the age criteria used across the included studies. While most studies defined the elderly population as individuals aged 65 years or older, some studies enrolled participants beginning at younger ages (e.g., 40, 45, or 50 years). This variation may introduce variability in the fall risk profiles and could potentially affect the generalizability of our findings to a uniformly defined elderly population.
Conclusions
This scoping review found that machine learning models using commonly recorded data can predict falls with high accuracy, as reported by the individual studies. The results highlight the importance of understanding the diverse factors influencing the model’s performance. The predominance of high-income countries in both research and participant representation underscores the need for a broader global perspective in this field. The variation in study populations, from community dwellers to inpatients with specific diagnoses, highlights the complexity of fall prediction and the necessity for tailored approaches. Moreover, the success of neural network models incorporating psychological and public factors demonstrates the potential for multifaceted models to enhance accuracy.
These findings offer a valuable overview of fall prediction models’ performance and variability, setting the stage for further advancements in this crucial area of healthcare research.
Data availability
No datasets were generated or analysed during the current study.
Abbreviations
- AI:
-
Artificial intelligence
- AUC:
-
Area under the receiving curve
- AUPRC:
-
Area under the precision-recall curve
- AUROC:
-
Area under the receiver operating characteristic curve
- CI:
-
Confidence interval
- CMS:
-
Center for Medicare and Medicaid Services
- ML:
-
Machine learning
- MLP:
-
Multilayer perceptron
- MCMC:
-
Markov Chain Monte Carlo
- NN:
-
Neural network
- PRISMA:
-
Preferred Reporting Items for Systematic reviews and Meta-Analysis
- ROC:
-
Receiver operating characteristic
- SES:
-
Socioeconomic status
- SVM:
-
Support vector machine
- USA:
-
United States of America
References
Lusardi MM, Fritz S, Middleton A, Allison L, Wingood M, Phillips E, et al. Determining risk of falls in community dwelling older adults: a systematic review and meta-analysis using posttest probability. J Geriatr Phys Ther. 2017;40:1–36. https://doiorg.publicaciones.saludcastillayleon.es/10.1519/JPT.0000000000000099.
Centers for Disease Control and Prevention, CDC, United States of America. n.d. Retreived at: https://www.cdc.gov/injury/wisqars/index.html.
Centers for Disease Control and Prevention, CDC - Costs of Older Adults Falls. n.d. Accessed at: https://www.cdc.gov/falls/data/fall-cost.html.
Houry D, Florence C, Baldwin G, Stevens J, McClure R. The CDC Injury Center’s response to the growing public health problem of falls among older adults. Am J Lifestyle Med 2016;10. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/1559827615600137.
Mattie AS, Webster BL. Centers for medicare and medicaid services’ “never events”: an analysis and recommendations to hospitals. Health Care Manag. 2008;27:338–49. https://doiorg.publicaciones.saludcastillayleon.es/10.1097/HCM.0b013e31818c8037.
Miake-Lye IM, Hempel S, Ganz DA, Shekelle PG. Inpatient fall prevention programs as a patient safety strategy: a systematic review. Ann Intern Med. 2013;158:390–6. https://doiorg.publicaciones.saludcastillayleon.es/10.7326/0003-4819-158-5-201303051-00005.
Grubel F. Falls: a principal patient incident. Hosp Manage. 1959;88(37–8):137.
Bergen G, Stevens MR, Burns ER. Falls and fall injuries among adults aged ≥65 years — United States, 2014. MMWR Morb Mortal Wkly Rep. 2016;65:993–8. https://doiorg.publicaciones.saludcastillayleon.es/10.15585/mmwr.mm6537a2.
Doktorchik C, Lu M, Quan H, Ringham C, Eastwood C. A qualitative evaluation of clinically coded data quality from health information manager perspectives. Health Inf Manag J. 2020;49:19–27. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/1833358319855031.
Skinner CS, Halm EA, Bishop WP, Ahn C, Gupta S, Farrell D, et al. Impact of risk assessment and tailored versus nontailored risk information on colorectal cancer testing in primary care: a randomized controlled trial. Cancer Epidemiol Biomark Prev Publ Am Assoc Cancer Res Cosponsored Am Soc Prev Oncol. 2015;24:1523–30. https://doiorg.publicaciones.saludcastillayleon.es/10.1158/1055-9965.EPI-15-0122.
Lockhart TE, Soangra R, Yoon H, Wu T, Frames CW, Weaver R, et al. Prediction of fall risk among community-dwelling older adults using a wearable system. Sci Rep. 2021;11:20976. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41598-021-00458-5.
Usmani S, Saboor A, Haris M, Khan MA, Park H. Latest research trends in fall detection and prevention using machine learning: a systematic review. Sensors. 2021;21: 5134. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/s21155134.
Noh B, Youm C, Goh E, Lee M, Park H, Jeon H, et al. XGBoost based machine learning approach to predict the risk of fall in older adults using gait outcomes. Sci Rep. 2021;11:12183. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41598-021-91797-w.
Aisbitt GM, Nolte T, Fonagy P. Editorial perspective: the digital divide – inequalities in remote therapy for children and adolescents. Child Adolesc Ment Health. 2022:camh.12545. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/camh.12545.
Sun N, Esom K, Dhaliwal M, Amon JJ. Human rights and digital health technologies. Health Hum Rights. 2020;22:21–32.
Rosales A, Fernández-Ardèvol M. Ageism in the era of digital platforms. Converg Int J Res New Media Technol. 2020;26:1074–87. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/1354856520930905.
Chau D, Osborne T. Using technology to improve care of older adults. n.d.
Moher D. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement (Chinese edition). J Chin Integr Med. 2009;7:889–96. https://doiorg.publicaciones.saludcastillayleon.es/10.3736/jcim20090918.
Razmara J, Zaboli MH, Hassankhani H. Elderly fall risk prediction based on a physiological profile approach using artificial neural networks. Health Informatics J. 2018;24:410–8. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/1460458216677841.
Kabeshova A, Launay CP, Gromov VA, Fantino B, Levinoff EJ, Allali G, et al. Falling in the elderly: do statistical models matter for performance criteria of fall prediction? Results from two large population-based studies. Eur J Intern Med. 2016;27:48–56. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.ejim.2015.11.019.
Kabeshova A, Launay CP, Gromov VA, Annweiler C, Fantino B, Beauchet O. Artificial neural network and falls in community-dwellers: a new approach to identify the risk of recurrent falling? J Am Med Dir Assoc. 2015;16:277–81. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jamda.2014.09.013.
Womack JA, Murphy TE, Bathulapalli H, Smith A, Bates J, Jarad S, et al. Serious falls in middle-aged veterans: development and validation of a predictive risk model. J Am Geriatr Soc. 2020;68:2847–54. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/jgs.16773.
Nakatani H, Nakao M, Uchiyama H, Toyoshiba H, Ochiai C. Predicting inpatient falls using natural language processing of nursing records obtained from Japanese Electronic Medical Records: case-control study. JMIR Med Inform. 2020;8: e16970. https://doiorg.publicaciones.saludcastillayleon.es/10.2196/16970.
Homer ML, Palmer NP, Fox KP, Armstrong J, Mandl KD. Predicting falls in people aged 65 years and older from insurance claims. Am J Med. 2017;130:744.e17-744.e23. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.amjmed.2017.01.003.
Liu C-H, Hu Y-H, Lin Y-H. A machine learning–based fall risk assessment model for inpatients. CIN Comput Inform Nurs. 2021;39:450–9. https://doiorg.publicaciones.saludcastillayleon.es/10.1097/CIN.0000000000000727.
Deschamps T, Le Goff CG, Berrut G, Cornu C, Mignardot J-B. A decision model to predict the risk of the first fall onset. Exp Gerontol. 2016;81:51–5. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.exger.2016.04.016.
Makino K, Lee S, Bae S, Chiba I, Harada K, Katayama O, et al. Simplified decision-tree algorithm to predict falls for community-dwelling older adults. J Clin Med. 2021;10: 5184. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/jcm10215184.
Suzuki M, Yamamoto R, Ishiguro Y, Sasaki H, Kotaki H. Deep learning prediction of falls among nursing home residents with Alzheimer’s disease. Geriatr Gerontol Int. 2020;20:589–94. https://doiorg.publicaciones.saludcastillayleon.es/10.1111/ggi.13920.
Patterson BW, Engstrom CJ, Sah V, Smith MA, Mendonça EA, Pulia MS, et al. Training and interpreting machine learning algorithms to evaluate fall risk after emergency department visits. Med Care. 2019;57:560–6. https://doiorg.publicaciones.saludcastillayleon.es/10.1097/MLR.0000000000001140.
Thapa R, Garikipati A, Shokouhi S, Hurtado M, Barnes G, Hoffman J, et al. Predicting falls in long-term care facilities: machine learning study. JMIR Aging. 2022;5: e35373. https://doiorg.publicaciones.saludcastillayleon.es/10.2196/35373.
Chu W-M, Kristiani E, Wang Y-C, Lin Y-R, Lin S-Y, Chan W-C, et al. A model for predicting fall risks of hospitalized elderly in Taiwan-a machine learning approach based on both electronic health records and comprehensive geriatric assessment. Front Med. 2022;9: 937216. https://doiorg.publicaciones.saludcastillayleon.es/10.3389/fmed.2022.937216.
Suzuki Y, Suzuki H, Ishikawa T, Yamada Y, Yatoh S, Sugano Y, et al. Exploratory analysis using machine learning of predictive factors for falls in type 2 diabetes. Sci Rep. 2022;12:11965. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41598-022-15224-4.
Fong KNK, Chung RCK, Sze PPC, Ng CKM. Factors associated with fall risk of community-dwelling older people: a decision tree analysis. Digit Health. 2023;9:20552076231181200. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/20552076231181202.
Donoghue OA, Hernandez B, O’Connell MDL, Kenny RA. Using conditional inference forests to examine predictive ability for future falls and syncope in older adults: results from The Irish Longitudinal Study on Ageing. J Gerontol Ser A. 2023;78:673–82. https://doiorg.publicaciones.saludcastillayleon.es/10.1093/gerona/glac156.
Choi JH, Choi ES, Park D. In-hospital fall prediction using machine learning algorithms and the Morse fall scale in patients with acute stroke: a nested case-control study. BMC Med Inform Decis Mak. 2023;23:246. https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-023-02330-0.
Sharma V, Kulkarni V, Joon T, Eurich DT, Simpson SH, Voaklander D, et al. Predicting falls-related admissions in older adults in Alberta, Canada: a machine-learning falls prevention tool developed using population administrative health data. BMJ Open. 2023;13: e071321. https://doiorg.publicaciones.saludcastillayleon.es/10.1136/bmjopen-2022-071321.
Soylemez E, Tokgoz-Yilmaz S. Predicting fall risk in elderly individuals: a comparative analysis of machine learning models using patient characteristics, functional balance tests, and computerized dynamic posturography. J Laryngol Otol 2024:1–31. https://doiorg.publicaciones.saludcastillayleon.es/10.1017/S0022215124002160.
The World Bank, World Bank Country and Lending Groups (2021). n.d. Retrieved from: https://datahelpdesk.worldbank.org/knowledgebase/articles/906519-world-bank-country-and-lending-groups.
Wiseman JM, Stamper DS, Sheridan E, Caterino JM, Quatman-Yates CC, Quatman CE. Barriers to the initiation of home modifications for older adults for fall prevention. Geriatr Orthop Surg Rehabil. 2021;12: 21514593211002161. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/21514593211002161.
Bahri Y, Dyer E, Kaplan J, Lee J, Sharma U. Explaining neural scaling laws. Proc Natl Acad Sci. 2024;121: e2311878121. https://doiorg.publicaciones.saludcastillayleon.es/10.1073/pnas.2311878121.
Hestness J, Narang S, Ardalani N, Diamos G, Jun H, Kianinejad H, et al. Deep learning scaling is predictable, Empirically 2017. https://doiorg.publicaciones.saludcastillayleon.es/10.48550/ARXIV.1712.00409.
Rosenfeld JS, Rosenfeld A, Belinkov Y, Shavit N. A constructive prediction of the generalization error across scales. 2019. https://doiorg.publicaciones.saludcastillayleon.es/10.48550/ARXIV.1909.12673.
Röösli E, Bozkurt S, Hernandez-Boussard T. Peeking into a black box, the fairness and generalizability of a MIMIC-III benchmarking model. Sci Data. 2022;9:24. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/s41597-021-01110-7.
Hernandez-Boussard T, Siddique SM, Bierman AS, Hightower M, Burstin H. Promoting equity in clinical decision making: dismantling race-based medicine: commentary examines promoting equity in clinical decision-making. Health Aff (Millwood). 2023;42:1369–73. https://doiorg.publicaciones.saludcastillayleon.es/10.1377/hlthaff.2023.00545.
Steyerberg EW, Vickers AJ, Cook NR, Gerds T, Gonen M, Obuchowski N, et al. Assessing the performance of prediction models: a framework for traditional and novel measures. Epidemiology. 2010;21:128–38. https://doiorg.publicaciones.saludcastillayleon.es/10.1097/EDE.0b013e3181c30fb2.
Acknowledgements
None.
Funding
This project was supported by grant number R01HS024096 from the Agency for Healthcare Research and Quality. The content is solely the responsibility of the authors and does not necessarily represent the official views of the Agency for Healthcare Research and Quality.
Author information
Authors and Affiliations
Contributions
Conceptualization, A.C, C.F., and T.H.B.; methodology, A.C., A.S. C.F.; validation, A.C. A.S. C.C. F.N. A.G. T.H.B.; formal analysis, A.C., C.F., A.S., investigation, A.C., C.F., A.S., C.C.; data curation, A.C., A.S., C.F.; writing—original draft preparation, A.C.; writing—review and editing, C.F., C.C., A.S., F.N., A.G., T.H.B.; supervision, A.G., T.H.B.; project administration, T.H.B.; All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not pertinent, as the present is a review.
Consent for publication
Not pertinent, as the present is a review.
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
41512_2025_190_MOESM3_ESM.docx
Supplementary Material 3. Supplemental Material S3: All variables used and the highest performing used in the included studies
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Capodici, A., Fanconi, C., Curtin, C. et al. A scoping review of machine learning models to predict risk of falls in elders, without using sensor data. Diagn Progn Res 9, 11 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s41512-025-00190-y
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s41512-025-00190-y