Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
261
result(s) for
"Williams, Ross D"
Sort by:
Feasibility and evaluation of a large-scale external validation approach for patient-level prediction in an international data network: validation of models predicting stroke in female patients newly diagnosed with atrial fibrillation
by
Park, Rae Woong
,
You, Seng Chan
,
Falconer, Thomas
in
Atrial fibrillation
,
Atrial Fibrillation - diagnosis
,
Atrial Fibrillation - epidemiology
2020
Background
To demonstrate how the Observational Healthcare Data Science and Informatics (OHDSI) collaborative network and standardization can be utilized to scale-up external validation of patient-level prediction models by enabling validation across a large number of heterogeneous observational healthcare datasets.
Methods
Five previously published prognostic models (ATRIA, CHADS
2
, CHADS
2
VASC, Q-Stroke and Framingham) that predict future risk of stroke in patients with atrial fibrillation were replicated using the OHDSI frameworks. A network study was run that enabled the five models to be externally validated across nine observational healthcare datasets spanning three countries and five independent sites.
Results
The five existing models were able to be integrated into the OHDSI framework for patient-level prediction and they obtained mean c-statistics ranging between 0.57–0.63 across the 6 databases with sufficient data to predict stroke within 1 year of initial atrial fibrillation diagnosis for females with atrial fibrillation. This was comparable with existing validation studies. The validation network study was run across nine datasets within 60 days once the models were replicated. An R package for the study was published at
https://github.com/OHDSI/StudyProtocolSandbox/tree/master/ExistingStrokeRiskExternalValidation
.
Conclusion
This study demonstrates the ability to scale up external validation of patient-level prediction models using a collaboration of researchers and a data standardization that enable models to be readily shared across data sites. External validation is necessary to understand the transportability or reproducibility of a prediction model, but without collaborative approaches it can take three or more years for a model to be validated by one independent researcher. In this paper we show it is possible to both scale-up and speed-up external validation by showing how validation can be done across multiple databases in less than 2 months. We recommend that researchers developing new prediction models use the OHDSI network to externally validate their models.
Journal Article
Development and validation of a patient-level model to predict dementia across a network of observational databases
by
Fridgeirsson, Egill A.
,
Ryan, Patrick B.
,
John, Luis H.
in
Age groups
,
Aged
,
Aged, 80 and over
2024
Background
A prediction model can be a useful tool to quantify the risk of a patient developing dementia in the next years and take risk-factor-targeted intervention. Numerous dementia prediction models have been developed, but few have been externally validated, likely limiting their clinical uptake. In our previous work, we had limited success in externally validating some of these existing models due to inadequate reporting. As a result, we are compelled to develop and externally validate novel models to predict dementia in the general population across a network of observational databases. We assess regularization methods to obtain parsimonious models that are of lower complexity and easier to implement.
Methods
Logistic regression models were developed across a network of five observational databases with electronic health records (EHRs) and claims data to predict 5-year dementia risk in persons aged 55–84. The regularization methods L1 and Broken Adaptive Ridge (BAR) as well as three candidate predictor sets to optimize prediction performance were assessed. The predictor sets include a baseline set using only age and sex, a full set including all available candidate predictors, and a phenotype set which includes a limited number of clinically relevant predictors.
Results
BAR can be used for variable selection, outperforming L1 when a parsimonious model is desired. Adding candidate predictors for disease diagnosis and drug exposure generally improves the performance of baseline models using only age and sex. While a model trained on German EHR data saw an increase in AUROC from 0.74 to 0.83 with additional predictors, a model trained on US EHR data showed only minimal improvement from 0.79 to 0.81 AUROC. Nevertheless, the latter model developed using BAR regularization on the clinically relevant predictor set was ultimately chosen as best performing model as it demonstrated more consistent external validation performance and improved calibration.
Conclusions
We developed and externally validated patient-level models to predict dementia. Our results show that although dementia prediction is highly driven by demographic age, adding predictors based on condition diagnoses and drug exposures further improves prediction performance. BAR regularization outperforms L1 regularization to yield the most parsimonious yet still well-performing prediction model for dementia.
Journal Article
Learning patient-level prediction models across multiple healthcare databases: evaluation of ensembles for increasing model transportability
by
Rijnbeek, Peter R.
,
Schuemie, Martijn J.
,
Reps, Jenna Marie
in
Algorithms
,
Calibration
,
Communication
2022
Background
Prognostic models that are accurate could help aid medical decision making. Large observational databases often contain temporal medical data for large and diverse populations of patients. It may be possible to learn prognostic models using the large observational data. Often the performance of a prognostic model undesirably worsens when transported to a different database (or into a clinical setting). In this study we investigate different ensemble approaches that combine prognostic models independently developed using different databases (a simple federated learning approach) to determine whether ensembles that combine models developed across databases can improve model transportability (perform better in new data than single database models)?
Methods
For a given prediction question we independently trained five single database models each using a different observational healthcare database. We then developed and investigated numerous ensemble models (fusion, stacking and mixture of experts) that combined the different database models. Performance of each model was investigated via discrimination and calibration using a leave one dataset out technique, i.e., hold out one database to use for validation and use the remaining four datasets for model development. The internal validation of a model developed using the hold out database was calculated and presented as the ‘internal benchmark’ for comparison.
Results
In this study the fusion ensembles generally outperformed the single database models when transported to a previously unseen database and the performances were more consistent across unseen databases. Stacking ensembles performed poorly in terms of discrimination when the labels in the unseen database were limited. Calibration was consistently poor when both ensembles and single database models were applied to previously unseen databases.
Conclusion
A simple federated learning approach that implements ensemble techniques to combine models independently developed across different databases for the same prediction question may improve the discriminative performance in new data (new database or clinical setting) but will need to be recalibrated using the new data. This could help medical decision making by improving prognostic model performance.
Journal Article
Finding a constrained number of predictor phenotypes for multiple outcome prediction
by
Wong, Jenna
,
Fisher, Renae R
,
Reps, Jenna M
in
Cardiac arrhythmia
,
Clinical decision making
,
Clinical Decision-Making - methods
2025
BackgroundPrognostic models help aid medical decision-making. Various prognostic models are available via websites such as MDCalc, but these models typically predict one outcome, for example, stroke risk. Each model requires individual predictors, for example, age, lab results and comorbidities. There is no clinical tool available to predict multiple outcomes from a list of common medical predictors.ObjectiveIdentify a constrained set of outcome-agnostic predictors.MethodsWe proposed a novel technique aggregating the standardised mean difference across hundreds of outcomes to learn a constrained set of predictors that appear to be predictive of many outcomes. Model performance was evaluated using the constrained set of predictors across eight prediction tasks. We compared against existing models, models using only age/sex predictors and models without any predictor constraints.ResultsWe identified 67 predictors in our constrained set, plus age/sex. Our predictors included illnesses in the following categories: cardiovascular, kidney/liver, mental health, gastrointestinal, infectious and oncologic. Models developed using the constrained set of predictors achieved comparable discrimination compared with models using hundreds or thousands of predictors for five of the eight prediction tasks and slightly lower discrimination for three of the eight tasks. The constrained predictor models performed as good or better than all existing clinical models.ConclusionsIt is possible to develop models for hundreds or thousands of outcomes that use the same small set of predictors. This makes it feasible to implement many prediction models via a single website form. Our set of predictors can also be used for future models and prognostic model research.
Journal Article
Unlocking efficiency in real-world collaborative studies: a multi-site international study with one-shot lossless GLMM algorithm
by
Simon, Katherine R.
,
DuVall, Scott L.
,
Matheny, Michael E.
in
631/114/2401
,
631/114/2415
,
639/705/531
2025
The widespread adoption of real-world data has given rise to numerous healthcare-distributed research networks, but multi-site analyses still face administrative burdens and data privacy challenges. In response, we developed a Collaborative One-shot Lossless Algorithm for Generalized Linear Mixed Models (COLA-GLMM), the first-ever algorithm that achieves both
lossless
and
one-shot
properties. COLA-GLMM ensures accuracy against the gold standard of pooled data while requiring only summary statistics and completes within a single communication round, eliminating the usual back-and-forth overhead. We further introduced an enhanced version that employs homomorphic encryption to reduce the risks of summary statistics misuse at the coordinating center. The simulation studies showed near-exact agreement with the gold standard in parameter estimation, with relative differences of 7.8 × 10
−6
%–3.0% under various cell suppression settings. We also validated COLA‑GLMM on eight international decentralized databases to identify risk factors for COVID‑19 mortality. Together, these results show that COLA‑GLMM enables accurate, low‑burden, and privacy-preserving multi‑site research.
Journal Article
90-Day all-cause mortality can be predicted following a total knee replacement: an international, network study to develop and validate a prediction model
by
Prieto-Alhambra, Daniel
,
Rijnbeek, Peter R.
,
Ryan, Patrick B.
in
Comorbidity
,
Complications
,
Decision making
2022
Purpose
The purpose of this study was to develop and validate a prediction model for 90-day mortality following a total knee replacement (TKR). TKR is a safe and cost-effective surgical procedure for treating severe knee osteoarthritis (OA). Although complications following surgery are rare, prediction tools could help identify high-risk patients who could be targeted with preventative interventions. The aim was to develop and validate a simple model to help inform treatment choices.
Methods
A mortality prediction model for knee OA patients following TKR was developed and externally validated using a US claims database and a UK general practice database. The target population consisted of patients undergoing a primary TKR for knee OA, aged ≥ 40 years and registered for ≥ 1 year before surgery. LASSO logistic regression models were developed for post-operative (90-day) mortality. A second mortality model was developed with a reduced feature set to increase interpretability and usability.
Results
A total of 193,615 patients were included, with 40,950 in The Health Improvement Network (THIN) database and 152,665 in Optum. The full model predicting 90-day mortality yielded AUROC of 0.78 when trained in OPTUM and 0.70 when externally validated on THIN. The 12 variable model achieved internal AUROC of 0.77 and external AUROC of 0.71 in THIN.
Conclusions
A simple prediction model based on sex, age, and 10 comorbidities that can identify patients at high risk of short-term mortality following TKR was developed that demonstrated good, robust performance. The 12-feature mortality model is easily implemented and the performance suggests it could be used to inform evidence based shared decision-making prior to surgery and targeting prophylaxis for those at high risk.
Level of evidence
III.
Journal Article
Using Iterative Pairwise External Validation to Contextualize Prediction Model Performance: A Use Case Predicting 1-Year Heart Failure Risk in Patients with Diabetes Across Five Data Sources
by
Ryan, Patrick B.
,
Rijnbeek, Peter R.
,
Steyerberg, Ewout
in
Best practice
,
Calibration
,
Congestive heart failure
2022
Introduction
External validation of prediction models is increasingly being seen as a minimum requirement for acceptance in clinical practice. However, the lack of interoperability of healthcare databases has been the biggest barrier to this occurring on a large scale. Recent improvements in database interoperability enable a standardized analytical framework for model development and external validation. External validation of a model in a new database lacks context, whereby the external validation can be compared with a benchmark in this database. Iterative pairwise external validation (IPEV) is a framework that uses a rotating model development and validation approach to contextualize the assessment of performance across a network of databases. As a use case, we predicted 1-year risk of heart failure in patients with type 2 diabetes mellitus.
Methods
The method follows a two-step process involving (1) development of baseline and data-driven models in each database according to best practices and (2) validation of these models across the remaining databases. We introduce a heatmap visualization that supports the assessment of the internal and external model performance in all available databases. As a use case, we developed and validated models to predict 1-year risk of heart failure in patients initializing a second pharmacological intervention for type 2 diabetes mellitus. We leveraged the power of the Observational Medical Outcomes Partnership common data model to create an open-source software package to increase the consistency, speed, and transparency of this process.
Results
A total of 403,187 patients from five databases were included in the study. We developed five models that, when assessed internally, had a discriminative performance ranging from 0.73 to 0.81 area under the receiver operating characteristic curve with acceptable calibration. When we externally validated these models in a new database, three models achieved consistent performance and in context often performed similarly to models developed in the database itself. The visualization of IPEV provided valuable insights. From this, we identified the model developed in the Commercial Claims and Encounters (CCAE) database as the best performing model overall.
Conclusion
Using IPEV lends weight to the model development process. The rotation of development through multiple databases provides context to model assessment, leading to improved understanding of transportability and generalizability. The inclusion of a baseline model in all modelling steps provides further context to the performance gains of increasing model complexity. The CCAE model was identified as a candidate for clinical use. The use case demonstrates that IPEV provides a huge opportunity in a new era of standardised data and analytics to improve insight into and trust in prediction models at an unprecedented scale.
Journal Article
Development and validation of a prognostic model predicting symptomatic hemorrhagic transformation in acute ischemic stroke at scale in the OHDSI network
by
Shah, Nigam H.
,
Van Zandt, Mui
,
You, Seng Chan
in
Analysis
,
Biomedical engineering
,
Brain Ischemia - complications
2020
Hemorrhagic transformation (HT) after cerebral infarction is a complex and multifactorial phenomenon in the acute stage of ischemic stroke, and often results in a poor prognosis. Thus, identifying risk factors and making an early prediction of HT in acute cerebral infarction contributes not only to the selections of therapeutic regimen but also, more importantly, to the improvement of prognosis of acute cerebral infarction. The purpose of this study was to develop and validate a model to predict a patient's risk of HT within 30 days of initial ischemic stroke.
We utilized a retrospective multicenter observational cohort study design to develop a Lasso Logistic Regression prediction model with a large, US Electronic Health Record dataset which structured to the Observational Medical Outcomes Partnership (OMOP) Common Data Model (CDM). To examine clinical transportability, the model was externally validated across 10 additional real-world healthcare datasets include EHR records for patients from America, Europe and Asia.
In the database the model was developed, the target population cohort contained 621,178 patients with ischemic stroke, of which 5,624 patients had HT within 30 days following initial ischemic stroke. 612 risk predictors, including the distance a patient travels in an ambulance to get to care for a HT, were identified. An area under the receiver operating characteristic curve (AUC) of 0.75 was achieved in the internal validation of the risk model. External validation was performed across 10 databases totaling 5,515,508 patients with ischemic stroke, of which 86,401 patients had HT within 30 days following initial ischemic stroke. The mean external AUC was 0.71 and ranged between 0.60-0.78.
A HT prognostic predict model was developed with Lasso Logistic Regression based on routinely collected EMR data. This model can identify patients who have a higher risk of HT than the population average with an AUC of 0.78. It shows the OMOP CDM is an appropriate data standard for EMR secondary use in clinical multicenter research for prognostic prediction model development and validation. In the future, combining this model with clinical information systems will assist clinicians to make the right therapy decision for patients with acute ischemic stroke.
Journal Article
Prediction of sustained biologic and targeted synthetic DMARD-free remission in rheumatoid arthritis patients
by
Hügle, Thomas
,
Finckh, Axel
,
Burkard, Theresa
in
Adalimumab
,
Algorithms
,
Anti-inflammatory agents
2021
Abstract
Objectives
The aim was to develop a prediction model of sustained remission after cessation of biologic or targeted synthetic DMARD (b/tsDMARD) in RA.
Methods
We conducted an explorative cohort study among b/tsDMARD RA treatment episode courses stopped owing to remission in the Swiss Clinical Quality Management registry (SCQM; 2008–2019). The outcome was sustained b/tsDMARD-free remission of ≥12 months. We applied logistic regression model selection algorithms using stepwise, forward selection, backward selection and penalized regression to identify patient characteristics predictive of sustained b/tsDMARD-free remission. We compared c-statistics corrected for optimism between models. The three models with the highest c-statistics were validated in new SCQM data until 2020 (validation dataset).
Results
We identified 302 eligible episodes, of which 177 episodes (59%) achieved sustained b/tsDMARD-free remission. Two backward and one forward selection model, with eight, four and seven variables, respectively, obtained the highest c-statistics corrected for optimism of c = 0.72, c = 0.70 and c = 0.69, respectively. In the validation dataset (47 eligible episodes), the models performed with c = 0.99, c = 0.80 and c = 0.74, respectively, and excellent calibration. The best model included the following eight variables (measured at b/tsDMARD stop): RA duration, b/tsDMARD duration, other pain/anti-inflammatory drug use, quality of life (EuroQol), DAS28-ESR score, HAQ score, education, and interactions of RA duration and other pain/anti-inflammatory drug use and of b/tsDMARD duration and HAQ score.
Conclusion
Our results suggest that models with up to eight unique variables may predict sustained b/tsDMARD-free remission with good efficiency. External validation is warranted.
Journal Article
Seek COVER: using a disease proxy to rapidly develop and validate a personalized risk calculator for COVID-19 outcomes in an international network
by
Prats-Uribe, Albert
,
Duarte-Salles, Talita
,
Drakos, Iannis
in
Analysis
,
Coronaviruses
,
COVID-19
2022
Background
We investigated whether we could use influenza data to develop prediction models for COVID-19 to increase the speed at which prediction models can reliably be developed and validated early in a pandemic. We developed COVID-19 Estimated Risk (COVER) scores that quantify a patient’s risk of hospital admission with pneumonia (COVER-H), hospitalization with pneumonia requiring intensive services or death (COVER-I), or fatality (COVER-F) in the 30-days following COVID-19 diagnosis using historical data from patients with influenza or flu-like symptoms and tested this in COVID-19 patients.
Methods
We analyzed a federated network of electronic medical records and administrative claims data from 14 data sources and 6 countries containing data collected on or before 4/27/2020. We used a 2-step process to develop 3 scores using historical data from patients with influenza or flu-like symptoms any time prior to 2020. The first step was to create a data-driven model using LASSO regularized logistic regression, the covariates of which were used to develop aggregate covariates for the second step where the COVER scores were developed using a smaller set of features. These 3 COVER scores were then externally validated on patients with 1) influenza or flu-like symptoms and 2) confirmed or suspected COVID-19 diagnosis across 5 databases from South Korea, Spain, and the United States. Outcomes included i) hospitalization with pneumonia, ii) hospitalization with pneumonia requiring intensive services or death, and iii) death in the 30 days after index date.
Results
Overall, 44,507 COVID-19 patients were included for model validation. We identified 7 predictors (history of cancer, chronic obstructive pulmonary disease, diabetes, heart disease, hypertension, hyperlipidemia, kidney disease) which combined with age and sex discriminated which patients would experience any of our three outcomes. The models achieved good performance in influenza and COVID-19 cohorts. For COVID-19 the AUC ranges were, COVER-H: 0.69–0.81, COVER-I: 0.73–0.91, and COVER-F: 0.72–0.90. Calibration varied across the validations with some of the COVID-19 validations being less well calibrated than the influenza validations.
Conclusions
This research demonstrated the utility of using a proxy disease to develop a prediction model. The 3 COVER models with 9-predictors that were developed using influenza data perform well for COVID-19 patients for predicting hospitalization, intensive services, and fatality. The scores showed good discriminatory performance which transferred well to the COVID-19 population. There was some miscalibration in the COVID-19 validations, which is potentially due to the difference in symptom severity between the two diseases. A possible solution for this is to recalibrate the models in each location before use.
Journal Article