Skip to main content

Development, validation and recalibration of a prediction model for prediabetes: an EHR and NHANES-based study

Abstract

Background

A prediction model that estimates the risk of elevated glycated hemoglobin (HbA1c) was developed from electronic health record (EHR) data to identify adult patients at risk for prediabetes who may otherwise go undetected. We aimed to assess the internal performance of a new penalized regression model using the same EHR data and compare it to the previously developed stepdown approximation for predicting HbA1c ≥ 5.7%, the cut-off for prediabetes. Additionally, we sought to externally validate and recalibrate the approximation model using 2017–2020 pre-pandemic National Health and Nutrition Examination Survey (NHANES) data.

Methods

We developed logistic regression models using EHR data through two approaches: the Least Absolute Shrinkage and Selection Operator (LASSO) and stepdown approximation. Internal validation was performed using the bootstrap method, with internal performance evaluated by the Brier score, C-statistic, calibration intercept and slope, and the integrated calibration index. We externally validated the approximation model by applying original model coefficients to NHANES, and we examined the approximation model’s performance after recalibration in NHANES.

Results

The EHR cohort included 22,635 patients, with 26% identified as having prediabetes. Both the LASSO and approximation models demonstrated similar discrimination in the EHR cohort, with optimism-corrected C-statistics of 0.760 and 0.763, respectively. The LASSO model included 23 predictor variables, while the approximation model contained 8. Among the 2,348 NHANES participants who met the inclusion criteria, 30.1% had prediabetes. External validation of the LASSO model was not possible due to the unavailability of some predictor variables. The approximation model discriminated well in the NHANES dataset, achieving a C-statistic of 0.787.

Conclusion

The approximation method demonstrated comparable performance to LASSO in the EHR development cohort, making it a viable option for healthcare organizations with limited resources to collect a comprehensive set of candidate predictor variables. NHANES data may be suitable for externally validating a clinical prediction model developed with EHR data to assess generalizability to a nationally representative sample, depending on the model’s intended use and the alignment of predictor variable definitions with those used in the model’s original development.

Peer Review reports

Background

Over 1 in 3 American adults have prediabetes [1], defined as having an elevated glycated hemoglobin (HbA1c) ≥ 5.7%, and 80% are unaware that they have it [1]. Additionally, adults with undiagnosed diabetes comprise 23.0% of prevalent diabetes cases in the United States [2]. Patients who go on to develop diabetes are at increased risk of cardiovascular complications [3], and have, on average, healthcare expenses that are 2.3 times higher than persons not diagnosed with diabetes [4]. As such, screening patients for early detection of elevated HbA1c facilitates early intervention, which may prevent or delay disease progression, prevent micro- and macrovascular complications, mitigate unnecessary healthcare expenditures attributable to diabetes, and improve patient outcomes [5]. Diagnostic prediction models that estimate the risk of elevated HbA1c can be implemented into the electronic health record (EHR) to identify those at risk for prediabetes who may otherwise go undetected.

The American Diabetes Association (ADA) and Centers for Disease Control and Prevention (CDC) offer a prediabetes risk test that includes history of gestational diabetes if female, family history of diabetes, physical activity, history of hypertension, patient age, sex, height, and weight [6]. The CDC Risk Score model, originally developed with NHANES 1999–2004 and validated with NHANES 2005–2006 [7], demonstrated poor performance in NHANES 2013–2014 [8]. This decline may be due to differences in predictor variable definitions and temporal changes in prediabetes prevalence. Despite it hovering just above the conventional level of significance, physical inactivity was included in the CDC model because of its protective and modifiable nature [7]. However, variables collected in health surveys, such as physical activity, are not always routinely captured in EHRs, complicating the application of survey-developed models to clinical settings.

Previously, Wells et al. [9] built a logistic regression model from EHR data using stepdown approximation to predict prediabetes. The model-development cohort comprised adult patients of Atrium Health Wake Forest Baptist Medical Center (AHWFBMC) in Winston-Salem, North Carolina who had undergone HbA1c testing, had prior evidence of hyperglycemia, or had a prescription for an antihyperglycemic medication between September 2012 and September 2016 [9]. The selection of candidate predictors was guided by their theoretical relationship to hyperglycemia [9]. Harrell’s model approximation method was used to derive the most parsimonious model [10]. The following predictors were selected from a larger subset of candidate variables in order from most to least importance: age, body mass index (BMI), random glucose, race, serum non–high-density lipoprotein (non-HDL), serum total cholesterol, estimated glomerular filtration rate (eGFR), and smoking status [9]. The approximation model was internally validated using tenfold cross-validation and outperformed alternative models with a C-statistic of 0.765 [9].

Alhassan et al. [11] replicated the Wells et al. approximation model using EHR data from Saudi Arabia by building three models using identical predictors with the exception of race, which was uniform across their patient population, and smoking status, which was absent from their dataset. While the omission of race and smoking status limited the external validation of the original model, they replicated the logistic regression equation used in the original model and validated it through tenfold cross-validation, which yielded commendable accuracy and calibration [11]. Alhassan et al. found that the model with fewer predictors performed the best and that the order of variable importance (most to least important: random glucose, age, eGFR, cholesterol, non-HDL, and BMI) differed from Wells et al. [11] Systematic population differences (e.g., location, data collection processes, and individual characteristics) likely contributed to dissimilarities in variable importance.

Generally, clinical prediction models perform better on the dataset used for development than on new patient populations [12, 13]. External validation is essential for evaluating a model’s predictive performance on a separate dataset that was not part of the model’s development [14, 15]. Assessing a prediction model's performance on new data is crucial for testing its generalizability and transportability, ensuring that it can reliably support decision-making in new patient populations before widespread implementation [12,13,14,15]. Although the Wells et al. model was replicated, it has not been compared to a model using a different variable selection technique within the same development cohort. First, we aimed to compare the performance of the prediabetes diagnostic model developed using the original approximation stepdown procedure to the LASSO method, as these two approaches offer distinct advantages: the approximation method focuses on creating a parsimonious model by retaining the most significant predictors and may be easier to interpret, while LASSO shrinks less relevant predictors toward zero, potentially enhancing model simplicity and performance [10, 16]. By comparing these two methods, we sought to ensure the robustness and potential advantages of each approach in improving predictive performance. Additionally, we aimed to externally validate and recalibrate the approximation model using NHANES data to assess its generalizability and transportability.

Methods

The Wake Forest University Health Sciences institutional review board approved this study (IRB00031798) and waived informed consent. This study conformed to the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD + AI) checklist [17]. The development cohort was identical to that used in Wells et al. [9], and 35 candidate predictor variables were considered (sirolimus was not considered due to low prevalence). The outcome of interest was defined as HbA1c ≥ 5.7%, or prediabetes, per the original study [9] and the ADA’s diagnostic criteria [18]. Given that we used the same dataset previously utilized by Wells et al. for model development, we verified the appropriateness of the sample size using the “pmsampsize” package [19, 20]. Based on an outcome prevalence of 26% [9], a C-statistic of 0.76 [9], and 35 candidate predictors, we calculated that a minimum sample size of 1,789 with 446 events was required for model development that ensured a shrinkage factor of ≥ 0.9, an absolute difference of ≤ 0.05 between the model's apparent and adjusted proportion of variance explained, and a margin of error ≤ 0.05 in the estimate of average outcome risk. LASSO and approximation logistic regression models were built to predict the probability of prediabetes. For the LASSO logistic regression, tenfold cross-validation was used to select the largest lambda at which the deviance was within one standard error of the minimal deviance. For the stepdown approximation procedure, we first fit a full logistic regression model in which continuous variables were fit using the restricted cubic splines function with 3 knots. Then, an ordinary least squares model was used to approximate the linear predictor of the full model. Variables were removed using backward elimination until the R-squared value reached 0.95 [10]. Both the LASSO and approximation models were internally validated using 2,000 bootstrap resamples whereby the entire modeling process (including tuning parameter selection via tenfold cross-validation for the LASSO model and backward elimination for the approximation model) was repeated for each resample to obtain an optimism-corrected C-statistic, calibration intercept, calibration slope, Brier score, integrated calibration index (ICI) [21, 22], and a bias-corrected calibration curve. The stability of the LASSO and approximation models was also assessed by performing 2,000 bootstrap resamples, after which we (1) calculated the average mean absolute difference between individuals' original predictions and those from the bootstrap models, and (2) generated mean absolute predictor error plots, prediction instability plots, and calibration instability plots for the two models [23].

The 2017–2020 pre-pandemic cycle of NHANES was used for external validation [24]. NHANES is a yearly survey of a nationally representative sample, consisting of interviews and physical examinations, designed to evaluate the health of adults and children in the United States. Validating the approximation model in NHANES data allowed us to assess (1) its transportability to a population with a different case-mix than the development data, and (2) its generalizability to a nationally representative sample. We selected data necessary to derive the predictor variables that corresponded to the original set of candidate variables used to build the original model [9]. Due to the limited data availability in NHANES, however, we were not able to identify peripheral vascular disease and neuropathy.

We included adult participants (≥ 18 years of age) with an HbA1c. To focus on those with prediabetes who may have been missed and would likely benefit from identification, we excluded participants who indicated that a doctor told them they had prediabetes or diabetes and those who took a medication indicated for diabetes management. Additional eligibility criteria included an indication of fasting status as this conferred that they were more likely to undergo the panel of laboratory tests necessary to derive many of the candidate variables, which was slightly different from the development data definition since fasting status was not reliably documented in the AHWFBMC EHR. Laboratory values for non-HDL were calculated by subtracting HDL from total cholesterol. A binary variable was created for obesity (BMI ≥ 30). We calculated eGFR using the Chronic Kidney Disease Epidemiology Collaboration (CKD-EPI) Creatinine Equation (2021) based on SCr, age, and sex [25]. We identified participants as taking medications only if the prescription container or pharmacy print-out was observed by the interviewer to limit recall bias and accurately capture prescription data. Only complete cases were used since the investigators felt that imputation would not be appropriate at model deployment [9]. NHANES variable definitions are in Additional File 1. The approximation model was externally validated by applying the original regression coefficients to NHANES and calculating the predicted probability of prediabetes. Additionally, we assessed the approximation model’s performance after recalibrating the intercept and overall calibration slope by fitting a logistic regression model to NHANES with the original approximation model linear predictor as the only covariable [26,27,28]. Predictive performance of the approximation model in NHANES was measured using discrimination (C-statistic), calibration (calibration intercept and slope, ICI), visually by a calibration curve, and simultaneous discrimination and calibration (Brier score). NHANES fasting subsample weights were used to estimate population totals, and external performance metrics were weighted using normalized fasting subsample weights, which were calculated by dividing the fasting subsample weight by the mean of all fasting subsample weights for the NHANES validation cohort [29]. Due to missingness of many candidate variables in NHANES we were not able to externally validate the LASSO model. We verified the appropriateness of the NHANES sample size for external validation using the “pmvalsampsize” package [30]. Based on the observed distribution of the linear predictor for the approximation model in the NHANES sample (mean -1.145, standard deviation 1.374), and an outcome prevalence of 30.1%, a minimum sample size of 2,204 with 664 events was required to externally validate the approximation model that precisely estimated an outcome event proportion of 1 (confidence interval width of 0.2), calibration slope of 1 (confidence interval width of 0.2), and a C-statistic of 0.79 (confidence interval width of 0.1). Statistical analyses were performed in R version 4.4.1 (R Foundation for Statistical Computing) using the “nephro” [31], “glmnet” [32], “rms” (Regression Modeling Strategies), and “CalibrationCurves” [33] packages.

Results

Descriptive statistics for the approximation model predictor variables across the development EHR and NHANES cohorts are in Table 1 (full descriptive statistics are in supplemental Tables 3–5 of Additional File 2). The LASSO model had 23 non-zero coefficients (Table 2, formula is in supplementary Table 6), with an optimism-corrected C-statistic of 0.760, intercept of -0.011, and slope of 0.987. In comparison, the approximation model had 16 coefficients (Table 2, supplemental Tables 7–10), with an optimism-corrected C-statistic of 0.763, intercept of -0.007 and a slope of 0.992. The calibration curve for LASSO revealed some underestimation for patients at moderate risk (Fig. 1). Whereas the calibration curve for the approximation model indicated overestimation for patients at high risk, the approximation was still well calibrated in the lower risk patients where the majority of the population lies (Fig. 2). The average mean absolute difference between the original LASSO model and bootstrapped predictions was 1.1%, and between the original approximation model and bootstrapped predictions was 1.8% (see additional file 2 for model instability plots).

Table 1 Descriptive statistics for approximation model predictor variables in EHR cohort and 2017–2020 (pre-pandemic) NHANES cohort
Table 2 Model performance metrics and 95% confidence intervals for the LASSO and approximation models in the development cohort. Performance of the approximation model in 2017–2020 (pre-pandemic) NHANES is also shown
Fig. 1
figure 1

Calibration curve for LASSO model fit on development EHR cohort

Fig. 2
figure 2

Calibration curve for approximation model fit on development EHR cohort

The NHANES validation cohort included 2,348 participants (Fig. 3). A larger proportion of participants (30.1%) had an elevated HbA1c in NHANES compared to 26.0% of patients in the development cohort. The means for continuous approximation predictors among participants with prediabetes between the development and NHANES cohort were similar (age 54.8 years vs 55.5 years; BMI 33.0% vs 31.1%; non-HDL 144 mg/dL vs 138.7 mg/dL; total cholesterol 192 mg/dL vs 191 mg/dL). The proportion of Black or African American participants was slightly higher (37.1% vs 35%), and the proportion of current smokers was similar (23.6% vs 21.1%), respectively. The approximation model had a C-statistic of 0.787 when applied to NHANES and showed miscalibration and overestimation (Table 2 and Fig. 4; Intercept 0.102; Slope 1.097; ICI 0.020), but the calibration improved after recalibration (Table 2 and Fig. 5; Intercept -0.000; Slope 1; ICI 0.004). Calibration of the approximation model in NHANES modestly improved with use of normalized fasting subsample weights (Table 2 and Fig. 6; Intercept 0.033; Slope 1.103; ICI 0.013). The formulas for the approximation model recalibrated in NHANES with and without sampling weights are available in supplemental Tables 9 and 10. (Fig. 7).

Fig. 3
figure 3

Flow diagram for 2017–2020 (pre-pandemic) NHANES external validation cohort

Fig. 4
figure 4

Calibration curve after applying the approximation model coefficients to NHANES

Fig. 5
figure 5

Calibration curve after recalibrating the approximation model to NHANES

Fig. 6
figure 6

Calibration curve after applying the approximation model coefficients to NHANES with normalized fasting subsample weights

Fig. 7
figure 7

Calibration curve after recalibrating the approximation model to NHANES with normalized fasting subsample weights

Discussion

Both the LASSO and approximation models demonstrated moderate discriminative performance in the development cohort, with the approximation model showing better calibration than the LASSO in development, and improved discrimination in the external validation cohort. The approximation model performed similarly to the LASSO model in the development cohort despite differences in variable selection methods. The LASSO uses shrinkage for variable selection by constraining the absolute values of the regression coefficients to be less than a bound determined by a penalty parameter lambda, effectively zeroing out coefficients of less relevant predictors to simplify the model [16, 32]. In contrast, the stepdown/backwards elimination method approximates the full logistic regression model by retaining the most significant predictors that explain the majority of the variance. This process results in a parsimonious model that effectively captures the key predictive relationships, maintaining similar performance to the full model. The approximation method can lead to a model that is easier to interpret which is beneficial in settings where simplicity and clarity are prioritized. The LASSO method can lead to a more simplified model that reduces overfitting, improves generalizability, and may enhance predictive performance, especially when dealing with large datasets with many predictors.

When applied to NHANES data, the approximation model showed good discriminative ability in predicting elevated HbA1c levels. Recalibration further improved its performance, indicating that some tuning to specific populations can enhance the model's utility for identifying individuals with an HbA1c ≥ 5.7%. The logistic recalibration allowed us to maintain the knowledge collected from the development data while incorporating knowledge from new patient data. We did not see the need for refitting the approximation model in NHANES, as refitting potentially disregards the knowledge gained from the development data [34, 35]. The use of normalized fasting subsample weights had negligible impact on the approximation model’s performance in NHANES, as normalization scales the weights so that their sum equals the sample size, which does not significantly influence individual predictions.

Using surveys or epidemiological cohort studies for EHR model validation, and vice versa, can be challenging due to temporal changes in disease prevalence and differences in variable measurement [36]. For instance, the CHARGE-AF risk score for atrial fibrillation was developed and validated using multiple prospective cohort studies, yet performed poorly when it was validated with single-institution EHR [36, 37]. Cohort studies may also have strict inclusion and exclusion criteria, and models developed using such data may have limited generalizability [27]. Moreover, models developed from routinely-collected EHR data may not perform well on cohort data because of differences in data collection, in patient catchment, and in overall outcome and predictor variable occurrence [38,39,40]. Accordingly, it is recommended that external validation be performed in new data that is relevant to the model’s intended use [14, 15]. While the approximation model is designed to predict elevated HbA1c in a clinical setting, NHANES participants, though originating from a different context than the development data, are relevant because they represent the underlying target superpopulation of patients with prediabetes.

Healthcare organizations with limited resources or smaller populations may benefit from applying the original coefficients described by Wells et al., as this can be implemented using a calculator outside the EHR, while those needing a predefined parsimonious model but lacking the capacity to capture all original variables may prefer recalibrating the approximation model. If data extraction and refitting are feasible, updating the approximation model's coefficients could enhance its discriminative ability. Complete full model refitting and approximation might be desirable for organizations wanting to deploy a prediction model that is more rigorously fit to local patient characteristics and available data, especially if the population is considerably different from the AHWFBMC EHR and NHANES cohorts as demonstrated in Alhassan et al. [11].

Strengths and Limitations

Strengths of our study include the use of a relatively large sample size for model development and the application of bootstrap resampling for internal validation and evaluation of predictive performance. The external validation was limited by a smaller sample size of the NHANES validation cohort, differences in variable definitions between the development cohort and NHANES, and the inability to validate the LASSO model. Although NHANES includes many of the structured data elements found in EHRs, it lacked some of the disease states associated with diabetes risk. To ensure consistent classification of variables, we used International Classification of Diseases 10th Revision (ICD-10) codes in the NHANES drug files to capture diagnoses whenever possible and minimized the use of questionnaire variables.

Conclusion

We showed that an approximation model intended to identify patients with an elevated HbA1c exhibited adequate predictive performance among an external population. The results indicate that the model may be transportable across different settings, making it a valuable and resource-efficient tool in clinical practice for identifying patients with prediabetes who could benefit from early intervention to prevent disease progression and adverse outcomes. Further validating the approximation model with external EHR data would strengthen evidence of the model’s transportability and generalizability, and provide insights into optimizing its integration within different EHR systems for use in routine clinical practice.

Data availability

The data files analyzed in this study are publicly available on the NHANES website at https://wwwn.cdc.gov/nchs/nhanes/Default.aspx. The development dataset and code used for this study is available from the corresponding author upon reasonable request. A study protocol was not prepared.

Abbreviations

ADA:

American Diabetes Association

BMI:

Body mass index

CDC:

Centers for Disease Control and Prevention

CKD-EPI:

Chronic Kidney Disease Epidemiology Collaboration

C-statistic:

Concordance statistic

eGFR:

Estimated glomerular filtration rate

EHR:

Electronic health record

HbA1c:

Glycated hemoglobin

HDL:

High-density lipoprotein

NHANES:

National Health and Nutrition Examination Survey

non-HDL:

Non–high-density lipoprotein

SBP:

Systolic blood pressure

SCr:

Serum creatinine

ICI:

Integrated calibration index

References

  1. Centers for Disease Control and Prevention. National Diabetes Statistics Report website: prevalence of prediabetes among adults. https://www.cdc.gov/diabetes/data/statistics-report/prevalence-of-prediabetes.html. Accessed 20 July 2022.

  2. Centers for Disease Control and Prevention. National Diabetes Statistics Report website: prevalence of both diagnosed and undiagnosed diabetes. https://www.cdc.gov/diabetes/data/statistics-report/diagnosed-undiagnosed-diabetes.html. Accessed 20 July 2022.

  3. Disease C, Management R. Standards of medical care in diabetes-2022. Diab Care. 2022;45(Suppl 1):S144-s174.

    Google Scholar 

  4. American Diabetes Association. Economic costs of diabetes in the U.S. in 2017. Diabetes Care. 2018;41(5):917–928.

  5. Jonas DE, Crotty K, Yun JDY, et al. Screening for prediabetes and type 2 diabetes: updated evidence report and systematic review for the US Preventive Services Task Force. JAMA. 2021;326(8):744–60.

    Article  PubMed  Google Scholar 

  6. American Diabetes Association and Centers for Disease Control and Prevention. Prediabetes risk test. https://www.cdc.gov/diabetes/prevention/pdf/Prediabetes-Risk-Test-Final.pdf. Accessed 20 July 2022.

  7. Bang H, Edwards AM, Bomback AS, et al. Development and validation of a patient self-assessment score for diabetes risk. Ann Intern Med. 2009;151(11):775–83.

    Article  PubMed  PubMed Central  Google Scholar 

  8. De Silva K, Jönsson D, Demmer RT. A combined strategy of feature selection and machine learning to identify predictors of prediabetes. J Am Med Inform Assoc. 2020;27(3):396–406.

    Article  PubMed  Google Scholar 

  9. Wells BJ, Lenoir KM, Diaz-Garelli JF, et al. Predicting current glycated hemoglobin values in adults: development of an algorithm from the electronic health record. JMIR Med Inform. 2018;6(4):e10780.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Harrell FE, Jr. Regression modeling strategies: with applications to linear models, logistic and ordinal regression, and survival analysis. 2nd ed. Cham, Switzerland: Springer International Publishing; 2015.

  11. Alhassan Z, Budgen D, Alshammari R, et al. Predicting current glycated hemoglobin levels in adults from electronic health records: validation of multiple logistic regression algorithm. JMIR Med Inform. 2020;8(7):e18963.

    Article  PubMed  PubMed Central  Google Scholar 

  12. Bleeker SE, Moll HA, Steyerberg EW, et al. External validation is necessary in prediction research: a clinical example. J Clin Epidemiol. 2003;56(9):826–32.

    Article  CAS  PubMed  Google Scholar 

  13. Toll DB, Janssen KJ, Vergouwe Y, et al. Validation, updating and impact of clinical prediction rules: a review. J Clin Epidemiol. 2008;61(11):1085–94.

    Article  CAS  PubMed  Google Scholar 

  14. Collins GS, Dhiman P, Ma J, et al. Evaluation of clinical prediction models (part 1): from development to external validation. Bmj. 2024;384:e074819.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Riley RD, Archer L, Snell KIE, et al. Evaluation of clinical prediction models (part 2): how to undertake an external validation study. Bmj. 2024;384:e074820.

    Article  PubMed  PubMed Central  Google Scholar 

  16. Tibshirani R. Regression shrinkage and selection via the Lasso. J R Stat Soc Series B Stat Methodol. 1996;58(1):267–88.

    Article  Google Scholar 

  17. Collins GS, Moons KGM, Dhiman P, et al. TRIPOD+AI statement: updated guidance for reporting clinical prediction models that use regression or machine learning methods. Bmj. 2024;385:e078378.

  18. Diagnosis and Classification of Diabetes. Standards of care in diabetes-2024. Diabetes Care. 2024;47(Suppl 1):S20-s42.

    Google Scholar 

  19. Riley RD, Snell KI, Ensor J, et al. Minimum sample size for developing a multivariable prediction model: PART II - binary and time-to-event outcomes. Stat Med. 2019;38(7):1276–96.

    Article  PubMed  Google Scholar 

  20. Riley RD, Van Calster B, Collins GS. A note on estimating the Cox-Snell R(2) from a reported C statistic (AUROC) to inform sample size calculations for developing a prediction model with a binary outcome. Stat Med. 2021;40(4):859–64.

    Article  PubMed  Google Scholar 

  21. Austin PC, Steyerberg EW. The Integrated Calibration Index (ICI) and related metrics for quantifying the calibration of logistic regression models. Stat Med. 2019;38(21):4051–65.

    Article  PubMed  PubMed Central  Google Scholar 

  22. Austin PC, Putter H, Giardiello D, et al. Graphical calibration curves and the integrated calibration index (ICI) for competing risk models. Diagn Progn Res. 2022;6(1):2.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Riley RD, Collins GS. Stability of clinical prediction models developed using statistical or machine learning methods. Biom J. 2023;65(8):e2200302.

    Article  PubMed  Google Scholar 

  24. Akinbami LJ, Chen TC, Davy O, et al. National Health and Nutrition examination survey, 2017-March 2020 prepandemic file: sample design, estimation, and analytic guidelines. Vital Health Stat 1. 2022;(190):1–36.

  25. Inker LA, Eneanya ND, Coresh J, et al. New Creatinine- and Cystatin C-Based equations to estimate GFR without race. N Engl J Med. 2021;385(19):1737–49.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  26. Steyerberg EW, Vergouwe Y. Towards better clinical prediction models: seven steps for development and an ABCD for validation. Eur Heart J. 2014;35(29):1925–31.

    Article  PubMed  PubMed Central  Google Scholar 

  27. Steyerberg EW. Clinical prediction models: a practical approach to development, validation, and updating. 2nd ed. Cham, Switzerland: Springer Nature Switzerland AG; 2019.

  28. Vergouwe Y, Nieboer D, Oostenbrink R, et al. A closed testing procedure to select an appropriate method for updating prediction models. Stat Med. 2017;36(28):4529–39.

    Article  PubMed  Google Scholar 

  29. Kaplan D, Ferguson AJ. On the utilization of sample weights in latent variable models. Struct Equ Modeling. 1999;6(4):305–21.

    Article  Google Scholar 

  30. Riley RD, Debray TPA, Collins GS, et al. Minimum sample size for external validation of a clinical prediction model with a binary outcome. Stat Med. 2021;40(19):4230–51.

    Article  PubMed  Google Scholar 

  31. Pattaro C, Riegler P, Stifter G, et al. Estimating the glomerular filtration rate in the general population using different equations: effects on classification and association. Nephron Clin Pract. 2013;123(1–2):102–11.

    Article  PubMed  Google Scholar 

  32. Friedman J, Hastie T, Tibshirani R. Regularization paths for generalized linear models via coordinate descent. J Stat Softw. 2010;33(1):1–22.

    Article  PubMed  PubMed Central  Google Scholar 

  33. Van Calster B, Nieboer D, Vergouwe Y, et al. A calibration hierarchy for risk models was defined: from utopia to empirical data. J Clin Epidemiol. 2016;74:167–76.

    Article  PubMed  Google Scholar 

  34. Janssen KJ, Moons KG, Kalkman CJ, et al. Updating methods improved the performance of a clinical prediction model in new patients. J Clin Epidemiol. J 2008;61(1):76–86.

    Article  CAS  PubMed  Google Scholar 

  35. Davis SE, Greevy RA, Fonnesbeck C, et al. A nonparametric updating method to correct clinical prediction model drift. J Am Med Inform Assoc. 2019;26(12):1448–57.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Goldstein BA, Navar AM, Pencina MJ. Risk prediction with electronic health records: the importance of model validation and clinical context. JAMA Cardiol. 2016;1(9):976–7.

    Article  PubMed  PubMed Central  Google Scholar 

  37. Kolek MJ, Graves AJ, Xu M, et al. Evaluation of a Prediction model for the development of atrial fibrillation in a repository of electronic medical records. JAMA Cardiol. 2016;1(9):1007–13.

    Article  PubMed  PubMed Central  Google Scholar 

  38. Gianfrancesco MA, Goldstein ND. A narrative review on the validity of electronic health record-based research in epidemiology. BMC Med Res Methodol. 2021;21(1):234.

    Article  PubMed  PubMed Central  Google Scholar 

  39. la Roi-Teeuw HM, van Royen FS, de Hond A, et al. Don’t be misled: 3 misconceptions about external validation of clinical prediction models. J Clin Epidemiol. 2024;172:111387.

    Article  PubMed  Google Scholar 

  40. Debray TP, Vergouwe Y, Koffijberg H, et al. A new framework to enhance the interpretation of external validation studies of clinical prediction models. J Clin Epidemiol. 2015;68(3):279–89.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

This work is dedicated to the memory of Dr. Michael W. Kattan, whose insight and passion have inspired advancements in medical risk prediction. His cheerful demeanor, kindness, and unwavering support left a lasting impact on all who had the privilege of working with him.

Funding

This work was supported by the National Center for Advancing Translational Sciences (NCATS) of the National Institutes of Health under Award Number TL1TR003136, and the Wake Forest Clinical and Translational Science Institute's NCATS Grant UL1TR001420. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Author information

Authors and Affiliations

Authors

Contributions

N.J.C., K.M.L., and B.J.W. conceptualized and designed the study. All authors contributed to data acquisition, analysis, and interpretation. All authors contributed to drafting and approving the final version of the manuscript.

Authors’ information

Dr. Nicholas J. Casacchia, PharmD, MS, is an analyst at Cleveland Clinic. The work described in this manuscript was started while Dr. Casacchia was a Postdoctoral Fellow at Wake Forest University School of Medicine and does not necessarily represent the views of Cleveland Clinic.

Corresponding author

Correspondence to Nicholas J. Casacchia.

Ethics declarations

Ethics approval and consent to participate

The Wake Forest University Health Sciences institutional review board approved this study and waived informed consent (IRB00031798).

All methods were conducted in accordance with the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD + AI) checklist (Additional File 3).

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Casacchia, N.J., Lenoir, K.M., Rigdon, J. et al. Development, validation and recalibration of a prediction model for prediabetes: an EHR and NHANES-based study. BMC Med Inform Decis Mak 24, 387 (2024). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-024-02803-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12911-024-02803-w

Keywords