Abstract
A significant minority of individuals develop trauma- and stressor-related disorders (TSRD) after surviving sepsis, a life-threatening immune response to infections. Accurate prediction of risk for TSRD can facilitate targeted early intervention strategies, but many existing models rely on research measures that are impractical to incorporate to standard emergency department workflows. To increase the feasibility of implementation, we developed models that predict TSRD in the year after survival from sepsis using only electronic health records from the hospitalization (n = 217,122 hospitalizations from 2012-2015). The optimal model was evaluated in a temporally independent prospective test sample (n = 128,783 hospitalizations from 2016-2017), where patients in the highest-risk decile accounted for nearly one-third of TSRD cases. Our approach demonstrates that risk for TSRD after sepsis can be stratified without additional assessment burden on clinicians and patients, which increases the likelihood of model implementation in hospital settings.
Similar content being viewed by others
Every year millions of patients develop sepsis, a life-threatening, dysregulated immune response to infection [1]. Advances in sepsis care over the prior two decades have improved survival rates, thus requiring an increased focus on the prevention of adverse long-term mental health outcomes, including posttraumatic stress disorder (PTSD) and other trauma-related psychiatric outcomes [1]. A systematic review [2] documented a wide range of PTSD prevalence among survivors of sepsis and other conditions treated at intensive care units (8% to 51%). Accurate and efficient prediction of PTSD risk may facilitate targeted screening and delivery of preventive and early interventions after hospital discharge [3]. While prior PTSD research typically focused on the identification of key risk factors [4], machine learning models aim to increase predictive accuracy by incorporating a large set of individually weak predictors and complex interactions [3]. Among individuals treated at intensive care units, systematic reviews document several predictors of PTSD, including pre-existing psychopathology, benzodiazepine administration, female sex, age, sedation, and trait anxiety [2, 5]. However, machine learning models that could potentially incorporate these and other variables to predict trauma-related psychiatric outcomes, including PTSD after survival from sepsis have not been developed.
Recent reviews summarize research on machine learning models to predict PTSD after a variety of potentially traumatic experiences, including military combat, natural disasters, transportation accidents, physical and sexual assault, and concussion [3, 6]. Some of these studies demonstrated that machine learning models could accurately identify the emergency department (ED) admits at high risk for PTSD after a variety of injuries resulting from motor vehicle accidents, falls, work-related accidents, physical and sexual assault, gunshot wounds, and terrorist attacks [7,8,9,10,11,12]. However, several limitations may impede real-world implementation. Most of the existing models have been developed from extensive research assessment batteries that are not part of standard ED workflows. Although some of these models included information that can be found in patient charts, including patient characteristics (e.g., age, sex, body mass index), injury characteristics, vital signs, lab results, medications, and pre-existing diagnoses, many of the most important predictors were from study surveys that assessed acute psychological responses, social support, and resilience [7,8,9,10,11,12]. Another limitation of past studies is that models were developed with relatively small samples ranging from 152 to 1003 participants [7,8,9,10,11,12]. Moreover, many of the samples were substantially reduced as a result of stringent inclusion/exclusion criteria, which may limit the generalizability of models to patient populations that meet similar research criteria. Finally, most studies relied on self-reported symptom measures to assess PTSD outcomes and did not consider other trauma- and stressor-related disorders. Our aim was to overcome these limitations by developing and validating a predictive model that relied solely on electronic health records (EHR) data collected at the time of hospitalization to predict clinician-diagnosed trauma- and stressor-related disorders (TSRD) in the year post-discharge from hospitalization for sepsis in a large, multi-site sample that included all survivors.
Methods
We used the Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD) guidelines (https://www.equator-network.org/reporting-guidelines/tripod-statement/). The TRIPOD Checklist for Prediction Model Development and Validation is provided in the Supplementary Information (Appendix 1). This study was approved by the Kaiser Permanente Northern California Institutional Review Board. The data from this study cannot be shared publicly as we do not have permission from patients to share their data outside of the Kaiser Permanente Northern California healthcare system.
Study Sample
Inclusion criteria were hospitalizations for sepsis or suspected infection that were admitted through emergency departments across 21 hospitals within the Kaiser Permanente Northern California integrated healthcare delivery system between January 1, 2012, and December 31, 2017 (Fig. 1). This resulted in the inclusion of 364,964 hospitalizations for sepsis or suspected infection (221,358 unique patients). We excluded 19,059 cases (5.2%) in which the patient died during the hospitalization. No other exclusion criteria were applied, resulting in a final sample of 345,905 hospitalizations (210,946 unique patients).
Predictors
To extract predictors, we used EHR data collected during the hospitalization, including the types of variables that have been used in prior studies that developed predictive models of psychiatric outcomes after hospitalization (e.g., patient characteristics, vital signs, lab results, medications, and pre-existing psychiatric diagnoses) [7,8,9,10,11,12]. Additional variables were selected based on clinical considerations of sepsis-related factors that could potentially contribute to long-term medical outcomes, including cardiovascular disease [13,14,15]. We also used natural language processing to extract predictors from clinician notes. Across all data types, one-hot encoding was applied to categorical variables (i.e., each category was converted to a binary indicator where 1 = present, 0 = absent), and variables with near-zero variance were excluded. This resulted in a final set of 607 predictors, including medications, clinician notes, laboratory values, mental status assessments, admission and hospitalization information, vital signs, demographics, and mental health diagnoses; data preparation procedures for each of these data types are described below.
Two-hundred and forty-six of the predictors (40.5%) were derived from medications that were administered during hospitalizations. For each medication, several indicators were created that varied in their level of specificity: 1) therapeutic class, 2) pharmaceutical class, 3) pharmaceutical subclass, and 4) National Drug Code (NDC; the Food and Drug Administration’s drug identifier). For example, the administration of olanzapine resulted in indicators for 1) psychotherapeutic drugs (therapeutic class), 2) atypical antipsychotic, dopamine-serotonin antagonist (pharmaceutical class), 3) thienobenzodiazepine (pharmaceutical subclass), and 4) 00002411633 (NDC). Each of the medication variables captured the number of times a patient received each class of (or specific) medication during the hospitalization. For example, a value of 3 for the variable “00002411633 (NDC)” would indicate that the specific medication Olanzapine (the drug that corresponds with the NDC code 00002411633) was administered 3 times during the hospitalization, a value of 12 for the variable “Psychotherapeutic medication class” would indicate psychotherapeutic medications (which is a class that includes many specific medications) were administered 12 times, and so on.
Ninety-eight of the predictors (16.1%) were presenting symptoms from clinician notes in the first 24 hours of hospitalization. The rationale for focusing on notes from the first 24 hours of hospitalization was to capture disease severity at the onset. The I2E natural language processing (NLP) software (Linguamatics I2E 5.4.1R13, Cambridge, UK) was used to identify and process presenting symptoms from raw text based on existing medical ontologies [16]. The most frequent terms were grouped into categories based on similarity and binary variables were created for each category to indicate the presence/absence of a symptom for each hospitalization. Additional details about these NLP-derived variables are published elsewhere [16].
Seventy-five of the predictors (12.4%) were derived from laboratory values for creatinine, hematocrit, total white blood cell count, glucose, sodium, bicarbonate, blood urea nitrogen, albumin, anion gap, troponin, total serum bilirubin, lactate, arterial PaCO2, arterial PaO2, and pH. Since laboratory analyses could be repeated during a hospitalization, variables were created to indicate the first, last, minimum, maximum, and mean of each lab value.
Seventy-five of the predictors (12.4%) were derived from assessments indicating mental status, level of consciousness (e.g., alert, awake, lethargic, stuporous), orientation (e.g., person, place, time, unable to assess), Schmid assessment (e.g., periodic confusion, comatose/unresponsive), speech (e.g., clear, slurred, rambling, no verbal response), pupil response, and Glasgow Coma Scale (GCS) total. For each of these, variables were created to indicate the first and last assessment values; means, minimums, and maximums were also created for the GCS total.
Forty-six of the predictors (7.6%) were derived from admission and hospitalization information, including the admission category; Charlson Comorbidity Index; Comorbidity Point Score (COPS2); Laboratory-based Acute Physiology Score (LAPS2); height, weight, and body mass index; whether the patient was admitted to the intensive care unit (direct, late, or otherwise) and if so the length of stay; whether the patient was transported in; whether the patient was an observation-only admission; time to antibiotic administration; the principal diagnosis; and the unit for the first inpatient stay.
Thirty of the predictors (4.9%) were derived from vital signs indicating heart rate, oxygen saturation, respiratory rate, temperature, and blood pressure. Variables were created to indicate the first, last, minimum, maximum, and mean of each vital sign value.
Twenty-five of the predictors (4.1%) were derived from demographic information such as age, race, ethnicity, and sex.
Twelve of the predictors (2.0%) were derived from psychiatric diagnostic codes recorded during the hospitalization (indicating that the patient received services related to this diagnosis during the hospitalization) or present in the patient’s problem list (indicating that the pre-existing diagnosis may be a prominent concern in the patient’s ongoing health care): mood disorder, alcohol use disorder, substance use disorder, anxiety disorder, major depressive affective disorder (single episode unspecified, recurrent episode unspecified, or recurrent episode in full remission), alcohol abuse, depressive disorder not elsewhere classified, anxiety state unspecified, and trauma- and stressor-related disorder (see Outcome section below).
Outcome
The target outcome was defined as the presence of a diagnostic code for a trauma- and stressor-related disorder (TSRD) in at least one encounter of any type (e.g., inpatient or outpatient) within 12 months postdischarge from the sepsis hospitalization. This outcome definition is based on the DSM-5 [17], which created TSRD as a distinct cluster that includes the diagnoses post-traumatic stress disorder (PTSD), acute stress disorder, adjustment disorder, other specified trauma- and stressor-related disorder, and unspecified trauma- and stressor-related disorder due to overlapping etiologies and symptoms [18]. Although the clinical impairment associated with acute stress disorder and adjustment disorder may be milder than what is typically observed in PTSD, meta-analyses suggest that it is not uncommon for these diagnoses to precede more severe mental health outcomes, including PTSD [19, 20]. Therefore, predicting risk for any of these disorders may facilitate the development and evaluation of preventive and early interventions, including targeted screening. Given systematic reviews and meta-analyses found long-term rates of prevalence of psychiatric outcomes among critical-illness survivors [2, 21, 22], a 12-month follow-up window was selected to capture cases that may have delayed emergence (or reporting) of symptoms. Supplementary Table 1 includes the full list of International Classification of Diseases (ICD) codes that were included in the outcome.
Among the 345,905 hospitalizations that were included in this study, 112,086 (32.4%) had censored (unobserved) outcomes in the one-year follow-up window due to either death (n = 11,940; 3.5%) or to lapses in health insurance coverage (n = 100,146; 29.0%). Previous prediction models excluded cases from the model development and evaluation if the outcome was missing [7, 9, 10, 12, 23, 24]. This is problematic because complete case analyses are only valid when data are missing completely at random (MCAR) [25], which is an unrealistically strong assumption that ignores the potential for attrition bias and can result in inaccurate performance estimates. Consistent with more recent prediction models [11, 26], we accounted for these encounters by using inverse probability of censoring weights, a robust approach for including cases when outcomes are missing at random (MAR; a less stringent assumption than MCAR); this approach leads to less biased performance estimates compared to complete-case approaches [27]. A stacked ensemble of ML algorithms (i.e., the approach used for the TSRD outcome model described in Statistical Analyses) was used to estimate the probability of having an observed outcome using all the pre-discharge predictors, and the inverse of these probabilities were used as weights in the TSRD prediction models. This non-parametric approach can capture complex high-dimension interactions among variables and does not make distributional assumptions about relations among variables and missingness. These weights were used across all analyses so that missing outcome data were always taken into account.
Statistical analysis
All analyses were conducted with the open-source statistical software R (version 4.1.1) [28]. Supplementary Figure 1 provides a schematic of the modeling pipeline, which is described in detail below. To protect against overfitting, ML models are often developed on a random subset of the data (i.e., training or development sample) and tested on the held-out portion (i.e., test sample). However, random subsampling may also lead to inaccurate estimates of performance given that the development dataset may contain participants whose data were collected after participants used to validate the model. In the real world, models need to generalize to future cases. To get closer to this ideal we used a prospective validation approach that is designed to assess how the prediction model may perform in future cases. All data used to develop and compare candidate models were from hospitalizations with discharge dates between 2012 and 2015 (n = 217,122). The remaining data (discharge dates between 2016-2017; n = 128,783) were used to test the best-performing model selected in the development phase.
In real-world healthcare settings, predictive models can be embedded within EHR systems to make predictions at the encounter level; therefore, models are typically developed and tested using encounter-level data [29,30,31]. In large samples that cover a wide span of time, it is common for some patients to have multiple encounters. To reduce the influence of patients with multiple encounters in the model development phase, cross-validation folds were stratified such that encounters from the same patient were all contained within the same fold. The performance of the final model was evaluated in the temporally independent prospective test sample, which could contain new encounters from patients who also had encounters in the development sample; therefore, sensitivity analyses were conducted to examine the performance of the model on the sample of encounters from new patients only.
Models were developed with the h2o machine learning platform [32] using approaches with varying levels of complexity. The least complex model was logistic regression with elastic net regularization, which can yield a model with fewer predictors because some predictor coefficients may be reduced to zero. Next in order of complexity was a core-predictor gradient boosting machine (GBM) model, which is a non-parametric decision tree approach that can capture complex high-dimensional interactions among predictors. An information-theoretic approach [33] was used to identify a set of core predictors using 20% of the development sample and the remaining 80% of the sample was used to develop a GBM model to predict TSRD using only core predictors. The most complex model was a stacked ensemble of various decision-tree algorithms, including GBM, XGBoost, distributed random forests, and extremely randomized trees [34]. In this approach, predictions from these base-learners were integrated by a meta-learner to generate a final prediction. Across all approaches, cross-validation was used to train models by optimizing log-loss, which captures the discrepancy between predicted probabilities and true outcomes.
To select a final model, we considered log-loss (lower is better) and area under the receiver operating characteristics (ROC) curve (AUC; higher is better, and an AUC of 0.50 indicates random or chance-level performance); these performance metrics were estimated in the development sample using 10-fold cross-validation (8 folds for the core-predictor GBM since two of the folds were used to identify core predictors). Consistent with prior prediction models, patients who had a pre-existing trauma- and stressor-related psychiatric diagnoses were not excluded from the sample [11, 12, 24, 26]. Instead, a univariate generalized linear model that used pre-existing TSRD diagnosis as the sole predictor of TSRD in the follow-up phase served as an informative benchmark: Yi = β0 + β×(pre-existing PTSD)i + εi The purpose of testing whether models outperformed this benchmark is two-fold. First, it is a more rigorous benchmark than prediction better than chance [35]. Second, this benchmark model captures how much of the predictive accuracy is driven by pre-existing diagnosis. Similar benchmarks have been used in other predictive models of psychiatric outcomes after trauma exposure [11, 12, 24, 26]. Given that simpler models may be easier to implement, we also took predictor parsimony into account such that a model with fewer predictors and comparable performance would be favored.
Once an optimal model was selected, it was applied to the independent test dataset. We assessed model discrimination with an ROC curve and AUC statistic, and model calibration with a logistic calibration curve and expected calibration error (ECE). We also measured sensitivity, specificity, and positive and negative predictive value in the test sample across deciles based on the predicted risk distribution in the development sample. The mean of each performance metric was estimated with 95% confidence intervals (CI) using weighted bootstrapping (1000 replications).
In addition to these performance metrics, we examined the relative importance of predictors in the optimal model with the Shapley Additive Explanations (SHAP) approach [36]. This method produces a plot that summarizes each predictor’s impact on model predictions across values of the predictor. As such, it provides information about the predictor’s relative importance (the magnitude of its impact on the final model prediction) and potential directionality (whether high or low values of the predictor are associated with increased or decreased probability of TSRD). Another advantage of this approach is that it is model-agnostic, which means it can be applied to any model (e.g., elastic net, GBM, or stacked ensemble). These analyses used the development sample to provide insight into what the optimal model learned. Although these analyses should not be causally interpreted, they may indicate which variables were most useful for predicting the outcome.
Finally, we examined whether model implementation may potentially lead to unfair outcomes across sociodemographic categories. Algorithmic fairness is a highly active area of research and there is currently no consensus on the optimal criteria to assess fairness [37]. We estimated ROC curves and AUC statistics across the self-reported categories of race, Hispanic/Latino ethnicity, and sex. ROC curves can reveal threshold regions where differences in sensitivity and specificity may exist across subgroups, and the AUC statistic can indicate whether there may be differences in overall discrimination across all thresholds. These analyses used the test sample because their purpose is to assess whether implementation of the final model could potentially lead to unfair outcomes.
Results
Table 1 displays basic demographic characteristics of the sample; no meaningful differences were observed across the development and test samples. The prevalence of TSRD in the year post-discharge was 7.1% (6.8% in the development sample and 7.4% in the test sample). This was considerably higher than the prevalence of pre-existing TSRD in the medical charts at the time of hospitalization (1.5% in the development sample and 1.6% in the test sample).
Model performance in the development phase was best for a stacked ensemble of decision-tree algorithms (log-loss = 0.219, AUC = 0.73, 607 predictors), followed by a gradient-boosted machine (GBM) model with core predictors (log-loss = 0.219, AUC = 0.71, 29 predictors), and an elastic net model (log-loss = 0.221, AUC = 0.71, 487 predictors). All models outperformed a benchmark model that used pre-existing TSRD diagnosis as the sole predictor of TSRD in the follow-up phase (log-loss = 0.229, AUC = 0.56). Moreover, the benchmark model results suggest that simply predicting follow-up TSRD on the basis of pre-existing TSRD is only slightly better than chance prediction, whereas incorporating predictors from the hospitalization resulted in meaningful increases in predictive performance. Specifically, when compared to the benchmark model, the other models showed a 27-30% increase in AUC, and a 3.5-4.4% reduction in log-loss.
We selected the GBM for validation because it achieved comparable performance to the stacked ensemble and elastic net using only 29 predictors (as opposed to hundreds). In the temporally independent prospective test sample, the GBM had similar classification performance, AUC = 0.72, 95% CI [0.71, 0.72], and low expected calibration error, ECE = 0.009, 95% CI [0.008, 0.011], suggesting good generalization (Fig. 2). Sensitivity analyses showed that results were similar after removing encounters in the test dataset from patients who also had encounters in the development dataset (AUC = 0.72, 95% CI [0.71, 0.73], ECE = 0.009, 95% CI [0.007, 0.011]; Supplementary Fig. 2). Table 2 provides the threshold-dependent metrics of the GBM model in the test sample across deciles based on the predicted risk distribution in the development sample. The 8.7% of patients whose predicted risk for TSRD fell in the highest risk decile accounted for nearly one third of the TSRD cases (Table 2).
Fig. 2C illustrates the relative importance of the 29 core predictors included in the final GBM arranged in order of the magnitude of their total contribution to predictions. Each point in the SHAP plot represents the impact of that predictor on an individual’s prediction, with positive and negative values denoting increases and decreases in the probability of TSRD, respectively. A key finding from the SHAP analyses is that while pre-existing TSRD made the strongest contribution to increased probability of TSRD, it did not make the strongest contribution overall. This is because only a small fraction of the development sample (1.5%) had a pre-existing TSRD diagnosis; for the remaining 98.5% that did not, this predictor was not very informative. This is consistent with the results of the benchmark model that used pre-existing TSRD as the sole predictor, which performed poorly (AUC = 0.56). This suggests that all variables in the model were weak as individual predictors, and that the final model’s predictive power arised from the collective contribution of variables and their interactions. The SHAP plot also provides information about the potential directionality of impact on predictions (note that the GBM is a nonparametric model that can capture non-linear relations). For most predictors, higher values (indicated in blue) were associated with higher predicted probability of TSRD, whereas lower values (indicated in yellow) were associated with a lower predicted probability of TSRD. Two clear exceptions to this pattern were age and heart rate. Table 3 contains a complete list of core predictors which cut across nine domains, including pre-existing TSRD diagnosis, other mental health diagnoses, patient characteristics, medications, lab values, vital signs, medical comorbidity, mental status, and admissions data. Weighted descriptive statistics are provided for the development and test samples, including the median and interquartile range (IQR) of each continuous predictor and proportions for the binary predictors. The standardized total- and net-predictive values from the core predictor analysis are also provided; these reflect the strength of each predictor’s association with the outcome, and its unique contribution to the prediction of the outcome, respectively.
Fifteen predictors (51.7%) had missing data. Arterial PaO2 and lactate are lab values that are not necessarily collected for every patient. Accordingly, 84.6% of the development sample (86.0% in the test sample) did not have Arterial PaO2 data, and 30.6% of both the development and test sample did not have lactate data. The remaining lab and vital sign values (bicarbonate, creatinine, white blood cell count, temperature, systolic blood pressure, hematocrit, diastolic blood pressure, heart rate, oxygen saturation), and mental status (alert and oriented X 3) were missing for 5.3–5.9% of the development sample and 9.7–10.1% of the test sample. The only remaining variable with missing data was height (0.5% in the development sample, and 0.4% in the test sample). These were missing due to the application of preprocessing steps in the data extraction that remove typos and clinically unrealistic values (e.g., negative values, extreme values). As noted in the Methods, all of the modeling approaches dealt with missing predictor data so that all hospitalizations could be included in the analysis. ROC curves and corresponding AUC statistics were nearly identical in the subsample of participants who had or did not have missing predictor data related to typos or clinically unrealistic values (Supplementary Fig. 3).
Model performance in the independent test sample was also examined across the self-reported demographic characteristics of ethnicity, race, and sex. Across these subgroups, TSRD prevalence ranged from 4.8% among the Asian subgroup to 10.1% among the Black or African American subgroup. Mean classification performance across subgroups, indexed by AUC, ranged from 0.70 to 0.76 (Supplementary Table 2). Figure 3 displays the ROC curves for subgroups across ethnicity, race, and sex, which showed considerable overlap in sensitivity across different levels of specificity. This suggests that model performance (in terms of sensitivity and specificity) is comparable across subgroups at a wide range of potential thresholds.
Discussion
Our study makes several contributions to the literature on prediction models of trauma-related psychiatric outcomes developed in hospital settings. This is the first study to focus on the life-threatening trauma of sepsis, a leading cause of death among hospitalized individuals [1]. Our sample size is over 300 times larger than the largest prior study [11]; this facilitated the comparison of model performance across sociodemographic subgroups, which is a critical step in ensuring that models can be implemented fairly across diverse patients. Our only exclusion criterion was death during the hospitalization, which may increase generalizability by ensuring that the model development and validation process included patients with conditions that often form part of exclusion criteria in the prior studies. In contrast to the use of a random sample for model validation, we used a more rigorous temporal validation approach which is more closely aligned to the goal of estimating model performance in future cases. While the AUC of our model (0.72) was in the lower end of the range of performance measured in previous models (AUC range 0.71 to 0.89) [7,8,9,10,11,12], our model is the first to solely rely on data from the electronic health records for predictors and outcomes. This increases the feasibility of model implementation because our model does not add any assessment burden to clinicians or patients. Together, this highlights the clinical significance of our results, which suggest that risk for TSRD after survival from sepsis can be stratified by our model. These findings are relevant to real-world clinicians because they indicate potential for a tool that may augment decision-making around prevention and early intervention for TSRD without increasing assessment burden during the hospitalization.
Variable importance analyses should not be causally interpreted, but they provide several insights about the prediction model. First, it is not surprising that the final set of predictors cut across a wide range of domains given that the core predictors approach is designed to identify variables that make unique contributions to the prediction. In fact, the only category of predictors that was not used by the optimal model was the clinician note variables. Although it is plausible the type and severity of presenting symptoms captured in the clinician notes are useful predictors of TSRD, this information may have been also captured by other variables that are associated with sepsis severity, including labs, vital signs, and medications administered. It is also possible that the NLP-identified symptoms were less consistent as predictors due to clinician variability in recording practices. Second, despite its status as one of the strongest predictors, pre-existing TSRD diagnosis on its own was a poor predictor of post-discharge TSRD. This highlights a strength of the GBM algorithm, which can make good predictions from individually weak predictors by capturing complex, high-dimensional interactions.
Our use of real-world EHR data resulted in a pool of predictors that have not been examined in previous predictive models of trauma-related psychiatric outcomes after hospitalization. Nevertheless, 10 of the 29 predictors in our model overlap with key predictors in previous studies. For example, all of the previous studies we reviewed (which mostly focused on PTSD outcomes) found that a measure of PTSD sequelae in the hospital (e.g., pre-existing condition, symptom levels from previous trauma, acute PTSD symptoms in the ED) was a strong predictor of PTSD in the follow-up phase [7,8,9,10,11,12]. Moreover, the studies that tested benchmark models similar to ours also found poorer performance relative to complex models with more predictors, with most AUCs in the 0.56 to 0.66 range [8,9,10,11], and one smaller study finding an AUC in the good range (0.78) [12]. Most of the reviewed studies also found age to be a key predictor [7,8,9,10, 12]; however, the direction of the association has been mixed, which highlights the need for caution when interpreting results at the level of individual predictors. The remaining predictors in our model that overlap with predictors identified in other studies were pre-existing mood disorder [8, 9, 11, 12], heart rate [7, 9, 11, 12], systolic blood pressure [7, 8], mental status [10, 12], length of stay [10, 12], pre-existing anxiety disorder [12], creatinine [8], and hematocrit [8]. Together this suggests that the predictive power of our model, as well as prior models, arises from a combination of information related to mental health (e.g., pre-existing diagnoses, psychiatric medications) alongside variables that capture the severity of the acute trauma (e.g., lab values, length of stay, administration of analgesics). Although none of these predictive models were designed to test causal hypotheses, our study replicated the predictive value of the aforementioned variables and also identified novel predictors, which is an important contribution to improving predictive models for trauma-related psychiatric outcomes. Future studies that include patients with other conditions in addition to sepsis can determine whether some of the predictors we found are unique to sepsis patients.
We note several limitations. We performed an exhaustive search for TSRD diagnostic codes across all healthcare system encounters in the year post-discharge; however, it is possible that some patients did not report symptoms to their providers and were therefore not assessed for TSRD. Although several algorithms have been developed and tested within VA healthcare systems to identify PTSD cases using a combination of diagnostic codes and other EHR data [38, 39], future research is necessary to improve the identification of TSRDs in civilian healthcare systems. This may improve the identification of cases that may otherwise go undetected, including individuals with subthreshold symptoms of PTSD, which have been linked to clinically significant distress and impairment [40]. Additionally, future studies based on real-world data would benefit from large-scale implementation of self-report assessments of TSRD symptoms in order to improve the identification of positive (and negative) cases, given that there are differences in access to care across patients, and diagnostic practices across clinicians. This would also provide an avenue to assess whether other trauma besides (or in addition to) the sepsis hospitalization played a role in the diagnosis. To our knowledge, this is the first prediction model for TSRD that incorporated natural language predictors from clinician notes. Nevertheless, a limitation is that we focused on clinician notes from the first 24 hours of hospitalization in order to capture signs and symptoms of the presenting illness. Future studies can incorporate notes from the complete hospitalization period, and apply recent advances in natural language processing to extract information. Although we found comparable classification performance across sociodemographic subgroups, algorithmic fairness is a highly active area of research and there is currently no consensus on the optimal criteria to assess fairness [37]. Moreover, we selected the sociodemographic characteristics that are most commonly examined and that are well characterized in the EHR; future research can examine additional factors, including social determinants of health, which are increasingly integrated into EHR. Although we used inverse probability weighting to reduce the potential for attrition bias in performance estimates, unmeasured confounding is a possibility. Predictive models are not designed to test causal hypotheses; therefore, the core predictors identified during the model development phase should not be interpreted as causal factors [41]. Additionally, predictors that were not selected by the core-predictor analysis should not be assumed to be unrelated to the outcome. Finally, a key strength of this study is the minimal exclusion criteria; although we excluded individuals with lapses in health insurance coverage, we did not impose criteria related to the number of encounters with the healthcare system during the follow-up window, which may have associations with the presence of TSRD diagnosis in the EHR.
Conclusions
The predictive model we developed demonstrates that risk for a TSRD diagnosis in the year after survival from sepsis can be stratified without additional assessment burden on clinicians and patients because it relies on a parsimonious set of predictors that are a standard part of hospital workflows. Future research is needed to clarify whether our model attains a level of accuracy needed for cost-effective implementation of targeted prevention and early intervention strategies.
Data availability
The data from this study cannot be shared publicly as we do not have permission from patients to share their data outside of the Kaiser Permanente Northern California healthcare system.
References
Prescott HC, Angus DC. Enhancing Recovery From Sepsis: A Review. JAMA. 2018;319:62–75.
Davydow DS, Gifford JM, Desai SV, Needham DM, Bienvenu OJ. Posttraumatic stress disorder in general intensive care unit survivors: A systematic review. Gen Hosp Psychiatry. 2008;30:421–34.
Schultebraucks K, Chang BP. The opportunities and challenges of machine learning in the acute care setting for precision prevention of posttraumatic stress sequelae. Exp Neurol. 2021;336:113526.
Tortella-Feliu M, Fullana MA, Pérez-Vigil A, Torres X, Chamorro J, Littarelli SA, et al. Risk factors for posttraumatic stress disorder: An umbrella review of systematic reviews and meta-analyses. Neurosci Biobehav Rev. 2019;107:154–65.
Tedstone JE, Tarrier N. Posttraumatic stress disorder following medical illness and treatment. Clin Psychol Rev. 2003;23:409–48.
Ramos-Lima LF, Waikamp V, Antonelli-Salgado T, Passos IC, Freitas LHM. The use of machine learning techniques in trauma-related disorders: A systematic review. J Psychiatr Res. 2020;121:159–72.
Schultebraucks K, Sijbrandij M, Galatzer-Levy I, Mouthaan J, Olff M, van Zuiden M. Forecasting individual risk for long-term Posttraumatic Stress Disorder in emergency medical settings using biomedical data: A machine learning multicenter cohort study. Neurobiol Stress. 2021;14:100297.
Schultebraucks K, Shalev AY, Michopoulos V, Grudzen CR, Shin S-M, Stevens JS, et al. A validated predictive algorithm of post-traumatic stress course following emergency department admission after a traumatic stressor. Nat Med. 2020;26:1084–8.
Galatzer-Levy IR, Ma S, Statnikov A, Yehuda R, Shalev AY. Utilization of machine learning for prediction of post-traumatic stress: a re-examination of cortisol in the prediction and pathways to non-remitting PTSD. Transl Psychiatry. 2017;7:e0.
Galatzer-Levy IR, Karstoft K-I, Statnikov A, Shalev AY. Quantitative forecasting of PTSD from early trauma responses: a Machine Learning application. J Psychiatr Res. 2014;59:68–76.
Ziobrowski HN, Kennedy CJ, Ustun B, House SL, Beaudoin FL, An X, et al. Development and validation of a model to predict posttraumatic stress disorder and major depression after a motor vehicle collision. JAMA Psychiatry. 2021;78:1228–37.
Papini S, Pisner D, Shumake J, Powers MB, Beevers CG, Rainey EE, et al. Ensemble machine learning prediction of posttraumatic stress disorder screening status after emergency room hospitalization. J Anxiety Disord. 2018;60:35–42.
Walkey AJ, Knox DB, Myers LC, Thai KK, Jacobs JR, Kipnis P, et al. Prognostic Accuracy of Presepsis and Intrasepsis Characteristics for Prediction of Cardiovascular Events After a Sepsis Hospitalization. Crit Care Explor. 2022;4:e0674.
Walkey AJ, Myers LC, Thai KK, Kipnis P, Desai M, Go AS, et al. Practice Patterns and Outcomes Associated With Anticoagulation Use Following Sepsis Hospitalizations With New-Onset Atrial Fibrillation. Circ Cardiovasc Qual Outcomes. 2023;16:e009494.
Myers LC, Knox D, Thai KK, Kipnis P, Jacobs J, Lee C, et al. Predicting Cardiovascular Events after Sepsis with Death as a Competing Risk. Ann Am Thorac Soc. 2023;20:145–8.
Liu VX, Bhimarao M, Greene JD, Manickam RN, Martinez A, Schuler A, et al. The presentation, pace, and profile of infection and sepsis patients hospitalized through the emergency department: An exploratory analysis. Crit Care Explor. 2021;3:e0344.
American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders (DSM-5®). https://doi.org/10.1176/appi. books.9780890425596 (2013).
Fulton JJ, Calhoun PS, Beckham JC Trauma- and Stressor-Related Disorders: Posttraumatic Stress Disorder, Acute Stress Disorder, and Adjustment Disorder. In: Psychiatry. John Wiley & Sons, Ltd, pp 1142–69 (2015).
Bryant RA. Acute stress disorder. Curr Opin Psychol. 2017;14:127–31.
Morgan MA, Kelber MS, Bellanti DM, Beech EH, Boyd C, Galloway L, et al. Outcomes and prognosis of adjustment disorder in adults: A systematic review. J Psychiatr Res. 2022;156:498–510.
Righy C, Rosa RG, da Silva RTA, Kochhann R, Migliavaca CB, Robinson CC, et al. Prevalence of post-traumatic stress disorder symptoms in adult critical care survivors: a systematic review and meta-analysis. Crit Care Lond Engl. 2019;23:213.
Nikayin S, Rabiee A, Hashem MD, Huang M, Bienvenu OJ, Turnbull AE, et al. Anxiety symptoms in survivors of critical illness: a systematic review and meta-analysis. Gen Hosp Psychiatry. 2016;43:23–9.
Karstoft K-I, Galatzer-Levy IR, Statnikov A, Li Z, Shalev AY. members of Jerusalem Trauma Outreach and Prevention Study (J-TOPS) group. Bridging a translational gap: using machine learning to improve the prediction of PTSD. BMC Psychiatry. 2015;15:30.
Schultebraucks K, Qian M, Abu-Amara D, Dean K, Laska E, Siegel C, et al. Pre-deployment risk factors for PTSD in active-duty personnel deployed to Afghanistan: a machine-learning approach for analyzing multivariate predictors. Mol Psychiatry. 2021;26:5011–22.
Bhaskaran K, Smeeth L. What is the difference between missing completely at random and missing at random? Int J Epidemiol. 2014;43:1336–9.
Papini S, Norman SB, Campbell-Sills L, Sun X, He F, Kessler RC, et al. Development and Validation of a Machine Learning Prediction Model of Posttraumatic Stress Disorder After Military Deployment. JAMA Netw Open. 2023;6:e2321273.
Seaman SR, White IR. Review of inverse probability weighting for dealing with missing data. Stat Methods Med Res. 2013;22:278–95.
R Core Team. R: A language and environment for statistical computing. 2013).
Escobar GJ, Liu VX, Schuler A, Lawson B, Greene JD, Kipnis P. Automated Identification of Adults at Risk for In-Hospital Clinical Deterioration. N. Engl J Med. 2020;383:1951–60.
Papini S, Hsin H, Kipnis P, Liu VX, Lu Y, Sterling SA, et al. Performance of a Prediction Model of Suicide Attempts Across Race and Ethnicity. JAMA Psychiatry. 2023;80:399–400.
Shortreed SM, Walker RL, Johnson E, Wellman R, Cruz M, Ziebell R, et al. Complex modeling with detailed temporal predictors does not improve health records-based suicide risk prediction. NPJ Digit Med. 2023;6:47.
Aiello S, Eckstrand E, Fu A, Landry M, Aboyoun P. Machine Learning with R and H2O. H2O Bookl. 2016;550.
Mukhopadhyay S. InfoGram and admissible machine learning. Mach Learn. 2022;111:205–42.
LeDell E, Poirier S. H2o automl: Scalable automatic machine learning. In: Proceedings of the AutoML Workshop at ICML. 2020.
DeMasi O, Kording K, Recht B. Meaningless comparisons lead to false optimism in medical machine learning. PloS One. 2017;12:e0184604.
Lundberg SM, Lee S-I. A unified approach to interpreting model predictions. Adv Neural Inf Process Syst. 2017;30.
Mitchell S, Potash E, Barocas S, D’Amour A, Lum K. Algorithmic fairness: Choices, assumptions, and definitions. Annu Rev Stat Its Appl. 2021;8:141–63.
Harrington KM, Quaden R, Stein MB, Honerlaw JP, Cissell S, Pietrzak RH, et al. Validation of an Electronic Medical Record-Based Algorithm for Identifying Posttraumatic Stress Disorder in U.S. Veterans. J Trauma Stress. 2019;32:226–37.
Gravely AA, Cutting A, Nugent S, Grill J, Carlson K, Spoont M. Validity of PTSD diagnoses in VA administrative data: comparison of VA administrative PTSD diagnoses to self-reported PTSD Checklist scores. J Rehabil Res Dev. 2011;48:21–30.
Marshall RD, Olfson M, Hellman F, Blanco C, Guardino M, Struening EL. Comorbidity, impairment, and suicidality in subthreshold PTSD. Am J Psychiatry. 2001;158:1467–73.
Ramspek CL, Steyerberg EW, Riley RD, Rosendaal FR, Dekkers OM, Dekker FW, et al. Prediction or causality? A scoping review of their conflation within current observational research. Eur J Epidemiol. 2021;36:889–98.
Acknowledgements
The study was supported by a grant from NIMH (5T32MH125792) and grants from The Permanente Medical Group (Pilot Award and Delivery Science Fellowship); the funding sources were not involved in the design, conduct, analysis, or interpretation of the data or manuscript.
Author information
Authors and Affiliations
Contributions
Author S.P. conceived and designed the study, obtained funding, conducted the analyses, and wrote the first draft; authors Y.L. and J.D.G. extracted the data; author F.B., S.A.S., and V.X.L. provided administrative support; authors E.I., S.A.S., and V.X.L. provided supervision. All authors reviewed and edited the manuscript and approved submission of the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Papini, S., Iturralde, E., Lu, Y. et al. Development and validation of a machine learning model using electronic health records to predict trauma- and stressor-related psychiatric disorders after hospitalization with sepsis. Transl Psychiatry 13, 400 (2023). https://doi.org/10.1038/s41398-023-02699-6
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41398-023-02699-6