BACKGROUND AND OBJECTIVES

Early recognition and treatment of pediatric sepsis remain mainstay approaches to improve outcomes. Although most children with sepsis are diagnosed in the emergency department, some are admitted with unrecognized sepsis or develop sepsis while hospitalized. Our objective was to develop and validate a prediction model of pediatric sepsis to improve recognition in the inpatient setting.

METHODS

Patients with sepsis were identified using intention-to-treat criteria. Encounters from 2012 to 2018 were used as a derivation to train a prediction model using variables from an existing model. A 2-tier threshold was determined using a precision-recall curve: an “Alert” tier with high positive predictive value to prompt bedside evaluation and an “Aware” tier with high sensitivity to increase situational awareness. The model was prospectively validated in the electronic health record in silent mode during 2019.

RESULTS

A total of 55 980 encounters and 793 (1.4%) episodes of sepsis were used for derivation and prospective validation. The final model consisted of 13 variables with an area under the curve of 0.96 (95% confidence interval 0.95–0.97) in the validation set. The Aware tier had 100% sensitivity and the Alert tier had a positive predictive value of 14% (number needed to alert of 7) in the validation set.

CONCLUSIONS

We derived and prospectively validated a 2-tiered prediction model of inpatient pediatric sepsis designed to have a high sensitivity Aware threshold to enable situational awareness and a low number needed to Alert threshold to minimize false alerts. Our model was embedded in our electronic health record and implemented as clinical decision support, which is presented in a companion article.

Sepsis is a leading cause of death and morbidity in hospitalized children.1,2  Over the past decade, there has been significant effort to raise awareness for pediatric sepsis and create sepsis guidelines and quality improvement efforts for early sepsis recognition and timely initiation of treatment.3,4  Systematic screening is now recommended in the international management guidelines for pediatric sepsis-associated organ dysfunction and septic shock,4  and multiple states have adopted sepsis policies and protocols that focus on sepsis recognition.5  As a result, many institutions have implemented sepsis screening processes in the inpatient setting, with variable success.68 

During this study, our institution was a member of the Improving Pediatric Sepsis Outcomes (IPSO) Collaborative of the Children’s Hospital Association,9  where 1 of the key process metrics is the presence of sepsis recognition tools in both the emergency department and inpatient settings.10  After a failed trial of our emergency department sepsis screening algorithm and clinical decision support (CDS) tool in our inpatient wards in 2017 because of excessive false alerts, a team of experts was assembled to create a new approach to sepsis screening in hospitalized children. Our objective was to develop an actionable, data-driven prediction model of sepsis to augment shared situational awareness and improve early recognition of sepsis in non-ICU inpatient settings while minimizing alert burden on the clinical teams. The design of the CDS tool, the implementation of the prediction model in our institution’s electronic health record (EHR), and the design of the associated clinical workflows are presented in a separate, companion article.11 

This study describes development of a single-center prediction model and prospective validation of that model. All patients admitted to the acute care, non-ICU inpatient floors in the hospital between July 1, 2012, and December 31, 2019, were included in the study. Our institution is a 364-bed academic, quaternary-care, freestanding children’s hospital. Information from patients admitted to 1 of the hospital’s ICUs was excluded from the analysis while the patient was in the ICU, but information of any inpatient floor stays before or after an ICU stay was included to reflect real-world conditions.

The study was reviewed and approved by our institutional review board (2019–2465, approved December 10, 2018). The reporting of this study followed the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis statement and checklist.12  Data were analyzed using MATLAB 2019b (MathWorks, Natick, MA) and R version 4.0 (R Project for Statistical Computing, Vienna, Austria). A P value of <.05 was considered statistically significant.

To perform the model development and later implementation, we considered the input from a multidisciplinary team a key requirement for success. Therefore, we convened a team of clinicians, data scientists, quality improvement experts, and clinical informaticians. The team met regularly throughout the planning and execution of the project.

Data from 2012 to 2018 were analyzed retrospectively to evaluate the baseline model and develop the final prediction model. Data were extracted from the institutional data warehouse using structured queries and underwent quality checks for conformity, completeness, and plausibility.

Patients with sepsis on the inpatient units were identified using the IPSO Collaborative intention-to-treat criteria.13  Specifically, the IPSO intention-to-treat criteria were defined as treatment consistent with sepsis: intravenous antibiotics, ≥2 intravenous fluid boluses, or ≥1 bolus and a vasoactive agent within 6 hours and a blood culture within 72 hours.13  The IPSO criteria also include screening tools, huddles, and order sets as optional, site-specific criteria, but because we had no preexisting screening or huddle workflow in the inpatient setting, and there was very inconsistent use of related order sets, these were not included as criteria.

We first evaluated the performance of the inpatient sepsis prediction model developed at Nationwide Children’s Hospital, which is now available in the Epic Foundation system.14  This model includes routinely collected data in the EHR such as vital signs, nursing assessments, laboratory test results, presence of indwelling lines, and specific high-risk conditions ascertained by Systematized Nomenclature of Medicine concepts.15  Retrospective application of the Nationwide model for our institution yielded a sensitivity of 83% and a number needed to alert (NNA, calculated as 1/positive predictive value [PPV]) of 23 at the encounter level and did not meet our goals for implementation for an NNA of 10 or less for bedside evaluations.16 

We chose to leverage the variables identified by the Nationwide model and train a new model. To better capture the idiosyncrasies of our patient population, which includes high degree of medical complexity, the high-risk conditions included in the original Nationwide model (ie, history of solid organ transplant, cancer, sickle cell, encephalopathy, cerebral palsy, severe intellectual disability, immunodeficiency, and medical technology dependency) were disaggregated and evaluated as separate candidate predictors instead of as a composite single variable. The final list of predictor variables evaluated included: lethargy, prolonged capillary refill, fever or hypothermia, systolic hypotension, tachycardia, tachypnea, weak pulse exam, cool/cyanotic/pale skin exam, leukocytosis or leukopenia, bandemia, elevated alanine aminotransferase, history of solid organ transplant, immunodeficiency, immunosuppression, chemotherapy, medical technology dependency, presence of a central line or urinary drain, and admission within the last 12 hours. Lethargy was based on the Pediatric Early Warning Score Behavior assessment charted by the bedside nurse. Of note, none of the variables used in the model included any of the ones used as criteria for the IPSO intention-to-treat sepsis definition to avoid biasing the model (ie, simply detecting when the medical team has started treating sepsis). Following the approach of the Nationwide model, continuous variables were dichotomized or categorized to facilitate implementation in the EHR. The high-risk conditions (history of solid organ transplant, cancer, sickle cell, encephalopathy, cerebral palsy, severe intellectual disability, immunodeficiency, and medical technology dependency) were based on diagnosis groupers. Immunosuppression and chemotherapy were based on prescribed medications.14 

Inpatient encounters from 2012 to 2018 were used as a derivation set to train the model. The derivation set was structured into a right alignment framework for sepsis cases with controls matched by length of stay.17  In a right-aligned framework, samples are usually considered until the event of interest (for the positive cases) or the end of admission (for the negative cases).17  In our approach, variables were collected for the 24-hour time window preceding the time 0 for sepsis cases (on the basis of the IPSO criteria) and the corresponding time window for controls with similar length of stay matched in a 1:30 ratio. For example, a patient with sepsis at 35 hours after admission would have the values of the predictor variables extracted for the time window between 9 hours and 34 hours of admission. At the same time, 30 control patients with a length of stay of 35 or more hours would have the values for the same predictor variables extracted for the same time windows. This approach was taken to train the model by more closely mimicking real-world conditions and avoid training an overly optimistic model by using the predischarge time window for control patients, as is often performed in right alignment frameworks.17  Missing values were assumed to be normal (eg, if the alanine aminotransferase level was not measured, it was imputed as a normal value) or the condition absent (eg, if there was no diagnosis code for solid organ transplantation, the patient was assumed to not have that high-risk condition). In each 24-hour prediction time window, the worst value for each variable (eg, lowest systolic blood pressure, highest heart rate, etc) was used in the model.

Using the right aligned framework, a logistic regression model was trained and assessed using 40-fold cross-validation. Variables with a nonsignificant association with the outcome in the full model were excluded from the final model. The final model coefficients were transformed into an integer score for implementation by multiplying the value by 5 and rounding to the nearest integer.

A 2-tier threshold approach was determined using a precision-recall curve of the model: an “Alert” tier with a goal of having a PPV ≥10% (and an NNA ≤10 or less) to prompt clinical team bedside evaluation, and an “Aware” tier with high sensitivity of ≥80% to increase situational awareness.

Prospective validation was considered a critical step in the process. The reproducibility of the model outputs may be affected by real-world conditions such as erroneous or delayed data entry in the EHR by bedside clinicians and timeliness of variables, such as the diagnosis groupers used for classifying high-risk conditions.18  Therefore, the model was implemented in silent mode (ie, with no results provided to bedside clinicians) in the EHR, and scores from the year 2019 were used as prospective validation before live implementation.

In the prospective validation set, the performance was assessed on the basis of the maximum sepsis score at 2 levels: the encounter level, where the time periods were the time from admission to sepsis for cases or until discharge for controls; and at the 24-hour level, where time periods were all the 24-hour periods from admission until sepsis for cases or until discharge for controls.

A total of 55 980 inpatient encounters and 793 (1.4%) episodes of sepsis on the inpatient units between 2012 and 2019 were used for derivation (2012–2018) and prospective validation (2019) of the model. Patient and encounter characteristics including age, sex, race/ethnicity, preferred language, payer, and specific high-risk conditions for sepsis were similar between the training and validation periods (Table 1). The cumulative incidence of sepsis in the derivation set was slightly higher than in the prospective validation set (1.5% [n = 725] vs 1% [n = 68]).

TABLE 1

Clinical Characteristics of Patients in the Derivation and Prospective Validation Sets

Derivation Set (N = 47 372)Prospective Validation Set (N = 8608)
Age (y)   
 <1 8848 (18.7) 1489 (17.3) 
 1–4 12 446 (26.3) 2502 (29.1) 
 5–12 13 758 (29.0) 2509 (29.1) 
 13–17 9696 (20.5) 1748 (20.3) 
 ≥18 2624 (5.5) 360 (4.2) 
Male 25 443 (53.7) 4660 (54.1) 
Race/ethnicity   
 Hispanic 15 899 (33.6) 3039 (35.3) 
 Asian American 2237 (4.7) 404 (4.7) 
 Black, non-Hispanic 8450 (17.8) 1446 (16.8) 
 White, non-Hispanic 18 046 (38.1) 3114 (36.2) 
 All others 2740 (5.8) 605 (7.0) 
Preferred language English 40 090 (84.6) 7371 (85.6) 
Insurance   
 Medicaid or government 25 230 (53.3) 4704 (54.6) 
 Commercial 21 678 (45.8) 3810 (44.3) 
 Other 464 (1.0) 93 (1) 
Medical technology dependence 2169 (4.6) 337 (3.9) 
Immunosuppression 5433 (11.5) 1149 (13.3) 
Solid organ transplant 1556 (3.3) 144 (1.7) 
Sepsis 725 (1.5) 68 (1) 
Derivation Set (N = 47 372)Prospective Validation Set (N = 8608)
Age (y)   
 <1 8848 (18.7) 1489 (17.3) 
 1–4 12 446 (26.3) 2502 (29.1) 
 5–12 13 758 (29.0) 2509 (29.1) 
 13–17 9696 (20.5) 1748 (20.3) 
 ≥18 2624 (5.5) 360 (4.2) 
Male 25 443 (53.7) 4660 (54.1) 
Race/ethnicity   
 Hispanic 15 899 (33.6) 3039 (35.3) 
 Asian American 2237 (4.7) 404 (4.7) 
 Black, non-Hispanic 8450 (17.8) 1446 (16.8) 
 White, non-Hispanic 18 046 (38.1) 3114 (36.2) 
 All others 2740 (5.8) 605 (7.0) 
Preferred language English 40 090 (84.6) 7371 (85.6) 
Insurance   
 Medicaid or government 25 230 (53.3) 4704 (54.6) 
 Commercial 21 678 (45.8) 3810 (44.3) 
 Other 464 (1.0) 93 (1) 
Medical technology dependence 2169 (4.6) 337 (3.9) 
Immunosuppression 5433 (11.5) 1149 (13.3) 
Solid organ transplant 1556 (3.3) 144 (1.7) 
Sepsis 725 (1.5) 68 (1) 

The final model consisted of 13 routinely collected clinical variables including vital signs, nursing assessments, underlying high-risk conditions, and laboratory test results. Table 2 presents the variables in the final mode, the criteria for each variable, the model coefficients, and the integer score used for implementation. A precision-recall curve was used to set actionable thresholds (Fig 1). In the derivation set, the model had an area under the curve of 0.82, the Aware threshold had 87% sensitivity for sepsis, and the Alert threshold had a PPV of 10% (NNA of 10).

TABLE 2

Sepsis Prediction Score Variables, Criteria, and Weights in the Final Model

Sepsis Screening Score VariablesCriteriaModel CoefficientInteger Score
Lethargy (PEWS behavior) Yes/no 4.07 20 
Bandemia >10% immature neutrophils 3.25 16 
Prolonged capillary refill >2 s 2.79 14 
Fever or hypothermia By age: 2.09 10 
<90 d: <36.0°C or ≥38.0°C 
≥90 d: <36°C or ≥38.5°C 
Systolic hypotension By age: 1.58 
0–3 mo: <50 mmHg 
4–11 mo: <70 mmHg 
1–3 y: <75 mmHg 
4–11 y: <80 mmHg 
12–17 y: <85 mmHg 
≥18 y: <90 mmHg 
Solid organ transplant Diagnosis grouper, yes/no 1.66 
Elevated ALT By age: 1.39 
<90 d: >156 IU/L 
≥90 d: >72 IU/L 
Tachycardia By age: 1.19 
0–2 mo: >180 beats per min 
3–11 mo: >170 beats per min 
1–3 y: >150 beats per min 
4–11 y: >130 beats per min 
12–17 y: >120 beats per min 
≥18 y: >90 beats per min 
Immunodeficiency Diagnosis grouper, yes/no 0.95 
Medical technology Diagnosis grouper, yes/no 0.92 
First 12 h of admission Yes/no 0.81 
Central line or drain Yes/no 0.77 
Tachypnea By age 0.49 
0–11 mo: >60 breaths per min 
1–3 y: >40 breaths per min 
4–5 y: >34 breaths per min 
6–12 y: >30 breaths per min 
≥13 y: >24 breaths per min 
Sepsis Screening Score VariablesCriteriaModel CoefficientInteger Score
Lethargy (PEWS behavior) Yes/no 4.07 20 
Bandemia >10% immature neutrophils 3.25 16 
Prolonged capillary refill >2 s 2.79 14 
Fever or hypothermia By age: 2.09 10 
<90 d: <36.0°C or ≥38.0°C 
≥90 d: <36°C or ≥38.5°C 
Systolic hypotension By age: 1.58 
0–3 mo: <50 mmHg 
4–11 mo: <70 mmHg 
1–3 y: <75 mmHg 
4–11 y: <80 mmHg 
12–17 y: <85 mmHg 
≥18 y: <90 mmHg 
Solid organ transplant Diagnosis grouper, yes/no 1.66 
Elevated ALT By age: 1.39 
<90 d: >156 IU/L 
≥90 d: >72 IU/L 
Tachycardia By age: 1.19 
0–2 mo: >180 beats per min 
3–11 mo: >170 beats per min 
1–3 y: >150 beats per min 
4–11 y: >130 beats per min 
12–17 y: >120 beats per min 
≥18 y: >90 beats per min 
Immunodeficiency Diagnosis grouper, yes/no 0.95 
Medical technology Diagnosis grouper, yes/no 0.92 
First 12 h of admission Yes/no 0.81 
Central line or drain Yes/no 0.77 
Tachypnea By age 0.49 
0–11 mo: >60 breaths per min 
1–3 y: >40 breaths per min 
4–5 y: >34 breaths per min 
6–12 y: >30 breaths per min 
≥13 y: >24 breaths per min 

ALT, alanine aminotransferase; PEWS, Pediatric Early Warning Score.

FIGURE 1

Precision-recall curve and thresholds chosen in the derivation set.

FIGURE 1

Precision-recall curve and thresholds chosen in the derivation set.

Close modal

In the prospective validation set, the model had an area under the curve of 0.96 (95% confidence interval 0.95–0.97) at the encounter level and 0.91 (95% confidence interval 0.87–0.94) within 24 hours of sepsis onset (Table 3). The Aware threshold had 100% sensitivity for sepsis at the encounter and 24-hour levels, and the Alert threshold had a PPV of 14% (NNA of 7) at the encounter level and a PPV of 5% (NNA of 20) at the 24-hour level (Table 3).

TABLE 3

Performance Measure of the Sepsis Risk Score in the Prospective Validation Set

Timing of Sepsis Risk ScoreaAnytimeWithin 24 h
Total periods, No. 8608 35 394 
Sepsis prevalence, % 0.8 0.2 
Area under the curve (95% CI) 0.96 (0.95–0.97) 0.91 (0.87–0.94) 
Threshold type Alert Aware Alert Aware 
Diagnostic performance     
 Sensitivity, % (95% CI) 72 (60–81) 100 (93–100) 49 (37–61) 85 (74–92) 
 Specificity, % (95% CI) 96 (96–96) 75 (75–75) 98 (98–98) 80 (80–80) 
 PPV, % (95% CI) 14 (12–16) 3 (3–3) 5 (4–6) 1 (1–1) 
 NPV, % (95% CI) 99 (99–100) 100 (100–100) 99 (99–100) 100 (99–100) 
NNA 33 20 100 
Timing of Sepsis Risk ScoreaAnytimeWithin 24 h
Total periods, No. 8608 35 394 
Sepsis prevalence, % 0.8 0.2 
Area under the curve (95% CI) 0.96 (0.95–0.97) 0.91 (0.87–0.94) 
Threshold type Alert Aware Alert Aware 
Diagnostic performance     
 Sensitivity, % (95% CI) 72 (60–81) 100 (93–100) 49 (37–61) 85 (74–92) 
 Specificity, % (95% CI) 96 (96–96) 75 (75–75) 98 (98–98) 80 (80–80) 
 PPV, % (95% CI) 14 (12–16) 3 (3–3) 5 (4–6) 1 (1–1) 
 NPV, % (95% CI) 99 (99–100) 100 (100–100) 99 (99–100) 100 (99–100) 
NNA 33 20 100 

CI, confidence interval.

a

Performance was based on the maximum sepsis risk score in a given period: Anytime, time period from admission to the sepsis episode for cases or until discharge for controls; within 24 hours, each 24-hour period before the sepsis episode for cases or until discharge for controls. In the within 24-hours framework, any period >24 hours before sepsis for cases and any 24-hour period for controls were counted as negative instances.

We developed and prospectively validated a prediction model of sepsis in the non-ICU inpatient setting. Our model was designed to have 2 tiers of actionable thresholds, a high sensitivity threshold to increase situational awareness and a low NNA threshold to reduce the risk of alert fatigue.

The low incidence of sepsis in the non-ICU inpatient population, counterbalanced by the possible severe consequences of a missed or delayed diagnosis of severe sepsis or septic shock, make the development and implementation of sepsis prediction models challenging in the pediatric inpatient setting. EHR-based sepsis prediction tools have been shown to have poor predictive performance when used without validation before implementation.1921  Our multidisciplinary team employed a novel approach to create an automated, data-driven, EHR-based, 2-tiered sepsis prediction model to implement as a data-driven CDS tool with tailored responses and specific workflows, as presented in the companion paper.11 

By creating a 2-tiered sepsis prediction model, we aimed to maintain high situational awareness for sepsis while minimizing alert fatigue and stewarding resources. Alert fatigue is a known barrier to sepsis screening and is a pervasive threat to CDS implementation in health care.19,22  Additionally, although bedside sepsis huddles have been impactful, they do come at a cost of the time clinical teams spend performing and documenting huddles.23,24  Thresholding the Alert tier by NNA allowed for consideration of false alerts, alert fatigue, and resources required for a sepsis huddle when designing our model.16  The Aware tier with high sensitivity was designed to enable situational awareness and prompt discussions about sepsis risk at the individual patient, clinical team, and unit level. Situational awareness or “watcher” initiatives have demonstrated success at decreasing emergency ICU transfers and reducing serious safety events using proactive risk identification, mitigation, and escalation systems.2527  Applying the fundamental theorem of biomedical informatics, where the data-driven decision support tools serve to augment rather than replace clinician judgement, was instrumental when planning how to derive and implement our sepsis prediction model.28 

Our study has several strengths. First, we used a large, single-center derivation set to develop our model. Although this will likely reduce the generalizability of the model to other institutions (eg, our population has a high degree of medical complexity compared with other pediatric settings), we believe it maximizes its internal validity and effectiveness for our local implementation, which was the intended goal. Additionally, we believe that prospectively validating the model in real time represents a particular strength of our approach. The reproducibility of prediction models can be affected by the real-world conditions.18  For example, erroneous or noisy data can significantly affect model performance, particularly if the data used for derivation underwent significant data cleaning procedures. Furthermore, the timing when variables are available in real time (eg, diagnosis codes) may also affect model performance.

Our study also has several limitations. First, we used the IPSO intention-to-treat criteria for sepsis for model development, which may overestimate true sepsis incidence (ie, by capturing “overtreated” patients). Although the IPSO intention-to-treat sepsis criteria have known limitations, we believed it was important for us to stay aligned with the IPSO Collaborative approach given that the evaluation of our implementation efforts would be centered around the same sepsis criteria and the associated metrics used by the IPSO Collaborative. One risk of taking such approach is that the model is likely skewed toward greater sensitivity and lower specificity of true sepsis, but given that the goal of the prediction model is to increase shared situational awareness and leverage the team-based huddles and clinician evaluation of alerted patients to ascertain the presence of sepsis, this is a tradeoff that we deemed acceptable. It is also possible that the intention-to-treat sepsis criteria could miss cases of true sepsis on the acute care floors in which none of the treatment-based criteria were met while the patient was still on the acute care floor. Again, we believed that it was important for us to stay aligned with the IPSO Collaborative approach. Second, when evaluating our model and defining sensitivity and NNA thresholds, we did this by using the right-aligned, encounter-level data framework. Although this approach presents a true ratio of cases and controls that meet the model thresholds, it does not account for all the 24-hour periods when sepsis criteria are not being met in both cases and controls, as reflected in Table 3. From a practical standpoint, the right-aligned, encounter-level framework answers the question, “How many patients who subsequently met sepsis criteria had a score above the alert or aware thresholds compared to patients without sepsis?” whereas the 24-hour framework answers the question, “How many patients had a score above the Alert or Aware thresholds in 24-hour periods when no sepsis occurred subsequently?” The first question is primarily focused on the fundamentals of the model performance; that is, whether the predictors included do in fact predict sepsis or not, whereas the second question is primarily focused on the implementation practicalities, such as how many alerts would be expected with a 24-hour lockout period before a sepsis case is subsequently detected. Although the 24-hour period was chosen for pragmatic reasons given our a priori plan to have a 24-hour lockout period after an alert was triggered during clinical implementation, it is unclear what the risk–benefit ratio may be for alerts that are triggered before 24 hours in patients who eventually develop sepsis. Perhaps earlier awareness and vigilance may reduce the incidence of sepsis overall, or could potentially lead to precocious or unnecessary interventions. Unfortunately, such questions can only start to be answered after clinical implementation. Finally, we chose to leverage the same approach as the Nationwide model, which is now used in the Epic Foundation, to facilitate model development and implementation in the EHR. For example, we evaluated the same list of candidate predictors, we also used the worse values for each variable in the 24-hour prediction time window preceding the prediction time point, and categorized continuous variables into the same bins. However, it is possible that other sets of variables or features of the same variables (eg, change in heart rate over time) may be more predictive of sepsis.

We derived and prospectively validated a 2-tiered prediction model of pediatric sepsis designed to have a high sensitivity Aware threshold to enable situational awareness and a low NNA Alert threshold to minimize the rate of false alerts. Our model was embedded in our EHR and underwent live implementation in the form of a CDS tool and associated clinical workflows, which are presented in the companion article.11 

COMPANION PAPERS: Companions to this article can be found online at www.hosppeds.org/cgi/doi/10.1542/hpeds.2023-007218 and www.hosppeds.org/cgi/doi/10.1542/hpeds.2023-007285.

Dr Stephen conceptualized and designed the study, coordinated and supervised data collection, and drafted the manuscript; Dr Sanchez-Pinto conceptualized and designed the study, coordinated and supervised data collection, drafted the manuscript, and contributed to the acquisition of data, statistical analyses, and interpretation of the data; Dr Carroll and Mr Jones contributed to the acquisition of data, statistical analyses, and interpretation of the data; Mr Hoge, Ms Maciorowski, Ms O’Connell, Ms Schwab, Ms Rojas, and Dr Lucey contributed to the improvement project leadership, and execution and design of data collection instruments; and all authors reviewed and revised the manuscript, approved the final manuscript as submitted, and agree to be accountable for all aspects of the work.

FUNDING: No external funding.

CONFLICT OF INTEREST DISCLOSURES: The authors have indicated they have no conflicts of interest relevant to this article to disclose.

1
Matics
TJ
,
Sanchez-Pinto
LN
.
Adaptation and validation of a Pediatric Sequential Organ Failure Assessment Score and evaluation of the sepsis-3 definitions in critically ill children
.
JAMA Pediatr
.
2017
;
171
(
10
):
e172352
2
Tan
B
,
Wong
JJ
,
Sultana
R
, et al
.
Global case-fatality rates in pediatric severe sepsis and septic shock: a systematic review and meta-analysis
.
JAMA Pediatr
.
2019
;
173
(
4
):
352
362
3
Davis
AL
,
Carcillo
JA
,
Aneja
RK
, et al
.
American College of Critical Care Medicine clinical practice parameters for hemodynamic support of pediatric and neonatal septic shock
.
Crit Care Med
.
2017
;
45
(
6
):
1061
1093
4
Weiss
SL
,
Peters
MJ
,
Alhazzani
W
, et al
.
Surviving sepsis campaign international guidelines for the management of septic shock and sepsis-associated organ dysfunction in children
.
Intensive Care Med
.
2020
;
46
(
Suppl 1
):
10
67
5
Center for Disease Control and Prevention
.
State policy approaches to sepsis prevention and early treatment
.
6
Bradshaw
C
,
Goodman
I
,
Rosenberg
R
,
Bandera
C
,
Fierman
A
,
Rudy
B
.
Implementation of an inpatient pediatric sepsis identification pathway
.
Pediatrics
.
2016
;
137
(
3
):
e20144082
7
Eisenberg
M
,
Madden
K
,
Christianson
JR
,
Melendez
E
,
Harper
MB
.
Performance of an automated screening algorithm for early detection of pediatric severe sepsis
.
Pediatr Crit Care Med
.
2019
;
20
(
12
):
e516
e523
8
Ginestra
JC
,
Giannini
HM
,
Schweickert
WD
, et al
.
Clinician perception of a machine learning-based early warning system designed to predict severe sepsis and septic shock
.
Crit Care Med
.
2019
;
47
(
11
):
1477
1484
9
Children’s Hospital Association
.
Sepsis collaborative
.
10
Paul
R
,
Niedner
M
,
Brilli
R
, et al
.
IPSO Collaborative Investigators
.
Metric development for the multicenter Improving Pediatric Sepsis Outcomes (IPSO) Collaborative
.
Pediatrics
.
2021
;
147
(
5
):
e2020017889
11
Stephen
RJ
,
Lucey
K
,
Carroll
MS
, et al
.
Sepsis prediction in hospitalized children: clinical decision support design and deployment
.
Hosp Pediatr
.
2023
;
13
(
9
):
e2023007218
12
Collins
GS
,
Reitsma
JB
,
Altman
DG
,
Moons
KG
.
Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): the TRIPOD statement
.
BMJ
.
2015
;
350
:
g7594
13
Scott
HF
,
Brilli
RJ
,
Paul
R
, et al
.
Improving Pediatric Sepsis Outcomes (IPSO) Collaborative Investigators
.
Evaluating pediatric sepsis definitions designed for electronic health record extraction and multicenter quality improvement
.
Crit Care Med
.
2020
;
48
(
10
):
e916
e926
14
Nuss
KE
,
Kunar
JS
,
Ahrens
EA
.
Plan-Do-Study-Act Methodology: refining an inpatient pediatric sepsis screening process
.
Pediatr Qual Saf
.
2020
;
5
(
5
):
e338
15
SNOMED International
.
Available at: https://www.snomed.org. Accessed October 10, 2019
16
Dewan
M
,
Sanchez-Pinto
LN
.
Crystal balls and magic eight balls: the art of developing and implementing automated algorithms in acute care pediatrics
.
Pediatr Crit Care Med
.
2019
;
20
(
12
):
1197
1199
17
Lauritsen
SM
,
Thiesson
B
,
Jørgensen
MJ
, et al
.
The framing of machine learning risk prediction models illustrated by evaluation of sepsis in general wards
.
NPJ Digit Med
.
2021
;
4
(
1
):
158
18
Sanchez-Pinto
LN
,
Bennett
TD
.
Evaluation of machine learning models for clinical prediction problems
.
Pediatr Crit Care Med
.
2022
;
23
(
5
):
405
408
19
Ancker
JS
,
Edwards
A
,
Nosal
S
,
Hauser
D
,
Mauer
E
,
Kaushal
R
.
HITEC Investigators
.
Effects of workload, work complexity, and repeated alerts on alert fatigue in a clinical decision support system
.
BMC Med Inform Decis Mak
.
2017
;
17
(
1
):
36
20
Wong
A
,
Otles
E
,
Donnelly
JP
, et al
.
External validation of a widely implemented proprietary sepsis prediction model in hospitalized patients
.
JAMA Intern Med
.
2021
;
181
(
8
):
1065
1070
21
Wong
A
,
Cao
J
,
Lyons
PG
, et al
.
Quantification of sepsis model alerts in 24 US hospitals before and during the COVID-19 pandemic
.
JAMA Netw Open
.
2021
;
4
(
11
):
e2135286
22
Harrison
WN
,
Workman
JK
,
Bonafide
CP
,
Lockwood
JM
.
Surviving sepsis screening: the unintended consequences of continuous surveillance
.
Hosp Pediatr
.
2020
;
10
(
12
):
e14
e17
23
Balamuth
F
,
Alpern
ER
,
Abbadessa
MK
, et al
.
Improving recognition of pediatric severe sepsis in the emergency department: contributions of a vital sign-based electronic alert and bedside clinician identification
.
Ann Emerg Med
.
2017
;
70
(
6
):
759
768.e2
24
Paul
R
,
Melendez
E
,
Stack
A
,
Capraro
A
,
Monuteaux
M
,
Neuman
MI
.
Improving adherence to PALS septic shock guidelines
.
Pediatrics
.
2014
;
133
(
5
):
e1358
e1366
25
Brady
PW
,
Muething
S
,
Kotagal
U
, et al
.
Improving situation awareness to reduce unrecognized clinical deterioration and serious safety events
.
Pediatrics
.
2013
;
131
(
1
):
e298
e308
26
McClain Smith
M
,
Chumpia
M
,
Wargo
L
,
Nicol
J
,
Bugnitz
M
.
Watcher initiative associated with decrease in failure to rescue events in pediatric population
.
Hosp Pediatr
.
2017
;
7
(
12
):
710
715
27
Sosa
T
,
Sitterding
M
,
Dewan
M
, et al
.
Optimizing situation awareness to reduce emergency transfers in hospitalized children
.
Pediatrics
.
2021
;
148
(
4
):
e2020034603
28
Friedman
CPA
.
A “fundamental theorem” of biomedical informatics
.
J Am Med Inform Assoc
.
2009
;
16
(
2
):
169
170