LayerRx Mapping ID
364
Slot System
Featured Buckets
Featured Buckets Admin

Trends and Variation in the Use of Observation Stays at Children’s Hospitals

Article Type
Changed
Mon, 11/01/2021 - 12:34
Display Headline
Trends and Variation in the Use of Observation Stays at Children’s Hospitals

Payors have been refining reimbursement policies for observation and inpatient stays over the past decade, and the effects on the healthcare payment system are significant.1-4 Advocates claim that observation status could improve efficiency in the use of healthcare resources by reducing emergency department (ED) crowding and lowering costs for inpatient care.5,6 Critics consider observation status to be a cost-shifting strategy that could lead to financial burdens for patients and hospitals.7,8

Although reimbursement policies for observation stays traditionally have been set by the Centers for Medicare and Medicaid Services (CMS) in a uniform manner,4,8 state Medicaid programs and commercial health insurers have developed a variety of policies for using observation status in broader populations and hospitals.9-15 Coverage criteria and implementation timelines of these policies vary by states and commercial insurers.11-15 For example, the California Department of Health Care Services did not have a specific reimbursement rate for observation stays in 2020, while some state Medicaid programs have had reimbursement policies for observation services in place since 2010.11-15 These inconsistencies likely result in greater variation in use of observation stays across children’s hospitals than general hospitals.

Previous studies have shown rising trends in use of observation stays among adult patient populations and related implications for patients and general hospitals,16-19 but few studies have reported the trends for pediatric populations. In this study, we sought to (1) describe recent trends of observation stays for pediatric populations at children’s hospitals from 2010 through 2019 and (2) investigate features of this shifting pattern for pediatric populations and hospital-level use of observation stays.

METHODS

Study Design, Data, and Populations

We performed a retrospective analysis of the Pediatric Health Information System (PHIS), an administrative database that contains inpatient, observation, ambulatory, and ED encounter-level data from 50 not-for-profit, tertiary care children’s hospitals affiliated with the Children’s Hospital Association (CHA).20 PHIS has an indicator to classify patient types (inpatient, observation, ED visits, ambulatory surgery, clinic visit, and others). The data are de-identified at the time of submission and subjected to validity and reliability checks by CHA and Truven Health Analytics (Ann Arbor, MI) before being included in PHIS. Each encounter in PHIS has only one patient type; therefore, encounters that transition to a higher level of care are assigned to their highest level of care (eg, a patient transitions from observation to inpatient status is classified as an inpatient encounter) to avoid duplicate counting.

To ensure consistent evaluations over time, we included 29 children’s hospitals that consistently reported both inpatient and observation data to PHIS across all quarters from 2010 through 2019. We identified the 20 most common clinical conditions using the All Patients Refined Diagnosis Related Groups (APR-DRGs; 3M Corporation) based upon their total frequencies of observation and inpatient stays over the study period. Regression analyses were conducted using all encounters within the 20 most common APR-DRGs.

Because all data have been de-identified in the PHIS database, the institutional review board at Ann and Robert H. Lurie Children’s Hospital of Chicago granted this study institutional review board–exempt status.

Main Outcome and Measures

We first presented longitudinal trends of observation stays for children’s hospitals using annual percentage of observation stays defined as:

To determine whether different pediatric populations have different trends of observation stays, we measured the growth rates of observation stays for each APR-DRG. Specifically, we first calculated the percentage of observation stays by APR-DRGs and years as described, and then calculated the growth rate of observation stays for each APR-DRG:

Next, we employed prolonged length of stay (LOS) and hospitalization resource-intensity scores for kids (H-RISK) to further investigate the shifting pattern of observation stays. Because most state Medicaid and commercial policies dictate that observation stays should not last longer than 48 hours, we defined prolonged LOS as >2 days.11-15 We defined the annual percentage of observation stays with prolonged LOS for each year as:

Numerators and denominators of the three measures were obtained by pooling all children’s hospitals included in this study. H-RISK is a continuous variable developed by CHA to measure use of intensive care for children, which is comparable across various APR-DRGs.21 Changes in the empirical distribution of H-RISK from observation stays were presented over years using percentiles.

Other measures included sex, age, race, payor, and LOS. To investigate the use of observation stays among payors, we categorized payors into five groups: private, in-state Medicaid (managed care), in-state Medicaid (Children’s Health Insurance Program [CHIP]/others), other government, and all others, according to the data availability. The “private” group consisted of commercial preferred provider organizations, commercial health maintenance organizations, and commercial others. We combined both CHIP and in-state Medicaid (others), including Medicaid fee-for-service or unspecified Medicaid together as “in-state Medicaid (CHIP/others).” Detailed categorization information is summarized in Appendix Table 1. LOS was classified into four groups: 1 day (24 hours), 2 days (48 hours), 3 to 4 days, and >4 days.

Statistical Analysis

Descriptive statistics were stratified by inpatient and observation status and were summarized using frequency, percent, median, and interquartile range (IQR). Chi-square or Wilcoxon rank-sum tests were performed to examine differences between observation and inpatient status. Trends in annual percentage of observation stays and annual percentage of observation stays with prolonged LOS were estimated using first-order autoregressive models, in which year was considered a continuous variable. A nonparametric measure of rank correlation (Spearman’s rank correlation coefficient) was employed to evaluate the correlation between year and H-RISK from observation stays.

The risk-adjusted probability of being admitted as an observation stay was estimated using generalized linear mixed models by adjusting for year, age, sex, race, payor, LOS, H-RISK, and a random intercept for each hospital to control for patient clustering within a hospital (Appendix Model). Hospital-level use of observation stays was measured by risk-adjusted percent use of observation stays for each hospital using the predicted values from generalized linear mixed models. All analyses were performed using SAS software, version 9.4 (SAS Institute) and R (R Core Team, 2019), and P < .05 was considered statistically significant.

RESULTS

Increasing Trend of Observation Stays

Over the study period, there were 5,611,001 encounters, including 3,901,873 (69.5%) inpatient and 1,709,128 (30.5%) observation stays (Appendix Table 1). The number of observation stays increased from 117,246 in 2010 to 207,842 in 2019, and the number of inpatient stays slightly increased from 378,433 to 397,994 over the 10 years (Appendix Table 1). Because of different growth rates between observation and inpatient status, the annual percentage of observation stays increased from 23.7% in 2010 to 34.3% in 2019, while the annual percentage of inpatient stays decreased from 76.3% in 2010 to 65.7% in 2019 (Appendix Table 1; Figure 1, P < .001).

Number of Inpatient and Observation Stays and Annual Percentage of Observation Stays at Children’s Hospitals, 2010 to 2019

Different Growth Rates of Observation Stays for Various Pediatric Populations

As shown in the Table, growth rates of observation stays increased for 19 of the 20 most common APR-DRGs. The four APR-DRGs having the highest growth rates in observation stays were appendectomy, diabetes mellitus, kidney and urinary tract infections, and cellulitis and other bacterial skin infections (Appendix Figure). In particular, the annual percentage of observation stays for appendectomy increased from 19.8% in 2010 to 54.7% in 2019, with the number of observation stays growing from 2,321 to 7,876, while the number of inpatient stays decreased from 9,384 to 6,535 (Appendix Figure). The annual percentage of observation stays for diabetes mellitus increased from 8.16% in 2010 to 22.74% in 2019. Tonsil and adenoid procedures consistently held the largest numbers of observation stays across the 10 years among all the APR-DRGs, with 115,207 and 31,125 total observation and inpatient stays, respectively (Table).

Numbers and Growth Rates of Inpatient and Observation Stays for the 20 Most Common All Patients Refined Diagnosis Related Groups, 2010 to 2019

Characteristics of Observation and Inpatient Stays

Patient characteristics are summarized in Appendix Table 1. There were 542,344 (32.9%) observation stays among patients with in-state Medicaid (managed care), and 241,157 (27.4%) observation stays among in-state Medicaid (CHIP/others). The percentages of observation and inpatient stays were 29.8% and 70.2% for private payor, as well as 29.6% and 70.4% for other government payor. Overall, the median (IQR) of H-RISK among observation stays was 0.79 (0.57-1.19) vs 1.23 (0.72-2.43) for inpatient stays. There were 1,410,694 (82.5%) observation stays discharged within 1 day and 243,972 (14.3%) observation stays discharged within 2 days. However, there were 47,413 (2.8%) and 7,049 (0.4%) observation stays with LOS 3 to 4 days or >4 days, respectively.

Shifting Pattern in Observation Stays

The annual percentage of observation stays with prolonged LOS (>2 days) rose from 1.1% in 2010 to 4.6% in 2019 (P < .001; Figure 2). The empirical distribution of H-RISK from observation stays by years further suggests a slightly increasing trend in intensity of care under observation stays. As shown in Appendix Table 2, although the 1st, 5th, 10th, 25th, and 99th percentiles of H-RISK were relatively stable, the 50th, 75th, 90th, and 95th percentiles of H-RISK were increasing over time. The correlation between year and intensity of care used under observation stays (H-RISK from observation stays) was found to be weak but significantly positive (Spearman correlation coefficients = 0.04; P < .001).

Number and Annual Percentage of Observation Stays with Prolonged Length of Stay

Interaction coefficients from our regression model demonstrate that the existing inverse association between H-RISK and odds of admission as an observation stay became less negative over the years. In 2010, the adjusted odds ratio (OR) of H-RISK was 0.57 (95% CI, 0.55-0.59). By 2017, the adjusted OR had increased to 0.65 (95% CI, 0.64-0.66). Compared with 2010, the seven adjusted ORs of H-RISK at years 2012 through 2018 were observed to be higher and statistically significant (P < .001, Appendix Table 3).

Hospitals-Level Use of Observation Stays

After adjusting for all covariates and hospital random effects, hospital-level use of observation stays increased between 2010 and 2019 for 26 out of 29 children’s hospitals. Although observation status essentially was not used at two children’s hospitals over the study period, the median hospital-level use of observation stays was 26% in 2010 (IQR, 3%-36%) and increased to 46% (IQR: 39%; 55%) in 2019. As shown in Figure 3, the number of hospitals with a low percentage of observation stays (<26%) decreased from 15 in 2010 to 4 in 2019. The number of hospitals with a high percentage of observation stays (≥51%) increased from 5 in 2010 to 10 in 2019. Nevertheless, there remained significant variation in the use of observation stays, and the hospital-level use ranged from 0% to 67% in 2019.

Risk-Adjusted Hospital-Level Use of Observation Stays at Children’s Hospitals, 2010 to 2019

DISCUSSION

By 2020, observation status has become a key component of healthcare for pediatric patients, and its relevance for children’s hospitals recently has been described.22,23 However, trends in observation stays for pediatric populations are not known. This represents the first study showing temporal trends of observation stays at children’s hospitals after 2010. Our results confirm that the increase in observation stays for pediatric populations is not attributable to decreasing patient acuity at children’s hospitals. We found a weak but significantly positive correlation between year and intensity of care used under observation stays. Although this correlation might not be clinically important, it demonstrates that patient acuity in observation stays is not decreasing. Regression results suggest that observation stays now encompass patients who need relatively higher intensity of care compared with those admitted under observation status in 2010.

This study also identifies a unique pattern in the use of observation stays among pediatric populations. Earlier studies exclusively focused on observation stays that were admitted from EDs.24 Our results indicate that observation status has been used beyond a bridge from ED care to inpatient admission. In particular, observation status has expanded to include pediatric populations with more diverse clinical conditions (eg, appendicitis and diabetes mellitus), and has become a substantial component of postprocedural admissions (Appendix Figure). Looking forward, it is likely that the use of observation stays might surpass inpatient admissions for more conditions that primarily involve short-term stays.

Observation status originally was designed as a reimbursement strategy for patients who needed short stays in dedicated ED units or hospitals, but did not qualify for inpatient services.5,25 After several changes in reimbursement policies, CMS released the “two midnight rule” for Medicare beneficiaries in 2013, which replaced condition-based criteria with time-based criteria to determine an inpatient or observation stay.1 Some Medicaid programs and commercial payors have developed similar policies. Unlike the universal policy for Medicare populations, the regulations for pediatric populations vary by states and health insurers.11-15,26-28 This might partially explain the wide variation observed among children’s hospital-level use of observation stays. For example, the California Medicaid program did not have a reimbursement rate for observation services as of 2020, while the Texas Medicaid program has had a policy for observation stays since 2010.12,13 We found that two children’s hospitals in California had the lowest use of observation stays (almost zero), whereas the hospital-level use of observation stays was more than 50% for three out of four children’s hospitals in Texas. In addition to reimbursement policies, individual hospitals also might have different strategies for observation status designation. An earlier survey showed that there was lack of consistency in billing and payor-based designations of observation status at children’s hospitals.29 These findings suggest that children’s hospital-level use of observation stays likely is influenced by reimbursement policy and practical strategy for observation status determination.

Earlier studies reported that observation status could be a more efficient use of healthcare resources.5,6 However, there are still at least two concerns relevant to children’s hospitals during the last decade. The first is whether the use of observation stays can promote cost-saving or if it is just a cost-shifting strategy. An earlier study demonstrated that observation stays with prolonged LOS might increase risk of cost-sharing among adult patients.29 Our study reveals an increasing trend of observation stays with prolonged LOS for pediatric patients. Similar to adult patients, LOS exceeding 24 or 48 hours could lead to uncovered healthcare costs and financial burdens on families.30-32 Meanwhile, children’s hospitals also might take on a higher financial liability by implementing observation status. Earlier studies have indicated that resource use between observation and inpatient stays at children’s hospitals is similar, and increasing use of observation stays might lead to financial risk rather than cost effectiveness.33 Further, administrative costs of observation determination are considerably high.34 Medicaid is the major payor for pediatric patients in children’s hospitals. In this study, more than 50% of encounters were paid through Medicaid programs. It is well known that Medicaid reimbursement rates are lower than Medicare and commercial plans.35 Therefore, the cost-saving conclusion drawn from Medicare patients cannot be generalized to pediatric populations at children’s hospitals without cautious reevaluation.

A second concern with increasing use of observation stays is selection bias in public reporting and comparisons of hospital performance. Presently, four main categories of quality indicators established by the Agency for Healthcare Research and Quality rely heavily on inpatient encounters.36 In this study, we found that the range of hospital-level use of observation stays was large. In 2019, the risk-adjusted percent use of observation stays was less than 5% at three hospitals, while the percent use was greater than 60% in another three hospitals. Therefore, comparisons made without uniform accounting of observation stays might have significant implications for national rankings of children’s hospitals across the United States. These consequences have been investigated in several published studies.22,23,37-39

There are several limitations to our study. First, the study sample was limited to children’s hospitals that consistently reported inpatient and observation data over the entire study period. Eighteen hospitals (86%) excluded from this study did not consistently submit inpatient and observation data to PHIS from 2010 through 2019. The primary purpose of this study was to present temporal trends of observation stays for children’s hospitals, and it was important to build the hospital cohort based on valid and consistent data during the study period. Appendix Table 4 presents differences of hospital characteristics by included and excluded groups of hospitals. Excluded hospitals might have fewer resources (eg, fewer pediatric intensive care beds). Nonetheless, the selection of hospitals was optimized based on data availability. Second, this study was a retrospective review of an administrative database of children’s hospitals and units. The sample does not represent all children’s hospitals or pediatric patients in the United States, but there are no available data sources—that we know of—that can generate national estimates for both inpatient and observation stays. Third, we did not attempt to conclusively infer any causal effects, and several factors could explain the increasing trends, such as reimbursement policies, hospital-level implementation strategies, determination guidelines for observation status designation, as well as changes in clinical care. Further studies should investigate impact of these factors on the use of observation stays for pediatric patients and children’s hospitals.

CONCLUSION

Observation status has been increasingly used for pediatric patients with more diverse clinical conditions, and there is a rising trend of prolonged LOS among observation stays since 2010. Considerable variation exists in hospital-level use of observation stays across children’s hospitals. Observation status could be an opportunity to improve efficiency of healthcare resource use or could lead to a financial risk for patients with prolonged LOS. Future studies should explore appropriateness of observation care in clinical practice through leveraging efficient care and alleviating financial risk.

Files
References

1. Centers for Medicare & Medicaid Services. Fact Sheet: Two-Midnight Rule. Accessed April 11, 2021. https://www.cms.gov/newsroom/fact-sheets/fact-sheet-two-midnight-rule-0
2. BlueCross BlueShield of Rhode Island. Payment Policy Outpaient Observation. Accessed April 11, 2021. https://www.bcbsri.com/sites/default/files/polices/Outpatient-Observation.pdf
3. Blue Cross Blue Shield of Illinois. Observation Services Tool for Applying MCG Care Guidelines Clinical Payment and Coding Policy. Accessed April 11, 2021. https://www.bcbsil.com/pdf/standards/observation_services_cpcp.pdf
4. Medicare.gov. Inpatient or outpatient hospital status affects your costs. Accessed April 11, 2021. https://www.medicare.gov/what-medicare-covers/what-part-a-covers/inpatient-or-outpatient-hospital-status
5. Ross MA, Hockenberry JM, Mutter R, Barrett M, Wheatley M, Pitts SR. Protocol-driven emergency department observation units offer savings, shorter stays, and reduced admissions. Health Aff (Millwood). 2013;32(12):2149-2156. https://doi.org/10.1377/hlthaff.2013.0662
6. Baugh CW, Venkatesh AK, Hilton JA, Samuel PA, Schuur JD, Bohan JS. Making greater use of dedicated hospital observation units for many short-stay patients could save $3.1 billion a year. Health Aff (Millwood). 2012;31(10):2314-2323. https://doi.org/10.1377/hlthaff.2011.0926
7. Sheehy AM, Graf B, Gangireddy S, et al. Hospitalized but not admitted: characteristics of patients with “observation status” at an academic medical center. JAMA Intern Med. 2013;173(21):1991-1998. https://doi.org/10.1001/jamainternmed.2013.8185
8. Baugh CW, Schuur JD. Observation care—high-value care or a cost-shifting loophole? N Engl J Med. 2013;369(4):302-305. https://doi.org/10.1056/NEJMp1304493
9. Missouri Hospital Association. A patient’s guide to observation care. Accessed April 11, 2021. https://www.mhanet.com/mhaimages/PatientsGuideToObservationCareFlyer.pdf
10. Cigna. Employee-paid hospital care coverage- summary of benefits. Accessed April 11, 2021. https://www.cigna.com/iwov-resources/national-second-sale/docs/healthy-benefits/updated-HC-benefit-summary.pdf
11. BlueCross BlueShield of Minnesota. Reimbursement policy-observation care services. Accessed April 11, 2021. https://www.bluecrossmn.com/sites/default/files/DAM/2020-07/Evaluation%20and%20Management%20004_Observation%20Care%20Services%20_09.04.17.pdf
12. California Department of Health Care Services. Public Hospital Project Frequently Asked Questions. Accessed April 11, 2021. https://www.dhcs.ca.gov/provgovpart/Documents/Public%20Hospital%20Project/PHP_Final_FAQs_January2013ADA.pdf
13. Texas Medicaid & Healthcare Partnership. Inpatient and Outpatient Hospital Servicces Handbook. Accessed May 29, 2021. https://www.tmhp.com/sites/default/files/microsites/provider-manuals/tmppm/html/TMPPM/2_Inpatient_Outpatient_Hosp_Srvs/2_Inpatient_Outpatient_Hosp_Srvs.htm
14. Alabama Medicaid. Outpatient observation. Accessed April 11, 2021. https://medicaid.alabama.gov/news_detail.aspx?ID=5121
15. NC Medicaid. Medicaid and Health Choice Clinical Coverage Policy No: 2A-1. Accessed April 11, 2021. https://files.nc.gov/ncdma/documents/files/2A-1_0.pdf
16. Feng Z, Wright B, Mor V. Sharp rise in Medicare enrollees being held in hospitals for observation raises concerns about causes and consequences. Health Aff (Millwood). 2012;31(6):1251-1259. https://doi.org/10.1377/hlthaff.2012.0129
17. Wright B, O’Shea AM, Ayyagari P, Ugwi PG, Kaboli P, Vaughan Sarrazin M. Observation rates at veterans’ hospitals more than doubled during 2005-13, similar to Medicare trends. Health Aff (Millwood). 2015;34(10):1730-1737. https://doi.org/10.1377/hlthaff.2014.1474
18. Wright B, Jung HY, Feng Z, Mor V. Hospital, patient, and local health system characteristics associated with the prevalence and duration of observation care. Health Serv Res. 2014;49(4):1088-1107. https://doi.org/10.1111/1475-6773.12166
19. Sabbatini AK, Wright B, Hall MK, Basu A. The cost of observation care for commercially insured patients visiting the emergency department. Am J Emerg Med. 2018;36(9):1591-1596. https://doi.org/10.1016/j.ajem.2018.01.040
20. Children’s Hospital Association. Pediatric health information system. Accessed April 11, 2021. https://www.childrenshospitals.org/phis
21. Richardson T, Rodean J, Harris M, Berry J, Gay JC, Hall M. Development of hospitalization resource intensity scores for kids (H-RISK) and comparison across pediatric populations. J Hosp Med. 2018;13(9):602-608. https://doi.org/10.12788/jhm.2948
22. Gay JC, Hall M, Morse R, Fieldston ES, Synhorst DC, Macy ML.Observation encounters and length of stay benchmarking in children’s hospitals. Pediatrics. 2020;146(5):e20200120. https://doi.org/10.1542/peds.2020-0120
23. Synhorst DC, Hall M, Harris M, et al. Hospital observation status and readmission rates. Pediatrics. 2020;146(5):e2020003954. https://doi.org/10.1542/peds.2020-003954
24. Macy ML, Hall M, Shah SS, et al. Pediatric observation status: are we overlooking a growing population in children’s hospitals? J Hosp Med. 2012;7(7):530-536. https://doi.org/10.1002/jhm.1923
25. Macy ML, Kim CS, Sasson C, Lozon MM, Davis MM. Pediatric observation units in the United States: a systematic review. J Hosp Med. 2010;5(3):172-182. https://doi.org/10.1002/jhm.592
26. UnitedHealthcare. Observation services policy, facility. Accessed April 11, 2021. https://www.uhcprovider.com/content/dam/provider/docs/public/policies/medicaid-comm-plan-reimbursement/UHCCP-Facility-Observation-Services-Policy-(F7106).pdf
27. Cal SB-1076§1253.7. General acute care hospitals: observation services – Health and Safety. Accessed April 11, 2021. https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=201520160SB1076
28. Nebraska Total Care. 2021 Provider Billing Guide. Accessed April 11, 2021. https://www.nebraskatotalcare.com/content/dam/centene/Nebraska/PDFs/ProviderRelations/NTC_Nebraska_Total_Care_Provider_Billing_Guide_508.pdf
29. Macy ML, Hall M, Shah SS, et al. Differences in designations of observation care in US freestanding children’s hospitals: are they virtual or real? J Hosp Med. 2012;7(4):287-293. https://doi.org/10.1002/jhm.949
30. Hockenberry JM, Mutter R, Barrett M, Parlato J, Ross MA. Factors associated with prolonged observation services stays and the impact of long stays on patient cost. Health Serv Res. 2014;49(3):893-909. https://doi.org/10.1111/1475-6773.12143
31. Anthem BlueCross BlueShield. Ohio Provider Manual. Accessed April11, 2021. https://www11.anthem.com/provider/oh/f1/s0/t0/pw_g357368.pdf?refer=ahpprovider&state=oh
32. Humana. Provider manual for physicians, hospitals and healthcare providers. Accessed April 11, 2021. https://docushare-web.apps.cf.humana.com/Marketing/docushare-app?file=3932669
33. Fieldston ES, Shah SS, Hall M, et al. Resource utilization for observation-status stays at children’s hospitals. Pediatrics. 2013;131(6):1050-1058 https://doi.org/10.1542/peds.2012-249
34. Tejedor-Sojo J. Observation status-a name at what cost? Hosp Pediatr. 2014;4(5):321-323. https://doi.org/10.1542/hpeds.2014-0037.
35. Selden TM, Karaca Z, Keenan P, White C, Kronick R. The growing difference between public and private payment rates for inpatient hospital care. Health Aff (Millwood). 2015;34(12):2147-2150. https://doi.org/10.1377/hlthaff.2015.0706
36. Agency for Healthcare Research and Quality. AHRQ Quality Indicators. Accessed April 11, 2021. https://www.qualityindicators.ahrq.gov
37. Figueroa JF, Burke LG, Zheng J, Orav EJ, Jha AK. Trends in hospitalization vs observation stay for ambulatory care-sensitive conditions. JAMA Intern Med. 2019;179(12):1714-1716. https://doi.org/10.1001/jamainternmed.2019.3177
38. Markham JL, Hall M, Gay JC, Bettenhausen JL, Berry JG. Length of stay and cost of pediatric readmissions. Pediatrics. 2018;141(4):e20172934. https://doi.org/10.1542/peds.2017-2934.
39. Overman RA, Freburger JK, Assimon MM, Li X, Brookhart, MA. Observation stays in administrative claims databases: underestimation of hospitalized cases. Pharmacoepidemiol Drug Saf. 2014;23(9):902-910. https://doi.org/10.1002/pds.3647.

Article PDF
Author and Disclosure Information

1Surgical Outcomes Quality Improvement Center, Department of Surgery, Feinberg School of Medicine, Northwestern University, Chicago, Illinois; 2Children’s Hospital Association, Lenexa, Kansas; 3Division of Pediatric Surgery, Department of Surgery, Northwestern University Feinberg School of Medicine, Ann and Robert H Lurie Children’s Hospital, Chicago, Illinois.

Disclosures
The authors have no conflicts of interest to disclose.

Issue
Journal of Hospital Medicine 16(11)
Publications
Topics
Page Number
645-651. Published Online First July 21, 2021
Sections
Files
Files
Author and Disclosure Information

1Surgical Outcomes Quality Improvement Center, Department of Surgery, Feinberg School of Medicine, Northwestern University, Chicago, Illinois; 2Children’s Hospital Association, Lenexa, Kansas; 3Division of Pediatric Surgery, Department of Surgery, Northwestern University Feinberg School of Medicine, Ann and Robert H Lurie Children’s Hospital, Chicago, Illinois.

Disclosures
The authors have no conflicts of interest to disclose.

Author and Disclosure Information

1Surgical Outcomes Quality Improvement Center, Department of Surgery, Feinberg School of Medicine, Northwestern University, Chicago, Illinois; 2Children’s Hospital Association, Lenexa, Kansas; 3Division of Pediatric Surgery, Department of Surgery, Northwestern University Feinberg School of Medicine, Ann and Robert H Lurie Children’s Hospital, Chicago, Illinois.

Disclosures
The authors have no conflicts of interest to disclose.

Article PDF
Article PDF
Related Articles

Payors have been refining reimbursement policies for observation and inpatient stays over the past decade, and the effects on the healthcare payment system are significant.1-4 Advocates claim that observation status could improve efficiency in the use of healthcare resources by reducing emergency department (ED) crowding and lowering costs for inpatient care.5,6 Critics consider observation status to be a cost-shifting strategy that could lead to financial burdens for patients and hospitals.7,8

Although reimbursement policies for observation stays traditionally have been set by the Centers for Medicare and Medicaid Services (CMS) in a uniform manner,4,8 state Medicaid programs and commercial health insurers have developed a variety of policies for using observation status in broader populations and hospitals.9-15 Coverage criteria and implementation timelines of these policies vary by states and commercial insurers.11-15 For example, the California Department of Health Care Services did not have a specific reimbursement rate for observation stays in 2020, while some state Medicaid programs have had reimbursement policies for observation services in place since 2010.11-15 These inconsistencies likely result in greater variation in use of observation stays across children’s hospitals than general hospitals.

Previous studies have shown rising trends in use of observation stays among adult patient populations and related implications for patients and general hospitals,16-19 but few studies have reported the trends for pediatric populations. In this study, we sought to (1) describe recent trends of observation stays for pediatric populations at children’s hospitals from 2010 through 2019 and (2) investigate features of this shifting pattern for pediatric populations and hospital-level use of observation stays.

METHODS

Study Design, Data, and Populations

We performed a retrospective analysis of the Pediatric Health Information System (PHIS), an administrative database that contains inpatient, observation, ambulatory, and ED encounter-level data from 50 not-for-profit, tertiary care children’s hospitals affiliated with the Children’s Hospital Association (CHA).20 PHIS has an indicator to classify patient types (inpatient, observation, ED visits, ambulatory surgery, clinic visit, and others). The data are de-identified at the time of submission and subjected to validity and reliability checks by CHA and Truven Health Analytics (Ann Arbor, MI) before being included in PHIS. Each encounter in PHIS has only one patient type; therefore, encounters that transition to a higher level of care are assigned to their highest level of care (eg, a patient transitions from observation to inpatient status is classified as an inpatient encounter) to avoid duplicate counting.

To ensure consistent evaluations over time, we included 29 children’s hospitals that consistently reported both inpatient and observation data to PHIS across all quarters from 2010 through 2019. We identified the 20 most common clinical conditions using the All Patients Refined Diagnosis Related Groups (APR-DRGs; 3M Corporation) based upon their total frequencies of observation and inpatient stays over the study period. Regression analyses were conducted using all encounters within the 20 most common APR-DRGs.

Because all data have been de-identified in the PHIS database, the institutional review board at Ann and Robert H. Lurie Children’s Hospital of Chicago granted this study institutional review board–exempt status.

Main Outcome and Measures

We first presented longitudinal trends of observation stays for children’s hospitals using annual percentage of observation stays defined as:

To determine whether different pediatric populations have different trends of observation stays, we measured the growth rates of observation stays for each APR-DRG. Specifically, we first calculated the percentage of observation stays by APR-DRGs and years as described, and then calculated the growth rate of observation stays for each APR-DRG:

Next, we employed prolonged length of stay (LOS) and hospitalization resource-intensity scores for kids (H-RISK) to further investigate the shifting pattern of observation stays. Because most state Medicaid and commercial policies dictate that observation stays should not last longer than 48 hours, we defined prolonged LOS as >2 days.11-15 We defined the annual percentage of observation stays with prolonged LOS for each year as:

Numerators and denominators of the three measures were obtained by pooling all children’s hospitals included in this study. H-RISK is a continuous variable developed by CHA to measure use of intensive care for children, which is comparable across various APR-DRGs.21 Changes in the empirical distribution of H-RISK from observation stays were presented over years using percentiles.

Other measures included sex, age, race, payor, and LOS. To investigate the use of observation stays among payors, we categorized payors into five groups: private, in-state Medicaid (managed care), in-state Medicaid (Children’s Health Insurance Program [CHIP]/others), other government, and all others, according to the data availability. The “private” group consisted of commercial preferred provider organizations, commercial health maintenance organizations, and commercial others. We combined both CHIP and in-state Medicaid (others), including Medicaid fee-for-service or unspecified Medicaid together as “in-state Medicaid (CHIP/others).” Detailed categorization information is summarized in Appendix Table 1. LOS was classified into four groups: 1 day (24 hours), 2 days (48 hours), 3 to 4 days, and >4 days.

Statistical Analysis

Descriptive statistics were stratified by inpatient and observation status and were summarized using frequency, percent, median, and interquartile range (IQR). Chi-square or Wilcoxon rank-sum tests were performed to examine differences between observation and inpatient status. Trends in annual percentage of observation stays and annual percentage of observation stays with prolonged LOS were estimated using first-order autoregressive models, in which year was considered a continuous variable. A nonparametric measure of rank correlation (Spearman’s rank correlation coefficient) was employed to evaluate the correlation between year and H-RISK from observation stays.

The risk-adjusted probability of being admitted as an observation stay was estimated using generalized linear mixed models by adjusting for year, age, sex, race, payor, LOS, H-RISK, and a random intercept for each hospital to control for patient clustering within a hospital (Appendix Model). Hospital-level use of observation stays was measured by risk-adjusted percent use of observation stays for each hospital using the predicted values from generalized linear mixed models. All analyses were performed using SAS software, version 9.4 (SAS Institute) and R (R Core Team, 2019), and P < .05 was considered statistically significant.

RESULTS

Increasing Trend of Observation Stays

Over the study period, there were 5,611,001 encounters, including 3,901,873 (69.5%) inpatient and 1,709,128 (30.5%) observation stays (Appendix Table 1). The number of observation stays increased from 117,246 in 2010 to 207,842 in 2019, and the number of inpatient stays slightly increased from 378,433 to 397,994 over the 10 years (Appendix Table 1). Because of different growth rates between observation and inpatient status, the annual percentage of observation stays increased from 23.7% in 2010 to 34.3% in 2019, while the annual percentage of inpatient stays decreased from 76.3% in 2010 to 65.7% in 2019 (Appendix Table 1; Figure 1, P < .001).

Number of Inpatient and Observation Stays and Annual Percentage of Observation Stays at Children’s Hospitals, 2010 to 2019

Different Growth Rates of Observation Stays for Various Pediatric Populations

As shown in the Table, growth rates of observation stays increased for 19 of the 20 most common APR-DRGs. The four APR-DRGs having the highest growth rates in observation stays were appendectomy, diabetes mellitus, kidney and urinary tract infections, and cellulitis and other bacterial skin infections (Appendix Figure). In particular, the annual percentage of observation stays for appendectomy increased from 19.8% in 2010 to 54.7% in 2019, with the number of observation stays growing from 2,321 to 7,876, while the number of inpatient stays decreased from 9,384 to 6,535 (Appendix Figure). The annual percentage of observation stays for diabetes mellitus increased from 8.16% in 2010 to 22.74% in 2019. Tonsil and adenoid procedures consistently held the largest numbers of observation stays across the 10 years among all the APR-DRGs, with 115,207 and 31,125 total observation and inpatient stays, respectively (Table).

Numbers and Growth Rates of Inpatient and Observation Stays for the 20 Most Common All Patients Refined Diagnosis Related Groups, 2010 to 2019

Characteristics of Observation and Inpatient Stays

Patient characteristics are summarized in Appendix Table 1. There were 542,344 (32.9%) observation stays among patients with in-state Medicaid (managed care), and 241,157 (27.4%) observation stays among in-state Medicaid (CHIP/others). The percentages of observation and inpatient stays were 29.8% and 70.2% for private payor, as well as 29.6% and 70.4% for other government payor. Overall, the median (IQR) of H-RISK among observation stays was 0.79 (0.57-1.19) vs 1.23 (0.72-2.43) for inpatient stays. There were 1,410,694 (82.5%) observation stays discharged within 1 day and 243,972 (14.3%) observation stays discharged within 2 days. However, there were 47,413 (2.8%) and 7,049 (0.4%) observation stays with LOS 3 to 4 days or >4 days, respectively.

Shifting Pattern in Observation Stays

The annual percentage of observation stays with prolonged LOS (>2 days) rose from 1.1% in 2010 to 4.6% in 2019 (P < .001; Figure 2). The empirical distribution of H-RISK from observation stays by years further suggests a slightly increasing trend in intensity of care under observation stays. As shown in Appendix Table 2, although the 1st, 5th, 10th, 25th, and 99th percentiles of H-RISK were relatively stable, the 50th, 75th, 90th, and 95th percentiles of H-RISK were increasing over time. The correlation between year and intensity of care used under observation stays (H-RISK from observation stays) was found to be weak but significantly positive (Spearman correlation coefficients = 0.04; P < .001).

Number and Annual Percentage of Observation Stays with Prolonged Length of Stay

Interaction coefficients from our regression model demonstrate that the existing inverse association between H-RISK and odds of admission as an observation stay became less negative over the years. In 2010, the adjusted odds ratio (OR) of H-RISK was 0.57 (95% CI, 0.55-0.59). By 2017, the adjusted OR had increased to 0.65 (95% CI, 0.64-0.66). Compared with 2010, the seven adjusted ORs of H-RISK at years 2012 through 2018 were observed to be higher and statistically significant (P < .001, Appendix Table 3).

Hospitals-Level Use of Observation Stays

After adjusting for all covariates and hospital random effects, hospital-level use of observation stays increased between 2010 and 2019 for 26 out of 29 children’s hospitals. Although observation status essentially was not used at two children’s hospitals over the study period, the median hospital-level use of observation stays was 26% in 2010 (IQR, 3%-36%) and increased to 46% (IQR: 39%; 55%) in 2019. As shown in Figure 3, the number of hospitals with a low percentage of observation stays (<26%) decreased from 15 in 2010 to 4 in 2019. The number of hospitals with a high percentage of observation stays (≥51%) increased from 5 in 2010 to 10 in 2019. Nevertheless, there remained significant variation in the use of observation stays, and the hospital-level use ranged from 0% to 67% in 2019.

Risk-Adjusted Hospital-Level Use of Observation Stays at Children’s Hospitals, 2010 to 2019

DISCUSSION

By 2020, observation status has become a key component of healthcare for pediatric patients, and its relevance for children’s hospitals recently has been described.22,23 However, trends in observation stays for pediatric populations are not known. This represents the first study showing temporal trends of observation stays at children’s hospitals after 2010. Our results confirm that the increase in observation stays for pediatric populations is not attributable to decreasing patient acuity at children’s hospitals. We found a weak but significantly positive correlation between year and intensity of care used under observation stays. Although this correlation might not be clinically important, it demonstrates that patient acuity in observation stays is not decreasing. Regression results suggest that observation stays now encompass patients who need relatively higher intensity of care compared with those admitted under observation status in 2010.

This study also identifies a unique pattern in the use of observation stays among pediatric populations. Earlier studies exclusively focused on observation stays that were admitted from EDs.24 Our results indicate that observation status has been used beyond a bridge from ED care to inpatient admission. In particular, observation status has expanded to include pediatric populations with more diverse clinical conditions (eg, appendicitis and diabetes mellitus), and has become a substantial component of postprocedural admissions (Appendix Figure). Looking forward, it is likely that the use of observation stays might surpass inpatient admissions for more conditions that primarily involve short-term stays.

Observation status originally was designed as a reimbursement strategy for patients who needed short stays in dedicated ED units or hospitals, but did not qualify for inpatient services.5,25 After several changes in reimbursement policies, CMS released the “two midnight rule” for Medicare beneficiaries in 2013, which replaced condition-based criteria with time-based criteria to determine an inpatient or observation stay.1 Some Medicaid programs and commercial payors have developed similar policies. Unlike the universal policy for Medicare populations, the regulations for pediatric populations vary by states and health insurers.11-15,26-28 This might partially explain the wide variation observed among children’s hospital-level use of observation stays. For example, the California Medicaid program did not have a reimbursement rate for observation services as of 2020, while the Texas Medicaid program has had a policy for observation stays since 2010.12,13 We found that two children’s hospitals in California had the lowest use of observation stays (almost zero), whereas the hospital-level use of observation stays was more than 50% for three out of four children’s hospitals in Texas. In addition to reimbursement policies, individual hospitals also might have different strategies for observation status designation. An earlier survey showed that there was lack of consistency in billing and payor-based designations of observation status at children’s hospitals.29 These findings suggest that children’s hospital-level use of observation stays likely is influenced by reimbursement policy and practical strategy for observation status determination.

Earlier studies reported that observation status could be a more efficient use of healthcare resources.5,6 However, there are still at least two concerns relevant to children’s hospitals during the last decade. The first is whether the use of observation stays can promote cost-saving or if it is just a cost-shifting strategy. An earlier study demonstrated that observation stays with prolonged LOS might increase risk of cost-sharing among adult patients.29 Our study reveals an increasing trend of observation stays with prolonged LOS for pediatric patients. Similar to adult patients, LOS exceeding 24 or 48 hours could lead to uncovered healthcare costs and financial burdens on families.30-32 Meanwhile, children’s hospitals also might take on a higher financial liability by implementing observation status. Earlier studies have indicated that resource use between observation and inpatient stays at children’s hospitals is similar, and increasing use of observation stays might lead to financial risk rather than cost effectiveness.33 Further, administrative costs of observation determination are considerably high.34 Medicaid is the major payor for pediatric patients in children’s hospitals. In this study, more than 50% of encounters were paid through Medicaid programs. It is well known that Medicaid reimbursement rates are lower than Medicare and commercial plans.35 Therefore, the cost-saving conclusion drawn from Medicare patients cannot be generalized to pediatric populations at children’s hospitals without cautious reevaluation.

A second concern with increasing use of observation stays is selection bias in public reporting and comparisons of hospital performance. Presently, four main categories of quality indicators established by the Agency for Healthcare Research and Quality rely heavily on inpatient encounters.36 In this study, we found that the range of hospital-level use of observation stays was large. In 2019, the risk-adjusted percent use of observation stays was less than 5% at three hospitals, while the percent use was greater than 60% in another three hospitals. Therefore, comparisons made without uniform accounting of observation stays might have significant implications for national rankings of children’s hospitals across the United States. These consequences have been investigated in several published studies.22,23,37-39

There are several limitations to our study. First, the study sample was limited to children’s hospitals that consistently reported inpatient and observation data over the entire study period. Eighteen hospitals (86%) excluded from this study did not consistently submit inpatient and observation data to PHIS from 2010 through 2019. The primary purpose of this study was to present temporal trends of observation stays for children’s hospitals, and it was important to build the hospital cohort based on valid and consistent data during the study period. Appendix Table 4 presents differences of hospital characteristics by included and excluded groups of hospitals. Excluded hospitals might have fewer resources (eg, fewer pediatric intensive care beds). Nonetheless, the selection of hospitals was optimized based on data availability. Second, this study was a retrospective review of an administrative database of children’s hospitals and units. The sample does not represent all children’s hospitals or pediatric patients in the United States, but there are no available data sources—that we know of—that can generate national estimates for both inpatient and observation stays. Third, we did not attempt to conclusively infer any causal effects, and several factors could explain the increasing trends, such as reimbursement policies, hospital-level implementation strategies, determination guidelines for observation status designation, as well as changes in clinical care. Further studies should investigate impact of these factors on the use of observation stays for pediatric patients and children’s hospitals.

CONCLUSION

Observation status has been increasingly used for pediatric patients with more diverse clinical conditions, and there is a rising trend of prolonged LOS among observation stays since 2010. Considerable variation exists in hospital-level use of observation stays across children’s hospitals. Observation status could be an opportunity to improve efficiency of healthcare resource use or could lead to a financial risk for patients with prolonged LOS. Future studies should explore appropriateness of observation care in clinical practice through leveraging efficient care and alleviating financial risk.

Payors have been refining reimbursement policies for observation and inpatient stays over the past decade, and the effects on the healthcare payment system are significant.1-4 Advocates claim that observation status could improve efficiency in the use of healthcare resources by reducing emergency department (ED) crowding and lowering costs for inpatient care.5,6 Critics consider observation status to be a cost-shifting strategy that could lead to financial burdens for patients and hospitals.7,8

Although reimbursement policies for observation stays traditionally have been set by the Centers for Medicare and Medicaid Services (CMS) in a uniform manner,4,8 state Medicaid programs and commercial health insurers have developed a variety of policies for using observation status in broader populations and hospitals.9-15 Coverage criteria and implementation timelines of these policies vary by states and commercial insurers.11-15 For example, the California Department of Health Care Services did not have a specific reimbursement rate for observation stays in 2020, while some state Medicaid programs have had reimbursement policies for observation services in place since 2010.11-15 These inconsistencies likely result in greater variation in use of observation stays across children’s hospitals than general hospitals.

Previous studies have shown rising trends in use of observation stays among adult patient populations and related implications for patients and general hospitals,16-19 but few studies have reported the trends for pediatric populations. In this study, we sought to (1) describe recent trends of observation stays for pediatric populations at children’s hospitals from 2010 through 2019 and (2) investigate features of this shifting pattern for pediatric populations and hospital-level use of observation stays.

METHODS

Study Design, Data, and Populations

We performed a retrospective analysis of the Pediatric Health Information System (PHIS), an administrative database that contains inpatient, observation, ambulatory, and ED encounter-level data from 50 not-for-profit, tertiary care children’s hospitals affiliated with the Children’s Hospital Association (CHA).20 PHIS has an indicator to classify patient types (inpatient, observation, ED visits, ambulatory surgery, clinic visit, and others). The data are de-identified at the time of submission and subjected to validity and reliability checks by CHA and Truven Health Analytics (Ann Arbor, MI) before being included in PHIS. Each encounter in PHIS has only one patient type; therefore, encounters that transition to a higher level of care are assigned to their highest level of care (eg, a patient transitions from observation to inpatient status is classified as an inpatient encounter) to avoid duplicate counting.

To ensure consistent evaluations over time, we included 29 children’s hospitals that consistently reported both inpatient and observation data to PHIS across all quarters from 2010 through 2019. We identified the 20 most common clinical conditions using the All Patients Refined Diagnosis Related Groups (APR-DRGs; 3M Corporation) based upon their total frequencies of observation and inpatient stays over the study period. Regression analyses were conducted using all encounters within the 20 most common APR-DRGs.

Because all data have been de-identified in the PHIS database, the institutional review board at Ann and Robert H. Lurie Children’s Hospital of Chicago granted this study institutional review board–exempt status.

Main Outcome and Measures

We first presented longitudinal trends of observation stays for children’s hospitals using annual percentage of observation stays defined as:

To determine whether different pediatric populations have different trends of observation stays, we measured the growth rates of observation stays for each APR-DRG. Specifically, we first calculated the percentage of observation stays by APR-DRGs and years as described, and then calculated the growth rate of observation stays for each APR-DRG:

Next, we employed prolonged length of stay (LOS) and hospitalization resource-intensity scores for kids (H-RISK) to further investigate the shifting pattern of observation stays. Because most state Medicaid and commercial policies dictate that observation stays should not last longer than 48 hours, we defined prolonged LOS as >2 days.11-15 We defined the annual percentage of observation stays with prolonged LOS for each year as:

Numerators and denominators of the three measures were obtained by pooling all children’s hospitals included in this study. H-RISK is a continuous variable developed by CHA to measure use of intensive care for children, which is comparable across various APR-DRGs.21 Changes in the empirical distribution of H-RISK from observation stays were presented over years using percentiles.

Other measures included sex, age, race, payor, and LOS. To investigate the use of observation stays among payors, we categorized payors into five groups: private, in-state Medicaid (managed care), in-state Medicaid (Children’s Health Insurance Program [CHIP]/others), other government, and all others, according to the data availability. The “private” group consisted of commercial preferred provider organizations, commercial health maintenance organizations, and commercial others. We combined both CHIP and in-state Medicaid (others), including Medicaid fee-for-service or unspecified Medicaid together as “in-state Medicaid (CHIP/others).” Detailed categorization information is summarized in Appendix Table 1. LOS was classified into four groups: 1 day (24 hours), 2 days (48 hours), 3 to 4 days, and >4 days.

Statistical Analysis

Descriptive statistics were stratified by inpatient and observation status and were summarized using frequency, percent, median, and interquartile range (IQR). Chi-square or Wilcoxon rank-sum tests were performed to examine differences between observation and inpatient status. Trends in annual percentage of observation stays and annual percentage of observation stays with prolonged LOS were estimated using first-order autoregressive models, in which year was considered a continuous variable. A nonparametric measure of rank correlation (Spearman’s rank correlation coefficient) was employed to evaluate the correlation between year and H-RISK from observation stays.

The risk-adjusted probability of being admitted as an observation stay was estimated using generalized linear mixed models by adjusting for year, age, sex, race, payor, LOS, H-RISK, and a random intercept for each hospital to control for patient clustering within a hospital (Appendix Model). Hospital-level use of observation stays was measured by risk-adjusted percent use of observation stays for each hospital using the predicted values from generalized linear mixed models. All analyses were performed using SAS software, version 9.4 (SAS Institute) and R (R Core Team, 2019), and P < .05 was considered statistically significant.

RESULTS

Increasing Trend of Observation Stays

Over the study period, there were 5,611,001 encounters, including 3,901,873 (69.5%) inpatient and 1,709,128 (30.5%) observation stays (Appendix Table 1). The number of observation stays increased from 117,246 in 2010 to 207,842 in 2019, and the number of inpatient stays slightly increased from 378,433 to 397,994 over the 10 years (Appendix Table 1). Because of different growth rates between observation and inpatient status, the annual percentage of observation stays increased from 23.7% in 2010 to 34.3% in 2019, while the annual percentage of inpatient stays decreased from 76.3% in 2010 to 65.7% in 2019 (Appendix Table 1; Figure 1, P < .001).

Number of Inpatient and Observation Stays and Annual Percentage of Observation Stays at Children’s Hospitals, 2010 to 2019

Different Growth Rates of Observation Stays for Various Pediatric Populations

As shown in the Table, growth rates of observation stays increased for 19 of the 20 most common APR-DRGs. The four APR-DRGs having the highest growth rates in observation stays were appendectomy, diabetes mellitus, kidney and urinary tract infections, and cellulitis and other bacterial skin infections (Appendix Figure). In particular, the annual percentage of observation stays for appendectomy increased from 19.8% in 2010 to 54.7% in 2019, with the number of observation stays growing from 2,321 to 7,876, while the number of inpatient stays decreased from 9,384 to 6,535 (Appendix Figure). The annual percentage of observation stays for diabetes mellitus increased from 8.16% in 2010 to 22.74% in 2019. Tonsil and adenoid procedures consistently held the largest numbers of observation stays across the 10 years among all the APR-DRGs, with 115,207 and 31,125 total observation and inpatient stays, respectively (Table).

Numbers and Growth Rates of Inpatient and Observation Stays for the 20 Most Common All Patients Refined Diagnosis Related Groups, 2010 to 2019

Characteristics of Observation and Inpatient Stays

Patient characteristics are summarized in Appendix Table 1. There were 542,344 (32.9%) observation stays among patients with in-state Medicaid (managed care), and 241,157 (27.4%) observation stays among in-state Medicaid (CHIP/others). The percentages of observation and inpatient stays were 29.8% and 70.2% for private payor, as well as 29.6% and 70.4% for other government payor. Overall, the median (IQR) of H-RISK among observation stays was 0.79 (0.57-1.19) vs 1.23 (0.72-2.43) for inpatient stays. There were 1,410,694 (82.5%) observation stays discharged within 1 day and 243,972 (14.3%) observation stays discharged within 2 days. However, there were 47,413 (2.8%) and 7,049 (0.4%) observation stays with LOS 3 to 4 days or >4 days, respectively.

Shifting Pattern in Observation Stays

The annual percentage of observation stays with prolonged LOS (>2 days) rose from 1.1% in 2010 to 4.6% in 2019 (P < .001; Figure 2). The empirical distribution of H-RISK from observation stays by years further suggests a slightly increasing trend in intensity of care under observation stays. As shown in Appendix Table 2, although the 1st, 5th, 10th, 25th, and 99th percentiles of H-RISK were relatively stable, the 50th, 75th, 90th, and 95th percentiles of H-RISK were increasing over time. The correlation between year and intensity of care used under observation stays (H-RISK from observation stays) was found to be weak but significantly positive (Spearman correlation coefficients = 0.04; P < .001).

Number and Annual Percentage of Observation Stays with Prolonged Length of Stay

Interaction coefficients from our regression model demonstrate that the existing inverse association between H-RISK and odds of admission as an observation stay became less negative over the years. In 2010, the adjusted odds ratio (OR) of H-RISK was 0.57 (95% CI, 0.55-0.59). By 2017, the adjusted OR had increased to 0.65 (95% CI, 0.64-0.66). Compared with 2010, the seven adjusted ORs of H-RISK at years 2012 through 2018 were observed to be higher and statistically significant (P < .001, Appendix Table 3).

Hospitals-Level Use of Observation Stays

After adjusting for all covariates and hospital random effects, hospital-level use of observation stays increased between 2010 and 2019 for 26 out of 29 children’s hospitals. Although observation status essentially was not used at two children’s hospitals over the study period, the median hospital-level use of observation stays was 26% in 2010 (IQR, 3%-36%) and increased to 46% (IQR: 39%; 55%) in 2019. As shown in Figure 3, the number of hospitals with a low percentage of observation stays (<26%) decreased from 15 in 2010 to 4 in 2019. The number of hospitals with a high percentage of observation stays (≥51%) increased from 5 in 2010 to 10 in 2019. Nevertheless, there remained significant variation in the use of observation stays, and the hospital-level use ranged from 0% to 67% in 2019.

Risk-Adjusted Hospital-Level Use of Observation Stays at Children’s Hospitals, 2010 to 2019

DISCUSSION

By 2020, observation status has become a key component of healthcare for pediatric patients, and its relevance for children’s hospitals recently has been described.22,23 However, trends in observation stays for pediatric populations are not known. This represents the first study showing temporal trends of observation stays at children’s hospitals after 2010. Our results confirm that the increase in observation stays for pediatric populations is not attributable to decreasing patient acuity at children’s hospitals. We found a weak but significantly positive correlation between year and intensity of care used under observation stays. Although this correlation might not be clinically important, it demonstrates that patient acuity in observation stays is not decreasing. Regression results suggest that observation stays now encompass patients who need relatively higher intensity of care compared with those admitted under observation status in 2010.

This study also identifies a unique pattern in the use of observation stays among pediatric populations. Earlier studies exclusively focused on observation stays that were admitted from EDs.24 Our results indicate that observation status has been used beyond a bridge from ED care to inpatient admission. In particular, observation status has expanded to include pediatric populations with more diverse clinical conditions (eg, appendicitis and diabetes mellitus), and has become a substantial component of postprocedural admissions (Appendix Figure). Looking forward, it is likely that the use of observation stays might surpass inpatient admissions for more conditions that primarily involve short-term stays.

Observation status originally was designed as a reimbursement strategy for patients who needed short stays in dedicated ED units or hospitals, but did not qualify for inpatient services.5,25 After several changes in reimbursement policies, CMS released the “two midnight rule” for Medicare beneficiaries in 2013, which replaced condition-based criteria with time-based criteria to determine an inpatient or observation stay.1 Some Medicaid programs and commercial payors have developed similar policies. Unlike the universal policy for Medicare populations, the regulations for pediatric populations vary by states and health insurers.11-15,26-28 This might partially explain the wide variation observed among children’s hospital-level use of observation stays. For example, the California Medicaid program did not have a reimbursement rate for observation services as of 2020, while the Texas Medicaid program has had a policy for observation stays since 2010.12,13 We found that two children’s hospitals in California had the lowest use of observation stays (almost zero), whereas the hospital-level use of observation stays was more than 50% for three out of four children’s hospitals in Texas. In addition to reimbursement policies, individual hospitals also might have different strategies for observation status designation. An earlier survey showed that there was lack of consistency in billing and payor-based designations of observation status at children’s hospitals.29 These findings suggest that children’s hospital-level use of observation stays likely is influenced by reimbursement policy and practical strategy for observation status determination.

Earlier studies reported that observation status could be a more efficient use of healthcare resources.5,6 However, there are still at least two concerns relevant to children’s hospitals during the last decade. The first is whether the use of observation stays can promote cost-saving or if it is just a cost-shifting strategy. An earlier study demonstrated that observation stays with prolonged LOS might increase risk of cost-sharing among adult patients.29 Our study reveals an increasing trend of observation stays with prolonged LOS for pediatric patients. Similar to adult patients, LOS exceeding 24 or 48 hours could lead to uncovered healthcare costs and financial burdens on families.30-32 Meanwhile, children’s hospitals also might take on a higher financial liability by implementing observation status. Earlier studies have indicated that resource use between observation and inpatient stays at children’s hospitals is similar, and increasing use of observation stays might lead to financial risk rather than cost effectiveness.33 Further, administrative costs of observation determination are considerably high.34 Medicaid is the major payor for pediatric patients in children’s hospitals. In this study, more than 50% of encounters were paid through Medicaid programs. It is well known that Medicaid reimbursement rates are lower than Medicare and commercial plans.35 Therefore, the cost-saving conclusion drawn from Medicare patients cannot be generalized to pediatric populations at children’s hospitals without cautious reevaluation.

A second concern with increasing use of observation stays is selection bias in public reporting and comparisons of hospital performance. Presently, four main categories of quality indicators established by the Agency for Healthcare Research and Quality rely heavily on inpatient encounters.36 In this study, we found that the range of hospital-level use of observation stays was large. In 2019, the risk-adjusted percent use of observation stays was less than 5% at three hospitals, while the percent use was greater than 60% in another three hospitals. Therefore, comparisons made without uniform accounting of observation stays might have significant implications for national rankings of children’s hospitals across the United States. These consequences have been investigated in several published studies.22,23,37-39

There are several limitations to our study. First, the study sample was limited to children’s hospitals that consistently reported inpatient and observation data over the entire study period. Eighteen hospitals (86%) excluded from this study did not consistently submit inpatient and observation data to PHIS from 2010 through 2019. The primary purpose of this study was to present temporal trends of observation stays for children’s hospitals, and it was important to build the hospital cohort based on valid and consistent data during the study period. Appendix Table 4 presents differences of hospital characteristics by included and excluded groups of hospitals. Excluded hospitals might have fewer resources (eg, fewer pediatric intensive care beds). Nonetheless, the selection of hospitals was optimized based on data availability. Second, this study was a retrospective review of an administrative database of children’s hospitals and units. The sample does not represent all children’s hospitals or pediatric patients in the United States, but there are no available data sources—that we know of—that can generate national estimates for both inpatient and observation stays. Third, we did not attempt to conclusively infer any causal effects, and several factors could explain the increasing trends, such as reimbursement policies, hospital-level implementation strategies, determination guidelines for observation status designation, as well as changes in clinical care. Further studies should investigate impact of these factors on the use of observation stays for pediatric patients and children’s hospitals.

CONCLUSION

Observation status has been increasingly used for pediatric patients with more diverse clinical conditions, and there is a rising trend of prolonged LOS among observation stays since 2010. Considerable variation exists in hospital-level use of observation stays across children’s hospitals. Observation status could be an opportunity to improve efficiency of healthcare resource use or could lead to a financial risk for patients with prolonged LOS. Future studies should explore appropriateness of observation care in clinical practice through leveraging efficient care and alleviating financial risk.

References

1. Centers for Medicare & Medicaid Services. Fact Sheet: Two-Midnight Rule. Accessed April 11, 2021. https://www.cms.gov/newsroom/fact-sheets/fact-sheet-two-midnight-rule-0
2. BlueCross BlueShield of Rhode Island. Payment Policy Outpaient Observation. Accessed April 11, 2021. https://www.bcbsri.com/sites/default/files/polices/Outpatient-Observation.pdf
3. Blue Cross Blue Shield of Illinois. Observation Services Tool for Applying MCG Care Guidelines Clinical Payment and Coding Policy. Accessed April 11, 2021. https://www.bcbsil.com/pdf/standards/observation_services_cpcp.pdf
4. Medicare.gov. Inpatient or outpatient hospital status affects your costs. Accessed April 11, 2021. https://www.medicare.gov/what-medicare-covers/what-part-a-covers/inpatient-or-outpatient-hospital-status
5. Ross MA, Hockenberry JM, Mutter R, Barrett M, Wheatley M, Pitts SR. Protocol-driven emergency department observation units offer savings, shorter stays, and reduced admissions. Health Aff (Millwood). 2013;32(12):2149-2156. https://doi.org/10.1377/hlthaff.2013.0662
6. Baugh CW, Venkatesh AK, Hilton JA, Samuel PA, Schuur JD, Bohan JS. Making greater use of dedicated hospital observation units for many short-stay patients could save $3.1 billion a year. Health Aff (Millwood). 2012;31(10):2314-2323. https://doi.org/10.1377/hlthaff.2011.0926
7. Sheehy AM, Graf B, Gangireddy S, et al. Hospitalized but not admitted: characteristics of patients with “observation status” at an academic medical center. JAMA Intern Med. 2013;173(21):1991-1998. https://doi.org/10.1001/jamainternmed.2013.8185
8. Baugh CW, Schuur JD. Observation care—high-value care or a cost-shifting loophole? N Engl J Med. 2013;369(4):302-305. https://doi.org/10.1056/NEJMp1304493
9. Missouri Hospital Association. A patient’s guide to observation care. Accessed April 11, 2021. https://www.mhanet.com/mhaimages/PatientsGuideToObservationCareFlyer.pdf
10. Cigna. Employee-paid hospital care coverage- summary of benefits. Accessed April 11, 2021. https://www.cigna.com/iwov-resources/national-second-sale/docs/healthy-benefits/updated-HC-benefit-summary.pdf
11. BlueCross BlueShield of Minnesota. Reimbursement policy-observation care services. Accessed April 11, 2021. https://www.bluecrossmn.com/sites/default/files/DAM/2020-07/Evaluation%20and%20Management%20004_Observation%20Care%20Services%20_09.04.17.pdf
12. California Department of Health Care Services. Public Hospital Project Frequently Asked Questions. Accessed April 11, 2021. https://www.dhcs.ca.gov/provgovpart/Documents/Public%20Hospital%20Project/PHP_Final_FAQs_January2013ADA.pdf
13. Texas Medicaid & Healthcare Partnership. Inpatient and Outpatient Hospital Servicces Handbook. Accessed May 29, 2021. https://www.tmhp.com/sites/default/files/microsites/provider-manuals/tmppm/html/TMPPM/2_Inpatient_Outpatient_Hosp_Srvs/2_Inpatient_Outpatient_Hosp_Srvs.htm
14. Alabama Medicaid. Outpatient observation. Accessed April 11, 2021. https://medicaid.alabama.gov/news_detail.aspx?ID=5121
15. NC Medicaid. Medicaid and Health Choice Clinical Coverage Policy No: 2A-1. Accessed April 11, 2021. https://files.nc.gov/ncdma/documents/files/2A-1_0.pdf
16. Feng Z, Wright B, Mor V. Sharp rise in Medicare enrollees being held in hospitals for observation raises concerns about causes and consequences. Health Aff (Millwood). 2012;31(6):1251-1259. https://doi.org/10.1377/hlthaff.2012.0129
17. Wright B, O’Shea AM, Ayyagari P, Ugwi PG, Kaboli P, Vaughan Sarrazin M. Observation rates at veterans’ hospitals more than doubled during 2005-13, similar to Medicare trends. Health Aff (Millwood). 2015;34(10):1730-1737. https://doi.org/10.1377/hlthaff.2014.1474
18. Wright B, Jung HY, Feng Z, Mor V. Hospital, patient, and local health system characteristics associated with the prevalence and duration of observation care. Health Serv Res. 2014;49(4):1088-1107. https://doi.org/10.1111/1475-6773.12166
19. Sabbatini AK, Wright B, Hall MK, Basu A. The cost of observation care for commercially insured patients visiting the emergency department. Am J Emerg Med. 2018;36(9):1591-1596. https://doi.org/10.1016/j.ajem.2018.01.040
20. Children’s Hospital Association. Pediatric health information system. Accessed April 11, 2021. https://www.childrenshospitals.org/phis
21. Richardson T, Rodean J, Harris M, Berry J, Gay JC, Hall M. Development of hospitalization resource intensity scores for kids (H-RISK) and comparison across pediatric populations. J Hosp Med. 2018;13(9):602-608. https://doi.org/10.12788/jhm.2948
22. Gay JC, Hall M, Morse R, Fieldston ES, Synhorst DC, Macy ML.Observation encounters and length of stay benchmarking in children’s hospitals. Pediatrics. 2020;146(5):e20200120. https://doi.org/10.1542/peds.2020-0120
23. Synhorst DC, Hall M, Harris M, et al. Hospital observation status and readmission rates. Pediatrics. 2020;146(5):e2020003954. https://doi.org/10.1542/peds.2020-003954
24. Macy ML, Hall M, Shah SS, et al. Pediatric observation status: are we overlooking a growing population in children’s hospitals? J Hosp Med. 2012;7(7):530-536. https://doi.org/10.1002/jhm.1923
25. Macy ML, Kim CS, Sasson C, Lozon MM, Davis MM. Pediatric observation units in the United States: a systematic review. J Hosp Med. 2010;5(3):172-182. https://doi.org/10.1002/jhm.592
26. UnitedHealthcare. Observation services policy, facility. Accessed April 11, 2021. https://www.uhcprovider.com/content/dam/provider/docs/public/policies/medicaid-comm-plan-reimbursement/UHCCP-Facility-Observation-Services-Policy-(F7106).pdf
27. Cal SB-1076§1253.7. General acute care hospitals: observation services – Health and Safety. Accessed April 11, 2021. https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=201520160SB1076
28. Nebraska Total Care. 2021 Provider Billing Guide. Accessed April 11, 2021. https://www.nebraskatotalcare.com/content/dam/centene/Nebraska/PDFs/ProviderRelations/NTC_Nebraska_Total_Care_Provider_Billing_Guide_508.pdf
29. Macy ML, Hall M, Shah SS, et al. Differences in designations of observation care in US freestanding children’s hospitals: are they virtual or real? J Hosp Med. 2012;7(4):287-293. https://doi.org/10.1002/jhm.949
30. Hockenberry JM, Mutter R, Barrett M, Parlato J, Ross MA. Factors associated with prolonged observation services stays and the impact of long stays on patient cost. Health Serv Res. 2014;49(3):893-909. https://doi.org/10.1111/1475-6773.12143
31. Anthem BlueCross BlueShield. Ohio Provider Manual. Accessed April11, 2021. https://www11.anthem.com/provider/oh/f1/s0/t0/pw_g357368.pdf?refer=ahpprovider&state=oh
32. Humana. Provider manual for physicians, hospitals and healthcare providers. Accessed April 11, 2021. https://docushare-web.apps.cf.humana.com/Marketing/docushare-app?file=3932669
33. Fieldston ES, Shah SS, Hall M, et al. Resource utilization for observation-status stays at children’s hospitals. Pediatrics. 2013;131(6):1050-1058 https://doi.org/10.1542/peds.2012-249
34. Tejedor-Sojo J. Observation status-a name at what cost? Hosp Pediatr. 2014;4(5):321-323. https://doi.org/10.1542/hpeds.2014-0037.
35. Selden TM, Karaca Z, Keenan P, White C, Kronick R. The growing difference between public and private payment rates for inpatient hospital care. Health Aff (Millwood). 2015;34(12):2147-2150. https://doi.org/10.1377/hlthaff.2015.0706
36. Agency for Healthcare Research and Quality. AHRQ Quality Indicators. Accessed April 11, 2021. https://www.qualityindicators.ahrq.gov
37. Figueroa JF, Burke LG, Zheng J, Orav EJ, Jha AK. Trends in hospitalization vs observation stay for ambulatory care-sensitive conditions. JAMA Intern Med. 2019;179(12):1714-1716. https://doi.org/10.1001/jamainternmed.2019.3177
38. Markham JL, Hall M, Gay JC, Bettenhausen JL, Berry JG. Length of stay and cost of pediatric readmissions. Pediatrics. 2018;141(4):e20172934. https://doi.org/10.1542/peds.2017-2934.
39. Overman RA, Freburger JK, Assimon MM, Li X, Brookhart, MA. Observation stays in administrative claims databases: underestimation of hospitalized cases. Pharmacoepidemiol Drug Saf. 2014;23(9):902-910. https://doi.org/10.1002/pds.3647.

References

1. Centers for Medicare & Medicaid Services. Fact Sheet: Two-Midnight Rule. Accessed April 11, 2021. https://www.cms.gov/newsroom/fact-sheets/fact-sheet-two-midnight-rule-0
2. BlueCross BlueShield of Rhode Island. Payment Policy Outpaient Observation. Accessed April 11, 2021. https://www.bcbsri.com/sites/default/files/polices/Outpatient-Observation.pdf
3. Blue Cross Blue Shield of Illinois. Observation Services Tool for Applying MCG Care Guidelines Clinical Payment and Coding Policy. Accessed April 11, 2021. https://www.bcbsil.com/pdf/standards/observation_services_cpcp.pdf
4. Medicare.gov. Inpatient or outpatient hospital status affects your costs. Accessed April 11, 2021. https://www.medicare.gov/what-medicare-covers/what-part-a-covers/inpatient-or-outpatient-hospital-status
5. Ross MA, Hockenberry JM, Mutter R, Barrett M, Wheatley M, Pitts SR. Protocol-driven emergency department observation units offer savings, shorter stays, and reduced admissions. Health Aff (Millwood). 2013;32(12):2149-2156. https://doi.org/10.1377/hlthaff.2013.0662
6. Baugh CW, Venkatesh AK, Hilton JA, Samuel PA, Schuur JD, Bohan JS. Making greater use of dedicated hospital observation units for many short-stay patients could save $3.1 billion a year. Health Aff (Millwood). 2012;31(10):2314-2323. https://doi.org/10.1377/hlthaff.2011.0926
7. Sheehy AM, Graf B, Gangireddy S, et al. Hospitalized but not admitted: characteristics of patients with “observation status” at an academic medical center. JAMA Intern Med. 2013;173(21):1991-1998. https://doi.org/10.1001/jamainternmed.2013.8185
8. Baugh CW, Schuur JD. Observation care—high-value care or a cost-shifting loophole? N Engl J Med. 2013;369(4):302-305. https://doi.org/10.1056/NEJMp1304493
9. Missouri Hospital Association. A patient’s guide to observation care. Accessed April 11, 2021. https://www.mhanet.com/mhaimages/PatientsGuideToObservationCareFlyer.pdf
10. Cigna. Employee-paid hospital care coverage- summary of benefits. Accessed April 11, 2021. https://www.cigna.com/iwov-resources/national-second-sale/docs/healthy-benefits/updated-HC-benefit-summary.pdf
11. BlueCross BlueShield of Minnesota. Reimbursement policy-observation care services. Accessed April 11, 2021. https://www.bluecrossmn.com/sites/default/files/DAM/2020-07/Evaluation%20and%20Management%20004_Observation%20Care%20Services%20_09.04.17.pdf
12. California Department of Health Care Services. Public Hospital Project Frequently Asked Questions. Accessed April 11, 2021. https://www.dhcs.ca.gov/provgovpart/Documents/Public%20Hospital%20Project/PHP_Final_FAQs_January2013ADA.pdf
13. Texas Medicaid & Healthcare Partnership. Inpatient and Outpatient Hospital Servicces Handbook. Accessed May 29, 2021. https://www.tmhp.com/sites/default/files/microsites/provider-manuals/tmppm/html/TMPPM/2_Inpatient_Outpatient_Hosp_Srvs/2_Inpatient_Outpatient_Hosp_Srvs.htm
14. Alabama Medicaid. Outpatient observation. Accessed April 11, 2021. https://medicaid.alabama.gov/news_detail.aspx?ID=5121
15. NC Medicaid. Medicaid and Health Choice Clinical Coverage Policy No: 2A-1. Accessed April 11, 2021. https://files.nc.gov/ncdma/documents/files/2A-1_0.pdf
16. Feng Z, Wright B, Mor V. Sharp rise in Medicare enrollees being held in hospitals for observation raises concerns about causes and consequences. Health Aff (Millwood). 2012;31(6):1251-1259. https://doi.org/10.1377/hlthaff.2012.0129
17. Wright B, O’Shea AM, Ayyagari P, Ugwi PG, Kaboli P, Vaughan Sarrazin M. Observation rates at veterans’ hospitals more than doubled during 2005-13, similar to Medicare trends. Health Aff (Millwood). 2015;34(10):1730-1737. https://doi.org/10.1377/hlthaff.2014.1474
18. Wright B, Jung HY, Feng Z, Mor V. Hospital, patient, and local health system characteristics associated with the prevalence and duration of observation care. Health Serv Res. 2014;49(4):1088-1107. https://doi.org/10.1111/1475-6773.12166
19. Sabbatini AK, Wright B, Hall MK, Basu A. The cost of observation care for commercially insured patients visiting the emergency department. Am J Emerg Med. 2018;36(9):1591-1596. https://doi.org/10.1016/j.ajem.2018.01.040
20. Children’s Hospital Association. Pediatric health information system. Accessed April 11, 2021. https://www.childrenshospitals.org/phis
21. Richardson T, Rodean J, Harris M, Berry J, Gay JC, Hall M. Development of hospitalization resource intensity scores for kids (H-RISK) and comparison across pediatric populations. J Hosp Med. 2018;13(9):602-608. https://doi.org/10.12788/jhm.2948
22. Gay JC, Hall M, Morse R, Fieldston ES, Synhorst DC, Macy ML.Observation encounters and length of stay benchmarking in children’s hospitals. Pediatrics. 2020;146(5):e20200120. https://doi.org/10.1542/peds.2020-0120
23. Synhorst DC, Hall M, Harris M, et al. Hospital observation status and readmission rates. Pediatrics. 2020;146(5):e2020003954. https://doi.org/10.1542/peds.2020-003954
24. Macy ML, Hall M, Shah SS, et al. Pediatric observation status: are we overlooking a growing population in children’s hospitals? J Hosp Med. 2012;7(7):530-536. https://doi.org/10.1002/jhm.1923
25. Macy ML, Kim CS, Sasson C, Lozon MM, Davis MM. Pediatric observation units in the United States: a systematic review. J Hosp Med. 2010;5(3):172-182. https://doi.org/10.1002/jhm.592
26. UnitedHealthcare. Observation services policy, facility. Accessed April 11, 2021. https://www.uhcprovider.com/content/dam/provider/docs/public/policies/medicaid-comm-plan-reimbursement/UHCCP-Facility-Observation-Services-Policy-(F7106).pdf
27. Cal SB-1076§1253.7. General acute care hospitals: observation services – Health and Safety. Accessed April 11, 2021. https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=201520160SB1076
28. Nebraska Total Care. 2021 Provider Billing Guide. Accessed April 11, 2021. https://www.nebraskatotalcare.com/content/dam/centene/Nebraska/PDFs/ProviderRelations/NTC_Nebraska_Total_Care_Provider_Billing_Guide_508.pdf
29. Macy ML, Hall M, Shah SS, et al. Differences in designations of observation care in US freestanding children’s hospitals: are they virtual or real? J Hosp Med. 2012;7(4):287-293. https://doi.org/10.1002/jhm.949
30. Hockenberry JM, Mutter R, Barrett M, Parlato J, Ross MA. Factors associated with prolonged observation services stays and the impact of long stays on patient cost. Health Serv Res. 2014;49(3):893-909. https://doi.org/10.1111/1475-6773.12143
31. Anthem BlueCross BlueShield. Ohio Provider Manual. Accessed April11, 2021. https://www11.anthem.com/provider/oh/f1/s0/t0/pw_g357368.pdf?refer=ahpprovider&state=oh
32. Humana. Provider manual for physicians, hospitals and healthcare providers. Accessed April 11, 2021. https://docushare-web.apps.cf.humana.com/Marketing/docushare-app?file=3932669
33. Fieldston ES, Shah SS, Hall M, et al. Resource utilization for observation-status stays at children’s hospitals. Pediatrics. 2013;131(6):1050-1058 https://doi.org/10.1542/peds.2012-249
34. Tejedor-Sojo J. Observation status-a name at what cost? Hosp Pediatr. 2014;4(5):321-323. https://doi.org/10.1542/hpeds.2014-0037.
35. Selden TM, Karaca Z, Keenan P, White C, Kronick R. The growing difference between public and private payment rates for inpatient hospital care. Health Aff (Millwood). 2015;34(12):2147-2150. https://doi.org/10.1377/hlthaff.2015.0706
36. Agency for Healthcare Research and Quality. AHRQ Quality Indicators. Accessed April 11, 2021. https://www.qualityindicators.ahrq.gov
37. Figueroa JF, Burke LG, Zheng J, Orav EJ, Jha AK. Trends in hospitalization vs observation stay for ambulatory care-sensitive conditions. JAMA Intern Med. 2019;179(12):1714-1716. https://doi.org/10.1001/jamainternmed.2019.3177
38. Markham JL, Hall M, Gay JC, Bettenhausen JL, Berry JG. Length of stay and cost of pediatric readmissions. Pediatrics. 2018;141(4):e20172934. https://doi.org/10.1542/peds.2017-2934.
39. Overman RA, Freburger JK, Assimon MM, Li X, Brookhart, MA. Observation stays in administrative claims databases: underestimation of hospitalized cases. Pharmacoepidemiol Drug Saf. 2014;23(9):902-910. https://doi.org/10.1002/pds.3647.

Issue
Journal of Hospital Medicine 16(11)
Issue
Journal of Hospital Medicine 16(11)
Page Number
645-651. Published Online First July 21, 2021
Page Number
645-651. Published Online First July 21, 2021
Publications
Publications
Topics
Article Type
Display Headline
Trends and Variation in the Use of Observation Stays at Children’s Hospitals
Display Headline
Trends and Variation in the Use of Observation Stays at Children’s Hospitals
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Mehul V Raval, MD, MS; Email: [email protected].
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Page Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media
Media Files

Inpatient Glycemic Control With Sliding Scale Insulin in Noncritical Patients With Type 2 Diabetes: Who Can Slide?

Article Type
Changed
Thu, 07/29/2021 - 10:23
Display Headline
Inpatient Glycemic Control With Sliding Scale Insulin in Noncritical Patients With Type 2 Diabetes: Who Can Slide?

Sliding scale insulin (SSI) for inpatient glycemic control was first proposed by Elliott P Joslin in 1934 when he recommended titration of insulin based on urine glucose levels.1 As bedside glucose meters became widely available, physicians transitioned to dosing SSI based on capillary blood glucose (BG) levels,2,3 and SSI became widely used for the management of inpatient hyperglycemia.1 However, during the past decade, there has been strong opposition to the use of SSI in hospitals. Many authors oppose its use, highlighting the retrospective rather than prospective nature of SSI therapy and concerns about inadequate glycemic control.4-6 In 2004, the American College of Endocrinology first released a position statement discouraging the use of SSI alone and recommended basal-bolus insulin as the preferred method of glycemic control for inpatients with type 2 diabetes (T2D).7 The American Diabetes Association (ADA) inpatient guidelines in 20058 and the Endocrine Society guidelines in 20129 also opposed SSI monotherapy and reaffirmed that a basal-bolus insulin regimen should be used for most non–critically ill patients with diabetes. Those guidelines remain in place currently.

Several randomized controlled trials (RCTs) and meta-analyses have shown that basal-bolus insulin regimens provide superior glycemic control in non–critical inpatients when compared with SSI alone.10-14 In addition, the RABBIT 2 (Randomized Study of Basal-Bolus Insulin Therapy in the Inpatient Management of Patients With Type 2 Diabetes) trial showed a significant reduction in perioperative complications10 among surgical patients when treated with basal-bolus insulin therapy. Despite these studies and strong recommendations against its use, SSI continues to be widely used in the United States. According to a 2007 survey of 44 US hospitals, 41% of noncritical patients with hyperglycemia were treated with SSI alone.15 In addition, SSI remains one of the most commonly prescribed insulin regimens in many countries around the world.16-19 The persistence of SSI use raises questions as to why clinicians continue to use a therapy that has been strongly criticized. Some authors point to convenience and fear of hypoglycemia with a basal-bolus insulin regimen.20,21 Alternatively, it is possible that SSI usage remains so pervasive because it is effective in a subset of patients. In fact, a 2018 Cochrane review concluded that existing evidence is not sufficiently robust to definitively recommend basal-bolus insulin over SSI for inpatient diabetes management of non–critically ill patients despite existing guidelines.22

Owing to the ongoing controversy and widespread use of SSI, we designed an exploratory analysis to understand the rationale for such therapy by investigating whether a certain subpopulation of hospitalized patients with T2D may achieve target glycemic control with SSI alone. We hypothesized that noncritical patients with mild hyperglycemia and admission BG <180 mg/dL would do well with SSI alone and may not require intensive treatment with basal-bolus insulin regimens. To address this question, we used electronic health records with individual-level patient data to assess inpatient glycemic control of non–critically ill patients with T2D treated with SSI alone.

METHODS

Participants

Data from 25,813 adult noncritical inpatients with T2D, with an index admission between June 1, 2010, and June 30, 2018, were obtained through the Emory Healthcare Clinical Data Warehouse infrastructure program. All patients were admitted to Emory Healthcare hospitals, including Emory University Hospital, Emory University Hospital Midtown, and Emory Saint Joseph’s Hospital, in Atlanta, Georgia. Data were extracted for each patient during the index hospitalization, including demographics, anthropometrics, and admission and inpatient laboratory values. Information was collected on daily point-of-care glucose values, hemoglobin A1c (HbA1c), hypoglycemic events, insulin doses, hospital complications, comorbidities, and hospital setting (medical vs surgical admission). International Classification of Diseases, 9th and 10th Revisions (ICD-9/10) codes were used to determine diagnosis of T2D, comorbidities, and complications.

From our initial dataset, we identified 16,366 patients who were treated with SSI during hospitalization. We excluded patients who were admitted to the intensive care unit (ICU) or placed on intravenous insulin, patients with missing admission BG values, and patients with a length of stay less than 1 day. To prevent inclusion of patients presenting in diabetic ketoacidosis or hyperosmolar hyperglycemic syndrome, we excluded patients with an admission BG >500 mg/dL. We then excluded 6,739 patients who received basal insulin within the first 2 days of hospitalization, as well as 943 patients who were treated with noninsulin (oral or injectable) antidiabetic agents. Our final dataset included 8,095 patients (Appendix Figure).

Patients in the SSI cohort included all patients who were treated with short-acting insulin only (regular insulin or rapid-acting [lispro, aspart, glulisine] insulin analogs) during the first 2 days of hospitalization. Patients who remained on only short-acting insulin during the entire hospitalization were defined as continuous SSI patients. Patients who subsequently received basal insulin after day 2 of hospitalization were defined as patients who transitioned to basal. Patients were stratified according to admission BG levels (first BG available on day of admission) and HbA1c (when available during index admission). We compared the baseline characteristics and clinical outcomes of patients who remained on SSI alone throughout the entirety of hospitalization with those of patients who required transition to basal insulin. The mean hospital BG was calculated by taking the average of all BG measurements during the hospital stay. We defined hypoglycemia as a BG <70 mg/dL and severe hypoglycemia as BG <40 mg/dL. Repeated hypoglycemia values were excluded if they occurred within a period of 2 hours.

Outcome Measures

The primary outcome was the percentage of patients with T2D achieving target glycemic control with SSI therapy, defined as mean hospital BG between 70 and 180 mg/dL without hypoglycemia <70 mg/dL during hospital stay. This threshold was determined based on 2019 ADA recommendations targeting hospital BG <180 mg/dL and avoidance of hypoglycemia.23

Statistical Analysis

Patients were stratified according to continuous SSI versus transitioned to basal treatment. Patients who remained on continuous SSI were further categorized into four categories based on admission BG: <140 mg/dL, 140 to 180 mg/dL, 180 to 250 mg/dL, and ≥250 mg/dL. Clinical characteristics were compared using Wilcoxon rank-sum tests (if continuous) and chi-square tests or Fisher exact tests (if categorical). We then compared the clinical outcomes among continuous SSI patients with different admission BG levels (<140 mg/dL, 140-180 mg/dL, 180-250 mg/dL, and ≥250 mg/dL) and with different HbA1c levels (<7%, 7%-8%, 8%-9%, ≥9%). Within each scenario, logistic regression for the outcome of poor glycemic control, defined as mean hospital BG >180 mg/dL, was performed to evaluate the HbA1c levels and admission BG levels controlling for other factors (age, gender, body mass index [BMI], race, setting [medicine versus surgery] and Charlson Comorbidity Index score). A P value < .05 was regarded as statistically significant. All analyses were performed based on available cases and conducted in SAS version 9.4 (SAS Institute Inc.).

RESULTS

Among 25,813 adult patients with T2D, 8,095 patients (31.4%) were treated with SSI alone during the first 2 days of hospitalization. Of those patients treated with SSI, 6,903 (85%) remained on continuous SSI alone during the entire hospitalization, and 1,192 (15%) were transitioned to basal insulin. The clinical characteristics of these patients on continuous SSI and those who transitioned to basal insulin are shown in Table 1. Patients who transitioned to basal insulin had significantly higher mean (SD) admission BG (191.8 [88.2] mg/dL vs 156.4 [65.4] mg/dL, P < .001) and higher mean (SD) HbA1c (8.1% [2.0%] vs 7.01% [1.5%], P < .001), compared with those who remained on continuous SSI. Patients who transitioned to basal insulin were also younger and more likely to have chronic kidney disease (CKD), but less likely to have congestive heart failure, coronary artery disease, or chronic obstructive pulmonary disease (COPD). The Charlson Comorbidity Index score was significantly higher for patients who transitioned to basal (4.4 [2.5]) than for those who remained on continuous SSI (4.1 [2.5], P < .001). There were no significant differences among sex, BMI, or glomerular filtration rate (GFR) on admission. Of those transitioned to basal insulin, 53% achieved a mean hospitalization BG <180 mg/dL, compared with 82% of those on continuous SSI. The overall rate of hypoglycemia in the continuous SSI group was 8% compared with 18% in those transitioned to basal insulin.

Clinical Characteristics of Hospitalized Patients With Type 2 Diabetes Treated With SSI by Continuous SSI vs Transitioned to Basal Insulin

Of the patients who remained on continuous SSI throughout the hospitalization, 3,319 patients (48%) had admission BG <140 mg/dL, 1,671 patients (24%) had admission BG 140 to 180 mg/dL, and 1,913 patients (28%) had admission BG >180 mg/dL. Only 9% of patients who remained on continuous SSI had admission BG ≥250 mg/dL. Patients with admission BG <140 mg/dL were older, had lower BMI and HbA1c, had higher rates of COPD and CKD, and were more likely to be admitted to a surgical service compared with patients with admission BG >140 mg/dL (P < .05 for all; Table 2).

Clinical Characteristics of Patients Who Remained on Continuous SSI by Admission Blood Glucose Concentration

Hospital glycemic control for patients on continuous SSI according to admission BG is displayed in Table 3. Among patients who remained on continuous SSI, 96% of patients with admission BG <140 mg/dL had a mean hospital BG <180 mg/dL; of them, 86% achieved target control without hypoglycemia. Similar rates of target control were achieved in patients with admission BG 140 to 180 mg/dL (83%), in contrast to patients with admission BG ≥250 mg/dL, of whom only 18% achieved target control (P < .001). These findings parallel those seen in patients transitioned to basal insulin. Of patients in the transition group admitted with BG <140 mg/dL and <180 mg/dL, 88.5% and 84.6% had mean hospital BG <180 mg/dL, respectively, while 69.1% and 68.9% had mean BG between 70 and 180 mg/dL without hypoglycemia. The overall frequency of hypoglycemia <70 mg/dL among patients on continuous SSI was 8% and was more common in patients with admission BG <140 mg/dL (10%) compared with patients with higher admission glucose levels (BG 140-180 mg/dL [4%], 180-250 mg/dL [4%], or ≥250 mg/dL [6%], P < .001). There was no difference in rates of severe hypoglycemia <40 mg/dL among groups.

Glycemic Data of Patients on Continuous SSI by Admission Blood Glucose Concentration and Admission HbA1c

HbA1c data were available for 2,560 of the patients on continuous SSI (Table 3). Mean hospital BG increased significantly with increasing HbA1c values. Patients admitted with HbA1c <7% had lower mean (SD) hospital BG (132.2 [28.2] mg/dL) and were more likely to achieve target glucose control during hospitalization (85%) compared with those with HbA1c 7% to 8% (mean BG, 148.7 [30.8] mg/dL; 80% target control), HbA1c 8% to 9% (mean BG, 169.1 [37.9] mg/dL; 61% target control), or HbA1c ≥9% (mean BG, 194.9 [53.4] mg/dL; 38% target control) (P < .001).

In a logistic regression analysis adjusted for age, gender, BMI, race, setting (medicine vs surgery), and Charlson Comorbidity Index score, the odds of poor glycemic control increased with higher admission BG (admission BG 140-180 mg/dL: odds ratio [OR], 1.8; 95% CI, 1.5-2.2; admission BG 180-250 mg/dL: OR, 3.7; 95% CI, 3.1-4.4; admission BG ≥250 mg/dL: OR, 7.2; 95% CI, 5.8-9.0; reference admission BG <140 mg/dL; Figure). Similarly, the logistic regression analysis showed greater odds of poor in-hospital glycemic control with increasing HbA1c (OR, 6.1; 95% CI, 4.3-8.8 for HbA1c >9% compared with HbA1c <7%).

Results of Logistic Regression Analysis

DISCUSSION

This large retrospective cohort study examined the effectiveness of SSI for glycemic control in noncritical inpatients with T2D. Our results indicate that SSI is still widely used in our hospital system, with 31.4% of our initial cohort managed with SSI alone. We found that 86% of patients with BG <140 mg/dL and 83% of patients with BG 140 to 180 mg/dL achieved glycemic control without hypoglycemia when managed with SSI alone, compared with 53% of those admitted with BG 180 to 250 mg/dL and only 18% of those with admission BG ≥250 mg/dL. This high success rate of achieving optimal BG control with SSI alone is comparable to that seen with transition to basal insulin and may explain the prevalent use of SSI for the management of patients with T2D and mild to moderate hyperglycemia.

Published clinical guideline recommendations promoting the use of basal-bolus insulin treatment algorithms are based on the results of a few RCTs that compared the efficacy of SSI vs a basal-bolus insulin regimen. These studies reported significantly lower mean daily BG concentration with basal or basal-bolus insulin therapy compared with SSI.10,11,24 However, it is interesting to note that the mean admission BG of patients treated with SSI in these RCTs ranged from 184 to 225 mg/dL. Patients in these trials were excluded if admission BG was <140 mg/dL.10,11,24 This is in contrast to our study evaluating real-world data in non–critically ill settings in which we found that 48% of patients treated with SSI had admission BG <140 mg/dL, and nearly 75% had admission BG <180 mg/dL. This suggests that by nature of study design, most RCTs excluded the population of patients who do achieve good glycemic control with SSI and may have contributed to the perception that basal insulin is preferable in all populations.

Our analysis indicates that healthcare professionals should consider admission BG when selecting the type of insulin regimen to manage patients with T2D in the hospital. Our results suggest that SSI may be appropriate for many patients with admission BG <180 mg/dL and should be avoided as monotherapy in patients with admission BG ≥180 mg/dL, as the proportion of patients achieving target control decreased with increasing admission BG. More importantly, if a patient is not controlled with SSI alone, intensification of therapy with the addition of basal insulin is indicated to achieve glycemic control. In addition, we found that the admission HbA1c is an appropriate marker to consider as well, with hospital glycemic control deteriorating with increasing HbA1c values, paralleling the admission BG. The main limitation to widespread use of HbA1c for therapeutic decision-making is access to values at time of patient admission; in our population, only 37% of patients had an HbA1c value available during the index hospitalization.

Previous publications have reported that hypoglycemia carries significant safety concerns, especially among a hospitalized population.25-27 As such, we included hypoglycemia as an important metric in our definition of target glycemic control rather than simply using mean hospital BG or number of hyperglycemic events to define treatment effectiveness. We did find a higher rate of hypoglycemia in patients with moderate admission BG treated with SSI compared with those with higher admission BG; however, few patients overall experienced clinically significant (<54 mg/dL) or severe (<40 mg/dL) hypoglycemia.

In our population, only 15% of patients started on SSI received additional basal insulin during hospitalization. This finding is similar to data reported in the Rabbit 2 trial, in which 14% of patients failed SSI alone, with a higher failure rate among those with higher BG on admission.10 Given the observational nature of this study, we cannot definitively state why certain patients in our population required additional basal insulin, but we can hypothesize that these patients admitted with BG ≥180 mg/dL had higher treatment failure rates and greater rates of hyperglycemia, therefore receiving intensified insulin therapy as clinically indicated at the discretion of the treating physician. Patients who transitioned from SSI to basal insulin had significantly higher admission BG and HbA1c compared with patients who remained on SSI alone. We noted that the rates of hypoglycemia were higher in the group that transitioned to basal (18% vs 8%) and similar to rates reported in previous RCTs.11,24

This observational study takes advantage of a large, diverse study population and a combination of medicine and surgery patients in a real-world setting. We acknowledge several limitations in our study. Our primary data were observational in nature, and as such, some baseline patient characteristics were notably different between groups, suggesting selection bias for treatment allocation to SSI. We do not know which patients were managed by primary teams compared with specialized diabetes consult services, which may also influence treatment regimens. We did not have access to information about patients’ at-home diabetes medication regimens or duration of diabetes, both of which have been shown in prior publications to affect an individual’s overall hospital glycemic control. Data on HbA1c values were available for only approximately one-third of patients. In addition, our study did not include patients without a history of diabetes who developed stress-induced hyperglycemia, a population that may benefit from conservative therapy such as SSI.28 A diagnosis of CKD was defined based on ICD 9/10 codes and not on admission estimated GFR. More specific data regarding stage of CKD or changes in renal function over the duration of hospitalization are not available, which could influence insulin prescribing practice. In addition, we defined the basal group as patients prescribed any form of basal insulin (NPH, glargine, detemir or degludec), and we do not have information on the use of prandial versus correction doses of rapid-acting insulin in the basal insulin–treated group.

CONCLUSION

In conclusion, our observational study indicates that the use of SSI results in appropriate target glycemic control for most noncritical medicine and surgery patients with admission BG <180 mg/dL. In agreement with previous RCTs, our study confirms that SSI as monotherapy is frequently inadequate in patients with significant hyperglycemia >180 mg/dL.10,11,24,29 We propose that an individualized approach to inpatient glycemic management is imperative, and cautious use of SSI may be a viable option for certain patients with mild hyperglycemia and admission BG <180 mg/dL. Further observational and randomized studies are needed to confirm the efficacy of SSI therapy in T2D patients with mild hyperglycemia. By identifying which subset of patients can be safely managed with SSI alone, we can better understand which patients will require escalation of therapy with intensive glucose management.

Files
References

1. Umpierrez GE, Palacio A, Smiley D. Sliding scale insulin use: myth or insanity? Am J Med. 2007;120(7):563-567. https://doi.org/10.1016/j.amjmed.2006.05.070
2. Kitabchi AE, Ayyagari V, Guerra SM. The efficacy of low-dose versus conventional therapy of insulin for treatment of diabetic ketoacidosis. Ann Intern Med. 1976;84(6):633-638. https://doi.org/10.7326/0003-4819-84-6-633
3. Skyler JS, Skyler DL, Seigler DE, O’Sullivan MJ. Algorithms for adjustment of insulin dosage by patients who monitor blood glucose. Diabetes Care. 1981;4(2):311-318. https://doi.org/10.2337/diacare.4.2.311
4. Gearhart JG, Duncan JL 3rd, Replogle WH, Forbes RC, Walley EJ. Efficacy of sliding-scale insulin therapy: a comparison with prospective regimens. Fam Pract Res J. 1994;14(4):313-322.
5. Queale WS, Seidler AJ, Brancati FL. Glycemic control and sliding scale insulin use in medical inpatients with diabetes mellitus. Arch Intern Med. 1997;157(5):545-552.
6. Clement S, Braithwaite SS, Magee MF, et al. Management of diabetes and hyperglycemia in hospitals. Diabetes Care. 2004;27(2):553-591. https://doi.org/10.2337/diacare.27.2.553
7. Garber AJ, Moghissi ES, Bransome ED Jr, et al. American College of Endocrinology position statement on inpatient diabetes and metabolic control. Endocr Pract. 2004;10(1):78-82. https://doi.org/10.4158/EP.10.1.77
8. American Diabetes Association. Standards of medical care in diabetes. Diabetes Care. 2005;28(suppl 1):S4-S36.
9. Umpierrez GE, Hellman R, Korytkowski MT, , et al. Management of hyperglycemia in hospitalized patients in non-critical care setting: an Endocrine Society clinical practice guideline. J Clin Endocrinol Metab. 2012;97(1):16-38. https://doi.org/10.1210/jc.2011-2098
10. Umpierrez GE, Smiley D, Zisman A, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes. Diabetes Care. 2007;30(9):2181-2186. https://doi.org/10.2337/dc07-0295
11. Umpierrez GE, Smiley D, Jacobs S, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes undergoing general surgery (RABBIT 2 surgery). Diabetes Care. 2011;34(2):256-261. https://doi.org/10.2337/dc10-1407
12. Schroeder JE, Liebergall M, Raz I, Egleston R, Ben Sussan G, Peyser A. Benefits of a simple glycaemic protocol in an orthopaedic surgery ward: a randomized prospective study. Diabetes Metab Res Rev. 2012;28:71-75. https://doi.org/10.1002/dmrr.1217
13. Lee YY, Lin YM, Leu WJ, et al. Sliding-scale insulin used for blood glucose control: a meta-analysis of randomized controlled trials. Metabolism. 2015;64(9):1183-1192. https://doi.org/10.1016/j.metabol.2015.05.011
14. Christensen MB, Gotfredsen A, Nørgaard K. Efficacy of basal-bolus insulin regimens in the inpatient management of non-critically ill patients with type 2 diabetes: a systematic review and meta-analysis. Diabetes Metab Res Rev. 2017;33(5):e2885. https://doi.org/10.1002/dmrr.2885
15. Wexler DJ, Meigs JB, Cagliero E, Nathan DM, Grant RW. Prevalence of hyper- and hypoglycemia among inpatients with diabetes: a national survey of 44 U.S. hospitals. Diabetes Care. 2007;30(2):367-369. https://doi.org/10.2337/dc06-1715
16. Moreira ED Jr, Silveira PCB, Neves RCS, Souza C Jr, Nunes ZO, Almeida MdCC. Glycemic control and diabetes management in hospitalized patients in Brazil. Diabetol Metab Syndr. 2013;5(1):62. https://doi.org/10.1186/1758-5996-5-62
17. Akhtar ST, Mahmood K, Naqvi IH, Vaswani AS. Inpatient management of type 2 diabetes mellitus: does choice of insulin regimen really matter? Pakistan J Med Sci. 2014;30(4):895-898.
18. Gómez Cuervo C, Sánchez Morla A, Pérez-Jacoiste Asín MA, Bisbal Pardo O, Pérez Ordoño L, Vila Santos J. Effective adverse event reduction with bolus-basal versus sliding scale insulin therapy in patients with diabetes during conventional hospitalization: systematic review and meta-analysis. Endocrinol Nutr. 2016;63(4):145-156. https://doi.org/10.1016/j.endonu.2015.11.008
19. Bain A, Hasan SS, Babar ZUD. Interventions to improve insulin prescribing practice for people with diabetes in hospital: a systematic review. Diabet Med. 2019;36(8):948-960. https://doi.org/10.1111/dme.13982
20. Ambrus DB, O’Connor MJ. Things We Do For No Reason: sliding-scale insulin as monotherapy for glycemic control in hospitalized patients. J Hosp Med. 2019;14(2):114-116. https://doi.org/10.12788/jhm.3109
21. Nau KC, Lorenzetti RC, Cucuzzella M, Devine T, Kline J. Glycemic control in hospitalized patients not in intensive care: beyond sliding-scale insulin. Am Fam Physician. 2010;81(9):1130-1135.
22. Colunga-Lozano LE, Gonzalez Torres FJ, Delgado-Figueroa N, et al. Sliding scale insulin for non-critically ill hospitalised adults with diabetes mellitus. Cochrane Database Syst Rev. 2018;11(11):CD011296. https://doi.org/10.1002/14651858.CD011296.pub2
23. American Diabetes Association. Diabetes care in the hospital: Standards of Medical Care in Diabetes—2019. Diabetes Care. 2019;42(suppl 1):S173-S181. https://doi.org/10.2337/dc19-S015
24. Umpierrez GE, Smiley D, Hermayer K, et al. Randomized study comparing a basal-bolus with a basal plus correction management of medical and surgical patients with type 2 diabetes: basal plus trial. Diabetes Care. 2013;36(8):2169-2174. https://doi.org/10.2337/dc12-1988
25. Turchin A, Matheny ME, Shubina M, Scanlon SV, Greenwood B, Pendergrass ML. Hypoglycemia and clinical outcomes in patients with diabetes hospitalized in the general ward. Diabetes Care. 2009;32(7):1153-1157. https://doi.org/10.2337/dc08-2127
26. Garg R, Hurwitz S, Turchin A, Trivedi A. Hypoglycemia, with or without insulin therapy, is associated with increased mortality among hospitalized patients. Diabetes Care. 2013;36(5):1107-1110. https://doi.org/10.2337/dc12-1296
27. Zapatero A, Gómez-Huelgas R, González N, et al. Frequency of hypoglycemia and its impact on length of stay, mortality, and short-term readmission in patients with diabetes hospitalized in internal medicine wards. Endocr Pract. 2014;20(9):870-875. https://doi.org/10.4158/EP14006.OR
28. Umpierrez GE, Isaacs SD, Bazargan N, You X, Thaler LM, Kitabchi AE. Hyperglycemia: an independent marker of in-hospital mortality in patients with undiagnosed diabetes. J Clin Endocrinol Metab. 2002;87(3):978-982. https://doi.org/10.1210/jcem.87.3.8341
29. Dickerson LM, Ye X, Sack JL, Hueston WJ. Glycemic control in medical inpatients with type 2 diabetes mellitus receiving sliding scale insulin regimens versus routine diabetes medications: a multicenter randomized controlled trial. Ann Fam Med. 2003;1(1):29-35. https://doi.org/10.1370/afm.2

Article PDF
Author and Disclosure Information

1Department of Medicine, Emory University, Atlanta, Georgia; 2Rollins School of Public Health, Emory University, Atlanta, Georgia.

Disclosures
Dr Umpierrez is partly supported by research grants from the National Center for Advancing Translational Sciences of the National Institutes of Health (NIH) under Award Number UL1TR002378 from the Clinical and Translational Science Awards Program and an NIH grant U30, P30DK11102, and has received research grant support to Emory University for investigator-initiated studies from Novo Nordisk, AstraZeneca, and Dexcom. Dr Pasquel is partly supported by NIH/NIGMS grant 1K23GM128221-01A1, has received consulting fees from Merck, Boehringer Ingelheim, Eli Lilly and Company, and AstraZeneca, and research support from Merck and Dexcom.

Funding
This study was supported by the Jacobs Family Foundation Fund and by the Emory Endocrinology Division research funds (GEU).

Issue
Journal of Hospital Medicine 16(8)
Publications
Topics
Page Number
462-468. Published Online Only July 21, 2021
Sections
Files
Files
Author and Disclosure Information

1Department of Medicine, Emory University, Atlanta, Georgia; 2Rollins School of Public Health, Emory University, Atlanta, Georgia.

Disclosures
Dr Umpierrez is partly supported by research grants from the National Center for Advancing Translational Sciences of the National Institutes of Health (NIH) under Award Number UL1TR002378 from the Clinical and Translational Science Awards Program and an NIH grant U30, P30DK11102, and has received research grant support to Emory University for investigator-initiated studies from Novo Nordisk, AstraZeneca, and Dexcom. Dr Pasquel is partly supported by NIH/NIGMS grant 1K23GM128221-01A1, has received consulting fees from Merck, Boehringer Ingelheim, Eli Lilly and Company, and AstraZeneca, and research support from Merck and Dexcom.

Funding
This study was supported by the Jacobs Family Foundation Fund and by the Emory Endocrinology Division research funds (GEU).

Author and Disclosure Information

1Department of Medicine, Emory University, Atlanta, Georgia; 2Rollins School of Public Health, Emory University, Atlanta, Georgia.

Disclosures
Dr Umpierrez is partly supported by research grants from the National Center for Advancing Translational Sciences of the National Institutes of Health (NIH) under Award Number UL1TR002378 from the Clinical and Translational Science Awards Program and an NIH grant U30, P30DK11102, and has received research grant support to Emory University for investigator-initiated studies from Novo Nordisk, AstraZeneca, and Dexcom. Dr Pasquel is partly supported by NIH/NIGMS grant 1K23GM128221-01A1, has received consulting fees from Merck, Boehringer Ingelheim, Eli Lilly and Company, and AstraZeneca, and research support from Merck and Dexcom.

Funding
This study was supported by the Jacobs Family Foundation Fund and by the Emory Endocrinology Division research funds (GEU).

Article PDF
Article PDF
Related Articles

Sliding scale insulin (SSI) for inpatient glycemic control was first proposed by Elliott P Joslin in 1934 when he recommended titration of insulin based on urine glucose levels.1 As bedside glucose meters became widely available, physicians transitioned to dosing SSI based on capillary blood glucose (BG) levels,2,3 and SSI became widely used for the management of inpatient hyperglycemia.1 However, during the past decade, there has been strong opposition to the use of SSI in hospitals. Many authors oppose its use, highlighting the retrospective rather than prospective nature of SSI therapy and concerns about inadequate glycemic control.4-6 In 2004, the American College of Endocrinology first released a position statement discouraging the use of SSI alone and recommended basal-bolus insulin as the preferred method of glycemic control for inpatients with type 2 diabetes (T2D).7 The American Diabetes Association (ADA) inpatient guidelines in 20058 and the Endocrine Society guidelines in 20129 also opposed SSI monotherapy and reaffirmed that a basal-bolus insulin regimen should be used for most non–critically ill patients with diabetes. Those guidelines remain in place currently.

Several randomized controlled trials (RCTs) and meta-analyses have shown that basal-bolus insulin regimens provide superior glycemic control in non–critical inpatients when compared with SSI alone.10-14 In addition, the RABBIT 2 (Randomized Study of Basal-Bolus Insulin Therapy in the Inpatient Management of Patients With Type 2 Diabetes) trial showed a significant reduction in perioperative complications10 among surgical patients when treated with basal-bolus insulin therapy. Despite these studies and strong recommendations against its use, SSI continues to be widely used in the United States. According to a 2007 survey of 44 US hospitals, 41% of noncritical patients with hyperglycemia were treated with SSI alone.15 In addition, SSI remains one of the most commonly prescribed insulin regimens in many countries around the world.16-19 The persistence of SSI use raises questions as to why clinicians continue to use a therapy that has been strongly criticized. Some authors point to convenience and fear of hypoglycemia with a basal-bolus insulin regimen.20,21 Alternatively, it is possible that SSI usage remains so pervasive because it is effective in a subset of patients. In fact, a 2018 Cochrane review concluded that existing evidence is not sufficiently robust to definitively recommend basal-bolus insulin over SSI for inpatient diabetes management of non–critically ill patients despite existing guidelines.22

Owing to the ongoing controversy and widespread use of SSI, we designed an exploratory analysis to understand the rationale for such therapy by investigating whether a certain subpopulation of hospitalized patients with T2D may achieve target glycemic control with SSI alone. We hypothesized that noncritical patients with mild hyperglycemia and admission BG <180 mg/dL would do well with SSI alone and may not require intensive treatment with basal-bolus insulin regimens. To address this question, we used electronic health records with individual-level patient data to assess inpatient glycemic control of non–critically ill patients with T2D treated with SSI alone.

METHODS

Participants

Data from 25,813 adult noncritical inpatients with T2D, with an index admission between June 1, 2010, and June 30, 2018, were obtained through the Emory Healthcare Clinical Data Warehouse infrastructure program. All patients were admitted to Emory Healthcare hospitals, including Emory University Hospital, Emory University Hospital Midtown, and Emory Saint Joseph’s Hospital, in Atlanta, Georgia. Data were extracted for each patient during the index hospitalization, including demographics, anthropometrics, and admission and inpatient laboratory values. Information was collected on daily point-of-care glucose values, hemoglobin A1c (HbA1c), hypoglycemic events, insulin doses, hospital complications, comorbidities, and hospital setting (medical vs surgical admission). International Classification of Diseases, 9th and 10th Revisions (ICD-9/10) codes were used to determine diagnosis of T2D, comorbidities, and complications.

From our initial dataset, we identified 16,366 patients who were treated with SSI during hospitalization. We excluded patients who were admitted to the intensive care unit (ICU) or placed on intravenous insulin, patients with missing admission BG values, and patients with a length of stay less than 1 day. To prevent inclusion of patients presenting in diabetic ketoacidosis or hyperosmolar hyperglycemic syndrome, we excluded patients with an admission BG >500 mg/dL. We then excluded 6,739 patients who received basal insulin within the first 2 days of hospitalization, as well as 943 patients who were treated with noninsulin (oral or injectable) antidiabetic agents. Our final dataset included 8,095 patients (Appendix Figure).

Patients in the SSI cohort included all patients who were treated with short-acting insulin only (regular insulin or rapid-acting [lispro, aspart, glulisine] insulin analogs) during the first 2 days of hospitalization. Patients who remained on only short-acting insulin during the entire hospitalization were defined as continuous SSI patients. Patients who subsequently received basal insulin after day 2 of hospitalization were defined as patients who transitioned to basal. Patients were stratified according to admission BG levels (first BG available on day of admission) and HbA1c (when available during index admission). We compared the baseline characteristics and clinical outcomes of patients who remained on SSI alone throughout the entirety of hospitalization with those of patients who required transition to basal insulin. The mean hospital BG was calculated by taking the average of all BG measurements during the hospital stay. We defined hypoglycemia as a BG <70 mg/dL and severe hypoglycemia as BG <40 mg/dL. Repeated hypoglycemia values were excluded if they occurred within a period of 2 hours.

Outcome Measures

The primary outcome was the percentage of patients with T2D achieving target glycemic control with SSI therapy, defined as mean hospital BG between 70 and 180 mg/dL without hypoglycemia <70 mg/dL during hospital stay. This threshold was determined based on 2019 ADA recommendations targeting hospital BG <180 mg/dL and avoidance of hypoglycemia.23

Statistical Analysis

Patients were stratified according to continuous SSI versus transitioned to basal treatment. Patients who remained on continuous SSI were further categorized into four categories based on admission BG: <140 mg/dL, 140 to 180 mg/dL, 180 to 250 mg/dL, and ≥250 mg/dL. Clinical characteristics were compared using Wilcoxon rank-sum tests (if continuous) and chi-square tests or Fisher exact tests (if categorical). We then compared the clinical outcomes among continuous SSI patients with different admission BG levels (<140 mg/dL, 140-180 mg/dL, 180-250 mg/dL, and ≥250 mg/dL) and with different HbA1c levels (<7%, 7%-8%, 8%-9%, ≥9%). Within each scenario, logistic regression for the outcome of poor glycemic control, defined as mean hospital BG >180 mg/dL, was performed to evaluate the HbA1c levels and admission BG levels controlling for other factors (age, gender, body mass index [BMI], race, setting [medicine versus surgery] and Charlson Comorbidity Index score). A P value < .05 was regarded as statistically significant. All analyses were performed based on available cases and conducted in SAS version 9.4 (SAS Institute Inc.).

RESULTS

Among 25,813 adult patients with T2D, 8,095 patients (31.4%) were treated with SSI alone during the first 2 days of hospitalization. Of those patients treated with SSI, 6,903 (85%) remained on continuous SSI alone during the entire hospitalization, and 1,192 (15%) were transitioned to basal insulin. The clinical characteristics of these patients on continuous SSI and those who transitioned to basal insulin are shown in Table 1. Patients who transitioned to basal insulin had significantly higher mean (SD) admission BG (191.8 [88.2] mg/dL vs 156.4 [65.4] mg/dL, P < .001) and higher mean (SD) HbA1c (8.1% [2.0%] vs 7.01% [1.5%], P < .001), compared with those who remained on continuous SSI. Patients who transitioned to basal insulin were also younger and more likely to have chronic kidney disease (CKD), but less likely to have congestive heart failure, coronary artery disease, or chronic obstructive pulmonary disease (COPD). The Charlson Comorbidity Index score was significantly higher for patients who transitioned to basal (4.4 [2.5]) than for those who remained on continuous SSI (4.1 [2.5], P < .001). There were no significant differences among sex, BMI, or glomerular filtration rate (GFR) on admission. Of those transitioned to basal insulin, 53% achieved a mean hospitalization BG <180 mg/dL, compared with 82% of those on continuous SSI. The overall rate of hypoglycemia in the continuous SSI group was 8% compared with 18% in those transitioned to basal insulin.

Clinical Characteristics of Hospitalized Patients With Type 2 Diabetes Treated With SSI by Continuous SSI vs Transitioned to Basal Insulin

Of the patients who remained on continuous SSI throughout the hospitalization, 3,319 patients (48%) had admission BG <140 mg/dL, 1,671 patients (24%) had admission BG 140 to 180 mg/dL, and 1,913 patients (28%) had admission BG >180 mg/dL. Only 9% of patients who remained on continuous SSI had admission BG ≥250 mg/dL. Patients with admission BG <140 mg/dL were older, had lower BMI and HbA1c, had higher rates of COPD and CKD, and were more likely to be admitted to a surgical service compared with patients with admission BG >140 mg/dL (P < .05 for all; Table 2).

Clinical Characteristics of Patients Who Remained on Continuous SSI by Admission Blood Glucose Concentration

Hospital glycemic control for patients on continuous SSI according to admission BG is displayed in Table 3. Among patients who remained on continuous SSI, 96% of patients with admission BG <140 mg/dL had a mean hospital BG <180 mg/dL; of them, 86% achieved target control without hypoglycemia. Similar rates of target control were achieved in patients with admission BG 140 to 180 mg/dL (83%), in contrast to patients with admission BG ≥250 mg/dL, of whom only 18% achieved target control (P < .001). These findings parallel those seen in patients transitioned to basal insulin. Of patients in the transition group admitted with BG <140 mg/dL and <180 mg/dL, 88.5% and 84.6% had mean hospital BG <180 mg/dL, respectively, while 69.1% and 68.9% had mean BG between 70 and 180 mg/dL without hypoglycemia. The overall frequency of hypoglycemia <70 mg/dL among patients on continuous SSI was 8% and was more common in patients with admission BG <140 mg/dL (10%) compared with patients with higher admission glucose levels (BG 140-180 mg/dL [4%], 180-250 mg/dL [4%], or ≥250 mg/dL [6%], P < .001). There was no difference in rates of severe hypoglycemia <40 mg/dL among groups.

Glycemic Data of Patients on Continuous SSI by Admission Blood Glucose Concentration and Admission HbA1c

HbA1c data were available for 2,560 of the patients on continuous SSI (Table 3). Mean hospital BG increased significantly with increasing HbA1c values. Patients admitted with HbA1c <7% had lower mean (SD) hospital BG (132.2 [28.2] mg/dL) and were more likely to achieve target glucose control during hospitalization (85%) compared with those with HbA1c 7% to 8% (mean BG, 148.7 [30.8] mg/dL; 80% target control), HbA1c 8% to 9% (mean BG, 169.1 [37.9] mg/dL; 61% target control), or HbA1c ≥9% (mean BG, 194.9 [53.4] mg/dL; 38% target control) (P < .001).

In a logistic regression analysis adjusted for age, gender, BMI, race, setting (medicine vs surgery), and Charlson Comorbidity Index score, the odds of poor glycemic control increased with higher admission BG (admission BG 140-180 mg/dL: odds ratio [OR], 1.8; 95% CI, 1.5-2.2; admission BG 180-250 mg/dL: OR, 3.7; 95% CI, 3.1-4.4; admission BG ≥250 mg/dL: OR, 7.2; 95% CI, 5.8-9.0; reference admission BG <140 mg/dL; Figure). Similarly, the logistic regression analysis showed greater odds of poor in-hospital glycemic control with increasing HbA1c (OR, 6.1; 95% CI, 4.3-8.8 for HbA1c >9% compared with HbA1c <7%).

Results of Logistic Regression Analysis

DISCUSSION

This large retrospective cohort study examined the effectiveness of SSI for glycemic control in noncritical inpatients with T2D. Our results indicate that SSI is still widely used in our hospital system, with 31.4% of our initial cohort managed with SSI alone. We found that 86% of patients with BG <140 mg/dL and 83% of patients with BG 140 to 180 mg/dL achieved glycemic control without hypoglycemia when managed with SSI alone, compared with 53% of those admitted with BG 180 to 250 mg/dL and only 18% of those with admission BG ≥250 mg/dL. This high success rate of achieving optimal BG control with SSI alone is comparable to that seen with transition to basal insulin and may explain the prevalent use of SSI for the management of patients with T2D and mild to moderate hyperglycemia.

Published clinical guideline recommendations promoting the use of basal-bolus insulin treatment algorithms are based on the results of a few RCTs that compared the efficacy of SSI vs a basal-bolus insulin regimen. These studies reported significantly lower mean daily BG concentration with basal or basal-bolus insulin therapy compared with SSI.10,11,24 However, it is interesting to note that the mean admission BG of patients treated with SSI in these RCTs ranged from 184 to 225 mg/dL. Patients in these trials were excluded if admission BG was <140 mg/dL.10,11,24 This is in contrast to our study evaluating real-world data in non–critically ill settings in which we found that 48% of patients treated with SSI had admission BG <140 mg/dL, and nearly 75% had admission BG <180 mg/dL. This suggests that by nature of study design, most RCTs excluded the population of patients who do achieve good glycemic control with SSI and may have contributed to the perception that basal insulin is preferable in all populations.

Our analysis indicates that healthcare professionals should consider admission BG when selecting the type of insulin regimen to manage patients with T2D in the hospital. Our results suggest that SSI may be appropriate for many patients with admission BG <180 mg/dL and should be avoided as monotherapy in patients with admission BG ≥180 mg/dL, as the proportion of patients achieving target control decreased with increasing admission BG. More importantly, if a patient is not controlled with SSI alone, intensification of therapy with the addition of basal insulin is indicated to achieve glycemic control. In addition, we found that the admission HbA1c is an appropriate marker to consider as well, with hospital glycemic control deteriorating with increasing HbA1c values, paralleling the admission BG. The main limitation to widespread use of HbA1c for therapeutic decision-making is access to values at time of patient admission; in our population, only 37% of patients had an HbA1c value available during the index hospitalization.

Previous publications have reported that hypoglycemia carries significant safety concerns, especially among a hospitalized population.25-27 As such, we included hypoglycemia as an important metric in our definition of target glycemic control rather than simply using mean hospital BG or number of hyperglycemic events to define treatment effectiveness. We did find a higher rate of hypoglycemia in patients with moderate admission BG treated with SSI compared with those with higher admission BG; however, few patients overall experienced clinically significant (<54 mg/dL) or severe (<40 mg/dL) hypoglycemia.

In our population, only 15% of patients started on SSI received additional basal insulin during hospitalization. This finding is similar to data reported in the Rabbit 2 trial, in which 14% of patients failed SSI alone, with a higher failure rate among those with higher BG on admission.10 Given the observational nature of this study, we cannot definitively state why certain patients in our population required additional basal insulin, but we can hypothesize that these patients admitted with BG ≥180 mg/dL had higher treatment failure rates and greater rates of hyperglycemia, therefore receiving intensified insulin therapy as clinically indicated at the discretion of the treating physician. Patients who transitioned from SSI to basal insulin had significantly higher admission BG and HbA1c compared with patients who remained on SSI alone. We noted that the rates of hypoglycemia were higher in the group that transitioned to basal (18% vs 8%) and similar to rates reported in previous RCTs.11,24

This observational study takes advantage of a large, diverse study population and a combination of medicine and surgery patients in a real-world setting. We acknowledge several limitations in our study. Our primary data were observational in nature, and as such, some baseline patient characteristics were notably different between groups, suggesting selection bias for treatment allocation to SSI. We do not know which patients were managed by primary teams compared with specialized diabetes consult services, which may also influence treatment regimens. We did not have access to information about patients’ at-home diabetes medication regimens or duration of diabetes, both of which have been shown in prior publications to affect an individual’s overall hospital glycemic control. Data on HbA1c values were available for only approximately one-third of patients. In addition, our study did not include patients without a history of diabetes who developed stress-induced hyperglycemia, a population that may benefit from conservative therapy such as SSI.28 A diagnosis of CKD was defined based on ICD 9/10 codes and not on admission estimated GFR. More specific data regarding stage of CKD or changes in renal function over the duration of hospitalization are not available, which could influence insulin prescribing practice. In addition, we defined the basal group as patients prescribed any form of basal insulin (NPH, glargine, detemir or degludec), and we do not have information on the use of prandial versus correction doses of rapid-acting insulin in the basal insulin–treated group.

CONCLUSION

In conclusion, our observational study indicates that the use of SSI results in appropriate target glycemic control for most noncritical medicine and surgery patients with admission BG <180 mg/dL. In agreement with previous RCTs, our study confirms that SSI as monotherapy is frequently inadequate in patients with significant hyperglycemia >180 mg/dL.10,11,24,29 We propose that an individualized approach to inpatient glycemic management is imperative, and cautious use of SSI may be a viable option for certain patients with mild hyperglycemia and admission BG <180 mg/dL. Further observational and randomized studies are needed to confirm the efficacy of SSI therapy in T2D patients with mild hyperglycemia. By identifying which subset of patients can be safely managed with SSI alone, we can better understand which patients will require escalation of therapy with intensive glucose management.

Sliding scale insulin (SSI) for inpatient glycemic control was first proposed by Elliott P Joslin in 1934 when he recommended titration of insulin based on urine glucose levels.1 As bedside glucose meters became widely available, physicians transitioned to dosing SSI based on capillary blood glucose (BG) levels,2,3 and SSI became widely used for the management of inpatient hyperglycemia.1 However, during the past decade, there has been strong opposition to the use of SSI in hospitals. Many authors oppose its use, highlighting the retrospective rather than prospective nature of SSI therapy and concerns about inadequate glycemic control.4-6 In 2004, the American College of Endocrinology first released a position statement discouraging the use of SSI alone and recommended basal-bolus insulin as the preferred method of glycemic control for inpatients with type 2 diabetes (T2D).7 The American Diabetes Association (ADA) inpatient guidelines in 20058 and the Endocrine Society guidelines in 20129 also opposed SSI monotherapy and reaffirmed that a basal-bolus insulin regimen should be used for most non–critically ill patients with diabetes. Those guidelines remain in place currently.

Several randomized controlled trials (RCTs) and meta-analyses have shown that basal-bolus insulin regimens provide superior glycemic control in non–critical inpatients when compared with SSI alone.10-14 In addition, the RABBIT 2 (Randomized Study of Basal-Bolus Insulin Therapy in the Inpatient Management of Patients With Type 2 Diabetes) trial showed a significant reduction in perioperative complications10 among surgical patients when treated with basal-bolus insulin therapy. Despite these studies and strong recommendations against its use, SSI continues to be widely used in the United States. According to a 2007 survey of 44 US hospitals, 41% of noncritical patients with hyperglycemia were treated with SSI alone.15 In addition, SSI remains one of the most commonly prescribed insulin regimens in many countries around the world.16-19 The persistence of SSI use raises questions as to why clinicians continue to use a therapy that has been strongly criticized. Some authors point to convenience and fear of hypoglycemia with a basal-bolus insulin regimen.20,21 Alternatively, it is possible that SSI usage remains so pervasive because it is effective in a subset of patients. In fact, a 2018 Cochrane review concluded that existing evidence is not sufficiently robust to definitively recommend basal-bolus insulin over SSI for inpatient diabetes management of non–critically ill patients despite existing guidelines.22

Owing to the ongoing controversy and widespread use of SSI, we designed an exploratory analysis to understand the rationale for such therapy by investigating whether a certain subpopulation of hospitalized patients with T2D may achieve target glycemic control with SSI alone. We hypothesized that noncritical patients with mild hyperglycemia and admission BG <180 mg/dL would do well with SSI alone and may not require intensive treatment with basal-bolus insulin regimens. To address this question, we used electronic health records with individual-level patient data to assess inpatient glycemic control of non–critically ill patients with T2D treated with SSI alone.

METHODS

Participants

Data from 25,813 adult noncritical inpatients with T2D, with an index admission between June 1, 2010, and June 30, 2018, were obtained through the Emory Healthcare Clinical Data Warehouse infrastructure program. All patients were admitted to Emory Healthcare hospitals, including Emory University Hospital, Emory University Hospital Midtown, and Emory Saint Joseph’s Hospital, in Atlanta, Georgia. Data were extracted for each patient during the index hospitalization, including demographics, anthropometrics, and admission and inpatient laboratory values. Information was collected on daily point-of-care glucose values, hemoglobin A1c (HbA1c), hypoglycemic events, insulin doses, hospital complications, comorbidities, and hospital setting (medical vs surgical admission). International Classification of Diseases, 9th and 10th Revisions (ICD-9/10) codes were used to determine diagnosis of T2D, comorbidities, and complications.

From our initial dataset, we identified 16,366 patients who were treated with SSI during hospitalization. We excluded patients who were admitted to the intensive care unit (ICU) or placed on intravenous insulin, patients with missing admission BG values, and patients with a length of stay less than 1 day. To prevent inclusion of patients presenting in diabetic ketoacidosis or hyperosmolar hyperglycemic syndrome, we excluded patients with an admission BG >500 mg/dL. We then excluded 6,739 patients who received basal insulin within the first 2 days of hospitalization, as well as 943 patients who were treated with noninsulin (oral or injectable) antidiabetic agents. Our final dataset included 8,095 patients (Appendix Figure).

Patients in the SSI cohort included all patients who were treated with short-acting insulin only (regular insulin or rapid-acting [lispro, aspart, glulisine] insulin analogs) during the first 2 days of hospitalization. Patients who remained on only short-acting insulin during the entire hospitalization were defined as continuous SSI patients. Patients who subsequently received basal insulin after day 2 of hospitalization were defined as patients who transitioned to basal. Patients were stratified according to admission BG levels (first BG available on day of admission) and HbA1c (when available during index admission). We compared the baseline characteristics and clinical outcomes of patients who remained on SSI alone throughout the entirety of hospitalization with those of patients who required transition to basal insulin. The mean hospital BG was calculated by taking the average of all BG measurements during the hospital stay. We defined hypoglycemia as a BG <70 mg/dL and severe hypoglycemia as BG <40 mg/dL. Repeated hypoglycemia values were excluded if they occurred within a period of 2 hours.

Outcome Measures

The primary outcome was the percentage of patients with T2D achieving target glycemic control with SSI therapy, defined as mean hospital BG between 70 and 180 mg/dL without hypoglycemia <70 mg/dL during hospital stay. This threshold was determined based on 2019 ADA recommendations targeting hospital BG <180 mg/dL and avoidance of hypoglycemia.23

Statistical Analysis

Patients were stratified according to continuous SSI versus transitioned to basal treatment. Patients who remained on continuous SSI were further categorized into four categories based on admission BG: <140 mg/dL, 140 to 180 mg/dL, 180 to 250 mg/dL, and ≥250 mg/dL. Clinical characteristics were compared using Wilcoxon rank-sum tests (if continuous) and chi-square tests or Fisher exact tests (if categorical). We then compared the clinical outcomes among continuous SSI patients with different admission BG levels (<140 mg/dL, 140-180 mg/dL, 180-250 mg/dL, and ≥250 mg/dL) and with different HbA1c levels (<7%, 7%-8%, 8%-9%, ≥9%). Within each scenario, logistic regression for the outcome of poor glycemic control, defined as mean hospital BG >180 mg/dL, was performed to evaluate the HbA1c levels and admission BG levels controlling for other factors (age, gender, body mass index [BMI], race, setting [medicine versus surgery] and Charlson Comorbidity Index score). A P value < .05 was regarded as statistically significant. All analyses were performed based on available cases and conducted in SAS version 9.4 (SAS Institute Inc.).

RESULTS

Among 25,813 adult patients with T2D, 8,095 patients (31.4%) were treated with SSI alone during the first 2 days of hospitalization. Of those patients treated with SSI, 6,903 (85%) remained on continuous SSI alone during the entire hospitalization, and 1,192 (15%) were transitioned to basal insulin. The clinical characteristics of these patients on continuous SSI and those who transitioned to basal insulin are shown in Table 1. Patients who transitioned to basal insulin had significantly higher mean (SD) admission BG (191.8 [88.2] mg/dL vs 156.4 [65.4] mg/dL, P < .001) and higher mean (SD) HbA1c (8.1% [2.0%] vs 7.01% [1.5%], P < .001), compared with those who remained on continuous SSI. Patients who transitioned to basal insulin were also younger and more likely to have chronic kidney disease (CKD), but less likely to have congestive heart failure, coronary artery disease, or chronic obstructive pulmonary disease (COPD). The Charlson Comorbidity Index score was significantly higher for patients who transitioned to basal (4.4 [2.5]) than for those who remained on continuous SSI (4.1 [2.5], P < .001). There were no significant differences among sex, BMI, or glomerular filtration rate (GFR) on admission. Of those transitioned to basal insulin, 53% achieved a mean hospitalization BG <180 mg/dL, compared with 82% of those on continuous SSI. The overall rate of hypoglycemia in the continuous SSI group was 8% compared with 18% in those transitioned to basal insulin.

Clinical Characteristics of Hospitalized Patients With Type 2 Diabetes Treated With SSI by Continuous SSI vs Transitioned to Basal Insulin

Of the patients who remained on continuous SSI throughout the hospitalization, 3,319 patients (48%) had admission BG <140 mg/dL, 1,671 patients (24%) had admission BG 140 to 180 mg/dL, and 1,913 patients (28%) had admission BG >180 mg/dL. Only 9% of patients who remained on continuous SSI had admission BG ≥250 mg/dL. Patients with admission BG <140 mg/dL were older, had lower BMI and HbA1c, had higher rates of COPD and CKD, and were more likely to be admitted to a surgical service compared with patients with admission BG >140 mg/dL (P < .05 for all; Table 2).

Clinical Characteristics of Patients Who Remained on Continuous SSI by Admission Blood Glucose Concentration

Hospital glycemic control for patients on continuous SSI according to admission BG is displayed in Table 3. Among patients who remained on continuous SSI, 96% of patients with admission BG <140 mg/dL had a mean hospital BG <180 mg/dL; of them, 86% achieved target control without hypoglycemia. Similar rates of target control were achieved in patients with admission BG 140 to 180 mg/dL (83%), in contrast to patients with admission BG ≥250 mg/dL, of whom only 18% achieved target control (P < .001). These findings parallel those seen in patients transitioned to basal insulin. Of patients in the transition group admitted with BG <140 mg/dL and <180 mg/dL, 88.5% and 84.6% had mean hospital BG <180 mg/dL, respectively, while 69.1% and 68.9% had mean BG between 70 and 180 mg/dL without hypoglycemia. The overall frequency of hypoglycemia <70 mg/dL among patients on continuous SSI was 8% and was more common in patients with admission BG <140 mg/dL (10%) compared with patients with higher admission glucose levels (BG 140-180 mg/dL [4%], 180-250 mg/dL [4%], or ≥250 mg/dL [6%], P < .001). There was no difference in rates of severe hypoglycemia <40 mg/dL among groups.

Glycemic Data of Patients on Continuous SSI by Admission Blood Glucose Concentration and Admission HbA1c

HbA1c data were available for 2,560 of the patients on continuous SSI (Table 3). Mean hospital BG increased significantly with increasing HbA1c values. Patients admitted with HbA1c <7% had lower mean (SD) hospital BG (132.2 [28.2] mg/dL) and were more likely to achieve target glucose control during hospitalization (85%) compared with those with HbA1c 7% to 8% (mean BG, 148.7 [30.8] mg/dL; 80% target control), HbA1c 8% to 9% (mean BG, 169.1 [37.9] mg/dL; 61% target control), or HbA1c ≥9% (mean BG, 194.9 [53.4] mg/dL; 38% target control) (P < .001).

In a logistic regression analysis adjusted for age, gender, BMI, race, setting (medicine vs surgery), and Charlson Comorbidity Index score, the odds of poor glycemic control increased with higher admission BG (admission BG 140-180 mg/dL: odds ratio [OR], 1.8; 95% CI, 1.5-2.2; admission BG 180-250 mg/dL: OR, 3.7; 95% CI, 3.1-4.4; admission BG ≥250 mg/dL: OR, 7.2; 95% CI, 5.8-9.0; reference admission BG <140 mg/dL; Figure). Similarly, the logistic regression analysis showed greater odds of poor in-hospital glycemic control with increasing HbA1c (OR, 6.1; 95% CI, 4.3-8.8 for HbA1c >9% compared with HbA1c <7%).

Results of Logistic Regression Analysis

DISCUSSION

This large retrospective cohort study examined the effectiveness of SSI for glycemic control in noncritical inpatients with T2D. Our results indicate that SSI is still widely used in our hospital system, with 31.4% of our initial cohort managed with SSI alone. We found that 86% of patients with BG <140 mg/dL and 83% of patients with BG 140 to 180 mg/dL achieved glycemic control without hypoglycemia when managed with SSI alone, compared with 53% of those admitted with BG 180 to 250 mg/dL and only 18% of those with admission BG ≥250 mg/dL. This high success rate of achieving optimal BG control with SSI alone is comparable to that seen with transition to basal insulin and may explain the prevalent use of SSI for the management of patients with T2D and mild to moderate hyperglycemia.

Published clinical guideline recommendations promoting the use of basal-bolus insulin treatment algorithms are based on the results of a few RCTs that compared the efficacy of SSI vs a basal-bolus insulin regimen. These studies reported significantly lower mean daily BG concentration with basal or basal-bolus insulin therapy compared with SSI.10,11,24 However, it is interesting to note that the mean admission BG of patients treated with SSI in these RCTs ranged from 184 to 225 mg/dL. Patients in these trials were excluded if admission BG was <140 mg/dL.10,11,24 This is in contrast to our study evaluating real-world data in non–critically ill settings in which we found that 48% of patients treated with SSI had admission BG <140 mg/dL, and nearly 75% had admission BG <180 mg/dL. This suggests that by nature of study design, most RCTs excluded the population of patients who do achieve good glycemic control with SSI and may have contributed to the perception that basal insulin is preferable in all populations.

Our analysis indicates that healthcare professionals should consider admission BG when selecting the type of insulin regimen to manage patients with T2D in the hospital. Our results suggest that SSI may be appropriate for many patients with admission BG <180 mg/dL and should be avoided as monotherapy in patients with admission BG ≥180 mg/dL, as the proportion of patients achieving target control decreased with increasing admission BG. More importantly, if a patient is not controlled with SSI alone, intensification of therapy with the addition of basal insulin is indicated to achieve glycemic control. In addition, we found that the admission HbA1c is an appropriate marker to consider as well, with hospital glycemic control deteriorating with increasing HbA1c values, paralleling the admission BG. The main limitation to widespread use of HbA1c for therapeutic decision-making is access to values at time of patient admission; in our population, only 37% of patients had an HbA1c value available during the index hospitalization.

Previous publications have reported that hypoglycemia carries significant safety concerns, especially among a hospitalized population.25-27 As such, we included hypoglycemia as an important metric in our definition of target glycemic control rather than simply using mean hospital BG or number of hyperglycemic events to define treatment effectiveness. We did find a higher rate of hypoglycemia in patients with moderate admission BG treated with SSI compared with those with higher admission BG; however, few patients overall experienced clinically significant (<54 mg/dL) or severe (<40 mg/dL) hypoglycemia.

In our population, only 15% of patients started on SSI received additional basal insulin during hospitalization. This finding is similar to data reported in the Rabbit 2 trial, in which 14% of patients failed SSI alone, with a higher failure rate among those with higher BG on admission.10 Given the observational nature of this study, we cannot definitively state why certain patients in our population required additional basal insulin, but we can hypothesize that these patients admitted with BG ≥180 mg/dL had higher treatment failure rates and greater rates of hyperglycemia, therefore receiving intensified insulin therapy as clinically indicated at the discretion of the treating physician. Patients who transitioned from SSI to basal insulin had significantly higher admission BG and HbA1c compared with patients who remained on SSI alone. We noted that the rates of hypoglycemia were higher in the group that transitioned to basal (18% vs 8%) and similar to rates reported in previous RCTs.11,24

This observational study takes advantage of a large, diverse study population and a combination of medicine and surgery patients in a real-world setting. We acknowledge several limitations in our study. Our primary data were observational in nature, and as such, some baseline patient characteristics were notably different between groups, suggesting selection bias for treatment allocation to SSI. We do not know which patients were managed by primary teams compared with specialized diabetes consult services, which may also influence treatment regimens. We did not have access to information about patients’ at-home diabetes medication regimens or duration of diabetes, both of which have been shown in prior publications to affect an individual’s overall hospital glycemic control. Data on HbA1c values were available for only approximately one-third of patients. In addition, our study did not include patients without a history of diabetes who developed stress-induced hyperglycemia, a population that may benefit from conservative therapy such as SSI.28 A diagnosis of CKD was defined based on ICD 9/10 codes and not on admission estimated GFR. More specific data regarding stage of CKD or changes in renal function over the duration of hospitalization are not available, which could influence insulin prescribing practice. In addition, we defined the basal group as patients prescribed any form of basal insulin (NPH, glargine, detemir or degludec), and we do not have information on the use of prandial versus correction doses of rapid-acting insulin in the basal insulin–treated group.

CONCLUSION

In conclusion, our observational study indicates that the use of SSI results in appropriate target glycemic control for most noncritical medicine and surgery patients with admission BG <180 mg/dL. In agreement with previous RCTs, our study confirms that SSI as monotherapy is frequently inadequate in patients with significant hyperglycemia >180 mg/dL.10,11,24,29 We propose that an individualized approach to inpatient glycemic management is imperative, and cautious use of SSI may be a viable option for certain patients with mild hyperglycemia and admission BG <180 mg/dL. Further observational and randomized studies are needed to confirm the efficacy of SSI therapy in T2D patients with mild hyperglycemia. By identifying which subset of patients can be safely managed with SSI alone, we can better understand which patients will require escalation of therapy with intensive glucose management.

References

1. Umpierrez GE, Palacio A, Smiley D. Sliding scale insulin use: myth or insanity? Am J Med. 2007;120(7):563-567. https://doi.org/10.1016/j.amjmed.2006.05.070
2. Kitabchi AE, Ayyagari V, Guerra SM. The efficacy of low-dose versus conventional therapy of insulin for treatment of diabetic ketoacidosis. Ann Intern Med. 1976;84(6):633-638. https://doi.org/10.7326/0003-4819-84-6-633
3. Skyler JS, Skyler DL, Seigler DE, O’Sullivan MJ. Algorithms for adjustment of insulin dosage by patients who monitor blood glucose. Diabetes Care. 1981;4(2):311-318. https://doi.org/10.2337/diacare.4.2.311
4. Gearhart JG, Duncan JL 3rd, Replogle WH, Forbes RC, Walley EJ. Efficacy of sliding-scale insulin therapy: a comparison with prospective regimens. Fam Pract Res J. 1994;14(4):313-322.
5. Queale WS, Seidler AJ, Brancati FL. Glycemic control and sliding scale insulin use in medical inpatients with diabetes mellitus. Arch Intern Med. 1997;157(5):545-552.
6. Clement S, Braithwaite SS, Magee MF, et al. Management of diabetes and hyperglycemia in hospitals. Diabetes Care. 2004;27(2):553-591. https://doi.org/10.2337/diacare.27.2.553
7. Garber AJ, Moghissi ES, Bransome ED Jr, et al. American College of Endocrinology position statement on inpatient diabetes and metabolic control. Endocr Pract. 2004;10(1):78-82. https://doi.org/10.4158/EP.10.1.77
8. American Diabetes Association. Standards of medical care in diabetes. Diabetes Care. 2005;28(suppl 1):S4-S36.
9. Umpierrez GE, Hellman R, Korytkowski MT, , et al. Management of hyperglycemia in hospitalized patients in non-critical care setting: an Endocrine Society clinical practice guideline. J Clin Endocrinol Metab. 2012;97(1):16-38. https://doi.org/10.1210/jc.2011-2098
10. Umpierrez GE, Smiley D, Zisman A, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes. Diabetes Care. 2007;30(9):2181-2186. https://doi.org/10.2337/dc07-0295
11. Umpierrez GE, Smiley D, Jacobs S, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes undergoing general surgery (RABBIT 2 surgery). Diabetes Care. 2011;34(2):256-261. https://doi.org/10.2337/dc10-1407
12. Schroeder JE, Liebergall M, Raz I, Egleston R, Ben Sussan G, Peyser A. Benefits of a simple glycaemic protocol in an orthopaedic surgery ward: a randomized prospective study. Diabetes Metab Res Rev. 2012;28:71-75. https://doi.org/10.1002/dmrr.1217
13. Lee YY, Lin YM, Leu WJ, et al. Sliding-scale insulin used for blood glucose control: a meta-analysis of randomized controlled trials. Metabolism. 2015;64(9):1183-1192. https://doi.org/10.1016/j.metabol.2015.05.011
14. Christensen MB, Gotfredsen A, Nørgaard K. Efficacy of basal-bolus insulin regimens in the inpatient management of non-critically ill patients with type 2 diabetes: a systematic review and meta-analysis. Diabetes Metab Res Rev. 2017;33(5):e2885. https://doi.org/10.1002/dmrr.2885
15. Wexler DJ, Meigs JB, Cagliero E, Nathan DM, Grant RW. Prevalence of hyper- and hypoglycemia among inpatients with diabetes: a national survey of 44 U.S. hospitals. Diabetes Care. 2007;30(2):367-369. https://doi.org/10.2337/dc06-1715
16. Moreira ED Jr, Silveira PCB, Neves RCS, Souza C Jr, Nunes ZO, Almeida MdCC. Glycemic control and diabetes management in hospitalized patients in Brazil. Diabetol Metab Syndr. 2013;5(1):62. https://doi.org/10.1186/1758-5996-5-62
17. Akhtar ST, Mahmood K, Naqvi IH, Vaswani AS. Inpatient management of type 2 diabetes mellitus: does choice of insulin regimen really matter? Pakistan J Med Sci. 2014;30(4):895-898.
18. Gómez Cuervo C, Sánchez Morla A, Pérez-Jacoiste Asín MA, Bisbal Pardo O, Pérez Ordoño L, Vila Santos J. Effective adverse event reduction with bolus-basal versus sliding scale insulin therapy in patients with diabetes during conventional hospitalization: systematic review and meta-analysis. Endocrinol Nutr. 2016;63(4):145-156. https://doi.org/10.1016/j.endonu.2015.11.008
19. Bain A, Hasan SS, Babar ZUD. Interventions to improve insulin prescribing practice for people with diabetes in hospital: a systematic review. Diabet Med. 2019;36(8):948-960. https://doi.org/10.1111/dme.13982
20. Ambrus DB, O’Connor MJ. Things We Do For No Reason: sliding-scale insulin as monotherapy for glycemic control in hospitalized patients. J Hosp Med. 2019;14(2):114-116. https://doi.org/10.12788/jhm.3109
21. Nau KC, Lorenzetti RC, Cucuzzella M, Devine T, Kline J. Glycemic control in hospitalized patients not in intensive care: beyond sliding-scale insulin. Am Fam Physician. 2010;81(9):1130-1135.
22. Colunga-Lozano LE, Gonzalez Torres FJ, Delgado-Figueroa N, et al. Sliding scale insulin for non-critically ill hospitalised adults with diabetes mellitus. Cochrane Database Syst Rev. 2018;11(11):CD011296. https://doi.org/10.1002/14651858.CD011296.pub2
23. American Diabetes Association. Diabetes care in the hospital: Standards of Medical Care in Diabetes—2019. Diabetes Care. 2019;42(suppl 1):S173-S181. https://doi.org/10.2337/dc19-S015
24. Umpierrez GE, Smiley D, Hermayer K, et al. Randomized study comparing a basal-bolus with a basal plus correction management of medical and surgical patients with type 2 diabetes: basal plus trial. Diabetes Care. 2013;36(8):2169-2174. https://doi.org/10.2337/dc12-1988
25. Turchin A, Matheny ME, Shubina M, Scanlon SV, Greenwood B, Pendergrass ML. Hypoglycemia and clinical outcomes in patients with diabetes hospitalized in the general ward. Diabetes Care. 2009;32(7):1153-1157. https://doi.org/10.2337/dc08-2127
26. Garg R, Hurwitz S, Turchin A, Trivedi A. Hypoglycemia, with or without insulin therapy, is associated with increased mortality among hospitalized patients. Diabetes Care. 2013;36(5):1107-1110. https://doi.org/10.2337/dc12-1296
27. Zapatero A, Gómez-Huelgas R, González N, et al. Frequency of hypoglycemia and its impact on length of stay, mortality, and short-term readmission in patients with diabetes hospitalized in internal medicine wards. Endocr Pract. 2014;20(9):870-875. https://doi.org/10.4158/EP14006.OR
28. Umpierrez GE, Isaacs SD, Bazargan N, You X, Thaler LM, Kitabchi AE. Hyperglycemia: an independent marker of in-hospital mortality in patients with undiagnosed diabetes. J Clin Endocrinol Metab. 2002;87(3):978-982. https://doi.org/10.1210/jcem.87.3.8341
29. Dickerson LM, Ye X, Sack JL, Hueston WJ. Glycemic control in medical inpatients with type 2 diabetes mellitus receiving sliding scale insulin regimens versus routine diabetes medications: a multicenter randomized controlled trial. Ann Fam Med. 2003;1(1):29-35. https://doi.org/10.1370/afm.2

References

1. Umpierrez GE, Palacio A, Smiley D. Sliding scale insulin use: myth or insanity? Am J Med. 2007;120(7):563-567. https://doi.org/10.1016/j.amjmed.2006.05.070
2. Kitabchi AE, Ayyagari V, Guerra SM. The efficacy of low-dose versus conventional therapy of insulin for treatment of diabetic ketoacidosis. Ann Intern Med. 1976;84(6):633-638. https://doi.org/10.7326/0003-4819-84-6-633
3. Skyler JS, Skyler DL, Seigler DE, O’Sullivan MJ. Algorithms for adjustment of insulin dosage by patients who monitor blood glucose. Diabetes Care. 1981;4(2):311-318. https://doi.org/10.2337/diacare.4.2.311
4. Gearhart JG, Duncan JL 3rd, Replogle WH, Forbes RC, Walley EJ. Efficacy of sliding-scale insulin therapy: a comparison with prospective regimens. Fam Pract Res J. 1994;14(4):313-322.
5. Queale WS, Seidler AJ, Brancati FL. Glycemic control and sliding scale insulin use in medical inpatients with diabetes mellitus. Arch Intern Med. 1997;157(5):545-552.
6. Clement S, Braithwaite SS, Magee MF, et al. Management of diabetes and hyperglycemia in hospitals. Diabetes Care. 2004;27(2):553-591. https://doi.org/10.2337/diacare.27.2.553
7. Garber AJ, Moghissi ES, Bransome ED Jr, et al. American College of Endocrinology position statement on inpatient diabetes and metabolic control. Endocr Pract. 2004;10(1):78-82. https://doi.org/10.4158/EP.10.1.77
8. American Diabetes Association. Standards of medical care in diabetes. Diabetes Care. 2005;28(suppl 1):S4-S36.
9. Umpierrez GE, Hellman R, Korytkowski MT, , et al. Management of hyperglycemia in hospitalized patients in non-critical care setting: an Endocrine Society clinical practice guideline. J Clin Endocrinol Metab. 2012;97(1):16-38. https://doi.org/10.1210/jc.2011-2098
10. Umpierrez GE, Smiley D, Zisman A, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes. Diabetes Care. 2007;30(9):2181-2186. https://doi.org/10.2337/dc07-0295
11. Umpierrez GE, Smiley D, Jacobs S, et al. Randomized study of basal-bolus insulin therapy in the inpatient management of patients with type 2 diabetes undergoing general surgery (RABBIT 2 surgery). Diabetes Care. 2011;34(2):256-261. https://doi.org/10.2337/dc10-1407
12. Schroeder JE, Liebergall M, Raz I, Egleston R, Ben Sussan G, Peyser A. Benefits of a simple glycaemic protocol in an orthopaedic surgery ward: a randomized prospective study. Diabetes Metab Res Rev. 2012;28:71-75. https://doi.org/10.1002/dmrr.1217
13. Lee YY, Lin YM, Leu WJ, et al. Sliding-scale insulin used for blood glucose control: a meta-analysis of randomized controlled trials. Metabolism. 2015;64(9):1183-1192. https://doi.org/10.1016/j.metabol.2015.05.011
14. Christensen MB, Gotfredsen A, Nørgaard K. Efficacy of basal-bolus insulin regimens in the inpatient management of non-critically ill patients with type 2 diabetes: a systematic review and meta-analysis. Diabetes Metab Res Rev. 2017;33(5):e2885. https://doi.org/10.1002/dmrr.2885
15. Wexler DJ, Meigs JB, Cagliero E, Nathan DM, Grant RW. Prevalence of hyper- and hypoglycemia among inpatients with diabetes: a national survey of 44 U.S. hospitals. Diabetes Care. 2007;30(2):367-369. https://doi.org/10.2337/dc06-1715
16. Moreira ED Jr, Silveira PCB, Neves RCS, Souza C Jr, Nunes ZO, Almeida MdCC. Glycemic control and diabetes management in hospitalized patients in Brazil. Diabetol Metab Syndr. 2013;5(1):62. https://doi.org/10.1186/1758-5996-5-62
17. Akhtar ST, Mahmood K, Naqvi IH, Vaswani AS. Inpatient management of type 2 diabetes mellitus: does choice of insulin regimen really matter? Pakistan J Med Sci. 2014;30(4):895-898.
18. Gómez Cuervo C, Sánchez Morla A, Pérez-Jacoiste Asín MA, Bisbal Pardo O, Pérez Ordoño L, Vila Santos J. Effective adverse event reduction with bolus-basal versus sliding scale insulin therapy in patients with diabetes during conventional hospitalization: systematic review and meta-analysis. Endocrinol Nutr. 2016;63(4):145-156. https://doi.org/10.1016/j.endonu.2015.11.008
19. Bain A, Hasan SS, Babar ZUD. Interventions to improve insulin prescribing practice for people with diabetes in hospital: a systematic review. Diabet Med. 2019;36(8):948-960. https://doi.org/10.1111/dme.13982
20. Ambrus DB, O’Connor MJ. Things We Do For No Reason: sliding-scale insulin as monotherapy for glycemic control in hospitalized patients. J Hosp Med. 2019;14(2):114-116. https://doi.org/10.12788/jhm.3109
21. Nau KC, Lorenzetti RC, Cucuzzella M, Devine T, Kline J. Glycemic control in hospitalized patients not in intensive care: beyond sliding-scale insulin. Am Fam Physician. 2010;81(9):1130-1135.
22. Colunga-Lozano LE, Gonzalez Torres FJ, Delgado-Figueroa N, et al. Sliding scale insulin for non-critically ill hospitalised adults with diabetes mellitus. Cochrane Database Syst Rev. 2018;11(11):CD011296. https://doi.org/10.1002/14651858.CD011296.pub2
23. American Diabetes Association. Diabetes care in the hospital: Standards of Medical Care in Diabetes—2019. Diabetes Care. 2019;42(suppl 1):S173-S181. https://doi.org/10.2337/dc19-S015
24. Umpierrez GE, Smiley D, Hermayer K, et al. Randomized study comparing a basal-bolus with a basal plus correction management of medical and surgical patients with type 2 diabetes: basal plus trial. Diabetes Care. 2013;36(8):2169-2174. https://doi.org/10.2337/dc12-1988
25. Turchin A, Matheny ME, Shubina M, Scanlon SV, Greenwood B, Pendergrass ML. Hypoglycemia and clinical outcomes in patients with diabetes hospitalized in the general ward. Diabetes Care. 2009;32(7):1153-1157. https://doi.org/10.2337/dc08-2127
26. Garg R, Hurwitz S, Turchin A, Trivedi A. Hypoglycemia, with or without insulin therapy, is associated with increased mortality among hospitalized patients. Diabetes Care. 2013;36(5):1107-1110. https://doi.org/10.2337/dc12-1296
27. Zapatero A, Gómez-Huelgas R, González N, et al. Frequency of hypoglycemia and its impact on length of stay, mortality, and short-term readmission in patients with diabetes hospitalized in internal medicine wards. Endocr Pract. 2014;20(9):870-875. https://doi.org/10.4158/EP14006.OR
28. Umpierrez GE, Isaacs SD, Bazargan N, You X, Thaler LM, Kitabchi AE. Hyperglycemia: an independent marker of in-hospital mortality in patients with undiagnosed diabetes. J Clin Endocrinol Metab. 2002;87(3):978-982. https://doi.org/10.1210/jcem.87.3.8341
29. Dickerson LM, Ye X, Sack JL, Hueston WJ. Glycemic control in medical inpatients with type 2 diabetes mellitus receiving sliding scale insulin regimens versus routine diabetes medications: a multicenter randomized controlled trial. Ann Fam Med. 2003;1(1):29-35. https://doi.org/10.1370/afm.2

Issue
Journal of Hospital Medicine 16(8)
Issue
Journal of Hospital Medicine 16(8)
Page Number
462-468. Published Online Only July 21, 2021
Page Number
462-468. Published Online Only July 21, 2021
Publications
Publications
Topics
Article Type
Display Headline
Inpatient Glycemic Control With Sliding Scale Insulin in Noncritical Patients With Type 2 Diabetes: Who Can Slide?
Display Headline
Inpatient Glycemic Control With Sliding Scale Insulin in Noncritical Patients With Type 2 Diabetes: Who Can Slide?
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Guillermo E Umpierrez, MD, CDE; Email: [email protected]; Telephone: 404-778-1665.
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Page Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media
Media Files

Identifying the Sickest During Triage: Using Point-of-Care Severity Scores to Predict Prognosis in Emergency Department Patients With Suspected Sepsis

Article Type
Changed
Thu, 07/29/2021 - 10:19
Display Headline
Identifying the Sickest During Triage: Using Point-of-Care Severity Scores to Predict Prognosis in Emergency Department Patients With Suspected Sepsis

Sepsis is the leading cause of in-hospital mortality in the United States.1 Sepsis is present on admission in 85% of cases, and each hour delay in antibiotic treatment is associated with 4% to 7% increased odds of mortality.2,3 Prompt identification and treatment of sepsis is essential for reducing morbidity and mortality, but identifying sepsis during triage is challenging.2

Risk stratification scores that rely solely on data readily available at the bedside have been developed to quickly identify those at greatest risk of poor outcomes from sepsis in real time. The quick Sequential Organ Failure Assessment (qSOFA) score, the National Early Warning System (NEWS2), and the Shock Index are easy-to-calculate measures that use routinely collected clinical data that are not subject to laboratory delay. These scores can be incorporated into electronic health record (EHR)-based alerts and can be calculated longitudinally to track the risk of poor outcomes over time. qSOFA was developed to quantify patient risk at bedside in non-intensive care unit (ICU) settings, but there is no consensus about its ability to predict adverse outcomes such as mortality and ICU admission.4-6 The United Kingdom’s National Health Service uses NEWS2 to identify patients at risk for sepsis.7 NEWS has been shown to have similar or better sensitivity in identifying poorer outcomes in sepsis patients compared with systemic inflammatory response syndrome (SIRS) criteria and qSOFA.4,8-11 However, since the latest update of NEWS2 in 2017, there has been little study of its predictive ability. The Shock Index is a simple bedside score (heart rate divided by systolic blood pressure) that was developed to detect changes in cardiovascular performance before systemic shock onset. Although it was not developed for infection and has not been regularly applied in the sepsis literature, the Shock Index might be useful for identifying patients at increased risk of poor outcomes. Patients with higher and sustained Shock Index scores are more likely to experience morbidity, such as hyperlactatemia, vasopressor use, and organ failure, and also have an increased risk of mortality.12-14

Although the predictive abilities of these bedside risk stratification scores have been assessed individually using standard binary cut-points, the comparative performance of qSOFA, the Shock Index, and NEWS2 has not been evaluated in patients presenting to an emergency department (ED) with suspected sepsis. Our objective was to provide a head-to-head comparison of the test characteristics of qSOFA, the Shock Index, and NEWS2 calculated at ED triage for predicting in-hospital mortality and ED-to-ICU admission in patients with suspected sepsis to help health systems and providers select screening measures.

METHODS

Design and Setting

We conducted a retrospective cohort study of ED patients who presented with suspected sepsis to the University of California San Francisco (UCSF) Helen Diller Medical Center at Parnassus Heights between June 1, 2012, and December 31, 2018. Our institution is a 785-bed academic teaching hospital with approximately 30,000 ED encounters per year. The study was approved with a waiver of informed consent by the UCSF Human Research Protection Program.

Participants

We use an Epic-based EHR platform (Epic 2017, Epic Systems Corporation) for clinical care, which was implemented on June 1, 2012. All data elements were obtained from Clarity, the relational database that stores Epic’s inpatient data. The study included encounters for patients age ≥18 years who had blood cultures ordered within 24 hours of ED presentation and administration of intravenous antibiotics within 24 hours. Repeat encounters were treated independently in our analysis.

Outcomes and Measures

We compared the ability of qSOFA, the Shock Index, and NEWS2 to predict in-hospital mortality and admission to the ICU from the ED (ED-to-ICU admission). We used the most abnormal vital signs and clinical assessments gathered within 30 minutes of ED presentation to identify patients who were qSOFA-positive, Shock Index-positive, and NEWS2-positive based on standard cut-points of risk. Data elements used to calculate qSOFA, Shock Index, and NEWS2 included blood pressure, heart rate, respiratory rate, Glasgow Coma Scale (GCS) score, oxygen saturation, requirement for supplemental oxygen, and temperature (Table 1). Patients were considered positive if they had a qSOFAscore ≥2, Shock Index of >0.7, and NEWS2 ≥5 based on triage vital signs.7,15,16 We considered patients to have altered mental status, a criterion used for NEWS2, if they had a GCS score <15 instead of using the “alert, verbal, confusion, pain, unresponsive” scale, which is not captured in our EHR, a method that has been used in earlier studies.17,18 Missing assessments were considered normal. Although our primary analysis focused on the scores calculated within 30 minutes of ED presentation, we performed a sensitivity analysis examining scores calculated within 1 hour of ED presentation in the event of a delay in gathering triage vital sign data.

Data Elements Collected in the 72 Hours After Emergency Department Presentation to Identify Patients Meeting qSOFA, Shock Index, and NEWS2 Criteria

We compared demographic and clinical characteristics of patients who were positive for qSOFA, the Shock Index, and NEWS2. Demographic data were extracted from the EHR and included primary language, age, sex, and insurance status. All International Classification of Diseases (ICD)-9/10 diagnosis codes were pulled from Clarity billing tables. We used the Elixhauser comorbidity groupings19 of ICD-9/10 codes present on admission to identify preexisting comorbidities and underlying organ dysfunction. To estimate burden of comorbid illnesses, we calculated the validated van Walraven comorbidity index,20 which provides an estimated risk of in-hospital death based on documented Elixhauser comorbidities. Admission level of care (acute, stepdown, or intensive care) was collected for inpatient admissions to assess initial illness severity.21 We also evaluated discharge disposition and in-hospital mortality. Index blood culture results were collected, and dates and timestamps of mechanical ventilation, fluid, vasopressor, and antibiotic administration were obtained for the duration of the encounter.

UCSF uses an automated, real-time, algorithm-based severe sepsis alert that is triggered when a patient meets ≥2 SIRS criteria and again when the patient meets severe sepsis or septic shock criteria (ie, ≥2 SIRS criteria in addition to end-organ dysfunction and/or fluid nonresponsive hypotension). This sepsis screening alert was in use for the duration of our study.22

Statistical Analysis

We performed a subgroup analysis among those who were diagnosed with sepsis, according to the 2016 Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3) criteria. Sepsis is defined as a change in Sequential Organ Failure Assessment (SOFA) score of ≥2 points within the first 48 hours.23 Additionally, patients meeting Sepsis-3 criteria needed to (1) receive ≥4 days of sequential antibiotic therapy or experience death or discharge to hospice before 4 days of antibiotic therapy or (2) have a validated sepsis discharge billing code. These parameters were added to increase the specificity of our sample.24

All statistical analyses were conducted using Stata 14 (StataCorp). We summarized differences in demographic and clinical characteristics among the populations meeting each severity score but elected not to conduct hypothesis testing because patients could be positive for one or more scores. We calculated sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for each score to predict in-hospital mortality and ED-to-ICU admission. To allow comparison with other studies, we also created a composite outcome of either in-hospital mortality or ED-to-ICU admission. To assess score discrimination to predict in-hospital mortality and ED-to-ICU admission, we calculated the area under the receiver operating characteristic curve (AUROC) along with asymptotic normal 95% CI using the “roctab” command considering a binary cut-point, as well as the full range of scores measured in the cohort. The AUROC range from 0.50 to 1.00 and a score in the 0.70 to 0.80 range can be considered fair.25 We assessed significant differences between severity score AUROCs using the DeLong method26 implemented through Stata 14’s “roccomp” command. As a sensitivity analysis, we explored whether the standard cut-points for qSOFA, the Shock Index, and NEWS2 provided the highest AUROC in our population by calculating test characteristics for several score cut-points.

RESULTS

Within our sample 23,837 ED patients had blood cultures ordered within 24 hours of ED presentation and were considered to have suspected sepsis. The mean age of the cohort was 60.8 years, and 1,612 (6.8%) had positive blood cultures. A total of 12,928 patients (54.2%) were found to have sepsis. We documented 1,427 in-hospital deaths (6.0%) and 3,149 (13.2%) ED-to-ICU admissions. At ED triage 1,921 (8.1%) were qSOFA-positive, 4,273 (17.9%) were Shock Index-positive, and 11,832 (49.6%) were NEWS2-positive. At ED triage, blood pressure, heart rate, respiratory rate, and oxygen saturated were documented in >99% of patients, 93.5% had temperature documented, and 28.5% had GCS recorded. If the window of assessment was widened to 1 hour, GCS was only documented among 44.2% of those with suspected sepsis.

Demographic Characteristics and Clinical Course

We identified significant differences when comparing demographic and clinical characteristics among patients who scored positive for the three severity measures at triage (Table 2). Although no hypothesis testing was conducted because patients could meet one or more scores, qSOFA-positive patients were older (median 70, 66, and 64 years, respectively), more likely to have Medicare as the primary payor (67.6% vs 59.7% vs 56.6%), to have chronic renal failure (26.1%, 23.1%, and 23.3%, respectively), to have a greater degree of underlying comorbidities based on the van Walraven Comorbidity Index (median 15, 12, and 11, respectively), and to be admitted to the ICU from the ED (48.1%, 36.3%, and 21.0%, respectively) compared with those positive for the Shock Index or NEWS2.

Demographic Characteristics and Clinical Course of Patients With Suspected Infection and Populations Who Met Standard Cut-Points of qSOFA, Shock Index, and NEWS2

qSOFA-positive patients received antibiotics more quickly than those who were Shock Index-positive or NEWS2-positive (median 1.5, 1.8, and 2.8 hours after admission, respectively). In addition, those who were qSOFA-positive were more likely to have a positive blood culture (10.9%, 9.4%, and 8.5%, respectively) and to receive an EHR-based diagnosis of sepsis (77.0%, 69.6%, and 60.9%, respectively) than those who were Shock Index- or NEWS2-positive. Those who were qSOFA-positive also were more likely to be mechanically ventilated during their hospital stay (25.4%, 19.2%, and 10.8%, respectively) and to receive vasopressors (33.5%, 22.5%, and 12.2%, respectively). In-hospital mortality also was more common among those who were qSOFA-positive at triage (23.4%, 15.3%, and 9.2%, respectively).

Because both qSOFA and NEWS2 incorporate GCS, we explored baseline characteristics of patients with GCS documented at triage (n = 6,794). These patients were older (median age 63 and 61 years, P < .0001), more likely to be male (54.9% and 53.4%, P = .0031), more likely to have renal failure (22.8% and 20.1%, P < .0001), more likely to have liver disease (14.2% and 12.8%, P = .006), had a higher van Walraven comorbidity score on presentation (median 10 and 8, P < .0001), and were more likely to go directly to the ICU from the ED (20.2% and 10.6%, P < .0001). However, among the 6,397 GCS scores documented at triage, only 1,579 (24.7%) were abnormal.

Test Characteristics of qSOFA, Shock Index, and NEWS2 for Predicting In-hospital Mortality and ED-to-ICU Admission

Among 23,837 patients with suspected sepsis, NEWS2 had the highest sensitivity for predicting in-hospital mortality (76.0%; 95% CI, 73.7%-78.2%) and ED-to-ICU admission (78.9%; 95% CI, 77.5%-80.4%) but had the lowest specificity for in-hospital mortality (52.0%; 95% CI, 51.4%-52.7%) and for ED-to-ICU admission (54.8%; 95% CI, 54.1%-55.5%) (Table 3). qSOFA had the lowest sensitivity for in-hospital mortality (31.5%; 95% CI, 29.1%-33.9%) and ED-to-ICU admission (29.3%; 95% CI, 27.7%-30.9%) but the highest specificity for in-hospital mortality (93.4%; 95% CI, 93.1%-93.8%) and ED-to-ICU admission (95.2%; 95% CI, 94.9%-95.5%). The Shock Index had a sensitivity that fell between qSOFA and NEWS2 for in-hospital mortality (45.8%; 95% CI, 43.2%-48.5%) and ED-to-ICU admission (49.2%; 95% CI, 47.5%-51.0%). The specificity of the Shock Index also was between qSOFA and NEWS2 for in-hospital mortality (83.9%; 95% CI, 83.4%-84.3%) and ED-to-ICU admission (86.8%; 95% CI, 86.4%-87.3%). All three scores exhibited relatively low PPV, ranging from 9.2% to 23.4% for in-hospital mortality and 21.0% to 48.0% for ED-to-ICU triage. Conversely, all three scores exhibited relatively high NPV, ranging from 95.5% to 97.1% for in-hospital mortality and 89.8% to 94.5% for ED-to-ICU triage. The patterns in sensitivity and specificity for in-hospital mortality and ED-to-ICU admission were similar among the 12,928 patients who received an EHR-based sepsis diagnosis with the tests generally demonstrating lower specificities, higher PPVs, and lower NPVs (Table 3).

Test Characteristics of Point-of-Care Severity Scores for Predicting Prognosis in ED Patients Presenting With Suspected Sepsis and Those With EHR-Based Sepsis Diagnosis

When considering a binary cutoff, the Shock Index exhibited the highest AUROC for in-hospital mortality (0.648; 95% CI, 0.635-0.662) and had a significantly higher AUROC than qSOFA (AUROC, 0.625; 95% CI, 0.612-0.637; P = .0005), but there was no difference compared with NEWS2 (AUROC, 0.640; 95% CI, 0.628-0.652; P = .2112). NEWS2 had a significantly higher AUROC than qSOFA for predicting in-hospital mortality (P = .0227). The Shock Index also exhibited the highest AUROC for ED-to-ICU admission (0.680; 95% CI, 0.617-0.689), which was significantly higher than the AUROC for qSOFA (P < .0001) and NEWS2 (P = 0.0151). NEWS2 had a significantly higher AUROC than qSOFA for predicting ED-to-ICU admission (P < .0001). Similar findings were seen in patients found to have sepsis. When considering the range of possible scores measured in our cohort, qSOFA and NEWS2 exhibited higher AUROCs for in-hospital mortality and ED-to-ICU admission than the Shock Index among patients with suspected infection and the subgroup with a sepsis diagnosis (Figure).

Receiver Operator Characteristic Curves for qSOFA, Shock Index, and NEWS2
The AUROCs of the scores were statistically significantly different for both in-hospital mortality (P = .0026) and ED-to-ICU admission (P < .0001). As a sensitivity analysis, we varied the binary cut-points of qSOFA, the Shock Index, and NEWS2 and calculated test characteristics within our study cohort (Appendix). For our institution, the qSOFA cut-point with the highest AUROC would be qSOFA > 0 for both in-hospital mortality (AUROC, 0.699; 95% CI, 0.687-0.711) and ED-to-ICU admission (AUROC, 0.716; 95% CI, 0.707-0.724), with 36.5% of the cohort meeting qSOFA. The NEWS2 cut-point with the highest AUROC would be NEWS2 ≥7 for both in-hospital mortality (AUROC, 0.653; 95% CI, 0.640-0.666) and ED-to-ICU admission (AUROC, 0.677; 95% CI, 0.668-0.686), with 20.3% of the cohort meeting NEWS2 at this cut-point. The standard Shock Index cut-point ≥0.7 exhibited the highest AUROC for in-hospital mortality and ED-to-ICU admission at our institution.

DISCUSSION

In this retrospective cohort study of 23,837 patients who presented to the ED with suspected sepsis, the standard qSOFA threshold was met least frequently, followed by the Shock Index and NEWS2. NEWS2 had the highest sensitivity but the lowest specificity for predicting in-hospital mortality and ED-to-ICU admission, making it a challenging bedside risk stratification scale for identifying patients at risk of poor clinical outcomes. When comparing predictive performance among the three scales, qSOFA had the highest specificity and the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission in this cohort of patients with suspected sepsis. These trends in sensitivity, specificity, and AUROC were consistent among those who met EHR criteria for a sepsis diagnosis. In the analysis of the three scoring systems using all available cut-points, qSOFA and NEWS2 had the highest AUROCs, followed by the Shock Index.

Considering the rapid progression from organ dysfunction to death in sepsis patients, as well as the difficulty establishing a sepsis diagnosis at triage,23 providers must quickly identify patients at increased risk of poor outcomes when they present to the ED. Sepsis alerts often are built using SIRS criteria,27 including the one used for sepsis surveillance at UCSF since 2012,22 but the white blood cell count criterion is subject to a laboratory lag and could lead to a delay in identification. Implementation of a point-of-care bedside score alert that uses readily available clinical data could allow providers to identify patients at greatest risk of poor outcomes immediately at ED presentation and triage, which motivated us to explore the predictive performance of qSOFA, the Shock Index, and NEWS2.

Our study is the first to provide a head-to-head comparison of the predictive performance of qSOFA, the Shock Index, and NEWS2, three easy-to-calculate bedside risk scores that use EHR data collected among patients with suspected sepsis. The Sepsis-3 guidelines recommend qSOFA to quickly identify non-ICU patients at greatest risk of poor outcomes because the measure exhibited predictive performance similar to the more extensive SOFA score outside the ICU.16,23 Although some studies have confirmed qSOFA’s high predictive performance,28-31 our test characteristics and AUROC findings are in line with other published analyses.4,6,10,17 The UK National Health Service is using NEWS2 to screen for patients at risk of poor outcomes from sepsis. Several analyses that assessed the predictive ability of NEWS have reported estimates in line with our findings.4,10,32 The Shock Index was introduced in 1967 and provided a metric to evaluate hemodynamic stability based on heart rate and systolic blood pressure.33 The Shock Index has been studied in several contexts, including sepsis,34 and studies show that a sustained Shock Index is associated with increased odds of vasopressor administration, higher prevalence of hyperlactatemia, and increased risk of poor outcomes in the ICU.13,14

For our study, we were particularly interested in exploring how the Shock Index would compare with more frequently used severity scores such as qSOFA and NEWS2 among patients with suspected sepsis, given the simplicity of its calculation and the easy availability of required data. In our cohort of 23,837 patients, only 159 people had missing blood pressure and only 71 had omitted heart rate. In contrast, both qSOFA and NEWS2 include an assessment of level of consciousness that can be subject to variability in assessment methods and EHR documentation across institutions.11 In our cohort, GCS within 30 minutes of ED presentation was missing in 72 patients, which could have led to incomplete calculation of qSOFA and NEWS2 if a missing value was not actually within normal limits.

Several investigations relate qSOFA to NEWS but few compare qSOFA with the newer NEWS2, and even fewer evaluate the Shock Index with any of these scores.10,11,18,29,35-37 In general, studies have shown that NEWS exhibits a higher AUROC for predicting mortality, sepsis with organ dysfunction, and ICU admission, often as a composite outcome.4,11,18,37,38 A handful of studies compare the Shock Index to SIRS; however, little has been done to compare the Shock Index to qSOFA or NEWS2, scores that have been used specifically for sepsis and might be more predictive of poor outcomes than SIRS.33 In our study, the Shock Index had a higher AUROC than either qSOFA or NEWS2 for predicting in-hospital mortality and ED-to-ICU admission measured as separate outcomes and as a composite outcome using standard cut-points for these scores.

When selecting a severity score to apply in an institution, it is important to carefully evaluate the score’s test characteristics, in addition to considering the availability of reliable data. Tests with high sensitivity and NPV for the population being studied can be useful to rule out disease or risk of poor outcome, while tests with high specificity and PPV can be useful to rule in disease or risk of poor outcome.39 When considering specificity, qSOFA’s performance was superior to the Shock Index and NEWS2 in our study, but a small percentage of the population was identified using a cut-point of qSOFA ≥2. If we used qSOFA and applied this standard cut-point at our institution, we could be confident that those identified were at increased risk, but we would miss a significant number of patients who would experience a poor outcome. When considering sensitivity, performance of NEWS2 was superior to qSOFA and the Shock Index in our study, but one-half of the population was identified using a cut-point of NEWS2 ≥5. If we were to apply this standard NEWS2 cut-point at our institution, we would assume that one-half of our population was at risk, which might drive resource use towards patients who will not experience a poor outcome. Although none of the scores exhibited a robust AUROC measure, the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission when using the standard binary cut-point, and its sensitivity and specificity is between that of qSOFA and NEWS2, potentially making it a score to use in settings where qSOFA and NEWS2 score components, such as altered mentation, are not reliably collected. Finally, our sensitivity analysis varying the binary cut-point of each score within our population demonstrated that the standard cut-points might not be as useful within a specific population and might need to be tailored for implementation, balancing sensitivity, specificity, PPV, and NPV to meet local priorities and ICU capacity.

Our study has limitations. It is a single-center, retrospective analysis, factors that could reduce generalizability. However, it does include a large and diverse patient population spanning several years. Missing GCS data could have affected the predictive ability of qSOFA and NEWS2 in our cohort. We could not reliably perform imputation of GCS because of the high missingness and therefore we assumed missing was normal, as was done in the Sepsis-3 derivation studies.16 Previous studies have attempted to impute GCS and have not observed improved performance of qSOFA to predict mortality.40 Because manually collected variables such as GCS are less reliably documented in the EHR, there might be limitations in their use for triage risk scores.

Although the current analysis focused on the predictive performance of qSOFA, the Shock Index, and NEWS2 at triage, performance of these scores could affect the ED team’s treatment decisions before handoff to the hospitalist team and the expected level of care the patient will receive after in-patient admission. These tests also have the advantage of being easy to calculate at the bedside over time, which could provide an objective assessment of longitudinal predicted prognosis. Future work should assess the longitudinal performance of each of these scores among those with suspected sepsis and to determine the impact using these scores would have on clinical and resource utilization outcomes.

CONCLUSION

Local priorities should drive selection of a screening tool, balancing sensitivity, specificity, PPV, and NPV to achieve the institution’s goals. qSOFA, Shock Index, and NEWS2 are risk stratification tools that can be easily implemented at ED triage using data available at the bedside. Although none of these scores performed strongly when comparing AUROCs, qSOFA was highly specific for identifying patients with poor outcomes, and NEWS2 was the most sensitive for ruling out those at high risk among patients with suspected sepsis. The Shock Index exhibited a sensitivity and specificity that fell between qSOFA and NEWS2 and also might be considered to identify those at increased risk, given its ease of implementation, particularly in settings where altered mentation is unreliably or inconsistently documented.

Acknowledgment

The authors thank the UCSF Division of Hospital Medicine Data Core for their assistance with data acquisition.

Files
References

1. Jones SL, Ashton CM, Kiehne LB, et al. Outcomes and resource use of sepsis-associated stays by presence on admission, severity, and hospital type. Med Care. 2016;54(3):303-310. https://doi.org/10.1097/MLR.0000000000000481
2. Seymour CW, Gesten F, Prescott HC, et al. Time to treatment and mortality during mandated emergency care for sepsis. N Engl J Med. 2017;376(23):2235-2244. https://doi.org/10.1056/NEJMoa1703058
3. Kumar A, Roberts D, Wood KE, et al. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit Care Med. 2006;34(6):1589-1596. https://doi.org/10.1097/01.CCM.0000217961.75225.E9
4. Churpek MM, Snyder A, Sokol S, Pettit NN, Edelson DP. Investigating the impact of different suspicion of infection criteria on the accuracy of Quick Sepsis-Related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores. Crit Care Med. 2017;45(11):1805-1812. https://doi.org/10.1097/CCM.0000000000002648
5. Abdullah SMOB, Sørensen RH, Dessau RBC, Sattar SMRU, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
6. Kim KS, Suh GJ, Kim K, et al. Quick Sepsis-related Organ Failure Assessment score is not sensitive enough to predict 28-day mortality in emergency department patients with sepsis: a retrospective review. Clin Exp Emerg Med. 2019;6(1):77-83. HTTPS://DOI.ORG/ 10.15441/ceem.17.294
7. National Early Warning Score (NEWS) 2: Standardising the assessment of acute-illness severity in the NHS. Royal College of Physicians; 2017.
8. Brink A, Alsma J, Verdonschot RJCG, et al. Predicting mortality in patients with suspected sepsis at the emergency department: a retrospective cohort study comparing qSOFA, SIRS and National Early Warning Score. PLoS One. 2019;14(1):e0211133. https://doi.org/ 10.1371/journal.pone.0211133
9. Redfern OC, Smith GB, Prytherch DR, Meredith P, Inada-Kim M, Schmidt PE. A comparison of the Quick Sequential (Sepsis-Related) Organ Failure Assessment Score and the National Early Warning Score in non-ICU patients with/without infection. Crit Care Med. 2018;46(12):1923-1933. https://doi.org/10.1097/CCM.0000000000003359
10. Churpek MM, Snyder A, Han X, et al. Quick Sepsis-related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores for detecting clinical deterioration in infected patients outside the intensive care unit. Am J Respir Crit Care Med. 2017;195(7):906-911. https://doi.org/10.1164/rccm.201604-0854OC
11. Goulden R, Hoyle MC, Monis J, et al. qSOFA, SIRS and NEWS for predicting inhospital mortality and ICU admission in emergency admissions treated as sepsis. Emerg Med J. 2018;35(6):345-349. https://doi.org/10.1136/emermed-2017-207120
12. Biney I, Shepherd A, Thomas J, Mehari A. Shock Index and outcomes in patients admitted to the ICU with sepsis. Chest. 2015;148(suppl 4):337A. https://doi.org/https://doi.org/10.1378/chest.2281151
13. Wira CR, Francis MW, Bhat S, Ehrman R, Conner D, Siegel M. The shock index as a predictor of vasopressor use in emergency department patients with severe sepsis. West J Emerg Med. 2014;15(1):60-66. https://doi.org/10.5811/westjem.2013.7.18472
14. Berger T, Green J, Horeczko T, et al. Shock index and early recognition of sepsis in the emergency department: pilot study. West J Emerg Med. 2013;14(2):168-174. https://doi.org/10.5811/westjem.2012.8.11546
15. Middleton DJ, Smith TO, Bedford R, Neilly M, Myint PK. Shock Index predicts outcome in patients with suspected sepsis or community-acquired pneumonia: a systematic review. J Clin Med. 2019;8(8):1144. https://doi.org/10.3390/jcm8081144
16. Seymour CW, Liu VX, Iwashyna TJ, et al. Assessment of clinical criteria for sepsis: for the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):762-774. https://doi.org/ 10.1001/jama.2016.0288
17. Abdullah S, Sørensen RH, Dessau RBC, Sattar S, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
18. Usman OA, Usman AA, Ward MA. Comparison of SIRS, qSOFA, and NEWS for the early identification of sepsis in the Emergency Department. Am J Emerg Med. 2018;37(8):1490-1497. https://doi.org/10.1016/j.ajem.2018.10.058
19. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Med Care. 1998;36(1):8-27. https://doi.org/10.1097/00005650-199801000-00004
20. van Walraven C, Austin PC, Jennings A, Quan H, Forster AJ. A modification of the Elixhauser comorbidity measures into a point system for hospital death using administrative data. Med Care. 2009;47(6):626-633. https://doi.org/10.1097/MLR.0b013e31819432e5
21. Prin M, Wunsch H. The role of stepdown beds in hospital care. Am J Respir Crit Care Med. 2014;190(11):1210-1216. https://doi.org/10.1164/rccm.201406-1117PP
22. Narayanan N, Gross AK, Pintens M, Fee C, MacDougall C. Effect of an electronic medical record alert for severe sepsis among ED patients. Am J Emerg Med. 2016;34(2):185-188. https://doi.org/10.1016/j.ajem.2015.10.005
23. Singer M, Deutschman CS, Seymour CW, et al. The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):801-810. https://doi.org/10.1001/jama.2016.0287
24. Rhee C, Dantes R, Epstein L, et al. Incidence and trends of sepsis in US hospitals using clinical vs claims data, 2009-2014. JAMA. 2017;318(13):1241-1249. https://doi.org/10.1001/jama.2017.13836
25. Safari S, Baratloo A, Elfil M, Negida A. Evidence based emergency medicine; part 5 receiver operating curve and area under the curve. Emerg (Tehran). 2016;4(2):111-113.
26. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988;44(3):837-845.
27. Kangas C, Iverson L, Pierce D. Sepsis screening: combining Early Warning Scores and SIRS Criteria. Clin Nurs Res. 2021;30(1):42-49. https://doi.org/10.1177/1054773818823334.
28. Freund Y, Lemachatti N, Krastinova E, et al. Prognostic accuracy of Sepsis-3 Criteria for in-hospital mortality among patients with suspected infection presenting to the emergency department. JAMA. 2017;317(3):301-308. https://doi.org/10.1001/jama.2016.20329
29. Finkelsztein EJ, Jones DS, Ma KC, et al. Comparison of qSOFA and SIRS for predicting adverse outcomes of patients with suspicion of sepsis outside the intensive care unit. Crit Care. 2017;21(1):73. https://doi.org/10.1186/s13054-017-1658-5
30. Canet E, Taylor DM, Khor R, Krishnan V, Bellomo R. qSOFA as predictor of mortality and prolonged ICU admission in Emergency Department patients with suspected infection. J Crit Care. 2018;48:118-123. https://doi.org/10.1016/j.jcrc.2018.08.022
31. Anand V, Zhang Z, Kadri SS, Klompas M, Rhee C; CDC Prevention Epicenters Program. Epidemiology of Quick Sequential Organ Failure Assessment criteria in undifferentiated patients and association with suspected infection and sepsis. Chest. 2019;156(2):289-297. https://doi.org/10.1016/j.chest.2019.03.032
32. Hamilton F, Arnold D, Baird A, Albur M, Whiting P. Early Warning Scores do not accurately predict mortality in sepsis: A meta-analysis and systematic review of the literature. J Infect. 2018;76(3):241-248. https://doi.org/10.1016/j.jinf.2018.01.002
33. Koch E, Lovett S, Nghiem T, Riggs RA, Rech MA. Shock Index in the emergency department: utility and limitations. Open Access Emerg Med. 2019;11:179-199. https://doi.org/10.2147/OAEM.S178358
34. Yussof SJ, Zakaria MI, Mohamed FL, Bujang MA, Lakshmanan S, Asaari AH. Value of Shock Index in prognosticating the short-term outcome of death for patients presenting with severe sepsis and septic shock in the emergency department. Med J Malaysia. 2012;67(4):406-411.
35. Siddiqui S, Chua M, Kumaresh V, Choo R. A comparison of pre ICU admission SIRS, EWS and q SOFA scores for predicting mortality and length of stay in ICU. J Crit Care. 2017;41:191-193. https://doi.org/10.1016/j.jcrc.2017.05.017
36. Costa RT, Nassar AP, Caruso P. Accuracy of SOFA, qSOFA, and SIRS scores for mortality in cancer patients admitted to an intensive care unit with suspected infection. J Crit Care. 2018;45:52-57. https://doi.org/10.1016/j.jcrc.2017.12.024
37. Mellhammar L, Linder A, Tverring J, et al. NEWS2 is Superior to qSOFA in detecting sepsis with organ dysfunction in the emergency department. J Clin Med. 2019;8(8):1128. https://doi.org/10.3390/jcm8081128
38. Szakmany T, Pugh R, Kopczynska M, et al. Defining sepsis on the wards: results of a multi-centre point-prevalence study comparing two sepsis definitions. Anaesthesia. 2018;73(2):195-204. https://doi.org/10.1111/anae.14062
39. Newman TB, Kohn MA. Evidence-Based Diagnosis: An Introduction to Clinical Epidemiology. Cambridge University Press; 2009.
40. Askim Å, Moser F, Gustad LT, et al. Poor performance of quick-SOFA (qSOFA) score in predicting severe sepsis and mortality - a prospective study of patients admitted with infection to the emergency department. Scand J Trauma Resusc Emerg Med. 2017;25(1):56. https://doi.org/10.1186/s13049-017-0399-4

Article PDF
Author and Disclosure Information

1Division of Hospital Medicine, Department of Medicine, University of California, San Francisco, California; 2Divisions of Nephrology and Critical Care Medicine, Department of Medicine, University of California, San Francisco, San Francisco, California.

Disclosures
Dr Prasad reports personal fees from EpiExcellence, LLC, outside the submitted work. Dr Fang’s institution received funding from the National Heart, Lung, and Blood Institute K24HL141354 and Patient-Centered Outcomes Research Institute. Dr Liu’s institution received funding from the National Diabetes and Digestive and Kidney Diseases K24DK113381.

Issue
Journal of Hospital Medicine 16(8)
Publications
Topics
Page Number
453-461. Published Online Only July 21, 2021
Sections
Files
Files
Author and Disclosure Information

1Division of Hospital Medicine, Department of Medicine, University of California, San Francisco, California; 2Divisions of Nephrology and Critical Care Medicine, Department of Medicine, University of California, San Francisco, San Francisco, California.

Disclosures
Dr Prasad reports personal fees from EpiExcellence, LLC, outside the submitted work. Dr Fang’s institution received funding from the National Heart, Lung, and Blood Institute K24HL141354 and Patient-Centered Outcomes Research Institute. Dr Liu’s institution received funding from the National Diabetes and Digestive and Kidney Diseases K24DK113381.

Author and Disclosure Information

1Division of Hospital Medicine, Department of Medicine, University of California, San Francisco, California; 2Divisions of Nephrology and Critical Care Medicine, Department of Medicine, University of California, San Francisco, San Francisco, California.

Disclosures
Dr Prasad reports personal fees from EpiExcellence, LLC, outside the submitted work. Dr Fang’s institution received funding from the National Heart, Lung, and Blood Institute K24HL141354 and Patient-Centered Outcomes Research Institute. Dr Liu’s institution received funding from the National Diabetes and Digestive and Kidney Diseases K24DK113381.

Article PDF
Article PDF
Related Articles

Sepsis is the leading cause of in-hospital mortality in the United States.1 Sepsis is present on admission in 85% of cases, and each hour delay in antibiotic treatment is associated with 4% to 7% increased odds of mortality.2,3 Prompt identification and treatment of sepsis is essential for reducing morbidity and mortality, but identifying sepsis during triage is challenging.2

Risk stratification scores that rely solely on data readily available at the bedside have been developed to quickly identify those at greatest risk of poor outcomes from sepsis in real time. The quick Sequential Organ Failure Assessment (qSOFA) score, the National Early Warning System (NEWS2), and the Shock Index are easy-to-calculate measures that use routinely collected clinical data that are not subject to laboratory delay. These scores can be incorporated into electronic health record (EHR)-based alerts and can be calculated longitudinally to track the risk of poor outcomes over time. qSOFA was developed to quantify patient risk at bedside in non-intensive care unit (ICU) settings, but there is no consensus about its ability to predict adverse outcomes such as mortality and ICU admission.4-6 The United Kingdom’s National Health Service uses NEWS2 to identify patients at risk for sepsis.7 NEWS has been shown to have similar or better sensitivity in identifying poorer outcomes in sepsis patients compared with systemic inflammatory response syndrome (SIRS) criteria and qSOFA.4,8-11 However, since the latest update of NEWS2 in 2017, there has been little study of its predictive ability. The Shock Index is a simple bedside score (heart rate divided by systolic blood pressure) that was developed to detect changes in cardiovascular performance before systemic shock onset. Although it was not developed for infection and has not been regularly applied in the sepsis literature, the Shock Index might be useful for identifying patients at increased risk of poor outcomes. Patients with higher and sustained Shock Index scores are more likely to experience morbidity, such as hyperlactatemia, vasopressor use, and organ failure, and also have an increased risk of mortality.12-14

Although the predictive abilities of these bedside risk stratification scores have been assessed individually using standard binary cut-points, the comparative performance of qSOFA, the Shock Index, and NEWS2 has not been evaluated in patients presenting to an emergency department (ED) with suspected sepsis. Our objective was to provide a head-to-head comparison of the test characteristics of qSOFA, the Shock Index, and NEWS2 calculated at ED triage for predicting in-hospital mortality and ED-to-ICU admission in patients with suspected sepsis to help health systems and providers select screening measures.

METHODS

Design and Setting

We conducted a retrospective cohort study of ED patients who presented with suspected sepsis to the University of California San Francisco (UCSF) Helen Diller Medical Center at Parnassus Heights between June 1, 2012, and December 31, 2018. Our institution is a 785-bed academic teaching hospital with approximately 30,000 ED encounters per year. The study was approved with a waiver of informed consent by the UCSF Human Research Protection Program.

Participants

We use an Epic-based EHR platform (Epic 2017, Epic Systems Corporation) for clinical care, which was implemented on June 1, 2012. All data elements were obtained from Clarity, the relational database that stores Epic’s inpatient data. The study included encounters for patients age ≥18 years who had blood cultures ordered within 24 hours of ED presentation and administration of intravenous antibiotics within 24 hours. Repeat encounters were treated independently in our analysis.

Outcomes and Measures

We compared the ability of qSOFA, the Shock Index, and NEWS2 to predict in-hospital mortality and admission to the ICU from the ED (ED-to-ICU admission). We used the most abnormal vital signs and clinical assessments gathered within 30 minutes of ED presentation to identify patients who were qSOFA-positive, Shock Index-positive, and NEWS2-positive based on standard cut-points of risk. Data elements used to calculate qSOFA, Shock Index, and NEWS2 included blood pressure, heart rate, respiratory rate, Glasgow Coma Scale (GCS) score, oxygen saturation, requirement for supplemental oxygen, and temperature (Table 1). Patients were considered positive if they had a qSOFAscore ≥2, Shock Index of >0.7, and NEWS2 ≥5 based on triage vital signs.7,15,16 We considered patients to have altered mental status, a criterion used for NEWS2, if they had a GCS score <15 instead of using the “alert, verbal, confusion, pain, unresponsive” scale, which is not captured in our EHR, a method that has been used in earlier studies.17,18 Missing assessments were considered normal. Although our primary analysis focused on the scores calculated within 30 minutes of ED presentation, we performed a sensitivity analysis examining scores calculated within 1 hour of ED presentation in the event of a delay in gathering triage vital sign data.

Data Elements Collected in the 72 Hours After Emergency Department Presentation to Identify Patients Meeting qSOFA, Shock Index, and NEWS2 Criteria

We compared demographic and clinical characteristics of patients who were positive for qSOFA, the Shock Index, and NEWS2. Demographic data were extracted from the EHR and included primary language, age, sex, and insurance status. All International Classification of Diseases (ICD)-9/10 diagnosis codes were pulled from Clarity billing tables. We used the Elixhauser comorbidity groupings19 of ICD-9/10 codes present on admission to identify preexisting comorbidities and underlying organ dysfunction. To estimate burden of comorbid illnesses, we calculated the validated van Walraven comorbidity index,20 which provides an estimated risk of in-hospital death based on documented Elixhauser comorbidities. Admission level of care (acute, stepdown, or intensive care) was collected for inpatient admissions to assess initial illness severity.21 We also evaluated discharge disposition and in-hospital mortality. Index blood culture results were collected, and dates and timestamps of mechanical ventilation, fluid, vasopressor, and antibiotic administration were obtained for the duration of the encounter.

UCSF uses an automated, real-time, algorithm-based severe sepsis alert that is triggered when a patient meets ≥2 SIRS criteria and again when the patient meets severe sepsis or septic shock criteria (ie, ≥2 SIRS criteria in addition to end-organ dysfunction and/or fluid nonresponsive hypotension). This sepsis screening alert was in use for the duration of our study.22

Statistical Analysis

We performed a subgroup analysis among those who were diagnosed with sepsis, according to the 2016 Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3) criteria. Sepsis is defined as a change in Sequential Organ Failure Assessment (SOFA) score of ≥2 points within the first 48 hours.23 Additionally, patients meeting Sepsis-3 criteria needed to (1) receive ≥4 days of sequential antibiotic therapy or experience death or discharge to hospice before 4 days of antibiotic therapy or (2) have a validated sepsis discharge billing code. These parameters were added to increase the specificity of our sample.24

All statistical analyses were conducted using Stata 14 (StataCorp). We summarized differences in demographic and clinical characteristics among the populations meeting each severity score but elected not to conduct hypothesis testing because patients could be positive for one or more scores. We calculated sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for each score to predict in-hospital mortality and ED-to-ICU admission. To allow comparison with other studies, we also created a composite outcome of either in-hospital mortality or ED-to-ICU admission. To assess score discrimination to predict in-hospital mortality and ED-to-ICU admission, we calculated the area under the receiver operating characteristic curve (AUROC) along with asymptotic normal 95% CI using the “roctab” command considering a binary cut-point, as well as the full range of scores measured in the cohort. The AUROC range from 0.50 to 1.00 and a score in the 0.70 to 0.80 range can be considered fair.25 We assessed significant differences between severity score AUROCs using the DeLong method26 implemented through Stata 14’s “roccomp” command. As a sensitivity analysis, we explored whether the standard cut-points for qSOFA, the Shock Index, and NEWS2 provided the highest AUROC in our population by calculating test characteristics for several score cut-points.

RESULTS

Within our sample 23,837 ED patients had blood cultures ordered within 24 hours of ED presentation and were considered to have suspected sepsis. The mean age of the cohort was 60.8 years, and 1,612 (6.8%) had positive blood cultures. A total of 12,928 patients (54.2%) were found to have sepsis. We documented 1,427 in-hospital deaths (6.0%) and 3,149 (13.2%) ED-to-ICU admissions. At ED triage 1,921 (8.1%) were qSOFA-positive, 4,273 (17.9%) were Shock Index-positive, and 11,832 (49.6%) were NEWS2-positive. At ED triage, blood pressure, heart rate, respiratory rate, and oxygen saturated were documented in >99% of patients, 93.5% had temperature documented, and 28.5% had GCS recorded. If the window of assessment was widened to 1 hour, GCS was only documented among 44.2% of those with suspected sepsis.

Demographic Characteristics and Clinical Course

We identified significant differences when comparing demographic and clinical characteristics among patients who scored positive for the three severity measures at triage (Table 2). Although no hypothesis testing was conducted because patients could meet one or more scores, qSOFA-positive patients were older (median 70, 66, and 64 years, respectively), more likely to have Medicare as the primary payor (67.6% vs 59.7% vs 56.6%), to have chronic renal failure (26.1%, 23.1%, and 23.3%, respectively), to have a greater degree of underlying comorbidities based on the van Walraven Comorbidity Index (median 15, 12, and 11, respectively), and to be admitted to the ICU from the ED (48.1%, 36.3%, and 21.0%, respectively) compared with those positive for the Shock Index or NEWS2.

Demographic Characteristics and Clinical Course of Patients With Suspected Infection and Populations Who Met Standard Cut-Points of qSOFA, Shock Index, and NEWS2

qSOFA-positive patients received antibiotics more quickly than those who were Shock Index-positive or NEWS2-positive (median 1.5, 1.8, and 2.8 hours after admission, respectively). In addition, those who were qSOFA-positive were more likely to have a positive blood culture (10.9%, 9.4%, and 8.5%, respectively) and to receive an EHR-based diagnosis of sepsis (77.0%, 69.6%, and 60.9%, respectively) than those who were Shock Index- or NEWS2-positive. Those who were qSOFA-positive also were more likely to be mechanically ventilated during their hospital stay (25.4%, 19.2%, and 10.8%, respectively) and to receive vasopressors (33.5%, 22.5%, and 12.2%, respectively). In-hospital mortality also was more common among those who were qSOFA-positive at triage (23.4%, 15.3%, and 9.2%, respectively).

Because both qSOFA and NEWS2 incorporate GCS, we explored baseline characteristics of patients with GCS documented at triage (n = 6,794). These patients were older (median age 63 and 61 years, P < .0001), more likely to be male (54.9% and 53.4%, P = .0031), more likely to have renal failure (22.8% and 20.1%, P < .0001), more likely to have liver disease (14.2% and 12.8%, P = .006), had a higher van Walraven comorbidity score on presentation (median 10 and 8, P < .0001), and were more likely to go directly to the ICU from the ED (20.2% and 10.6%, P < .0001). However, among the 6,397 GCS scores documented at triage, only 1,579 (24.7%) were abnormal.

Test Characteristics of qSOFA, Shock Index, and NEWS2 for Predicting In-hospital Mortality and ED-to-ICU Admission

Among 23,837 patients with suspected sepsis, NEWS2 had the highest sensitivity for predicting in-hospital mortality (76.0%; 95% CI, 73.7%-78.2%) and ED-to-ICU admission (78.9%; 95% CI, 77.5%-80.4%) but had the lowest specificity for in-hospital mortality (52.0%; 95% CI, 51.4%-52.7%) and for ED-to-ICU admission (54.8%; 95% CI, 54.1%-55.5%) (Table 3). qSOFA had the lowest sensitivity for in-hospital mortality (31.5%; 95% CI, 29.1%-33.9%) and ED-to-ICU admission (29.3%; 95% CI, 27.7%-30.9%) but the highest specificity for in-hospital mortality (93.4%; 95% CI, 93.1%-93.8%) and ED-to-ICU admission (95.2%; 95% CI, 94.9%-95.5%). The Shock Index had a sensitivity that fell between qSOFA and NEWS2 for in-hospital mortality (45.8%; 95% CI, 43.2%-48.5%) and ED-to-ICU admission (49.2%; 95% CI, 47.5%-51.0%). The specificity of the Shock Index also was between qSOFA and NEWS2 for in-hospital mortality (83.9%; 95% CI, 83.4%-84.3%) and ED-to-ICU admission (86.8%; 95% CI, 86.4%-87.3%). All three scores exhibited relatively low PPV, ranging from 9.2% to 23.4% for in-hospital mortality and 21.0% to 48.0% for ED-to-ICU triage. Conversely, all three scores exhibited relatively high NPV, ranging from 95.5% to 97.1% for in-hospital mortality and 89.8% to 94.5% for ED-to-ICU triage. The patterns in sensitivity and specificity for in-hospital mortality and ED-to-ICU admission were similar among the 12,928 patients who received an EHR-based sepsis diagnosis with the tests generally demonstrating lower specificities, higher PPVs, and lower NPVs (Table 3).

Test Characteristics of Point-of-Care Severity Scores for Predicting Prognosis in ED Patients Presenting With Suspected Sepsis and Those With EHR-Based Sepsis Diagnosis

When considering a binary cutoff, the Shock Index exhibited the highest AUROC for in-hospital mortality (0.648; 95% CI, 0.635-0.662) and had a significantly higher AUROC than qSOFA (AUROC, 0.625; 95% CI, 0.612-0.637; P = .0005), but there was no difference compared with NEWS2 (AUROC, 0.640; 95% CI, 0.628-0.652; P = .2112). NEWS2 had a significantly higher AUROC than qSOFA for predicting in-hospital mortality (P = .0227). The Shock Index also exhibited the highest AUROC for ED-to-ICU admission (0.680; 95% CI, 0.617-0.689), which was significantly higher than the AUROC for qSOFA (P < .0001) and NEWS2 (P = 0.0151). NEWS2 had a significantly higher AUROC than qSOFA for predicting ED-to-ICU admission (P < .0001). Similar findings were seen in patients found to have sepsis. When considering the range of possible scores measured in our cohort, qSOFA and NEWS2 exhibited higher AUROCs for in-hospital mortality and ED-to-ICU admission than the Shock Index among patients with suspected infection and the subgroup with a sepsis diagnosis (Figure).

Receiver Operator Characteristic Curves for qSOFA, Shock Index, and NEWS2
The AUROCs of the scores were statistically significantly different for both in-hospital mortality (P = .0026) and ED-to-ICU admission (P < .0001). As a sensitivity analysis, we varied the binary cut-points of qSOFA, the Shock Index, and NEWS2 and calculated test characteristics within our study cohort (Appendix). For our institution, the qSOFA cut-point with the highest AUROC would be qSOFA > 0 for both in-hospital mortality (AUROC, 0.699; 95% CI, 0.687-0.711) and ED-to-ICU admission (AUROC, 0.716; 95% CI, 0.707-0.724), with 36.5% of the cohort meeting qSOFA. The NEWS2 cut-point with the highest AUROC would be NEWS2 ≥7 for both in-hospital mortality (AUROC, 0.653; 95% CI, 0.640-0.666) and ED-to-ICU admission (AUROC, 0.677; 95% CI, 0.668-0.686), with 20.3% of the cohort meeting NEWS2 at this cut-point. The standard Shock Index cut-point ≥0.7 exhibited the highest AUROC for in-hospital mortality and ED-to-ICU admission at our institution.

DISCUSSION

In this retrospective cohort study of 23,837 patients who presented to the ED with suspected sepsis, the standard qSOFA threshold was met least frequently, followed by the Shock Index and NEWS2. NEWS2 had the highest sensitivity but the lowest specificity for predicting in-hospital mortality and ED-to-ICU admission, making it a challenging bedside risk stratification scale for identifying patients at risk of poor clinical outcomes. When comparing predictive performance among the three scales, qSOFA had the highest specificity and the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission in this cohort of patients with suspected sepsis. These trends in sensitivity, specificity, and AUROC were consistent among those who met EHR criteria for a sepsis diagnosis. In the analysis of the three scoring systems using all available cut-points, qSOFA and NEWS2 had the highest AUROCs, followed by the Shock Index.

Considering the rapid progression from organ dysfunction to death in sepsis patients, as well as the difficulty establishing a sepsis diagnosis at triage,23 providers must quickly identify patients at increased risk of poor outcomes when they present to the ED. Sepsis alerts often are built using SIRS criteria,27 including the one used for sepsis surveillance at UCSF since 2012,22 but the white blood cell count criterion is subject to a laboratory lag and could lead to a delay in identification. Implementation of a point-of-care bedside score alert that uses readily available clinical data could allow providers to identify patients at greatest risk of poor outcomes immediately at ED presentation and triage, which motivated us to explore the predictive performance of qSOFA, the Shock Index, and NEWS2.

Our study is the first to provide a head-to-head comparison of the predictive performance of qSOFA, the Shock Index, and NEWS2, three easy-to-calculate bedside risk scores that use EHR data collected among patients with suspected sepsis. The Sepsis-3 guidelines recommend qSOFA to quickly identify non-ICU patients at greatest risk of poor outcomes because the measure exhibited predictive performance similar to the more extensive SOFA score outside the ICU.16,23 Although some studies have confirmed qSOFA’s high predictive performance,28-31 our test characteristics and AUROC findings are in line with other published analyses.4,6,10,17 The UK National Health Service is using NEWS2 to screen for patients at risk of poor outcomes from sepsis. Several analyses that assessed the predictive ability of NEWS have reported estimates in line with our findings.4,10,32 The Shock Index was introduced in 1967 and provided a metric to evaluate hemodynamic stability based on heart rate and systolic blood pressure.33 The Shock Index has been studied in several contexts, including sepsis,34 and studies show that a sustained Shock Index is associated with increased odds of vasopressor administration, higher prevalence of hyperlactatemia, and increased risk of poor outcomes in the ICU.13,14

For our study, we were particularly interested in exploring how the Shock Index would compare with more frequently used severity scores such as qSOFA and NEWS2 among patients with suspected sepsis, given the simplicity of its calculation and the easy availability of required data. In our cohort of 23,837 patients, only 159 people had missing blood pressure and only 71 had omitted heart rate. In contrast, both qSOFA and NEWS2 include an assessment of level of consciousness that can be subject to variability in assessment methods and EHR documentation across institutions.11 In our cohort, GCS within 30 minutes of ED presentation was missing in 72 patients, which could have led to incomplete calculation of qSOFA and NEWS2 if a missing value was not actually within normal limits.

Several investigations relate qSOFA to NEWS but few compare qSOFA with the newer NEWS2, and even fewer evaluate the Shock Index with any of these scores.10,11,18,29,35-37 In general, studies have shown that NEWS exhibits a higher AUROC for predicting mortality, sepsis with organ dysfunction, and ICU admission, often as a composite outcome.4,11,18,37,38 A handful of studies compare the Shock Index to SIRS; however, little has been done to compare the Shock Index to qSOFA or NEWS2, scores that have been used specifically for sepsis and might be more predictive of poor outcomes than SIRS.33 In our study, the Shock Index had a higher AUROC than either qSOFA or NEWS2 for predicting in-hospital mortality and ED-to-ICU admission measured as separate outcomes and as a composite outcome using standard cut-points for these scores.

When selecting a severity score to apply in an institution, it is important to carefully evaluate the score’s test characteristics, in addition to considering the availability of reliable data. Tests with high sensitivity and NPV for the population being studied can be useful to rule out disease or risk of poor outcome, while tests with high specificity and PPV can be useful to rule in disease or risk of poor outcome.39 When considering specificity, qSOFA’s performance was superior to the Shock Index and NEWS2 in our study, but a small percentage of the population was identified using a cut-point of qSOFA ≥2. If we used qSOFA and applied this standard cut-point at our institution, we could be confident that those identified were at increased risk, but we would miss a significant number of patients who would experience a poor outcome. When considering sensitivity, performance of NEWS2 was superior to qSOFA and the Shock Index in our study, but one-half of the population was identified using a cut-point of NEWS2 ≥5. If we were to apply this standard NEWS2 cut-point at our institution, we would assume that one-half of our population was at risk, which might drive resource use towards patients who will not experience a poor outcome. Although none of the scores exhibited a robust AUROC measure, the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission when using the standard binary cut-point, and its sensitivity and specificity is between that of qSOFA and NEWS2, potentially making it a score to use in settings where qSOFA and NEWS2 score components, such as altered mentation, are not reliably collected. Finally, our sensitivity analysis varying the binary cut-point of each score within our population demonstrated that the standard cut-points might not be as useful within a specific population and might need to be tailored for implementation, balancing sensitivity, specificity, PPV, and NPV to meet local priorities and ICU capacity.

Our study has limitations. It is a single-center, retrospective analysis, factors that could reduce generalizability. However, it does include a large and diverse patient population spanning several years. Missing GCS data could have affected the predictive ability of qSOFA and NEWS2 in our cohort. We could not reliably perform imputation of GCS because of the high missingness and therefore we assumed missing was normal, as was done in the Sepsis-3 derivation studies.16 Previous studies have attempted to impute GCS and have not observed improved performance of qSOFA to predict mortality.40 Because manually collected variables such as GCS are less reliably documented in the EHR, there might be limitations in their use for triage risk scores.

Although the current analysis focused on the predictive performance of qSOFA, the Shock Index, and NEWS2 at triage, performance of these scores could affect the ED team’s treatment decisions before handoff to the hospitalist team and the expected level of care the patient will receive after in-patient admission. These tests also have the advantage of being easy to calculate at the bedside over time, which could provide an objective assessment of longitudinal predicted prognosis. Future work should assess the longitudinal performance of each of these scores among those with suspected sepsis and to determine the impact using these scores would have on clinical and resource utilization outcomes.

CONCLUSION

Local priorities should drive selection of a screening tool, balancing sensitivity, specificity, PPV, and NPV to achieve the institution’s goals. qSOFA, Shock Index, and NEWS2 are risk stratification tools that can be easily implemented at ED triage using data available at the bedside. Although none of these scores performed strongly when comparing AUROCs, qSOFA was highly specific for identifying patients with poor outcomes, and NEWS2 was the most sensitive for ruling out those at high risk among patients with suspected sepsis. The Shock Index exhibited a sensitivity and specificity that fell between qSOFA and NEWS2 and also might be considered to identify those at increased risk, given its ease of implementation, particularly in settings where altered mentation is unreliably or inconsistently documented.

Acknowledgment

The authors thank the UCSF Division of Hospital Medicine Data Core for their assistance with data acquisition.

Sepsis is the leading cause of in-hospital mortality in the United States.1 Sepsis is present on admission in 85% of cases, and each hour delay in antibiotic treatment is associated with 4% to 7% increased odds of mortality.2,3 Prompt identification and treatment of sepsis is essential for reducing morbidity and mortality, but identifying sepsis during triage is challenging.2

Risk stratification scores that rely solely on data readily available at the bedside have been developed to quickly identify those at greatest risk of poor outcomes from sepsis in real time. The quick Sequential Organ Failure Assessment (qSOFA) score, the National Early Warning System (NEWS2), and the Shock Index are easy-to-calculate measures that use routinely collected clinical data that are not subject to laboratory delay. These scores can be incorporated into electronic health record (EHR)-based alerts and can be calculated longitudinally to track the risk of poor outcomes over time. qSOFA was developed to quantify patient risk at bedside in non-intensive care unit (ICU) settings, but there is no consensus about its ability to predict adverse outcomes such as mortality and ICU admission.4-6 The United Kingdom’s National Health Service uses NEWS2 to identify patients at risk for sepsis.7 NEWS has been shown to have similar or better sensitivity in identifying poorer outcomes in sepsis patients compared with systemic inflammatory response syndrome (SIRS) criteria and qSOFA.4,8-11 However, since the latest update of NEWS2 in 2017, there has been little study of its predictive ability. The Shock Index is a simple bedside score (heart rate divided by systolic blood pressure) that was developed to detect changes in cardiovascular performance before systemic shock onset. Although it was not developed for infection and has not been regularly applied in the sepsis literature, the Shock Index might be useful for identifying patients at increased risk of poor outcomes. Patients with higher and sustained Shock Index scores are more likely to experience morbidity, such as hyperlactatemia, vasopressor use, and organ failure, and also have an increased risk of mortality.12-14

Although the predictive abilities of these bedside risk stratification scores have been assessed individually using standard binary cut-points, the comparative performance of qSOFA, the Shock Index, and NEWS2 has not been evaluated in patients presenting to an emergency department (ED) with suspected sepsis. Our objective was to provide a head-to-head comparison of the test characteristics of qSOFA, the Shock Index, and NEWS2 calculated at ED triage for predicting in-hospital mortality and ED-to-ICU admission in patients with suspected sepsis to help health systems and providers select screening measures.

METHODS

Design and Setting

We conducted a retrospective cohort study of ED patients who presented with suspected sepsis to the University of California San Francisco (UCSF) Helen Diller Medical Center at Parnassus Heights between June 1, 2012, and December 31, 2018. Our institution is a 785-bed academic teaching hospital with approximately 30,000 ED encounters per year. The study was approved with a waiver of informed consent by the UCSF Human Research Protection Program.

Participants

We use an Epic-based EHR platform (Epic 2017, Epic Systems Corporation) for clinical care, which was implemented on June 1, 2012. All data elements were obtained from Clarity, the relational database that stores Epic’s inpatient data. The study included encounters for patients age ≥18 years who had blood cultures ordered within 24 hours of ED presentation and administration of intravenous antibiotics within 24 hours. Repeat encounters were treated independently in our analysis.

Outcomes and Measures

We compared the ability of qSOFA, the Shock Index, and NEWS2 to predict in-hospital mortality and admission to the ICU from the ED (ED-to-ICU admission). We used the most abnormal vital signs and clinical assessments gathered within 30 minutes of ED presentation to identify patients who were qSOFA-positive, Shock Index-positive, and NEWS2-positive based on standard cut-points of risk. Data elements used to calculate qSOFA, Shock Index, and NEWS2 included blood pressure, heart rate, respiratory rate, Glasgow Coma Scale (GCS) score, oxygen saturation, requirement for supplemental oxygen, and temperature (Table 1). Patients were considered positive if they had a qSOFAscore ≥2, Shock Index of >0.7, and NEWS2 ≥5 based on triage vital signs.7,15,16 We considered patients to have altered mental status, a criterion used for NEWS2, if they had a GCS score <15 instead of using the “alert, verbal, confusion, pain, unresponsive” scale, which is not captured in our EHR, a method that has been used in earlier studies.17,18 Missing assessments were considered normal. Although our primary analysis focused on the scores calculated within 30 minutes of ED presentation, we performed a sensitivity analysis examining scores calculated within 1 hour of ED presentation in the event of a delay in gathering triage vital sign data.

Data Elements Collected in the 72 Hours After Emergency Department Presentation to Identify Patients Meeting qSOFA, Shock Index, and NEWS2 Criteria

We compared demographic and clinical characteristics of patients who were positive for qSOFA, the Shock Index, and NEWS2. Demographic data were extracted from the EHR and included primary language, age, sex, and insurance status. All International Classification of Diseases (ICD)-9/10 diagnosis codes were pulled from Clarity billing tables. We used the Elixhauser comorbidity groupings19 of ICD-9/10 codes present on admission to identify preexisting comorbidities and underlying organ dysfunction. To estimate burden of comorbid illnesses, we calculated the validated van Walraven comorbidity index,20 which provides an estimated risk of in-hospital death based on documented Elixhauser comorbidities. Admission level of care (acute, stepdown, or intensive care) was collected for inpatient admissions to assess initial illness severity.21 We also evaluated discharge disposition and in-hospital mortality. Index blood culture results were collected, and dates and timestamps of mechanical ventilation, fluid, vasopressor, and antibiotic administration were obtained for the duration of the encounter.

UCSF uses an automated, real-time, algorithm-based severe sepsis alert that is triggered when a patient meets ≥2 SIRS criteria and again when the patient meets severe sepsis or septic shock criteria (ie, ≥2 SIRS criteria in addition to end-organ dysfunction and/or fluid nonresponsive hypotension). This sepsis screening alert was in use for the duration of our study.22

Statistical Analysis

We performed a subgroup analysis among those who were diagnosed with sepsis, according to the 2016 Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3) criteria. Sepsis is defined as a change in Sequential Organ Failure Assessment (SOFA) score of ≥2 points within the first 48 hours.23 Additionally, patients meeting Sepsis-3 criteria needed to (1) receive ≥4 days of sequential antibiotic therapy or experience death or discharge to hospice before 4 days of antibiotic therapy or (2) have a validated sepsis discharge billing code. These parameters were added to increase the specificity of our sample.24

All statistical analyses were conducted using Stata 14 (StataCorp). We summarized differences in demographic and clinical characteristics among the populations meeting each severity score but elected not to conduct hypothesis testing because patients could be positive for one or more scores. We calculated sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) for each score to predict in-hospital mortality and ED-to-ICU admission. To allow comparison with other studies, we also created a composite outcome of either in-hospital mortality or ED-to-ICU admission. To assess score discrimination to predict in-hospital mortality and ED-to-ICU admission, we calculated the area under the receiver operating characteristic curve (AUROC) along with asymptotic normal 95% CI using the “roctab” command considering a binary cut-point, as well as the full range of scores measured in the cohort. The AUROC range from 0.50 to 1.00 and a score in the 0.70 to 0.80 range can be considered fair.25 We assessed significant differences between severity score AUROCs using the DeLong method26 implemented through Stata 14’s “roccomp” command. As a sensitivity analysis, we explored whether the standard cut-points for qSOFA, the Shock Index, and NEWS2 provided the highest AUROC in our population by calculating test characteristics for several score cut-points.

RESULTS

Within our sample 23,837 ED patients had blood cultures ordered within 24 hours of ED presentation and were considered to have suspected sepsis. The mean age of the cohort was 60.8 years, and 1,612 (6.8%) had positive blood cultures. A total of 12,928 patients (54.2%) were found to have sepsis. We documented 1,427 in-hospital deaths (6.0%) and 3,149 (13.2%) ED-to-ICU admissions. At ED triage 1,921 (8.1%) were qSOFA-positive, 4,273 (17.9%) were Shock Index-positive, and 11,832 (49.6%) were NEWS2-positive. At ED triage, blood pressure, heart rate, respiratory rate, and oxygen saturated were documented in >99% of patients, 93.5% had temperature documented, and 28.5% had GCS recorded. If the window of assessment was widened to 1 hour, GCS was only documented among 44.2% of those with suspected sepsis.

Demographic Characteristics and Clinical Course

We identified significant differences when comparing demographic and clinical characteristics among patients who scored positive for the three severity measures at triage (Table 2). Although no hypothesis testing was conducted because patients could meet one or more scores, qSOFA-positive patients were older (median 70, 66, and 64 years, respectively), more likely to have Medicare as the primary payor (67.6% vs 59.7% vs 56.6%), to have chronic renal failure (26.1%, 23.1%, and 23.3%, respectively), to have a greater degree of underlying comorbidities based on the van Walraven Comorbidity Index (median 15, 12, and 11, respectively), and to be admitted to the ICU from the ED (48.1%, 36.3%, and 21.0%, respectively) compared with those positive for the Shock Index or NEWS2.

Demographic Characteristics and Clinical Course of Patients With Suspected Infection and Populations Who Met Standard Cut-Points of qSOFA, Shock Index, and NEWS2

qSOFA-positive patients received antibiotics more quickly than those who were Shock Index-positive or NEWS2-positive (median 1.5, 1.8, and 2.8 hours after admission, respectively). In addition, those who were qSOFA-positive were more likely to have a positive blood culture (10.9%, 9.4%, and 8.5%, respectively) and to receive an EHR-based diagnosis of sepsis (77.0%, 69.6%, and 60.9%, respectively) than those who were Shock Index- or NEWS2-positive. Those who were qSOFA-positive also were more likely to be mechanically ventilated during their hospital stay (25.4%, 19.2%, and 10.8%, respectively) and to receive vasopressors (33.5%, 22.5%, and 12.2%, respectively). In-hospital mortality also was more common among those who were qSOFA-positive at triage (23.4%, 15.3%, and 9.2%, respectively).

Because both qSOFA and NEWS2 incorporate GCS, we explored baseline characteristics of patients with GCS documented at triage (n = 6,794). These patients were older (median age 63 and 61 years, P < .0001), more likely to be male (54.9% and 53.4%, P = .0031), more likely to have renal failure (22.8% and 20.1%, P < .0001), more likely to have liver disease (14.2% and 12.8%, P = .006), had a higher van Walraven comorbidity score on presentation (median 10 and 8, P < .0001), and were more likely to go directly to the ICU from the ED (20.2% and 10.6%, P < .0001). However, among the 6,397 GCS scores documented at triage, only 1,579 (24.7%) were abnormal.

Test Characteristics of qSOFA, Shock Index, and NEWS2 for Predicting In-hospital Mortality and ED-to-ICU Admission

Among 23,837 patients with suspected sepsis, NEWS2 had the highest sensitivity for predicting in-hospital mortality (76.0%; 95% CI, 73.7%-78.2%) and ED-to-ICU admission (78.9%; 95% CI, 77.5%-80.4%) but had the lowest specificity for in-hospital mortality (52.0%; 95% CI, 51.4%-52.7%) and for ED-to-ICU admission (54.8%; 95% CI, 54.1%-55.5%) (Table 3). qSOFA had the lowest sensitivity for in-hospital mortality (31.5%; 95% CI, 29.1%-33.9%) and ED-to-ICU admission (29.3%; 95% CI, 27.7%-30.9%) but the highest specificity for in-hospital mortality (93.4%; 95% CI, 93.1%-93.8%) and ED-to-ICU admission (95.2%; 95% CI, 94.9%-95.5%). The Shock Index had a sensitivity that fell between qSOFA and NEWS2 for in-hospital mortality (45.8%; 95% CI, 43.2%-48.5%) and ED-to-ICU admission (49.2%; 95% CI, 47.5%-51.0%). The specificity of the Shock Index also was between qSOFA and NEWS2 for in-hospital mortality (83.9%; 95% CI, 83.4%-84.3%) and ED-to-ICU admission (86.8%; 95% CI, 86.4%-87.3%). All three scores exhibited relatively low PPV, ranging from 9.2% to 23.4% for in-hospital mortality and 21.0% to 48.0% for ED-to-ICU triage. Conversely, all three scores exhibited relatively high NPV, ranging from 95.5% to 97.1% for in-hospital mortality and 89.8% to 94.5% for ED-to-ICU triage. The patterns in sensitivity and specificity for in-hospital mortality and ED-to-ICU admission were similar among the 12,928 patients who received an EHR-based sepsis diagnosis with the tests generally demonstrating lower specificities, higher PPVs, and lower NPVs (Table 3).

Test Characteristics of Point-of-Care Severity Scores for Predicting Prognosis in ED Patients Presenting With Suspected Sepsis and Those With EHR-Based Sepsis Diagnosis

When considering a binary cutoff, the Shock Index exhibited the highest AUROC for in-hospital mortality (0.648; 95% CI, 0.635-0.662) and had a significantly higher AUROC than qSOFA (AUROC, 0.625; 95% CI, 0.612-0.637; P = .0005), but there was no difference compared with NEWS2 (AUROC, 0.640; 95% CI, 0.628-0.652; P = .2112). NEWS2 had a significantly higher AUROC than qSOFA for predicting in-hospital mortality (P = .0227). The Shock Index also exhibited the highest AUROC for ED-to-ICU admission (0.680; 95% CI, 0.617-0.689), which was significantly higher than the AUROC for qSOFA (P < .0001) and NEWS2 (P = 0.0151). NEWS2 had a significantly higher AUROC than qSOFA for predicting ED-to-ICU admission (P < .0001). Similar findings were seen in patients found to have sepsis. When considering the range of possible scores measured in our cohort, qSOFA and NEWS2 exhibited higher AUROCs for in-hospital mortality and ED-to-ICU admission than the Shock Index among patients with suspected infection and the subgroup with a sepsis diagnosis (Figure).

Receiver Operator Characteristic Curves for qSOFA, Shock Index, and NEWS2
The AUROCs of the scores were statistically significantly different for both in-hospital mortality (P = .0026) and ED-to-ICU admission (P < .0001). As a sensitivity analysis, we varied the binary cut-points of qSOFA, the Shock Index, and NEWS2 and calculated test characteristics within our study cohort (Appendix). For our institution, the qSOFA cut-point with the highest AUROC would be qSOFA > 0 for both in-hospital mortality (AUROC, 0.699; 95% CI, 0.687-0.711) and ED-to-ICU admission (AUROC, 0.716; 95% CI, 0.707-0.724), with 36.5% of the cohort meeting qSOFA. The NEWS2 cut-point with the highest AUROC would be NEWS2 ≥7 for both in-hospital mortality (AUROC, 0.653; 95% CI, 0.640-0.666) and ED-to-ICU admission (AUROC, 0.677; 95% CI, 0.668-0.686), with 20.3% of the cohort meeting NEWS2 at this cut-point. The standard Shock Index cut-point ≥0.7 exhibited the highest AUROC for in-hospital mortality and ED-to-ICU admission at our institution.

DISCUSSION

In this retrospective cohort study of 23,837 patients who presented to the ED with suspected sepsis, the standard qSOFA threshold was met least frequently, followed by the Shock Index and NEWS2. NEWS2 had the highest sensitivity but the lowest specificity for predicting in-hospital mortality and ED-to-ICU admission, making it a challenging bedside risk stratification scale for identifying patients at risk of poor clinical outcomes. When comparing predictive performance among the three scales, qSOFA had the highest specificity and the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission in this cohort of patients with suspected sepsis. These trends in sensitivity, specificity, and AUROC were consistent among those who met EHR criteria for a sepsis diagnosis. In the analysis of the three scoring systems using all available cut-points, qSOFA and NEWS2 had the highest AUROCs, followed by the Shock Index.

Considering the rapid progression from organ dysfunction to death in sepsis patients, as well as the difficulty establishing a sepsis diagnosis at triage,23 providers must quickly identify patients at increased risk of poor outcomes when they present to the ED. Sepsis alerts often are built using SIRS criteria,27 including the one used for sepsis surveillance at UCSF since 2012,22 but the white blood cell count criterion is subject to a laboratory lag and could lead to a delay in identification. Implementation of a point-of-care bedside score alert that uses readily available clinical data could allow providers to identify patients at greatest risk of poor outcomes immediately at ED presentation and triage, which motivated us to explore the predictive performance of qSOFA, the Shock Index, and NEWS2.

Our study is the first to provide a head-to-head comparison of the predictive performance of qSOFA, the Shock Index, and NEWS2, three easy-to-calculate bedside risk scores that use EHR data collected among patients with suspected sepsis. The Sepsis-3 guidelines recommend qSOFA to quickly identify non-ICU patients at greatest risk of poor outcomes because the measure exhibited predictive performance similar to the more extensive SOFA score outside the ICU.16,23 Although some studies have confirmed qSOFA’s high predictive performance,28-31 our test characteristics and AUROC findings are in line with other published analyses.4,6,10,17 The UK National Health Service is using NEWS2 to screen for patients at risk of poor outcomes from sepsis. Several analyses that assessed the predictive ability of NEWS have reported estimates in line with our findings.4,10,32 The Shock Index was introduced in 1967 and provided a metric to evaluate hemodynamic stability based on heart rate and systolic blood pressure.33 The Shock Index has been studied in several contexts, including sepsis,34 and studies show that a sustained Shock Index is associated with increased odds of vasopressor administration, higher prevalence of hyperlactatemia, and increased risk of poor outcomes in the ICU.13,14

For our study, we were particularly interested in exploring how the Shock Index would compare with more frequently used severity scores such as qSOFA and NEWS2 among patients with suspected sepsis, given the simplicity of its calculation and the easy availability of required data. In our cohort of 23,837 patients, only 159 people had missing blood pressure and only 71 had omitted heart rate. In contrast, both qSOFA and NEWS2 include an assessment of level of consciousness that can be subject to variability in assessment methods and EHR documentation across institutions.11 In our cohort, GCS within 30 minutes of ED presentation was missing in 72 patients, which could have led to incomplete calculation of qSOFA and NEWS2 if a missing value was not actually within normal limits.

Several investigations relate qSOFA to NEWS but few compare qSOFA with the newer NEWS2, and even fewer evaluate the Shock Index with any of these scores.10,11,18,29,35-37 In general, studies have shown that NEWS exhibits a higher AUROC for predicting mortality, sepsis with organ dysfunction, and ICU admission, often as a composite outcome.4,11,18,37,38 A handful of studies compare the Shock Index to SIRS; however, little has been done to compare the Shock Index to qSOFA or NEWS2, scores that have been used specifically for sepsis and might be more predictive of poor outcomes than SIRS.33 In our study, the Shock Index had a higher AUROC than either qSOFA or NEWS2 for predicting in-hospital mortality and ED-to-ICU admission measured as separate outcomes and as a composite outcome using standard cut-points for these scores.

When selecting a severity score to apply in an institution, it is important to carefully evaluate the score’s test characteristics, in addition to considering the availability of reliable data. Tests with high sensitivity and NPV for the population being studied can be useful to rule out disease or risk of poor outcome, while tests with high specificity and PPV can be useful to rule in disease or risk of poor outcome.39 When considering specificity, qSOFA’s performance was superior to the Shock Index and NEWS2 in our study, but a small percentage of the population was identified using a cut-point of qSOFA ≥2. If we used qSOFA and applied this standard cut-point at our institution, we could be confident that those identified were at increased risk, but we would miss a significant number of patients who would experience a poor outcome. When considering sensitivity, performance of NEWS2 was superior to qSOFA and the Shock Index in our study, but one-half of the population was identified using a cut-point of NEWS2 ≥5. If we were to apply this standard NEWS2 cut-point at our institution, we would assume that one-half of our population was at risk, which might drive resource use towards patients who will not experience a poor outcome. Although none of the scores exhibited a robust AUROC measure, the Shock Index had the highest AUROC for in-hospital mortality and ED-to-ICU admission when using the standard binary cut-point, and its sensitivity and specificity is between that of qSOFA and NEWS2, potentially making it a score to use in settings where qSOFA and NEWS2 score components, such as altered mentation, are not reliably collected. Finally, our sensitivity analysis varying the binary cut-point of each score within our population demonstrated that the standard cut-points might not be as useful within a specific population and might need to be tailored for implementation, balancing sensitivity, specificity, PPV, and NPV to meet local priorities and ICU capacity.

Our study has limitations. It is a single-center, retrospective analysis, factors that could reduce generalizability. However, it does include a large and diverse patient population spanning several years. Missing GCS data could have affected the predictive ability of qSOFA and NEWS2 in our cohort. We could not reliably perform imputation of GCS because of the high missingness and therefore we assumed missing was normal, as was done in the Sepsis-3 derivation studies.16 Previous studies have attempted to impute GCS and have not observed improved performance of qSOFA to predict mortality.40 Because manually collected variables such as GCS are less reliably documented in the EHR, there might be limitations in their use for triage risk scores.

Although the current analysis focused on the predictive performance of qSOFA, the Shock Index, and NEWS2 at triage, performance of these scores could affect the ED team’s treatment decisions before handoff to the hospitalist team and the expected level of care the patient will receive after in-patient admission. These tests also have the advantage of being easy to calculate at the bedside over time, which could provide an objective assessment of longitudinal predicted prognosis. Future work should assess the longitudinal performance of each of these scores among those with suspected sepsis and to determine the impact using these scores would have on clinical and resource utilization outcomes.

CONCLUSION

Local priorities should drive selection of a screening tool, balancing sensitivity, specificity, PPV, and NPV to achieve the institution’s goals. qSOFA, Shock Index, and NEWS2 are risk stratification tools that can be easily implemented at ED triage using data available at the bedside. Although none of these scores performed strongly when comparing AUROCs, qSOFA was highly specific for identifying patients with poor outcomes, and NEWS2 was the most sensitive for ruling out those at high risk among patients with suspected sepsis. The Shock Index exhibited a sensitivity and specificity that fell between qSOFA and NEWS2 and also might be considered to identify those at increased risk, given its ease of implementation, particularly in settings where altered mentation is unreliably or inconsistently documented.

Acknowledgment

The authors thank the UCSF Division of Hospital Medicine Data Core for their assistance with data acquisition.

References

1. Jones SL, Ashton CM, Kiehne LB, et al. Outcomes and resource use of sepsis-associated stays by presence on admission, severity, and hospital type. Med Care. 2016;54(3):303-310. https://doi.org/10.1097/MLR.0000000000000481
2. Seymour CW, Gesten F, Prescott HC, et al. Time to treatment and mortality during mandated emergency care for sepsis. N Engl J Med. 2017;376(23):2235-2244. https://doi.org/10.1056/NEJMoa1703058
3. Kumar A, Roberts D, Wood KE, et al. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit Care Med. 2006;34(6):1589-1596. https://doi.org/10.1097/01.CCM.0000217961.75225.E9
4. Churpek MM, Snyder A, Sokol S, Pettit NN, Edelson DP. Investigating the impact of different suspicion of infection criteria on the accuracy of Quick Sepsis-Related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores. Crit Care Med. 2017;45(11):1805-1812. https://doi.org/10.1097/CCM.0000000000002648
5. Abdullah SMOB, Sørensen RH, Dessau RBC, Sattar SMRU, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
6. Kim KS, Suh GJ, Kim K, et al. Quick Sepsis-related Organ Failure Assessment score is not sensitive enough to predict 28-day mortality in emergency department patients with sepsis: a retrospective review. Clin Exp Emerg Med. 2019;6(1):77-83. HTTPS://DOI.ORG/ 10.15441/ceem.17.294
7. National Early Warning Score (NEWS) 2: Standardising the assessment of acute-illness severity in the NHS. Royal College of Physicians; 2017.
8. Brink A, Alsma J, Verdonschot RJCG, et al. Predicting mortality in patients with suspected sepsis at the emergency department: a retrospective cohort study comparing qSOFA, SIRS and National Early Warning Score. PLoS One. 2019;14(1):e0211133. https://doi.org/ 10.1371/journal.pone.0211133
9. Redfern OC, Smith GB, Prytherch DR, Meredith P, Inada-Kim M, Schmidt PE. A comparison of the Quick Sequential (Sepsis-Related) Organ Failure Assessment Score and the National Early Warning Score in non-ICU patients with/without infection. Crit Care Med. 2018;46(12):1923-1933. https://doi.org/10.1097/CCM.0000000000003359
10. Churpek MM, Snyder A, Han X, et al. Quick Sepsis-related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores for detecting clinical deterioration in infected patients outside the intensive care unit. Am J Respir Crit Care Med. 2017;195(7):906-911. https://doi.org/10.1164/rccm.201604-0854OC
11. Goulden R, Hoyle MC, Monis J, et al. qSOFA, SIRS and NEWS for predicting inhospital mortality and ICU admission in emergency admissions treated as sepsis. Emerg Med J. 2018;35(6):345-349. https://doi.org/10.1136/emermed-2017-207120
12. Biney I, Shepherd A, Thomas J, Mehari A. Shock Index and outcomes in patients admitted to the ICU with sepsis. Chest. 2015;148(suppl 4):337A. https://doi.org/https://doi.org/10.1378/chest.2281151
13. Wira CR, Francis MW, Bhat S, Ehrman R, Conner D, Siegel M. The shock index as a predictor of vasopressor use in emergency department patients with severe sepsis. West J Emerg Med. 2014;15(1):60-66. https://doi.org/10.5811/westjem.2013.7.18472
14. Berger T, Green J, Horeczko T, et al. Shock index and early recognition of sepsis in the emergency department: pilot study. West J Emerg Med. 2013;14(2):168-174. https://doi.org/10.5811/westjem.2012.8.11546
15. Middleton DJ, Smith TO, Bedford R, Neilly M, Myint PK. Shock Index predicts outcome in patients with suspected sepsis or community-acquired pneumonia: a systematic review. J Clin Med. 2019;8(8):1144. https://doi.org/10.3390/jcm8081144
16. Seymour CW, Liu VX, Iwashyna TJ, et al. Assessment of clinical criteria for sepsis: for the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):762-774. https://doi.org/ 10.1001/jama.2016.0288
17. Abdullah S, Sørensen RH, Dessau RBC, Sattar S, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
18. Usman OA, Usman AA, Ward MA. Comparison of SIRS, qSOFA, and NEWS for the early identification of sepsis in the Emergency Department. Am J Emerg Med. 2018;37(8):1490-1497. https://doi.org/10.1016/j.ajem.2018.10.058
19. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Med Care. 1998;36(1):8-27. https://doi.org/10.1097/00005650-199801000-00004
20. van Walraven C, Austin PC, Jennings A, Quan H, Forster AJ. A modification of the Elixhauser comorbidity measures into a point system for hospital death using administrative data. Med Care. 2009;47(6):626-633. https://doi.org/10.1097/MLR.0b013e31819432e5
21. Prin M, Wunsch H. The role of stepdown beds in hospital care. Am J Respir Crit Care Med. 2014;190(11):1210-1216. https://doi.org/10.1164/rccm.201406-1117PP
22. Narayanan N, Gross AK, Pintens M, Fee C, MacDougall C. Effect of an electronic medical record alert for severe sepsis among ED patients. Am J Emerg Med. 2016;34(2):185-188. https://doi.org/10.1016/j.ajem.2015.10.005
23. Singer M, Deutschman CS, Seymour CW, et al. The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):801-810. https://doi.org/10.1001/jama.2016.0287
24. Rhee C, Dantes R, Epstein L, et al. Incidence and trends of sepsis in US hospitals using clinical vs claims data, 2009-2014. JAMA. 2017;318(13):1241-1249. https://doi.org/10.1001/jama.2017.13836
25. Safari S, Baratloo A, Elfil M, Negida A. Evidence based emergency medicine; part 5 receiver operating curve and area under the curve. Emerg (Tehran). 2016;4(2):111-113.
26. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988;44(3):837-845.
27. Kangas C, Iverson L, Pierce D. Sepsis screening: combining Early Warning Scores and SIRS Criteria. Clin Nurs Res. 2021;30(1):42-49. https://doi.org/10.1177/1054773818823334.
28. Freund Y, Lemachatti N, Krastinova E, et al. Prognostic accuracy of Sepsis-3 Criteria for in-hospital mortality among patients with suspected infection presenting to the emergency department. JAMA. 2017;317(3):301-308. https://doi.org/10.1001/jama.2016.20329
29. Finkelsztein EJ, Jones DS, Ma KC, et al. Comparison of qSOFA and SIRS for predicting adverse outcomes of patients with suspicion of sepsis outside the intensive care unit. Crit Care. 2017;21(1):73. https://doi.org/10.1186/s13054-017-1658-5
30. Canet E, Taylor DM, Khor R, Krishnan V, Bellomo R. qSOFA as predictor of mortality and prolonged ICU admission in Emergency Department patients with suspected infection. J Crit Care. 2018;48:118-123. https://doi.org/10.1016/j.jcrc.2018.08.022
31. Anand V, Zhang Z, Kadri SS, Klompas M, Rhee C; CDC Prevention Epicenters Program. Epidemiology of Quick Sequential Organ Failure Assessment criteria in undifferentiated patients and association with suspected infection and sepsis. Chest. 2019;156(2):289-297. https://doi.org/10.1016/j.chest.2019.03.032
32. Hamilton F, Arnold D, Baird A, Albur M, Whiting P. Early Warning Scores do not accurately predict mortality in sepsis: A meta-analysis and systematic review of the literature. J Infect. 2018;76(3):241-248. https://doi.org/10.1016/j.jinf.2018.01.002
33. Koch E, Lovett S, Nghiem T, Riggs RA, Rech MA. Shock Index in the emergency department: utility and limitations. Open Access Emerg Med. 2019;11:179-199. https://doi.org/10.2147/OAEM.S178358
34. Yussof SJ, Zakaria MI, Mohamed FL, Bujang MA, Lakshmanan S, Asaari AH. Value of Shock Index in prognosticating the short-term outcome of death for patients presenting with severe sepsis and septic shock in the emergency department. Med J Malaysia. 2012;67(4):406-411.
35. Siddiqui S, Chua M, Kumaresh V, Choo R. A comparison of pre ICU admission SIRS, EWS and q SOFA scores for predicting mortality and length of stay in ICU. J Crit Care. 2017;41:191-193. https://doi.org/10.1016/j.jcrc.2017.05.017
36. Costa RT, Nassar AP, Caruso P. Accuracy of SOFA, qSOFA, and SIRS scores for mortality in cancer patients admitted to an intensive care unit with suspected infection. J Crit Care. 2018;45:52-57. https://doi.org/10.1016/j.jcrc.2017.12.024
37. Mellhammar L, Linder A, Tverring J, et al. NEWS2 is Superior to qSOFA in detecting sepsis with organ dysfunction in the emergency department. J Clin Med. 2019;8(8):1128. https://doi.org/10.3390/jcm8081128
38. Szakmany T, Pugh R, Kopczynska M, et al. Defining sepsis on the wards: results of a multi-centre point-prevalence study comparing two sepsis definitions. Anaesthesia. 2018;73(2):195-204. https://doi.org/10.1111/anae.14062
39. Newman TB, Kohn MA. Evidence-Based Diagnosis: An Introduction to Clinical Epidemiology. Cambridge University Press; 2009.
40. Askim Å, Moser F, Gustad LT, et al. Poor performance of quick-SOFA (qSOFA) score in predicting severe sepsis and mortality - a prospective study of patients admitted with infection to the emergency department. Scand J Trauma Resusc Emerg Med. 2017;25(1):56. https://doi.org/10.1186/s13049-017-0399-4

References

1. Jones SL, Ashton CM, Kiehne LB, et al. Outcomes and resource use of sepsis-associated stays by presence on admission, severity, and hospital type. Med Care. 2016;54(3):303-310. https://doi.org/10.1097/MLR.0000000000000481
2. Seymour CW, Gesten F, Prescott HC, et al. Time to treatment and mortality during mandated emergency care for sepsis. N Engl J Med. 2017;376(23):2235-2244. https://doi.org/10.1056/NEJMoa1703058
3. Kumar A, Roberts D, Wood KE, et al. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit Care Med. 2006;34(6):1589-1596. https://doi.org/10.1097/01.CCM.0000217961.75225.E9
4. Churpek MM, Snyder A, Sokol S, Pettit NN, Edelson DP. Investigating the impact of different suspicion of infection criteria on the accuracy of Quick Sepsis-Related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores. Crit Care Med. 2017;45(11):1805-1812. https://doi.org/10.1097/CCM.0000000000002648
5. Abdullah SMOB, Sørensen RH, Dessau RBC, Sattar SMRU, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
6. Kim KS, Suh GJ, Kim K, et al. Quick Sepsis-related Organ Failure Assessment score is not sensitive enough to predict 28-day mortality in emergency department patients with sepsis: a retrospective review. Clin Exp Emerg Med. 2019;6(1):77-83. HTTPS://DOI.ORG/ 10.15441/ceem.17.294
7. National Early Warning Score (NEWS) 2: Standardising the assessment of acute-illness severity in the NHS. Royal College of Physicians; 2017.
8. Brink A, Alsma J, Verdonschot RJCG, et al. Predicting mortality in patients with suspected sepsis at the emergency department: a retrospective cohort study comparing qSOFA, SIRS and National Early Warning Score. PLoS One. 2019;14(1):e0211133. https://doi.org/ 10.1371/journal.pone.0211133
9. Redfern OC, Smith GB, Prytherch DR, Meredith P, Inada-Kim M, Schmidt PE. A comparison of the Quick Sequential (Sepsis-Related) Organ Failure Assessment Score and the National Early Warning Score in non-ICU patients with/without infection. Crit Care Med. 2018;46(12):1923-1933. https://doi.org/10.1097/CCM.0000000000003359
10. Churpek MM, Snyder A, Han X, et al. Quick Sepsis-related Organ Failure Assessment, Systemic Inflammatory Response Syndrome, and Early Warning Scores for detecting clinical deterioration in infected patients outside the intensive care unit. Am J Respir Crit Care Med. 2017;195(7):906-911. https://doi.org/10.1164/rccm.201604-0854OC
11. Goulden R, Hoyle MC, Monis J, et al. qSOFA, SIRS and NEWS for predicting inhospital mortality and ICU admission in emergency admissions treated as sepsis. Emerg Med J. 2018;35(6):345-349. https://doi.org/10.1136/emermed-2017-207120
12. Biney I, Shepherd A, Thomas J, Mehari A. Shock Index and outcomes in patients admitted to the ICU with sepsis. Chest. 2015;148(suppl 4):337A. https://doi.org/https://doi.org/10.1378/chest.2281151
13. Wira CR, Francis MW, Bhat S, Ehrman R, Conner D, Siegel M. The shock index as a predictor of vasopressor use in emergency department patients with severe sepsis. West J Emerg Med. 2014;15(1):60-66. https://doi.org/10.5811/westjem.2013.7.18472
14. Berger T, Green J, Horeczko T, et al. Shock index and early recognition of sepsis in the emergency department: pilot study. West J Emerg Med. 2013;14(2):168-174. https://doi.org/10.5811/westjem.2012.8.11546
15. Middleton DJ, Smith TO, Bedford R, Neilly M, Myint PK. Shock Index predicts outcome in patients with suspected sepsis or community-acquired pneumonia: a systematic review. J Clin Med. 2019;8(8):1144. https://doi.org/10.3390/jcm8081144
16. Seymour CW, Liu VX, Iwashyna TJ, et al. Assessment of clinical criteria for sepsis: for the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):762-774. https://doi.org/ 10.1001/jama.2016.0288
17. Abdullah S, Sørensen RH, Dessau RBC, Sattar S, Wiese L, Nielsen FE. Prognostic accuracy of qSOFA in predicting 28-day mortality among infected patients in an emergency department: a prospective validation study. Emerg Med J. 2019;36(12):722-728. https://doi.org/10.1136/emermed-2019-208456
18. Usman OA, Usman AA, Ward MA. Comparison of SIRS, qSOFA, and NEWS for the early identification of sepsis in the Emergency Department. Am J Emerg Med. 2018;37(8):1490-1497. https://doi.org/10.1016/j.ajem.2018.10.058
19. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Med Care. 1998;36(1):8-27. https://doi.org/10.1097/00005650-199801000-00004
20. van Walraven C, Austin PC, Jennings A, Quan H, Forster AJ. A modification of the Elixhauser comorbidity measures into a point system for hospital death using administrative data. Med Care. 2009;47(6):626-633. https://doi.org/10.1097/MLR.0b013e31819432e5
21. Prin M, Wunsch H. The role of stepdown beds in hospital care. Am J Respir Crit Care Med. 2014;190(11):1210-1216. https://doi.org/10.1164/rccm.201406-1117PP
22. Narayanan N, Gross AK, Pintens M, Fee C, MacDougall C. Effect of an electronic medical record alert for severe sepsis among ED patients. Am J Emerg Med. 2016;34(2):185-188. https://doi.org/10.1016/j.ajem.2015.10.005
23. Singer M, Deutschman CS, Seymour CW, et al. The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA. 2016;315(8):801-810. https://doi.org/10.1001/jama.2016.0287
24. Rhee C, Dantes R, Epstein L, et al. Incidence and trends of sepsis in US hospitals using clinical vs claims data, 2009-2014. JAMA. 2017;318(13):1241-1249. https://doi.org/10.1001/jama.2017.13836
25. Safari S, Baratloo A, Elfil M, Negida A. Evidence based emergency medicine; part 5 receiver operating curve and area under the curve. Emerg (Tehran). 2016;4(2):111-113.
26. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988;44(3):837-845.
27. Kangas C, Iverson L, Pierce D. Sepsis screening: combining Early Warning Scores and SIRS Criteria. Clin Nurs Res. 2021;30(1):42-49. https://doi.org/10.1177/1054773818823334.
28. Freund Y, Lemachatti N, Krastinova E, et al. Prognostic accuracy of Sepsis-3 Criteria for in-hospital mortality among patients with suspected infection presenting to the emergency department. JAMA. 2017;317(3):301-308. https://doi.org/10.1001/jama.2016.20329
29. Finkelsztein EJ, Jones DS, Ma KC, et al. Comparison of qSOFA and SIRS for predicting adverse outcomes of patients with suspicion of sepsis outside the intensive care unit. Crit Care. 2017;21(1):73. https://doi.org/10.1186/s13054-017-1658-5
30. Canet E, Taylor DM, Khor R, Krishnan V, Bellomo R. qSOFA as predictor of mortality and prolonged ICU admission in Emergency Department patients with suspected infection. J Crit Care. 2018;48:118-123. https://doi.org/10.1016/j.jcrc.2018.08.022
31. Anand V, Zhang Z, Kadri SS, Klompas M, Rhee C; CDC Prevention Epicenters Program. Epidemiology of Quick Sequential Organ Failure Assessment criteria in undifferentiated patients and association with suspected infection and sepsis. Chest. 2019;156(2):289-297. https://doi.org/10.1016/j.chest.2019.03.032
32. Hamilton F, Arnold D, Baird A, Albur M, Whiting P. Early Warning Scores do not accurately predict mortality in sepsis: A meta-analysis and systematic review of the literature. J Infect. 2018;76(3):241-248. https://doi.org/10.1016/j.jinf.2018.01.002
33. Koch E, Lovett S, Nghiem T, Riggs RA, Rech MA. Shock Index in the emergency department: utility and limitations. Open Access Emerg Med. 2019;11:179-199. https://doi.org/10.2147/OAEM.S178358
34. Yussof SJ, Zakaria MI, Mohamed FL, Bujang MA, Lakshmanan S, Asaari AH. Value of Shock Index in prognosticating the short-term outcome of death for patients presenting with severe sepsis and septic shock in the emergency department. Med J Malaysia. 2012;67(4):406-411.
35. Siddiqui S, Chua M, Kumaresh V, Choo R. A comparison of pre ICU admission SIRS, EWS and q SOFA scores for predicting mortality and length of stay in ICU. J Crit Care. 2017;41:191-193. https://doi.org/10.1016/j.jcrc.2017.05.017
36. Costa RT, Nassar AP, Caruso P. Accuracy of SOFA, qSOFA, and SIRS scores for mortality in cancer patients admitted to an intensive care unit with suspected infection. J Crit Care. 2018;45:52-57. https://doi.org/10.1016/j.jcrc.2017.12.024
37. Mellhammar L, Linder A, Tverring J, et al. NEWS2 is Superior to qSOFA in detecting sepsis with organ dysfunction in the emergency department. J Clin Med. 2019;8(8):1128. https://doi.org/10.3390/jcm8081128
38. Szakmany T, Pugh R, Kopczynska M, et al. Defining sepsis on the wards: results of a multi-centre point-prevalence study comparing two sepsis definitions. Anaesthesia. 2018;73(2):195-204. https://doi.org/10.1111/anae.14062
39. Newman TB, Kohn MA. Evidence-Based Diagnosis: An Introduction to Clinical Epidemiology. Cambridge University Press; 2009.
40. Askim Å, Moser F, Gustad LT, et al. Poor performance of quick-SOFA (qSOFA) score in predicting severe sepsis and mortality - a prospective study of patients admitted with infection to the emergency department. Scand J Trauma Resusc Emerg Med. 2017;25(1):56. https://doi.org/10.1186/s13049-017-0399-4

Issue
Journal of Hospital Medicine 16(8)
Issue
Journal of Hospital Medicine 16(8)
Page Number
453-461. Published Online Only July 21, 2021
Page Number
453-461. Published Online Only July 21, 2021
Publications
Publications
Topics
Article Type
Display Headline
Identifying the Sickest During Triage: Using Point-of-Care Severity Scores to Predict Prognosis in Emergency Department Patients With Suspected Sepsis
Display Headline
Identifying the Sickest During Triage: Using Point-of-Care Severity Scores to Predict Prognosis in Emergency Department Patients With Suspected Sepsis
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Priya A Prasad, PhD, MPH; [email protected].
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Page Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media
Media Files

Drawing Down From Crisis: More Lessons From a Soldier

Article Type
Changed
Thu, 09/30/2021 - 14:48
Display Headline
Drawing Down From Crisis: More Lessons From a Soldier

Last year, I wrote an article for the Journal of Hospital Medicine offering tips to healthcare providers in what was then an expanding COVID-19 environment.1 These lessons were drawn from my experiences during the “tough fights” and crisis situations of my military career, situations similar to what healthcare providers experienced during the pandemic.

Now, as vaccination rates rise and hospitalization rates fall, the nation and healthcare profession begin the transition to “normalcy.” What should healthcare professionals expect as they transition from a year of operating in a crisis to resumption of the habitual? What memories and lessons will linger from a long, tough fight against COVID-19, and how might physicians best approach the many post-crisis challenges they will surely face?

My military experiences inform the tips I offer to those in the medical profession. Both professions depend on adeptly leading and building a functional and effective organizational culture under trying circumstances. It may seem strange, but the challenges healthcare workers (HCWs) faced in fighting COVID-19 are comparable to what soldiers experience on a battlefield. And now, as citizens return to “normal” (however normal is defined), only naïve HCWs will believe they can simply resume their previous habits and practices. This part of the journey will present new challenges and unique opportunities.

Healthcare has changed…and so have you! Just like soldiers coming home from the battlefield face a necessarily new and different world, HCWs will also face changing circumstances, environments, and organizational requirements. Given this new landscape, I offer some of my lessons learned coming out of combat to help you adapt.

REFLECTIONS

Heading home from my last combat tour in Iraq, I found myself gazing out the aircraft window and pondering my personal experiences during a very long combat tour commanding a multinational task force. Pulling out my green soldier’s notebook, I rapidly scratched out some reflections on where I was, what I had learned, and what I needed to address personally and professionally. In talking with physicians in the healthcare organization where I now work, this emotional checklist seems to mirror some of the same thoughts they face coming out of the COVID-19 crisis.

Expect exhaustion. There’s a military axiom that “fatigue can make cowards of us all,” and while I don’t think I had succumbed to cowardice in battle, after 15 months in combat I was exhausted. Commanders in combat—or HCWs fighting a pandemic—face unrelenting demands from a variety of audiences. Leaders are asked to solve unsolvable problems, be at the right place at the right time with the right answers, have more energy than others, be upbeat, and exhibit behaviors that will motivate the “troops.” That’s true even if they’re exhausted and weary to the bone, serving on multiple teams, and attending endless meetings. There is also the common and unfortunate expectation that leaders should not take any time for themselves.

During the pandemic, most HCWs reported sleeping less, having little time to interact casually with others, and having less time for personal reflection, exercise, personal growth, or even prayer. My solution for addressing exhaustion was to develop a personal plan to address each one of these areas—mental, emotional, physical, spiritual—with a detailed rest and recovery strategy. I wrote my plan down, knowing that I would need to discuss this blueprint with both my employer and my spouse, who I suspected would have different ideas on what my schedule should look like after returning “home.” Healthcare providers have been through the same kinds of stresses and need to ask themselves: What recovery plan have I designed to help me overcome the fatigue I feel, and have I talked about this plan with the people who will be affected by it?

Take pride in what your teams accomplished. I was proud of how my teams had accomplished the impossible and how they had adapted to continually changing situations. Whenever military organizations know they’ll face the enemy in combat, they feel heightened anxiety, increased fear, and concern about the preparedness of their team. The Army, like any successful team, attempts to mitigate those emotions through training. During my reflections, I remembered the teams that came together to accomplish very tough missions. Some of those teams were those I had concerns about prior to deployment, but fortunately they often surprised me with their adaptability and successes in combat.

Leaders in healthcare can likely relate. Even in normal situations, organizational fault lines exist between physicians, nurses, and administrators. These fault lines may manifest as communication disconnects and distrust between different members who may not completely trust one another due to differences in training, culture, or role within the organization. But during a crisis, rifts dissipate and trust evolves as different cultures are forced to work together. Many healthcare organizations report that, during the COVID crisis, most personality conflicts, communication disconnects, and organizational dysfunctions receded, and organizations saw more and greater coordination and collaboration. Extensive research on leadership demonstrates that crises drive teams to communicate better and become more effective and efficient in accomplishing stated goals, resulting in team members who relish “being there” for one another like never before. These positive changes must be reinforced to ensure these newly formed high-performing teams do not revert back to work silos, which usually occurs due to distrust.

Just as important as pride in teams is the pride in the accomplishment of specific individuals during times of crisis. Diverse members of any organization deliver some of the best solutions to the toughest problems when they are included in the discussion, allowed to bring their ideas to the table, and rewarded for their actions (and their courage)! Just one example is given by Dr Sasha Shillcut as she describes the innovations and adaptations of the women physicians she observed in her organization during the COVID-19 crisis,2 and there are many examples of other organizations citing similar transformation in areas like telemedicine, emergency department procedures, and equipment design and use.3,4

Anticipate “survivor’s guilt.” During my three combat tours, 253 soldiers under my command or in my organization sacrificed their lives for the mission, and many more were wounded in action. There are times when bad dreams remind me of some of the circumstances surrounding the incidents that took the lives of those who died, and I often wake with a start and in a sweat. The first question I always ask myself in the middle of the night when this happens is, “Why did they die, and why did I survive?” That question is always followed by, “What might I have done differently to prevent those deaths?”

As we draw down from treating patients during the COVID-19 crisis, healthcare providers must also be wary of “survivor’s guilt.” Survivor’s guilt is a strong emotion for anyone who has survived a crisis, especially when their friends or loved ones have not. Healthcare providers have lost many patients, but they have also lost colleagues, friends, and family members. Because you are in the healing profession, many of you will question what more you could have done to prevent the loss of life. You likely won’t ever be completely satisfied with the answer, but I have a recommendation that may assuage your emotions.

In combat, we continually memorialized our fallen comrades in ceremonies that are attended by the entire unit. One of my commanders had an idea to keep pictures of those who had made the ultimate sacrifice, and on my desk is a box with the 253 pictures of those dedicated individuals who were killed in action under my command or in my unit. On the top of the box are the words “Make It Matter.” I look at those pictures often to remember them and their selfless service to the nation, and I often ask myself whether I am “making it matter” in my daily activities. Does your healthcare facility have plans for a memorial service for all those who died while in your care? Is there a special tribute in your hospital to those healthcare providers who paid the ultimate sacrifice in caring for patients? Most importantly, have you rededicated yourself to your profession, knowing that what you learned during the pandemic will help you be a better physician in the future, and do you have the knowledge that you are making a meaningful difference every day you serve in healthcare?

Relish being home. On that flight back to family, my excitement was palpable. But there were challenges too, as I knew I had to continue to focus on my team, my organization, and my profession. While images on the internet often show soldiers returning from war rushing into the arms of their loved ones, soldiers never leave the demands associated with wearing the cloth of the country. As a result, many marriages and families are damaged when one member who has been so singularly focused returns home and is still caught up in the demands of the job. They find it is difficult to pick up where they’ve left off, forgetting their family has also been under a different kind of intense stress.

These same challenges will face HCWs. Many of you voluntarily distanced yourself from family and friends due to a fear of transmitting the disease. Spouses and children underwent traumatic challenges in their jobs, holding together the household and piloting kids through schooling. My biggest recommendation is this: strive for a return to a healthy balance, be wary of any sharp edges that appear in your personality or in your relationships, and be open in communicating with those you love. Relying on friends, counselors, and mentors who can provide trusted advice—as well as therapy, if necessary—is not a sign of weakness, but a sign of strength and courage. The pandemic has affected our lives more than we can imagine, and “coming out” of the crisis will continue to test our humanity and civility like never before. Trust me on this one. I’ve been there.

RECOMMENDATIONS FOR POST-CRISIS ACTIONS

These reflections open us to issues physicians must address in the months after your “redeployment” from dealing with the pandemic. When soldiers redeploy from combat, every unit develops a plan to address personal and professional growth for individual members of the team. Additionally, leaders develop a plan to sustain performance and improve teams and organizational approaches. The objective? Polish the diamond from what we learned during the crisis, while preparing for those things that might detract from effectiveness in future crises. It’s an SOP (standard operating procedure) for military units to do these things. Is this approach also advisable for healthcare professionals and teams in responding to crises?

Crises increase stress on individuals and disrupt the functioning of organizations, but crises also provide phenomenal opportunities for growth.5 Adaptive organizations, be they military or healthcare, must take time to understand how the crises affected people and the organizational framework, while also preparing for potential future disruptions. While HCWs and their respective organizations are usually adept at learning from short-term emergencies (eg, limited disease outbreaks, natural disasters, mass-casualty events), they are less practiced in addressing crises that affect the profession for months. It has been a century since the medical profession has been faced with a global pandemic, but experts suggest other pandemics may be on the short-term horizon.6 We ought to use this past year of experiences to prepare for them.

Pay attention to your personal needs and the conditions of others on your team. After returning from combat, I was exhausted and stressed intellectually, physically, emotionally, and spiritually. From what I’ve seen, healthcare providers fit that same description, and the fatigue is palpable. Many of you have experienced extreme stress. I have experienced extremepost-traumatic stress, and it is important to understand that this will affect some on your team.7 In addition to addressing stress—and this is advice I give to all the physicians I know—find the time to get a physical examination. While the Army requires yearly physicals for all soldiers (especially generals!), most healthcare providers I know are shockingly deficient in taking the time to get a checkup from one of their colleagues. Commit to fixing that.

Reflect on what you have learned during this period. Take an afternoon with an adult beverage (if that’s your style) and reflect on what you learned and what others might learn from your unique experiences. Then, take some notes and shape your ideas. What did you experience? What adaptations did you or your team make during the pandemic? What worked and what didn’t? What things do you want to sustain in your practice and what things do you want to eliminate? What did you learn about the medical arts…or even about your Hippocratic Oath? If you have a mentor, share these thoughts with them; if you don’t have a mentor, find one and then share your thoughts with them. Get some outside feedback.

Assess team strengths and weaknesses. If you’re a formal physician leader (someone with a title and a position on your team), it’s your responsibility to provide feedback on both people and processes. If you’re an informal leader (someone who is a member of the team but doesn’t have specific leadership responsibilities outside your clinical role) and you don’t see this happening, volunteer to run the session for your formal leader and your organization. This session should last several hours and be held in a comfortable setting. You should prepare your team so they aren’t defensive about the points that may arise. Determine strengths and opportunities by asking for feedback on communication, behaviors, medical knowledge, emotional intelligence, and execution of tasks. Determine which processes and systems either worked or didn’t work, and either polish the approaches or drive change to improve systems as you get back to normal. Crises provide an opportunity to fix what’s broken while also reinforcing the things that worked in the crisis that might not be normal procedure. Don’t go back to old ways if those weren’t the things or the approaches you were using under critical conditions.

Encourage completion of an organization-wide after-action review (AAR). As I started writing this article, I watched CNN’s Dr Sanjay Gupta conduct a review of actions with the key physicians who contributed to the last administration’s response to the pandemic. In watching that session—and having conducted hundreds of AARs in my military career—there was discussion of obvious good and bad leadership and management procedures, process issues that needed to be addressed, and decision-making that might be applauded or questioned. Every healthcare organization ought to conduct a similar AAR, with a review of the most important aspects of actions and teamwork, the hospital’s operations, logistical preparation, and leader and organization procedures that demand to be addressed.

The successful conduct of any AAR requires asking (and getting answers to) four questions: What happened?; Why did it happen the way it did?; What needs to be fixed or “polished” in the processes, systems, or leadership approach?; and Who is responsible for ensuring the fixes or adjustments occur? The facilitator (and the key leaders of the organization) must ask the right questions, must be deeply involved in getting the right people to comment on the issues, and must “pin the rose” on someone who will be responsible for carrying through on the fixes. At the end of the AAR, after the key topics are discussed, with a plan for addressing each, the person in charge of the organization must publish an action plan with details for ensuring the fixes.

Like all citizens across our nation, my family is grateful for the skill and professionalism exhibited by clinicians and healthcare providers during this devastating pandemic. While we are all breathing a sigh of relief as we see the end in sight, true professionals must take the opportunity to learn and grow from this crisis and adapt. Hopefully, the reflections and recommendations in this article—things I learned from a different profession—will provide ideas to my new colleagues in healthcare.

References

1. Hertling M. Ten tips for a crisis: lessons from a soldier. J Hosp Med. 2020;15(5): 275-276. https://doi.org/10.12788/jhm.3424
2. Shillcut S. The inspiring women physicians of the COVID-19 pandemic. MedPage Today. April 9, 2020. Accessed July 7, 2021. https://www.kevinmd.com/blog/2020/04/the-insiring-women-physicians-of-the-covid-19-pandemic.html
3. Daley B. Three medical innovations fueled by COVID-19 that will outlast the pandemic. The Conversation. March 9, 2021. Accessed July 7, 2021. https://theconversation.com/3-medical-innovations-fueled-by-covid-19-that-will-outlast-the-pandemic-156464
4. Drees J, Dyrda L, Adams K. Ten big advancements in healthcare tech during the pandemic. Becker’s Health IT. July 6, 2020. Accessed July 7, 2021. https://www.beckershospitalreview.com/digital-transformation/10-big-advancements-in-healthcare-tech-during-the-pandemic.html
5. Wang J. Developing organizational learning capacity in crisis management. Adv Developing Hum Resources. 10(3):425-445. https://doi.org/10.1177/1523422308316464
6. Morens DM, Fauci AS. Emerging pandemic diseases: how we got COVID-19. Cell. 2020;182(5):1077-1092. https://doi.org/10.1016/j.cell.2020.08.021
7. What is posttraumatic stress disorder? American Psychiatric Association. Reviewed August 2020. Accessed July 7, 2021. https://www.psychiatry.org/patients-families/ptsd/what-is-ptsd

Article PDF
Author and Disclosure Information

Lieutenant General, US Army (Retired); Consultant to Healthcare Organizations on Physician Leadership Development, Orlando, Florida; Military and International Affairs Analyst, CNN, Atlanta, Georgia; Adjunct Professor, Crummer School of Business, Rollins College, Orlando, Florida; Dean’s Alliance, School of Public Health, Indiana University, Bloomington.

Disclosures
The author has no conflicts to disclose.

Issue
Journal of Hospital Medicine 16(10)
Publications
Topics
Page Number
634-636. Published Online First July 14, 2021
Sections
Author and Disclosure Information

Lieutenant General, US Army (Retired); Consultant to Healthcare Organizations on Physician Leadership Development, Orlando, Florida; Military and International Affairs Analyst, CNN, Atlanta, Georgia; Adjunct Professor, Crummer School of Business, Rollins College, Orlando, Florida; Dean’s Alliance, School of Public Health, Indiana University, Bloomington.

Disclosures
The author has no conflicts to disclose.

Author and Disclosure Information

Lieutenant General, US Army (Retired); Consultant to Healthcare Organizations on Physician Leadership Development, Orlando, Florida; Military and International Affairs Analyst, CNN, Atlanta, Georgia; Adjunct Professor, Crummer School of Business, Rollins College, Orlando, Florida; Dean’s Alliance, School of Public Health, Indiana University, Bloomington.

Disclosures
The author has no conflicts to disclose.

Article PDF
Article PDF
Related Articles

Last year, I wrote an article for the Journal of Hospital Medicine offering tips to healthcare providers in what was then an expanding COVID-19 environment.1 These lessons were drawn from my experiences during the “tough fights” and crisis situations of my military career, situations similar to what healthcare providers experienced during the pandemic.

Now, as vaccination rates rise and hospitalization rates fall, the nation and healthcare profession begin the transition to “normalcy.” What should healthcare professionals expect as they transition from a year of operating in a crisis to resumption of the habitual? What memories and lessons will linger from a long, tough fight against COVID-19, and how might physicians best approach the many post-crisis challenges they will surely face?

My military experiences inform the tips I offer to those in the medical profession. Both professions depend on adeptly leading and building a functional and effective organizational culture under trying circumstances. It may seem strange, but the challenges healthcare workers (HCWs) faced in fighting COVID-19 are comparable to what soldiers experience on a battlefield. And now, as citizens return to “normal” (however normal is defined), only naïve HCWs will believe they can simply resume their previous habits and practices. This part of the journey will present new challenges and unique opportunities.

Healthcare has changed…and so have you! Just like soldiers coming home from the battlefield face a necessarily new and different world, HCWs will also face changing circumstances, environments, and organizational requirements. Given this new landscape, I offer some of my lessons learned coming out of combat to help you adapt.

REFLECTIONS

Heading home from my last combat tour in Iraq, I found myself gazing out the aircraft window and pondering my personal experiences during a very long combat tour commanding a multinational task force. Pulling out my green soldier’s notebook, I rapidly scratched out some reflections on where I was, what I had learned, and what I needed to address personally and professionally. In talking with physicians in the healthcare organization where I now work, this emotional checklist seems to mirror some of the same thoughts they face coming out of the COVID-19 crisis.

Expect exhaustion. There’s a military axiom that “fatigue can make cowards of us all,” and while I don’t think I had succumbed to cowardice in battle, after 15 months in combat I was exhausted. Commanders in combat—or HCWs fighting a pandemic—face unrelenting demands from a variety of audiences. Leaders are asked to solve unsolvable problems, be at the right place at the right time with the right answers, have more energy than others, be upbeat, and exhibit behaviors that will motivate the “troops.” That’s true even if they’re exhausted and weary to the bone, serving on multiple teams, and attending endless meetings. There is also the common and unfortunate expectation that leaders should not take any time for themselves.

During the pandemic, most HCWs reported sleeping less, having little time to interact casually with others, and having less time for personal reflection, exercise, personal growth, or even prayer. My solution for addressing exhaustion was to develop a personal plan to address each one of these areas—mental, emotional, physical, spiritual—with a detailed rest and recovery strategy. I wrote my plan down, knowing that I would need to discuss this blueprint with both my employer and my spouse, who I suspected would have different ideas on what my schedule should look like after returning “home.” Healthcare providers have been through the same kinds of stresses and need to ask themselves: What recovery plan have I designed to help me overcome the fatigue I feel, and have I talked about this plan with the people who will be affected by it?

Take pride in what your teams accomplished. I was proud of how my teams had accomplished the impossible and how they had adapted to continually changing situations. Whenever military organizations know they’ll face the enemy in combat, they feel heightened anxiety, increased fear, and concern about the preparedness of their team. The Army, like any successful team, attempts to mitigate those emotions through training. During my reflections, I remembered the teams that came together to accomplish very tough missions. Some of those teams were those I had concerns about prior to deployment, but fortunately they often surprised me with their adaptability and successes in combat.

Leaders in healthcare can likely relate. Even in normal situations, organizational fault lines exist between physicians, nurses, and administrators. These fault lines may manifest as communication disconnects and distrust between different members who may not completely trust one another due to differences in training, culture, or role within the organization. But during a crisis, rifts dissipate and trust evolves as different cultures are forced to work together. Many healthcare organizations report that, during the COVID crisis, most personality conflicts, communication disconnects, and organizational dysfunctions receded, and organizations saw more and greater coordination and collaboration. Extensive research on leadership demonstrates that crises drive teams to communicate better and become more effective and efficient in accomplishing stated goals, resulting in team members who relish “being there” for one another like never before. These positive changes must be reinforced to ensure these newly formed high-performing teams do not revert back to work silos, which usually occurs due to distrust.

Just as important as pride in teams is the pride in the accomplishment of specific individuals during times of crisis. Diverse members of any organization deliver some of the best solutions to the toughest problems when they are included in the discussion, allowed to bring their ideas to the table, and rewarded for their actions (and their courage)! Just one example is given by Dr Sasha Shillcut as she describes the innovations and adaptations of the women physicians she observed in her organization during the COVID-19 crisis,2 and there are many examples of other organizations citing similar transformation in areas like telemedicine, emergency department procedures, and equipment design and use.3,4

Anticipate “survivor’s guilt.” During my three combat tours, 253 soldiers under my command or in my organization sacrificed their lives for the mission, and many more were wounded in action. There are times when bad dreams remind me of some of the circumstances surrounding the incidents that took the lives of those who died, and I often wake with a start and in a sweat. The first question I always ask myself in the middle of the night when this happens is, “Why did they die, and why did I survive?” That question is always followed by, “What might I have done differently to prevent those deaths?”

As we draw down from treating patients during the COVID-19 crisis, healthcare providers must also be wary of “survivor’s guilt.” Survivor’s guilt is a strong emotion for anyone who has survived a crisis, especially when their friends or loved ones have not. Healthcare providers have lost many patients, but they have also lost colleagues, friends, and family members. Because you are in the healing profession, many of you will question what more you could have done to prevent the loss of life. You likely won’t ever be completely satisfied with the answer, but I have a recommendation that may assuage your emotions.

In combat, we continually memorialized our fallen comrades in ceremonies that are attended by the entire unit. One of my commanders had an idea to keep pictures of those who had made the ultimate sacrifice, and on my desk is a box with the 253 pictures of those dedicated individuals who were killed in action under my command or in my unit. On the top of the box are the words “Make It Matter.” I look at those pictures often to remember them and their selfless service to the nation, and I often ask myself whether I am “making it matter” in my daily activities. Does your healthcare facility have plans for a memorial service for all those who died while in your care? Is there a special tribute in your hospital to those healthcare providers who paid the ultimate sacrifice in caring for patients? Most importantly, have you rededicated yourself to your profession, knowing that what you learned during the pandemic will help you be a better physician in the future, and do you have the knowledge that you are making a meaningful difference every day you serve in healthcare?

Relish being home. On that flight back to family, my excitement was palpable. But there were challenges too, as I knew I had to continue to focus on my team, my organization, and my profession. While images on the internet often show soldiers returning from war rushing into the arms of their loved ones, soldiers never leave the demands associated with wearing the cloth of the country. As a result, many marriages and families are damaged when one member who has been so singularly focused returns home and is still caught up in the demands of the job. They find it is difficult to pick up where they’ve left off, forgetting their family has also been under a different kind of intense stress.

These same challenges will face HCWs. Many of you voluntarily distanced yourself from family and friends due to a fear of transmitting the disease. Spouses and children underwent traumatic challenges in their jobs, holding together the household and piloting kids through schooling. My biggest recommendation is this: strive for a return to a healthy balance, be wary of any sharp edges that appear in your personality or in your relationships, and be open in communicating with those you love. Relying on friends, counselors, and mentors who can provide trusted advice—as well as therapy, if necessary—is not a sign of weakness, but a sign of strength and courage. The pandemic has affected our lives more than we can imagine, and “coming out” of the crisis will continue to test our humanity and civility like never before. Trust me on this one. I’ve been there.

RECOMMENDATIONS FOR POST-CRISIS ACTIONS

These reflections open us to issues physicians must address in the months after your “redeployment” from dealing with the pandemic. When soldiers redeploy from combat, every unit develops a plan to address personal and professional growth for individual members of the team. Additionally, leaders develop a plan to sustain performance and improve teams and organizational approaches. The objective? Polish the diamond from what we learned during the crisis, while preparing for those things that might detract from effectiveness in future crises. It’s an SOP (standard operating procedure) for military units to do these things. Is this approach also advisable for healthcare professionals and teams in responding to crises?

Crises increase stress on individuals and disrupt the functioning of organizations, but crises also provide phenomenal opportunities for growth.5 Adaptive organizations, be they military or healthcare, must take time to understand how the crises affected people and the organizational framework, while also preparing for potential future disruptions. While HCWs and their respective organizations are usually adept at learning from short-term emergencies (eg, limited disease outbreaks, natural disasters, mass-casualty events), they are less practiced in addressing crises that affect the profession for months. It has been a century since the medical profession has been faced with a global pandemic, but experts suggest other pandemics may be on the short-term horizon.6 We ought to use this past year of experiences to prepare for them.

Pay attention to your personal needs and the conditions of others on your team. After returning from combat, I was exhausted and stressed intellectually, physically, emotionally, and spiritually. From what I’ve seen, healthcare providers fit that same description, and the fatigue is palpable. Many of you have experienced extreme stress. I have experienced extremepost-traumatic stress, and it is important to understand that this will affect some on your team.7 In addition to addressing stress—and this is advice I give to all the physicians I know—find the time to get a physical examination. While the Army requires yearly physicals for all soldiers (especially generals!), most healthcare providers I know are shockingly deficient in taking the time to get a checkup from one of their colleagues. Commit to fixing that.

Reflect on what you have learned during this period. Take an afternoon with an adult beverage (if that’s your style) and reflect on what you learned and what others might learn from your unique experiences. Then, take some notes and shape your ideas. What did you experience? What adaptations did you or your team make during the pandemic? What worked and what didn’t? What things do you want to sustain in your practice and what things do you want to eliminate? What did you learn about the medical arts…or even about your Hippocratic Oath? If you have a mentor, share these thoughts with them; if you don’t have a mentor, find one and then share your thoughts with them. Get some outside feedback.

Assess team strengths and weaknesses. If you’re a formal physician leader (someone with a title and a position on your team), it’s your responsibility to provide feedback on both people and processes. If you’re an informal leader (someone who is a member of the team but doesn’t have specific leadership responsibilities outside your clinical role) and you don’t see this happening, volunteer to run the session for your formal leader and your organization. This session should last several hours and be held in a comfortable setting. You should prepare your team so they aren’t defensive about the points that may arise. Determine strengths and opportunities by asking for feedback on communication, behaviors, medical knowledge, emotional intelligence, and execution of tasks. Determine which processes and systems either worked or didn’t work, and either polish the approaches or drive change to improve systems as you get back to normal. Crises provide an opportunity to fix what’s broken while also reinforcing the things that worked in the crisis that might not be normal procedure. Don’t go back to old ways if those weren’t the things or the approaches you were using under critical conditions.

Encourage completion of an organization-wide after-action review (AAR). As I started writing this article, I watched CNN’s Dr Sanjay Gupta conduct a review of actions with the key physicians who contributed to the last administration’s response to the pandemic. In watching that session—and having conducted hundreds of AARs in my military career—there was discussion of obvious good and bad leadership and management procedures, process issues that needed to be addressed, and decision-making that might be applauded or questioned. Every healthcare organization ought to conduct a similar AAR, with a review of the most important aspects of actions and teamwork, the hospital’s operations, logistical preparation, and leader and organization procedures that demand to be addressed.

The successful conduct of any AAR requires asking (and getting answers to) four questions: What happened?; Why did it happen the way it did?; What needs to be fixed or “polished” in the processes, systems, or leadership approach?; and Who is responsible for ensuring the fixes or adjustments occur? The facilitator (and the key leaders of the organization) must ask the right questions, must be deeply involved in getting the right people to comment on the issues, and must “pin the rose” on someone who will be responsible for carrying through on the fixes. At the end of the AAR, after the key topics are discussed, with a plan for addressing each, the person in charge of the organization must publish an action plan with details for ensuring the fixes.

Like all citizens across our nation, my family is grateful for the skill and professionalism exhibited by clinicians and healthcare providers during this devastating pandemic. While we are all breathing a sigh of relief as we see the end in sight, true professionals must take the opportunity to learn and grow from this crisis and adapt. Hopefully, the reflections and recommendations in this article—things I learned from a different profession—will provide ideas to my new colleagues in healthcare.

Last year, I wrote an article for the Journal of Hospital Medicine offering tips to healthcare providers in what was then an expanding COVID-19 environment.1 These lessons were drawn from my experiences during the “tough fights” and crisis situations of my military career, situations similar to what healthcare providers experienced during the pandemic.

Now, as vaccination rates rise and hospitalization rates fall, the nation and healthcare profession begin the transition to “normalcy.” What should healthcare professionals expect as they transition from a year of operating in a crisis to resumption of the habitual? What memories and lessons will linger from a long, tough fight against COVID-19, and how might physicians best approach the many post-crisis challenges they will surely face?

My military experiences inform the tips I offer to those in the medical profession. Both professions depend on adeptly leading and building a functional and effective organizational culture under trying circumstances. It may seem strange, but the challenges healthcare workers (HCWs) faced in fighting COVID-19 are comparable to what soldiers experience on a battlefield. And now, as citizens return to “normal” (however normal is defined), only naïve HCWs will believe they can simply resume their previous habits and practices. This part of the journey will present new challenges and unique opportunities.

Healthcare has changed…and so have you! Just like soldiers coming home from the battlefield face a necessarily new and different world, HCWs will also face changing circumstances, environments, and organizational requirements. Given this new landscape, I offer some of my lessons learned coming out of combat to help you adapt.

REFLECTIONS

Heading home from my last combat tour in Iraq, I found myself gazing out the aircraft window and pondering my personal experiences during a very long combat tour commanding a multinational task force. Pulling out my green soldier’s notebook, I rapidly scratched out some reflections on where I was, what I had learned, and what I needed to address personally and professionally. In talking with physicians in the healthcare organization where I now work, this emotional checklist seems to mirror some of the same thoughts they face coming out of the COVID-19 crisis.

Expect exhaustion. There’s a military axiom that “fatigue can make cowards of us all,” and while I don’t think I had succumbed to cowardice in battle, after 15 months in combat I was exhausted. Commanders in combat—or HCWs fighting a pandemic—face unrelenting demands from a variety of audiences. Leaders are asked to solve unsolvable problems, be at the right place at the right time with the right answers, have more energy than others, be upbeat, and exhibit behaviors that will motivate the “troops.” That’s true even if they’re exhausted and weary to the bone, serving on multiple teams, and attending endless meetings. There is also the common and unfortunate expectation that leaders should not take any time for themselves.

During the pandemic, most HCWs reported sleeping less, having little time to interact casually with others, and having less time for personal reflection, exercise, personal growth, or even prayer. My solution for addressing exhaustion was to develop a personal plan to address each one of these areas—mental, emotional, physical, spiritual—with a detailed rest and recovery strategy. I wrote my plan down, knowing that I would need to discuss this blueprint with both my employer and my spouse, who I suspected would have different ideas on what my schedule should look like after returning “home.” Healthcare providers have been through the same kinds of stresses and need to ask themselves: What recovery plan have I designed to help me overcome the fatigue I feel, and have I talked about this plan with the people who will be affected by it?

Take pride in what your teams accomplished. I was proud of how my teams had accomplished the impossible and how they had adapted to continually changing situations. Whenever military organizations know they’ll face the enemy in combat, they feel heightened anxiety, increased fear, and concern about the preparedness of their team. The Army, like any successful team, attempts to mitigate those emotions through training. During my reflections, I remembered the teams that came together to accomplish very tough missions. Some of those teams were those I had concerns about prior to deployment, but fortunately they often surprised me with their adaptability and successes in combat.

Leaders in healthcare can likely relate. Even in normal situations, organizational fault lines exist between physicians, nurses, and administrators. These fault lines may manifest as communication disconnects and distrust between different members who may not completely trust one another due to differences in training, culture, or role within the organization. But during a crisis, rifts dissipate and trust evolves as different cultures are forced to work together. Many healthcare organizations report that, during the COVID crisis, most personality conflicts, communication disconnects, and organizational dysfunctions receded, and organizations saw more and greater coordination and collaboration. Extensive research on leadership demonstrates that crises drive teams to communicate better and become more effective and efficient in accomplishing stated goals, resulting in team members who relish “being there” for one another like never before. These positive changes must be reinforced to ensure these newly formed high-performing teams do not revert back to work silos, which usually occurs due to distrust.

Just as important as pride in teams is the pride in the accomplishment of specific individuals during times of crisis. Diverse members of any organization deliver some of the best solutions to the toughest problems when they are included in the discussion, allowed to bring their ideas to the table, and rewarded for their actions (and their courage)! Just one example is given by Dr Sasha Shillcut as she describes the innovations and adaptations of the women physicians she observed in her organization during the COVID-19 crisis,2 and there are many examples of other organizations citing similar transformation in areas like telemedicine, emergency department procedures, and equipment design and use.3,4

Anticipate “survivor’s guilt.” During my three combat tours, 253 soldiers under my command or in my organization sacrificed their lives for the mission, and many more were wounded in action. There are times when bad dreams remind me of some of the circumstances surrounding the incidents that took the lives of those who died, and I often wake with a start and in a sweat. The first question I always ask myself in the middle of the night when this happens is, “Why did they die, and why did I survive?” That question is always followed by, “What might I have done differently to prevent those deaths?”

As we draw down from treating patients during the COVID-19 crisis, healthcare providers must also be wary of “survivor’s guilt.” Survivor’s guilt is a strong emotion for anyone who has survived a crisis, especially when their friends or loved ones have not. Healthcare providers have lost many patients, but they have also lost colleagues, friends, and family members. Because you are in the healing profession, many of you will question what more you could have done to prevent the loss of life. You likely won’t ever be completely satisfied with the answer, but I have a recommendation that may assuage your emotions.

In combat, we continually memorialized our fallen comrades in ceremonies that are attended by the entire unit. One of my commanders had an idea to keep pictures of those who had made the ultimate sacrifice, and on my desk is a box with the 253 pictures of those dedicated individuals who were killed in action under my command or in my unit. On the top of the box are the words “Make It Matter.” I look at those pictures often to remember them and their selfless service to the nation, and I often ask myself whether I am “making it matter” in my daily activities. Does your healthcare facility have plans for a memorial service for all those who died while in your care? Is there a special tribute in your hospital to those healthcare providers who paid the ultimate sacrifice in caring for patients? Most importantly, have you rededicated yourself to your profession, knowing that what you learned during the pandemic will help you be a better physician in the future, and do you have the knowledge that you are making a meaningful difference every day you serve in healthcare?

Relish being home. On that flight back to family, my excitement was palpable. But there were challenges too, as I knew I had to continue to focus on my team, my organization, and my profession. While images on the internet often show soldiers returning from war rushing into the arms of their loved ones, soldiers never leave the demands associated with wearing the cloth of the country. As a result, many marriages and families are damaged when one member who has been so singularly focused returns home and is still caught up in the demands of the job. They find it is difficult to pick up where they’ve left off, forgetting their family has also been under a different kind of intense stress.

These same challenges will face HCWs. Many of you voluntarily distanced yourself from family and friends due to a fear of transmitting the disease. Spouses and children underwent traumatic challenges in their jobs, holding together the household and piloting kids through schooling. My biggest recommendation is this: strive for a return to a healthy balance, be wary of any sharp edges that appear in your personality or in your relationships, and be open in communicating with those you love. Relying on friends, counselors, and mentors who can provide trusted advice—as well as therapy, if necessary—is not a sign of weakness, but a sign of strength and courage. The pandemic has affected our lives more than we can imagine, and “coming out” of the crisis will continue to test our humanity and civility like never before. Trust me on this one. I’ve been there.

RECOMMENDATIONS FOR POST-CRISIS ACTIONS

These reflections open us to issues physicians must address in the months after your “redeployment” from dealing with the pandemic. When soldiers redeploy from combat, every unit develops a plan to address personal and professional growth for individual members of the team. Additionally, leaders develop a plan to sustain performance and improve teams and organizational approaches. The objective? Polish the diamond from what we learned during the crisis, while preparing for those things that might detract from effectiveness in future crises. It’s an SOP (standard operating procedure) for military units to do these things. Is this approach also advisable for healthcare professionals and teams in responding to crises?

Crises increase stress on individuals and disrupt the functioning of organizations, but crises also provide phenomenal opportunities for growth.5 Adaptive organizations, be they military or healthcare, must take time to understand how the crises affected people and the organizational framework, while also preparing for potential future disruptions. While HCWs and their respective organizations are usually adept at learning from short-term emergencies (eg, limited disease outbreaks, natural disasters, mass-casualty events), they are less practiced in addressing crises that affect the profession for months. It has been a century since the medical profession has been faced with a global pandemic, but experts suggest other pandemics may be on the short-term horizon.6 We ought to use this past year of experiences to prepare for them.

Pay attention to your personal needs and the conditions of others on your team. After returning from combat, I was exhausted and stressed intellectually, physically, emotionally, and spiritually. From what I’ve seen, healthcare providers fit that same description, and the fatigue is palpable. Many of you have experienced extreme stress. I have experienced extremepost-traumatic stress, and it is important to understand that this will affect some on your team.7 In addition to addressing stress—and this is advice I give to all the physicians I know—find the time to get a physical examination. While the Army requires yearly physicals for all soldiers (especially generals!), most healthcare providers I know are shockingly deficient in taking the time to get a checkup from one of their colleagues. Commit to fixing that.

Reflect on what you have learned during this period. Take an afternoon with an adult beverage (if that’s your style) and reflect on what you learned and what others might learn from your unique experiences. Then, take some notes and shape your ideas. What did you experience? What adaptations did you or your team make during the pandemic? What worked and what didn’t? What things do you want to sustain in your practice and what things do you want to eliminate? What did you learn about the medical arts…or even about your Hippocratic Oath? If you have a mentor, share these thoughts with them; if you don’t have a mentor, find one and then share your thoughts with them. Get some outside feedback.

Assess team strengths and weaknesses. If you’re a formal physician leader (someone with a title and a position on your team), it’s your responsibility to provide feedback on both people and processes. If you’re an informal leader (someone who is a member of the team but doesn’t have specific leadership responsibilities outside your clinical role) and you don’t see this happening, volunteer to run the session for your formal leader and your organization. This session should last several hours and be held in a comfortable setting. You should prepare your team so they aren’t defensive about the points that may arise. Determine strengths and opportunities by asking for feedback on communication, behaviors, medical knowledge, emotional intelligence, and execution of tasks. Determine which processes and systems either worked or didn’t work, and either polish the approaches or drive change to improve systems as you get back to normal. Crises provide an opportunity to fix what’s broken while also reinforcing the things that worked in the crisis that might not be normal procedure. Don’t go back to old ways if those weren’t the things or the approaches you were using under critical conditions.

Encourage completion of an organization-wide after-action review (AAR). As I started writing this article, I watched CNN’s Dr Sanjay Gupta conduct a review of actions with the key physicians who contributed to the last administration’s response to the pandemic. In watching that session—and having conducted hundreds of AARs in my military career—there was discussion of obvious good and bad leadership and management procedures, process issues that needed to be addressed, and decision-making that might be applauded or questioned. Every healthcare organization ought to conduct a similar AAR, with a review of the most important aspects of actions and teamwork, the hospital’s operations, logistical preparation, and leader and organization procedures that demand to be addressed.

The successful conduct of any AAR requires asking (and getting answers to) four questions: What happened?; Why did it happen the way it did?; What needs to be fixed or “polished” in the processes, systems, or leadership approach?; and Who is responsible for ensuring the fixes or adjustments occur? The facilitator (and the key leaders of the organization) must ask the right questions, must be deeply involved in getting the right people to comment on the issues, and must “pin the rose” on someone who will be responsible for carrying through on the fixes. At the end of the AAR, after the key topics are discussed, with a plan for addressing each, the person in charge of the organization must publish an action plan with details for ensuring the fixes.

Like all citizens across our nation, my family is grateful for the skill and professionalism exhibited by clinicians and healthcare providers during this devastating pandemic. While we are all breathing a sigh of relief as we see the end in sight, true professionals must take the opportunity to learn and grow from this crisis and adapt. Hopefully, the reflections and recommendations in this article—things I learned from a different profession—will provide ideas to my new colleagues in healthcare.

References

1. Hertling M. Ten tips for a crisis: lessons from a soldier. J Hosp Med. 2020;15(5): 275-276. https://doi.org/10.12788/jhm.3424
2. Shillcut S. The inspiring women physicians of the COVID-19 pandemic. MedPage Today. April 9, 2020. Accessed July 7, 2021. https://www.kevinmd.com/blog/2020/04/the-insiring-women-physicians-of-the-covid-19-pandemic.html
3. Daley B. Three medical innovations fueled by COVID-19 that will outlast the pandemic. The Conversation. March 9, 2021. Accessed July 7, 2021. https://theconversation.com/3-medical-innovations-fueled-by-covid-19-that-will-outlast-the-pandemic-156464
4. Drees J, Dyrda L, Adams K. Ten big advancements in healthcare tech during the pandemic. Becker’s Health IT. July 6, 2020. Accessed July 7, 2021. https://www.beckershospitalreview.com/digital-transformation/10-big-advancements-in-healthcare-tech-during-the-pandemic.html
5. Wang J. Developing organizational learning capacity in crisis management. Adv Developing Hum Resources. 10(3):425-445. https://doi.org/10.1177/1523422308316464
6. Morens DM, Fauci AS. Emerging pandemic diseases: how we got COVID-19. Cell. 2020;182(5):1077-1092. https://doi.org/10.1016/j.cell.2020.08.021
7. What is posttraumatic stress disorder? American Psychiatric Association. Reviewed August 2020. Accessed July 7, 2021. https://www.psychiatry.org/patients-families/ptsd/what-is-ptsd

References

1. Hertling M. Ten tips for a crisis: lessons from a soldier. J Hosp Med. 2020;15(5): 275-276. https://doi.org/10.12788/jhm.3424
2. Shillcut S. The inspiring women physicians of the COVID-19 pandemic. MedPage Today. April 9, 2020. Accessed July 7, 2021. https://www.kevinmd.com/blog/2020/04/the-insiring-women-physicians-of-the-covid-19-pandemic.html
3. Daley B. Three medical innovations fueled by COVID-19 that will outlast the pandemic. The Conversation. March 9, 2021. Accessed July 7, 2021. https://theconversation.com/3-medical-innovations-fueled-by-covid-19-that-will-outlast-the-pandemic-156464
4. Drees J, Dyrda L, Adams K. Ten big advancements in healthcare tech during the pandemic. Becker’s Health IT. July 6, 2020. Accessed July 7, 2021. https://www.beckershospitalreview.com/digital-transformation/10-big-advancements-in-healthcare-tech-during-the-pandemic.html
5. Wang J. Developing organizational learning capacity in crisis management. Adv Developing Hum Resources. 10(3):425-445. https://doi.org/10.1177/1523422308316464
6. Morens DM, Fauci AS. Emerging pandemic diseases: how we got COVID-19. Cell. 2020;182(5):1077-1092. https://doi.org/10.1016/j.cell.2020.08.021
7. What is posttraumatic stress disorder? American Psychiatric Association. Reviewed August 2020. Accessed July 7, 2021. https://www.psychiatry.org/patients-families/ptsd/what-is-ptsd

Issue
Journal of Hospital Medicine 16(10)
Issue
Journal of Hospital Medicine 16(10)
Page Number
634-636. Published Online First July 14, 2021
Page Number
634-636. Published Online First July 14, 2021
Publications
Publications
Topics
Article Type
Display Headline
Drawing Down From Crisis: More Lessons From a Soldier
Display Headline
Drawing Down From Crisis: More Lessons From a Soldier
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Mark Hertling; Email: [email protected]; Twitter: @markhertling.
Content Gating
Open Access (article Unlocked/Open Access)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media

Preoperative Care Assessment of Need Scores Are Associated With Postoperative Mortality and Length of Stay in Veterans Undergoing Knee Replacement

Article Type
Changed
Wed, 10/27/2021 - 15:42

Risk calculators can be of great value in guiding clinical decision making, patient-centered precision medicine, and resource allocation.1 Several perioperative risk prediction models have emerged in recent decades that estimate specific hazards (eg, cardiovascular complications after noncardiac surgery) with varying accuracy and utility. In the perioperative sphere, the time windows are often limited to an index hospitalization or 30 days following surgery or discharge.2-9 Although longer periods are of interest to patients, families, and health systems, few widely used or validated models are designed to look beyond this very narrow window.10,11 In addition, perioperative risk prediction models do not routinely incorporate parameters of a wide variety of health or demographic domains, such as patterns of health care, health care utilization, or medication use.

In 2013, in response to the need for near real-time information to guide delivery of enhanced care management services, the Veterans Health Administration (VHA) Office of Informatics and Analytics developed automated risk prediction models that used detailed electronic health record (EHR) data. These models were used to report Care Assessment Need (CAN) scores each week for all VHA enrollees and include data from a wide array of health domains. These CAN scores predict the risk for hospitalization, death, or either event within 90 days and 1 year.12,13 Each score is reported as both a predicted probability (0-1) and as a percentile in relation to all other VHA enrollees (a value between 1 and 99).13 The data used to calculate CAN scores are listed in Table 1.12

Care Assessment Need Score 2.0 Variables


Surgical procedures or admissions would not be differentiated from nonsurgical admissions or other procedural clinic visits, and as such, it is not possible to isolate the effect of undergoing a surgical procedure from another health-related event on the CAN score. At the same time though, a short-term increase in system utilization caused by an elective surgical procedure such as a total knee replacement (TKR) would presumably be reflected in a change in CAN score, but this has not been studied.

Since their introduction, CAN scores have been routinely accessed by primary care teams and used to facilitate care coordination for thousands of VHA patients. However, these CAN scores are currently not available to VHA surgeons, anesthesiologists, or other perioperative clinicians. In this study, we examine the distributions of preoperative CAN scores and explore the relationships of preoperative CAN 1-year mortality scores with 1-year survival following discharge and length of stay (LOS) during index hospitalization in a cohort of US veterans who underwent TKR, the most common elective operation performed within the VHA system.

Methods

Following approval of the Durham Veterans Affairs Medical Center Institutional Review Board, all necessary data were extracted from the VHA Corporate Data Warehouse (CDW) repository.14 Informed consent was waived due to the minimal risk nature of the study.

We used Current Procedural Terminology codes (27438, 27446, 27447, 27486, 27487, 27488) and International Classification of Diseases, 9th edition clinical modification procedure codes (81.54, 81.55, 81.59, 00.80-00.84) to identify all veterans who had undergone primary or revision TKR between July 2014 and December 2015 in VHA Veterans Integrated Service Network 1 (Maine, Vermont, New Hampshire, Massachusetts, Connecticut, Rhode Island, New York, Pennsylvania, West Virginia, Virginia, North Carolina). Because we focused on outcomes following hospital discharge, patients who died before discharge were excluded from the analysis. Preoperative CAN 1-year mortality score was chosen as the measure under the assumption that long-term survival may be the most meaningful of the 4 possible CAN score measures.

Our primary objective was to determine distribution of preoperative CAN scores in the study population. Our secondary was to study relationships among the preoperative CAN 1-year mortality scores and 1-year mortality and hospital LOS.

Study Variables

For each patient, we extracted the date of index surgery. The primary exposure or independent variable was the CAN score in the week prior to this date. Because prior study has shown that CAN scores trajectories do not significantly change over time, the date-stamped CAN scores in the week before surgery represent what would have been available to clinicians in a preoperative setting.15 Since CAN scores are refreshed and overwritten every week, we extracted archived scores from the CDW.

 

 

For the 1-year survival outcome, the primary dependent variable, we queried the vital status files in the CDW for the date of death if applicable. We confirmed survival beyond 1 year by examining vital signs in the CDW for a minimum of 2 independent encounters beyond 1 year after the date of discharge. To compute the index LOS, the secondary outcome, we computed the difference between the date of admission and date of hospital discharge.

Statistical Methods

The parameters and performance of the multivariable logistic regression models developed to compute the various CAN mortality and hospitalization risk scores have been previously described.12 Briefly, Wang and colleagues created parsimonious regression models using backward selection. Model discrimination was evaluated using C (concordance)-statistic. Model calibration was assessed by comparing predicted vs observed event rates by risk deciles and performing Cox proportional hazards regression.

We plotted histograms to display preoperative CAN scores as a simple measure of distribution (Figure 1). We also examined the cumulative proportion of patients at each preoperative CAN 1-year mortality score.

Distribution of 1-Year Mortality CAN Scores


Using a conventional t test, we compared means of preoperative CAN 1-year mortality scores in patients who survived vs those who died within 1 year. We also constructed a plot of the proportion of patients who had died within 1 year vs preoperative CAN 1-year mortality scores. Kaplan-Meier curves were then constructed examining 1-year survival by CAN 1-year mortality score by terciles.

Finally, we examined the relationship between preoperative CAN 1-year mortality scores and index LOS in 2 ways: We plotted LOS across CAN scores, and we constructed a locally weighted regression and smoothing scatterplot. LOESS (locally weighted polynomial regression) models, involve fitting the polynomial using weighted least squares, giving more weight to points near the point whose response is being estimated and less weight to points further away. LOESS models help fit a smooth curve to data for which linear and nonlinear least squares regression models do not perform well and in situations where no theoretical model exists. LOESS are ideal for modeling complex processes. All analyses were performed using SAS, version 9.3.

Results

We identified 8206 patients who had undergone a TKR over the 18-month study period. The overall mean (SD) for age was 65 (8.41) years; 93% were male, and 78% were White veterans. Patient demographics are well described in a previous publication.16,17

In terms of model parameters for the CAN score models, C-statistics for the 90-day outcome models were as follows: 0.833 for the model predicting hospitalization (95% CI, 0.832-0.834); 0.865 for the model predicting death (95% CI, 0.863-0.876); and 0.811 for the model predicting either event (95% CI, 0.810-0.812). C-statistics for the 1-year outcome models were 0.809 for the model predicting hospitalization (95% CI, 0.808-0.810); 0.851 for the model predicting death (95% CI, 0.849-0.852); and 0.787 for the model predicting either event (95% CI, 0.786-0.787). Models were well calibrated with α = 0 and β = 1, demonstrating strong agreement between observed and predicted event rates.

Relationship Between 1-Year Mortality Rates and 1-Year CAN Scores

Distribution of Hospitalization CAN Scores top chart



The distribution of preoperative CAN 1-year mortality scores was close to normal (median, 50; interquartile range, 40; mean [SD] 48 [25.6]) (eTable). The original CAN score models were developed having an equal number of patients in each strata and as such, are normally distributed.12 Our cohort was similar in pattern of distribution. Distributions of the remaining preoperative CAN scores (90-day mortality, 1-year hospitalization, 90-day hospitalization) are shown in Figures 2, 3, and 4. Not surprisingly, histograms for both 90-day and 1-year hospitalization were skewed toward higher scores, indicating that these patients were expected to be hospitalized in the near future.

Distribution of CAN 1-Year Mortality Scores

Preoperative CAN 1-Year Mortality Scores for Veterans Who Underwent TKR table


Overall, 1.4% (110/8096) of patients died within 1 year of surgery. Comparing 1-year mortality CAN scores in survivors vs nonsurvivors, we found statistically significant differences in means (47 vs 66 respectively, P < .001) and medians (45 vs 75 respectively, P < .001) (Table 2). In the plot examining the relationship between preoperative 1-year mortality CAN scores and 1-year mortality, the percentage who died within 1 year increased initially for patients with CAN scores > 60 and again exponentially for patients with CAN scores > 80. Examining Kaplan-Meier curves, we found that survivors and nonsurvivors separated early after surgery, and the differences between the top tercile and the middle/lower terciles were statistically significant (P < .001). Mortality rates were about 0.5% in the lower and middle terciles but about 2% in the upper tercile (Figure 5).

Average Hospital LOS and 90-Day Mortality CAN Scores

Average Hospital LOS and 90-Day Mortality CAN Scores in Patients Undergoing Elective TKA

Kaplan-Meier Curves


In the plot examining the relationship between CAN scores and index LOS, the LOS rose significantly beyond a CAN score of 60 and dramatically beyond a CAN score of 80 (Figure 6). LOESS curves also showed 2 inflection points suggesting an incremental and sequential rise in the LOS with increasing CAN scores (Figure 7). Mean (SD) LOS in days for the lowest to highest terciles was 2.6 (1.7), 2.8 (2.1), and 3.6 (2.2), respectively.

 

 

Discussion

CAN scores are automatically generated each week by EHR-based multivariable risk models. These scores have excellent predictive accuracy for 90-day and 1-year mortality and hospitalization and are routinely used by VHA primary care teams to assist with clinical operations.13 We studied the distribution of CAN 1-year mortality scores in a preoperative context and examined relationships of the preoperative CAN 1-year mortality scores with postoperative mortality and LOS in 8206 veterans who underwent TKR.

There are several noteworthy findings. First, the overall 1-year mortality rate observed following TKR (1.4%) was similar to other published reports.18,19 Not surprisingly, preoperative CAN 1-year mortality scores were significantly higher in veterans who died compared with those of survivors. The majority of patients who died had a preoperative CAN 1-year mortality score > 75 while most who survived had a preoperative CAN 1-year mortality score < 45 (P < .001). Interestingly, the same scores showed a nonlinear correlation with LOS. Index LOS was about 4 days in patients in the highest tercile of CAN scores vs 2.5 days in the lowest tercile, but the initial increase in LOS was detected at a CAN score of about 55 to 60.

In addition, mortality rate varied widely in different segments of the population when grouped according to preoperative CAN scores. One-year mortality rates in the highest tercile reached 2%, about 4-fold higher than that of lower terciles (0.5%). Examination of the Kaplan-Meier curves showed that this difference in mortality between the highest tercile and the lower 2 groups appears soon after discharge and continues to increase over time, suggesting that the factors contributing to the increased mortality are present at the time of discharge and persist beyond the postoperative period. In summary, although CAN scores were not designed for use in the perioperative context, we found that preoperative CAN 1-year mortality scores are broadly predictive of mortality, but especially for increases in LOS following elective TKA, both increases in hospital LOS following elective TKA and mortality over the year after TKA.

Our findings raise several important questions. The decision to undergo elective surgery is complex. Arguably, individuals who undergo elective knee replacement should be healthy enough to undergo, recover, and reap the benefits from a procedure that does not extend life. The distribution of preoperative CAN 1-year mortality scores for our study population was similar to that of the general VHA enrollee population with similar measured mortality rates (≤ 0.5% vs ≥ 1.7% in the low and high terciles, respectively).1 Further study comparing outcomes in matched cohorts who did and did not undergo joint replacement would be of interest. In lieu of this, though, the association of high but not extreme CAN scores with increased hospital LOS may potentially be used to guide allocation of resources to this group, obviating the increased cost and risk to which this group is exposed. And the additional insight afforded by CAN scores may enhance shared decision-making models by identifying patients at the very highest risk (eg, 1-year mortality CAN score ≥ 90), patients who conceivably might not survive long enough to recover from and enjoy their reconstructed knee, who might in the long run be harmed by undergoing the procedure.

Many total joint arthroplasties are performed in older patients, a population in which frailty is increasingly recognized as a significant risk factor for poor outcomes.20,21 CAN scores reliably identify high-risk patients and have been shown to correlate with frailty in this group.22 Multiple authors have reported improved outcomes with cost reductions after implementation of programs targeting modifiable risk factors in high-risk surgical candidates.23-25 A preoperative assessment that includes the CAN score may be valuable in identifying patients who would benefit most from prehabilitation programs or other interventions designed to blunt the impact of frailty. It is true that many elements used to calculate the CAN score would not be considered modifiable, especially in the short term. However, specific contributors to frailty, such as nutritional status and polypharmacy might be potential candidates. As with all multivariable risk prediction models, there are multiple paths to a high CAN score, and further research to identify clinically relevant subgroups may help inform efforts to improve perioperative care within this population.

Hospital LOS is of intense interest for many reasons, not least its utility as a surrogate for cost and increased risk for immediate perioperative adverse events, such as multidrug-resistant hospital acquired infections, need for postacute facility-based rehabilitation, and deconditioning that increase risks of falls and fractures in the older population.26-29 In addition, its importance is magnified due to the COVID-19 pandemic context in which restarting elective surgery programs has changed traditional criteria by which patients are scheduled for surgery.

We have shown that elevated CAN scores are able to identify patients at risk for extended hospital stays and, as such, may be useful additional data in allocating scarce operating room time and other resources for optimal patient and health care provider safety.30,31 Individual surgeons and hospital systems would, of course, decide which patients should be triaged to go first, based on local priorities; however, choosing lower risk patients with minimal risk of morbidity and mortality while pursuing prehabilitation for higher risk patients is a reasonable approach.

 

 

Limitations

Our study has several limitations. Only a single surgical procedure was included, albeit the most common one performed in the VHA. In addition, no information was available concerning the precise clinical course for these patients, such as the duration of surgery, anesthetic technique, and management of acute, perioperative course. Although the assumption was made that patients received standard care in a manner such that these factors would not significantly affect either their mortality or their LOS out of proportion to their preoperative clinical status, confounding cannot be excluded. Therefore, further study is necessary to determine whether CAN scores can accurately predict mortality and/or LOS for patients undergoing other procedures. Further, a clinical trial is required to assess whether systematic provision of the CAN score at the point of surgery would impact care and, more important, impact outcomes. In addition, multivariable analyses were not performed, including and excluding various components of the CAN score models. Currently, CAN scores could be made available to the surgical/anesthesia communities at minimal or no cost and are updated automatically. Model calibration and discrimination in this particular setting were not validated.

Because our interest is in leveraging an existing resource to a current clinical and operational problem rather than in creating or validating a new tool, we chose to test the simple bivariate relationship between preoperative CAN scores and outcomes. We chose the preoperative 1-year mortality CAN score from among the 4 options under the assumption that long-term survival is the most meaningful of the 4 candidate outcomes. Finally, while the CAN scores are currently only calculated and generated for patients cared for within the VHA, few data elements are unavailable to civilian health systems. The most problematic would be documentation of actual prescription filling, but this is a topic of increasing interest to the medical and academic communities and access to such information we hope will improve.32-34

Conclusions

Although designed for use by VHA primary care teams, CAN scores also may have value for perioperative clinicians, predicting mortality and prolonged hospital LOS in those with elevated 1-year mortality scores. Advantages of CAN scores relative to other perioperative risk calculators lies in their ability to predict long-term rather than 30-day survival and that they are automatically generated on a near-real-time basis for all patients who receive care in VHA ambulatory clinics. Further study is needed to determine practical utility in shared decision making, preoperative evaluation and optimization, and perioperative resource allocation.

Acknowledgments

This work was supported by the US Department of Veterans Affairs (VA) National Center for Patient Safety, Field Office 10A4E, through the Patient Safety Center of Inquiry at the Durham VA Medical Center in North Carolina. The study also received support from the Center of Innovation to Accelerate Discovery and Practice Transformation (CIN 13-410) at the Durham VA Health Care System.

References

1. McNair AGK, MacKichan F, Donovan JL, et al. What surgeons tell patients and what patients want to know before major cancer surgery: a qualitative study. BMC Cancer. 2016;16:258. doi:10.1186/s12885-016-2292-3

2. Grover FL, Hammermeister KE, Burchfiel C. Initial report of the Veterans Administration Preoperative Risk Assessment Study for Cardiac Surgery. Ann Thorac Surg. 1990;50(1):12-26; discussion 27-18. doi:10.1016/0003-4975(90)90073-f

3. Khuri SF, Daley J, Henderson W, et al. The National Veterans Administration Surgical Risk Study: risk adjustment for the comparative assessment of the quality of surgical care. J Am Coll Surg. 1995;180(5):519-531.

4. Glance LG, Lustik SJ, Hannan EL, et al. The Surgical Mortality Probability Model: derivation and validation of a simple simple risk prediction rule for noncardiac surgery. Ann Surg. 2012;255(4):696-702. doi:10.1097/SLA.0b013e31824b45af

5. Keller DS, Kroll D, Papaconstantinou HT, Ellis CN. Development and validation of a methodology to reduce mortality using the veterans affairs surgical quality improvement program risk calculator. J Am Coll Surg. 2017;224(4):602-607. doi:10.1016/j.jamcollsurg.2016.12.033

6. Bilimoria KY, Liu Y, Paruch JL, et al. Development and evaluation of the universal ACS NSQIP surgical risk calculator: a decision aid and informed consent tool for patients and surgeons. J Am Coll Surg. 2013;217(5):833-842.e831-833. doi:10.1016/j.jamcollsurg.2013.07.385

7. Ford MK, Beattie WS, Wijeysundera DN. Systematic review: prediction of perioperative cardiac complications and mortality by the revised cardiac risk index. Ann Intern Med. 2010;152(1):26-35. doi:10.7326/0003-4819-152-1-201001050-00007

8. Gupta PK, Gupta H, Sundaram A, et al. Development and validation of a risk calculator for prediction of cardiac risk after surgery. Circulation. 2011;124(4):381-387. doi:10.1161/CIRCULATIONAHA.110.015701

9. Lee TH, Marcantonio ER, Mangione CM, et al. Derivation and prospective validation of a simple index for prediction of cardiac risk of major noncardiac surgery. Circulation. 1999;100(10):1043-1049. doi:10.1161/01.cir.100.10.1043

10. Smith T, Li X, Nylander W, Gunnar W. Thirty-day postoperative mortality risk estimates and 1-year survival in Veterans Health Administration surgery patients. JAMA Surg. 2016;151(5):417-422. doi:10.1001/jamasurg.2015.4882

11. Damhuis RA, Wijnhoven BP, Plaisier PW, Kirkels WJ, Kranse R, van Lanschot JJ. Comparison of 30-day, 90- day and in-hospital postoperative mortality for eight different cancer types. Br J Surg. 2012;99(8):1149-1154. doi:10.1002/bjs.8813

12. Wang L, Porter B, Maynard C, et al. Predicting risk of hospitalization or death among patients receiving primary care in the Veterans Health Administration. Med Care. 2013;51(4):368-373. doi:10.1016/j.amjcard.2012.06.038

13. Fihn SD, Francis J, Clancy C, et al. Insights from advanced analytics at the Veterans Health Administration. Health Aff (Millwood). 2014;33(7):1203-1211. doi:10.1377/hlthaff.2014.0054

14. Noël PH, Copeland LA, Perrin RA, et al. VHA Corporate Data Warehouse height and weight data: opportunities and challenges for health services research. J Rehabil Res Dev. 2010;47(8):739-750. doi:10.1682/jrrd.2009.08.0110

15. Wong ES, Yoon J, Piegari RI, Rosland AM, Fihn SD, Chang ET. Identifying latent subgroups of high-risk patients using risk score trajectories. J Gen Intern Med. 2018;33(12):2120-2126. doi:10.1007/s11606-018-4653-x

16. Chen Q, Hsia HL, Overman R, et al. Impact of an opioid safety initiative on patients undergoing total knee arthroplasty: a time series analysis. Anesthesiology. 2019;131(2):369-380. doi:10.1097/ALN.0000000000002771

17. Hsia HL, Takemoto S, van de Ven T, et al. Acute pain is associated with chronic opioid use after total knee arthroplasty. Reg Anesth Pain Med. 2018;43(7):705-711. doi:10.1097/AAP.0000000000000831

18. Inacio MCS, Dillon MT, Miric A, Navarro RA, Paxton EW. Mortality after total knee and total hip arthroplasty in a large integrated health care system. Perm J. 2017;21:16-171. doi:10.7812/TPP/16-171

19. Lee QJ, Mak WP, Wong YC. Mortality following primary total knee replacement in public hospitals in Hong Kong. Hong Kong Med J. 2016;22(3):237-241. doi:10.12809/hkmj154712

20. Lin HS, Watts JN, Peel NM, Hubbard RE. Frailty and post-operative outcomes in older surgical patients: a systematic review. BMC Geriatr. 2016;16(1):157. doi:10.1186/s12877-016-0329-8

21. Shinall MC Jr, Arya S, Youk A, et al. Association of preoperative patient frailty and operative stress with postoperative mortality. JAMA Surg. 2019;155(1):e194620. doi:10.1001/jamasurg.2019.4620

22. Ruiz JG, Priyadarshni S, Rahaman Z, et al. Validation of an automatically generated screening score for frailty: the care assessment need (CAN) score. BMC Geriatr. 2018;18(1):106. doi:10.1186/s12877-018-0802-7

23. Bernstein DN, Liu TC, Winegar AL, et al. Evaluation of a preoperative optimization protocol for primary hip and knee arthroplasty patients. J Arthroplasty. 2018;33(12):3642- 3648. doi:10.1016/j.arth.2018.08.018

24. Sodhi N, Anis HK, Coste M, et al. A nationwide analysis of preoperative planning on operative times and postoperative complications in total knee arthroplasty. J Knee Surg. 2019;32(11):1040-1045. doi:10.1055/s-0039-1677790

25. Krause A, Sayeed Z, El-Othmani M, Pallekonda V, Mihalko W, Saleh KJ. Outpatient total knee arthroplasty: are we there yet? (part 1). Orthop Clin North Am. 2018;49(1):1-6. doi:10.1016/j.ocl.2017.08.002

26. Barrasa-Villar JI, Aibar-Remón C, Prieto-Andrés P, Mareca- Doñate R, Moliner-Lahoz J. Impact on morbidity, mortality, and length of stay of hospital-acquired infections by resistant microorganisms. Clin Infect Dis. 2017;65(4):644-652. doi:10.1093/cid/cix411

27. Nikkel LE, Kates SL, Schreck M, Maceroli M, Mahmood B, Elfar JC. Length of hospital stay after hip fracture and risk of early mortality after discharge in New York state: retrospective cohort study. BMJ. 2015;351:h6246. doi:10.1136/bmj.h6246

28. Marfil-Garza BA, Belaunzarán-Zamudio PF, Gulias-Herrero A, et al. Risk factors associated with prolonged hospital length-of-stay: 18-year retrospective study of hospitalizations in a tertiary healthcare center in Mexico. PLoS One. 2018;13(11):e0207203. doi:10.1371/journal.pone.0207203

29. Hirsch CH, Sommers L, Olsen A, Mullen L, Winograd CH. The natural history of functional morbidity in hospitalized older patients. J Am Geriatr Soc. 1990;38(12):1296-1303. doi:10.1111/j.1532-5415.1990.tb03451.x

30. Iyengar KP, Jain VK, Vaish A, Vaishya R, Maini L, Lal H. Post COVID-19: planning strategies to resume orthopaedic surgery -challenges and considerations. J Clin Orthop Trauma. 2020;11(suppl 3):S291-S295. doi:10.1016/j.jcot.2020.04.028

31. O’Connor CM, Anoushiravani AA, DiCaprio MR, Healy WL, Iorio R. Economic recovery after the COVID-19 pandemic: resuming elective orthopedic surgery and total joint arthroplasty. J Arthroplasty. 2020;35(suppl 7):S32-S36. doi:10.1016/j.arth.2020.04.038.

32. Mauseth SA, Skurtveit S, Skovlund E, Langhammer A, Spigset O. Medication use and association with urinary incontinence in women: data from the Norwegian Prescription Database and the HUNT study. Neurourol Urodyn. 2018;37(4):1448-1457. doi:10.1002/nau.23473

33. Sultan RS, Correll CU, Schoenbaum M, King M, Walkup JT, Olfson M. National patterns of commonly prescribed psychotropic medications to young people. J Child Adolesc Psychopharmacol. 2018;28(3):158-165. doi:10.1089/cap.2017.0077

34. McCoy RG, Dykhoff HJ, Sangaralingham L, et al. Adoption of new glucose-lowering medications in the U.S.-the case of SGLT2 inhibitors: nationwide cohort study. Diabetes Technol Ther. 2019;21(12):702-712. doi:10.1089/dia.2019.0213

Article PDF
Author and Disclosure Information

Marissa Kuo is a Medical Student at Emory University School of Medicine in Atlanta, Georgia. Rebecca Schroeder is a Staff Physician at the Anesthesia Service, Durham Veterans Affairs Health Care System (DVAHCS) in North Carolina and an Associate Professor, Critical Care and Perioperative Epidemiologic Research (CAPER) Unit, Department of Anesthesiology, Duke University Medical Center (DUMC). Atilio Barbeito is a Staff Physician at the DVAHCS Anesthesia Service and Associate Professor CAPER Unit, DUMC. Carl Pieper is an Associate Professor of Biostatistics and Bioinformatics at Duke University Center for Aging and Duke University Department of Medicine and is a Senior Researcher at the Geriatrics Research Education and Clinical Center (GRECC) and Health Services Research and Development, DVAHCS. Vijay Krishnamoorthy is an Assistant Professor CAPER Unit DUMC. Samuel Wellman is Chief, Orthopedic Service, DVAHCS and an Associate Professor, Department of Orthopedic Surgery, Duke University Health System (DUHS). Susan Nicole Hastings is a Senior Researcher Duke University Center for Aging and Duke University Department of Medicine and is Director of the Center of Innovation to Accelerate Discovery and Practice Transformation; Senior Researcher GRECC and Health Services Research and Development, DVAHCS, and Professor, Department of Medicine and Department of Population Health DUHS. Karthik Raghunathan is a Staff Physician Anesthesia Service, DVAHCS and an Associate Professor with Tenure, and Codirector CAPER Unit, Department of Anesthesiology, DUMC.
Correspondence: Karthik Raghunathan ([email protected])

Author Contributions

Marissa Kuo contributed to interpretation of the results and wrote the paper with input from all authors. Rebecca Schroeder assisted in project supervision, study design, result interpretation, figure creation, manuscript creation, and final submission. Atilio Barbeito, Susan Nicole Hastings, Samuel Wellman, Vijay Krishnamoorthy, and Karthik Raghunathan were responsible for the main conceptual idea, interpretation of results, and manuscript preparation. Carl Pieper processed the data, performed the data analysis and calculations, assisted in interpretation of results, and reviewed and contributed to the final manuscript. Karthik Raghunathan was responsible for contributing to the main conceptual idea, development of the statistical plan, interpretation of results, and manuscript preparation.

Author disclosures
The authors report no actual or potential conflicts of interest with regard to this article.

Disclaimer
The opinions expressed herein are those of the authors and do not necessarily reflect those of Federal Practitioner, Frontline Medical Communications Inc., the US Government, or any of its agencies.

Issue
Federal Practitioner - 38(7)a
Publications
Topics
Page Number
316-324
Sections
Author and Disclosure Information

Marissa Kuo is a Medical Student at Emory University School of Medicine in Atlanta, Georgia. Rebecca Schroeder is a Staff Physician at the Anesthesia Service, Durham Veterans Affairs Health Care System (DVAHCS) in North Carolina and an Associate Professor, Critical Care and Perioperative Epidemiologic Research (CAPER) Unit, Department of Anesthesiology, Duke University Medical Center (DUMC). Atilio Barbeito is a Staff Physician at the DVAHCS Anesthesia Service and Associate Professor CAPER Unit, DUMC. Carl Pieper is an Associate Professor of Biostatistics and Bioinformatics at Duke University Center for Aging and Duke University Department of Medicine and is a Senior Researcher at the Geriatrics Research Education and Clinical Center (GRECC) and Health Services Research and Development, DVAHCS. Vijay Krishnamoorthy is an Assistant Professor CAPER Unit DUMC. Samuel Wellman is Chief, Orthopedic Service, DVAHCS and an Associate Professor, Department of Orthopedic Surgery, Duke University Health System (DUHS). Susan Nicole Hastings is a Senior Researcher Duke University Center for Aging and Duke University Department of Medicine and is Director of the Center of Innovation to Accelerate Discovery and Practice Transformation; Senior Researcher GRECC and Health Services Research and Development, DVAHCS, and Professor, Department of Medicine and Department of Population Health DUHS. Karthik Raghunathan is a Staff Physician Anesthesia Service, DVAHCS and an Associate Professor with Tenure, and Codirector CAPER Unit, Department of Anesthesiology, DUMC.
Correspondence: Karthik Raghunathan ([email protected])

Author Contributions

Marissa Kuo contributed to interpretation of the results and wrote the paper with input from all authors. Rebecca Schroeder assisted in project supervision, study design, result interpretation, figure creation, manuscript creation, and final submission. Atilio Barbeito, Susan Nicole Hastings, Samuel Wellman, Vijay Krishnamoorthy, and Karthik Raghunathan were responsible for the main conceptual idea, interpretation of results, and manuscript preparation. Carl Pieper processed the data, performed the data analysis and calculations, assisted in interpretation of results, and reviewed and contributed to the final manuscript. Karthik Raghunathan was responsible for contributing to the main conceptual idea, development of the statistical plan, interpretation of results, and manuscript preparation.

Author disclosures
The authors report no actual or potential conflicts of interest with regard to this article.

Disclaimer
The opinions expressed herein are those of the authors and do not necessarily reflect those of Federal Practitioner, Frontline Medical Communications Inc., the US Government, or any of its agencies.

Author and Disclosure Information

Marissa Kuo is a Medical Student at Emory University School of Medicine in Atlanta, Georgia. Rebecca Schroeder is a Staff Physician at the Anesthesia Service, Durham Veterans Affairs Health Care System (DVAHCS) in North Carolina and an Associate Professor, Critical Care and Perioperative Epidemiologic Research (CAPER) Unit, Department of Anesthesiology, Duke University Medical Center (DUMC). Atilio Barbeito is a Staff Physician at the DVAHCS Anesthesia Service and Associate Professor CAPER Unit, DUMC. Carl Pieper is an Associate Professor of Biostatistics and Bioinformatics at Duke University Center for Aging and Duke University Department of Medicine and is a Senior Researcher at the Geriatrics Research Education and Clinical Center (GRECC) and Health Services Research and Development, DVAHCS. Vijay Krishnamoorthy is an Assistant Professor CAPER Unit DUMC. Samuel Wellman is Chief, Orthopedic Service, DVAHCS and an Associate Professor, Department of Orthopedic Surgery, Duke University Health System (DUHS). Susan Nicole Hastings is a Senior Researcher Duke University Center for Aging and Duke University Department of Medicine and is Director of the Center of Innovation to Accelerate Discovery and Practice Transformation; Senior Researcher GRECC and Health Services Research and Development, DVAHCS, and Professor, Department of Medicine and Department of Population Health DUHS. Karthik Raghunathan is a Staff Physician Anesthesia Service, DVAHCS and an Associate Professor with Tenure, and Codirector CAPER Unit, Department of Anesthesiology, DUMC.
Correspondence: Karthik Raghunathan ([email protected])

Author Contributions

Marissa Kuo contributed to interpretation of the results and wrote the paper with input from all authors. Rebecca Schroeder assisted in project supervision, study design, result interpretation, figure creation, manuscript creation, and final submission. Atilio Barbeito, Susan Nicole Hastings, Samuel Wellman, Vijay Krishnamoorthy, and Karthik Raghunathan were responsible for the main conceptual idea, interpretation of results, and manuscript preparation. Carl Pieper processed the data, performed the data analysis and calculations, assisted in interpretation of results, and reviewed and contributed to the final manuscript. Karthik Raghunathan was responsible for contributing to the main conceptual idea, development of the statistical plan, interpretation of results, and manuscript preparation.

Author disclosures
The authors report no actual or potential conflicts of interest with regard to this article.

Disclaimer
The opinions expressed herein are those of the authors and do not necessarily reflect those of Federal Practitioner, Frontline Medical Communications Inc., the US Government, or any of its agencies.

Article PDF
Article PDF

Risk calculators can be of great value in guiding clinical decision making, patient-centered precision medicine, and resource allocation.1 Several perioperative risk prediction models have emerged in recent decades that estimate specific hazards (eg, cardiovascular complications after noncardiac surgery) with varying accuracy and utility. In the perioperative sphere, the time windows are often limited to an index hospitalization or 30 days following surgery or discharge.2-9 Although longer periods are of interest to patients, families, and health systems, few widely used or validated models are designed to look beyond this very narrow window.10,11 In addition, perioperative risk prediction models do not routinely incorporate parameters of a wide variety of health or demographic domains, such as patterns of health care, health care utilization, or medication use.

In 2013, in response to the need for near real-time information to guide delivery of enhanced care management services, the Veterans Health Administration (VHA) Office of Informatics and Analytics developed automated risk prediction models that used detailed electronic health record (EHR) data. These models were used to report Care Assessment Need (CAN) scores each week for all VHA enrollees and include data from a wide array of health domains. These CAN scores predict the risk for hospitalization, death, or either event within 90 days and 1 year.12,13 Each score is reported as both a predicted probability (0-1) and as a percentile in relation to all other VHA enrollees (a value between 1 and 99).13 The data used to calculate CAN scores are listed in Table 1.12

Care Assessment Need Score 2.0 Variables


Surgical procedures or admissions would not be differentiated from nonsurgical admissions or other procedural clinic visits, and as such, it is not possible to isolate the effect of undergoing a surgical procedure from another health-related event on the CAN score. At the same time though, a short-term increase in system utilization caused by an elective surgical procedure such as a total knee replacement (TKR) would presumably be reflected in a change in CAN score, but this has not been studied.

Since their introduction, CAN scores have been routinely accessed by primary care teams and used to facilitate care coordination for thousands of VHA patients. However, these CAN scores are currently not available to VHA surgeons, anesthesiologists, or other perioperative clinicians. In this study, we examine the distributions of preoperative CAN scores and explore the relationships of preoperative CAN 1-year mortality scores with 1-year survival following discharge and length of stay (LOS) during index hospitalization in a cohort of US veterans who underwent TKR, the most common elective operation performed within the VHA system.

Methods

Following approval of the Durham Veterans Affairs Medical Center Institutional Review Board, all necessary data were extracted from the VHA Corporate Data Warehouse (CDW) repository.14 Informed consent was waived due to the minimal risk nature of the study.

We used Current Procedural Terminology codes (27438, 27446, 27447, 27486, 27487, 27488) and International Classification of Diseases, 9th edition clinical modification procedure codes (81.54, 81.55, 81.59, 00.80-00.84) to identify all veterans who had undergone primary or revision TKR between July 2014 and December 2015 in VHA Veterans Integrated Service Network 1 (Maine, Vermont, New Hampshire, Massachusetts, Connecticut, Rhode Island, New York, Pennsylvania, West Virginia, Virginia, North Carolina). Because we focused on outcomes following hospital discharge, patients who died before discharge were excluded from the analysis. Preoperative CAN 1-year mortality score was chosen as the measure under the assumption that long-term survival may be the most meaningful of the 4 possible CAN score measures.

Our primary objective was to determine distribution of preoperative CAN scores in the study population. Our secondary was to study relationships among the preoperative CAN 1-year mortality scores and 1-year mortality and hospital LOS.

Study Variables

For each patient, we extracted the date of index surgery. The primary exposure or independent variable was the CAN score in the week prior to this date. Because prior study has shown that CAN scores trajectories do not significantly change over time, the date-stamped CAN scores in the week before surgery represent what would have been available to clinicians in a preoperative setting.15 Since CAN scores are refreshed and overwritten every week, we extracted archived scores from the CDW.

 

 

For the 1-year survival outcome, the primary dependent variable, we queried the vital status files in the CDW for the date of death if applicable. We confirmed survival beyond 1 year by examining vital signs in the CDW for a minimum of 2 independent encounters beyond 1 year after the date of discharge. To compute the index LOS, the secondary outcome, we computed the difference between the date of admission and date of hospital discharge.

Statistical Methods

The parameters and performance of the multivariable logistic regression models developed to compute the various CAN mortality and hospitalization risk scores have been previously described.12 Briefly, Wang and colleagues created parsimonious regression models using backward selection. Model discrimination was evaluated using C (concordance)-statistic. Model calibration was assessed by comparing predicted vs observed event rates by risk deciles and performing Cox proportional hazards regression.

We plotted histograms to display preoperative CAN scores as a simple measure of distribution (Figure 1). We also examined the cumulative proportion of patients at each preoperative CAN 1-year mortality score.

Distribution of 1-Year Mortality CAN Scores


Using a conventional t test, we compared means of preoperative CAN 1-year mortality scores in patients who survived vs those who died within 1 year. We also constructed a plot of the proportion of patients who had died within 1 year vs preoperative CAN 1-year mortality scores. Kaplan-Meier curves were then constructed examining 1-year survival by CAN 1-year mortality score by terciles.

Finally, we examined the relationship between preoperative CAN 1-year mortality scores and index LOS in 2 ways: We plotted LOS across CAN scores, and we constructed a locally weighted regression and smoothing scatterplot. LOESS (locally weighted polynomial regression) models, involve fitting the polynomial using weighted least squares, giving more weight to points near the point whose response is being estimated and less weight to points further away. LOESS models help fit a smooth curve to data for which linear and nonlinear least squares regression models do not perform well and in situations where no theoretical model exists. LOESS are ideal for modeling complex processes. All analyses were performed using SAS, version 9.3.

Results

We identified 8206 patients who had undergone a TKR over the 18-month study period. The overall mean (SD) for age was 65 (8.41) years; 93% were male, and 78% were White veterans. Patient demographics are well described in a previous publication.16,17

In terms of model parameters for the CAN score models, C-statistics for the 90-day outcome models were as follows: 0.833 for the model predicting hospitalization (95% CI, 0.832-0.834); 0.865 for the model predicting death (95% CI, 0.863-0.876); and 0.811 for the model predicting either event (95% CI, 0.810-0.812). C-statistics for the 1-year outcome models were 0.809 for the model predicting hospitalization (95% CI, 0.808-0.810); 0.851 for the model predicting death (95% CI, 0.849-0.852); and 0.787 for the model predicting either event (95% CI, 0.786-0.787). Models were well calibrated with α = 0 and β = 1, demonstrating strong agreement between observed and predicted event rates.

Relationship Between 1-Year Mortality Rates and 1-Year CAN Scores

Distribution of Hospitalization CAN Scores top chart



The distribution of preoperative CAN 1-year mortality scores was close to normal (median, 50; interquartile range, 40; mean [SD] 48 [25.6]) (eTable). The original CAN score models were developed having an equal number of patients in each strata and as such, are normally distributed.12 Our cohort was similar in pattern of distribution. Distributions of the remaining preoperative CAN scores (90-day mortality, 1-year hospitalization, 90-day hospitalization) are shown in Figures 2, 3, and 4. Not surprisingly, histograms for both 90-day and 1-year hospitalization were skewed toward higher scores, indicating that these patients were expected to be hospitalized in the near future.

Distribution of CAN 1-Year Mortality Scores

Preoperative CAN 1-Year Mortality Scores for Veterans Who Underwent TKR table


Overall, 1.4% (110/8096) of patients died within 1 year of surgery. Comparing 1-year mortality CAN scores in survivors vs nonsurvivors, we found statistically significant differences in means (47 vs 66 respectively, P < .001) and medians (45 vs 75 respectively, P < .001) (Table 2). In the plot examining the relationship between preoperative 1-year mortality CAN scores and 1-year mortality, the percentage who died within 1 year increased initially for patients with CAN scores > 60 and again exponentially for patients with CAN scores > 80. Examining Kaplan-Meier curves, we found that survivors and nonsurvivors separated early after surgery, and the differences between the top tercile and the middle/lower terciles were statistically significant (P < .001). Mortality rates were about 0.5% in the lower and middle terciles but about 2% in the upper tercile (Figure 5).

Average Hospital LOS and 90-Day Mortality CAN Scores

Average Hospital LOS and 90-Day Mortality CAN Scores in Patients Undergoing Elective TKA

Kaplan-Meier Curves


In the plot examining the relationship between CAN scores and index LOS, the LOS rose significantly beyond a CAN score of 60 and dramatically beyond a CAN score of 80 (Figure 6). LOESS curves also showed 2 inflection points suggesting an incremental and sequential rise in the LOS with increasing CAN scores (Figure 7). Mean (SD) LOS in days for the lowest to highest terciles was 2.6 (1.7), 2.8 (2.1), and 3.6 (2.2), respectively.

 

 

Discussion

CAN scores are automatically generated each week by EHR-based multivariable risk models. These scores have excellent predictive accuracy for 90-day and 1-year mortality and hospitalization and are routinely used by VHA primary care teams to assist with clinical operations.13 We studied the distribution of CAN 1-year mortality scores in a preoperative context and examined relationships of the preoperative CAN 1-year mortality scores with postoperative mortality and LOS in 8206 veterans who underwent TKR.

There are several noteworthy findings. First, the overall 1-year mortality rate observed following TKR (1.4%) was similar to other published reports.18,19 Not surprisingly, preoperative CAN 1-year mortality scores were significantly higher in veterans who died compared with those of survivors. The majority of patients who died had a preoperative CAN 1-year mortality score > 75 while most who survived had a preoperative CAN 1-year mortality score < 45 (P < .001). Interestingly, the same scores showed a nonlinear correlation with LOS. Index LOS was about 4 days in patients in the highest tercile of CAN scores vs 2.5 days in the lowest tercile, but the initial increase in LOS was detected at a CAN score of about 55 to 60.

In addition, mortality rate varied widely in different segments of the population when grouped according to preoperative CAN scores. One-year mortality rates in the highest tercile reached 2%, about 4-fold higher than that of lower terciles (0.5%). Examination of the Kaplan-Meier curves showed that this difference in mortality between the highest tercile and the lower 2 groups appears soon after discharge and continues to increase over time, suggesting that the factors contributing to the increased mortality are present at the time of discharge and persist beyond the postoperative period. In summary, although CAN scores were not designed for use in the perioperative context, we found that preoperative CAN 1-year mortality scores are broadly predictive of mortality, but especially for increases in LOS following elective TKA, both increases in hospital LOS following elective TKA and mortality over the year after TKA.

Our findings raise several important questions. The decision to undergo elective surgery is complex. Arguably, individuals who undergo elective knee replacement should be healthy enough to undergo, recover, and reap the benefits from a procedure that does not extend life. The distribution of preoperative CAN 1-year mortality scores for our study population was similar to that of the general VHA enrollee population with similar measured mortality rates (≤ 0.5% vs ≥ 1.7% in the low and high terciles, respectively).1 Further study comparing outcomes in matched cohorts who did and did not undergo joint replacement would be of interest. In lieu of this, though, the association of high but not extreme CAN scores with increased hospital LOS may potentially be used to guide allocation of resources to this group, obviating the increased cost and risk to which this group is exposed. And the additional insight afforded by CAN scores may enhance shared decision-making models by identifying patients at the very highest risk (eg, 1-year mortality CAN score ≥ 90), patients who conceivably might not survive long enough to recover from and enjoy their reconstructed knee, who might in the long run be harmed by undergoing the procedure.

Many total joint arthroplasties are performed in older patients, a population in which frailty is increasingly recognized as a significant risk factor for poor outcomes.20,21 CAN scores reliably identify high-risk patients and have been shown to correlate with frailty in this group.22 Multiple authors have reported improved outcomes with cost reductions after implementation of programs targeting modifiable risk factors in high-risk surgical candidates.23-25 A preoperative assessment that includes the CAN score may be valuable in identifying patients who would benefit most from prehabilitation programs or other interventions designed to blunt the impact of frailty. It is true that many elements used to calculate the CAN score would not be considered modifiable, especially in the short term. However, specific contributors to frailty, such as nutritional status and polypharmacy might be potential candidates. As with all multivariable risk prediction models, there are multiple paths to a high CAN score, and further research to identify clinically relevant subgroups may help inform efforts to improve perioperative care within this population.

Hospital LOS is of intense interest for many reasons, not least its utility as a surrogate for cost and increased risk for immediate perioperative adverse events, such as multidrug-resistant hospital acquired infections, need for postacute facility-based rehabilitation, and deconditioning that increase risks of falls and fractures in the older population.26-29 In addition, its importance is magnified due to the COVID-19 pandemic context in which restarting elective surgery programs has changed traditional criteria by which patients are scheduled for surgery.

We have shown that elevated CAN scores are able to identify patients at risk for extended hospital stays and, as such, may be useful additional data in allocating scarce operating room time and other resources for optimal patient and health care provider safety.30,31 Individual surgeons and hospital systems would, of course, decide which patients should be triaged to go first, based on local priorities; however, choosing lower risk patients with minimal risk of morbidity and mortality while pursuing prehabilitation for higher risk patients is a reasonable approach.

 

 

Limitations

Our study has several limitations. Only a single surgical procedure was included, albeit the most common one performed in the VHA. In addition, no information was available concerning the precise clinical course for these patients, such as the duration of surgery, anesthetic technique, and management of acute, perioperative course. Although the assumption was made that patients received standard care in a manner such that these factors would not significantly affect either their mortality or their LOS out of proportion to their preoperative clinical status, confounding cannot be excluded. Therefore, further study is necessary to determine whether CAN scores can accurately predict mortality and/or LOS for patients undergoing other procedures. Further, a clinical trial is required to assess whether systematic provision of the CAN score at the point of surgery would impact care and, more important, impact outcomes. In addition, multivariable analyses were not performed, including and excluding various components of the CAN score models. Currently, CAN scores could be made available to the surgical/anesthesia communities at minimal or no cost and are updated automatically. Model calibration and discrimination in this particular setting were not validated.

Because our interest is in leveraging an existing resource to a current clinical and operational problem rather than in creating or validating a new tool, we chose to test the simple bivariate relationship between preoperative CAN scores and outcomes. We chose the preoperative 1-year mortality CAN score from among the 4 options under the assumption that long-term survival is the most meaningful of the 4 candidate outcomes. Finally, while the CAN scores are currently only calculated and generated for patients cared for within the VHA, few data elements are unavailable to civilian health systems. The most problematic would be documentation of actual prescription filling, but this is a topic of increasing interest to the medical and academic communities and access to such information we hope will improve.32-34

Conclusions

Although designed for use by VHA primary care teams, CAN scores also may have value for perioperative clinicians, predicting mortality and prolonged hospital LOS in those with elevated 1-year mortality scores. Advantages of CAN scores relative to other perioperative risk calculators lies in their ability to predict long-term rather than 30-day survival and that they are automatically generated on a near-real-time basis for all patients who receive care in VHA ambulatory clinics. Further study is needed to determine practical utility in shared decision making, preoperative evaluation and optimization, and perioperative resource allocation.

Acknowledgments

This work was supported by the US Department of Veterans Affairs (VA) National Center for Patient Safety, Field Office 10A4E, through the Patient Safety Center of Inquiry at the Durham VA Medical Center in North Carolina. The study also received support from the Center of Innovation to Accelerate Discovery and Practice Transformation (CIN 13-410) at the Durham VA Health Care System.

Risk calculators can be of great value in guiding clinical decision making, patient-centered precision medicine, and resource allocation.1 Several perioperative risk prediction models have emerged in recent decades that estimate specific hazards (eg, cardiovascular complications after noncardiac surgery) with varying accuracy and utility. In the perioperative sphere, the time windows are often limited to an index hospitalization or 30 days following surgery or discharge.2-9 Although longer periods are of interest to patients, families, and health systems, few widely used or validated models are designed to look beyond this very narrow window.10,11 In addition, perioperative risk prediction models do not routinely incorporate parameters of a wide variety of health or demographic domains, such as patterns of health care, health care utilization, or medication use.

In 2013, in response to the need for near real-time information to guide delivery of enhanced care management services, the Veterans Health Administration (VHA) Office of Informatics and Analytics developed automated risk prediction models that used detailed electronic health record (EHR) data. These models were used to report Care Assessment Need (CAN) scores each week for all VHA enrollees and include data from a wide array of health domains. These CAN scores predict the risk for hospitalization, death, or either event within 90 days and 1 year.12,13 Each score is reported as both a predicted probability (0-1) and as a percentile in relation to all other VHA enrollees (a value between 1 and 99).13 The data used to calculate CAN scores are listed in Table 1.12

Care Assessment Need Score 2.0 Variables


Surgical procedures or admissions would not be differentiated from nonsurgical admissions or other procedural clinic visits, and as such, it is not possible to isolate the effect of undergoing a surgical procedure from another health-related event on the CAN score. At the same time though, a short-term increase in system utilization caused by an elective surgical procedure such as a total knee replacement (TKR) would presumably be reflected in a change in CAN score, but this has not been studied.

Since their introduction, CAN scores have been routinely accessed by primary care teams and used to facilitate care coordination for thousands of VHA patients. However, these CAN scores are currently not available to VHA surgeons, anesthesiologists, or other perioperative clinicians. In this study, we examine the distributions of preoperative CAN scores and explore the relationships of preoperative CAN 1-year mortality scores with 1-year survival following discharge and length of stay (LOS) during index hospitalization in a cohort of US veterans who underwent TKR, the most common elective operation performed within the VHA system.

Methods

Following approval of the Durham Veterans Affairs Medical Center Institutional Review Board, all necessary data were extracted from the VHA Corporate Data Warehouse (CDW) repository.14 Informed consent was waived due to the minimal risk nature of the study.

We used Current Procedural Terminology codes (27438, 27446, 27447, 27486, 27487, 27488) and International Classification of Diseases, 9th edition clinical modification procedure codes (81.54, 81.55, 81.59, 00.80-00.84) to identify all veterans who had undergone primary or revision TKR between July 2014 and December 2015 in VHA Veterans Integrated Service Network 1 (Maine, Vermont, New Hampshire, Massachusetts, Connecticut, Rhode Island, New York, Pennsylvania, West Virginia, Virginia, North Carolina). Because we focused on outcomes following hospital discharge, patients who died before discharge were excluded from the analysis. Preoperative CAN 1-year mortality score was chosen as the measure under the assumption that long-term survival may be the most meaningful of the 4 possible CAN score measures.

Our primary objective was to determine distribution of preoperative CAN scores in the study population. Our secondary was to study relationships among the preoperative CAN 1-year mortality scores and 1-year mortality and hospital LOS.

Study Variables

For each patient, we extracted the date of index surgery. The primary exposure or independent variable was the CAN score in the week prior to this date. Because prior study has shown that CAN scores trajectories do not significantly change over time, the date-stamped CAN scores in the week before surgery represent what would have been available to clinicians in a preoperative setting.15 Since CAN scores are refreshed and overwritten every week, we extracted archived scores from the CDW.

 

 

For the 1-year survival outcome, the primary dependent variable, we queried the vital status files in the CDW for the date of death if applicable. We confirmed survival beyond 1 year by examining vital signs in the CDW for a minimum of 2 independent encounters beyond 1 year after the date of discharge. To compute the index LOS, the secondary outcome, we computed the difference between the date of admission and date of hospital discharge.

Statistical Methods

The parameters and performance of the multivariable logistic regression models developed to compute the various CAN mortality and hospitalization risk scores have been previously described.12 Briefly, Wang and colleagues created parsimonious regression models using backward selection. Model discrimination was evaluated using C (concordance)-statistic. Model calibration was assessed by comparing predicted vs observed event rates by risk deciles and performing Cox proportional hazards regression.

We plotted histograms to display preoperative CAN scores as a simple measure of distribution (Figure 1). We also examined the cumulative proportion of patients at each preoperative CAN 1-year mortality score.

Distribution of 1-Year Mortality CAN Scores


Using a conventional t test, we compared means of preoperative CAN 1-year mortality scores in patients who survived vs those who died within 1 year. We also constructed a plot of the proportion of patients who had died within 1 year vs preoperative CAN 1-year mortality scores. Kaplan-Meier curves were then constructed examining 1-year survival by CAN 1-year mortality score by terciles.

Finally, we examined the relationship between preoperative CAN 1-year mortality scores and index LOS in 2 ways: We plotted LOS across CAN scores, and we constructed a locally weighted regression and smoothing scatterplot. LOESS (locally weighted polynomial regression) models, involve fitting the polynomial using weighted least squares, giving more weight to points near the point whose response is being estimated and less weight to points further away. LOESS models help fit a smooth curve to data for which linear and nonlinear least squares regression models do not perform well and in situations where no theoretical model exists. LOESS are ideal for modeling complex processes. All analyses were performed using SAS, version 9.3.

Results

We identified 8206 patients who had undergone a TKR over the 18-month study period. The overall mean (SD) for age was 65 (8.41) years; 93% were male, and 78% were White veterans. Patient demographics are well described in a previous publication.16,17

In terms of model parameters for the CAN score models, C-statistics for the 90-day outcome models were as follows: 0.833 for the model predicting hospitalization (95% CI, 0.832-0.834); 0.865 for the model predicting death (95% CI, 0.863-0.876); and 0.811 for the model predicting either event (95% CI, 0.810-0.812). C-statistics for the 1-year outcome models were 0.809 for the model predicting hospitalization (95% CI, 0.808-0.810); 0.851 for the model predicting death (95% CI, 0.849-0.852); and 0.787 for the model predicting either event (95% CI, 0.786-0.787). Models were well calibrated with α = 0 and β = 1, demonstrating strong agreement between observed and predicted event rates.

Relationship Between 1-Year Mortality Rates and 1-Year CAN Scores

Distribution of Hospitalization CAN Scores top chart



The distribution of preoperative CAN 1-year mortality scores was close to normal (median, 50; interquartile range, 40; mean [SD] 48 [25.6]) (eTable). The original CAN score models were developed having an equal number of patients in each strata and as such, are normally distributed.12 Our cohort was similar in pattern of distribution. Distributions of the remaining preoperative CAN scores (90-day mortality, 1-year hospitalization, 90-day hospitalization) are shown in Figures 2, 3, and 4. Not surprisingly, histograms for both 90-day and 1-year hospitalization were skewed toward higher scores, indicating that these patients were expected to be hospitalized in the near future.

Distribution of CAN 1-Year Mortality Scores

Preoperative CAN 1-Year Mortality Scores for Veterans Who Underwent TKR table


Overall, 1.4% (110/8096) of patients died within 1 year of surgery. Comparing 1-year mortality CAN scores in survivors vs nonsurvivors, we found statistically significant differences in means (47 vs 66 respectively, P < .001) and medians (45 vs 75 respectively, P < .001) (Table 2). In the plot examining the relationship between preoperative 1-year mortality CAN scores and 1-year mortality, the percentage who died within 1 year increased initially for patients with CAN scores > 60 and again exponentially for patients with CAN scores > 80. Examining Kaplan-Meier curves, we found that survivors and nonsurvivors separated early after surgery, and the differences between the top tercile and the middle/lower terciles were statistically significant (P < .001). Mortality rates were about 0.5% in the lower and middle terciles but about 2% in the upper tercile (Figure 5).

Average Hospital LOS and 90-Day Mortality CAN Scores

Average Hospital LOS and 90-Day Mortality CAN Scores in Patients Undergoing Elective TKA

Kaplan-Meier Curves


In the plot examining the relationship between CAN scores and index LOS, the LOS rose significantly beyond a CAN score of 60 and dramatically beyond a CAN score of 80 (Figure 6). LOESS curves also showed 2 inflection points suggesting an incremental and sequential rise in the LOS with increasing CAN scores (Figure 7). Mean (SD) LOS in days for the lowest to highest terciles was 2.6 (1.7), 2.8 (2.1), and 3.6 (2.2), respectively.

 

 

Discussion

CAN scores are automatically generated each week by EHR-based multivariable risk models. These scores have excellent predictive accuracy for 90-day and 1-year mortality and hospitalization and are routinely used by VHA primary care teams to assist with clinical operations.13 We studied the distribution of CAN 1-year mortality scores in a preoperative context and examined relationships of the preoperative CAN 1-year mortality scores with postoperative mortality and LOS in 8206 veterans who underwent TKR.

There are several noteworthy findings. First, the overall 1-year mortality rate observed following TKR (1.4%) was similar to other published reports.18,19 Not surprisingly, preoperative CAN 1-year mortality scores were significantly higher in veterans who died compared with those of survivors. The majority of patients who died had a preoperative CAN 1-year mortality score > 75 while most who survived had a preoperative CAN 1-year mortality score < 45 (P < .001). Interestingly, the same scores showed a nonlinear correlation with LOS. Index LOS was about 4 days in patients in the highest tercile of CAN scores vs 2.5 days in the lowest tercile, but the initial increase in LOS was detected at a CAN score of about 55 to 60.

In addition, mortality rate varied widely in different segments of the population when grouped according to preoperative CAN scores. One-year mortality rates in the highest tercile reached 2%, about 4-fold higher than that of lower terciles (0.5%). Examination of the Kaplan-Meier curves showed that this difference in mortality between the highest tercile and the lower 2 groups appears soon after discharge and continues to increase over time, suggesting that the factors contributing to the increased mortality are present at the time of discharge and persist beyond the postoperative period. In summary, although CAN scores were not designed for use in the perioperative context, we found that preoperative CAN 1-year mortality scores are broadly predictive of mortality, but especially for increases in LOS following elective TKA, both increases in hospital LOS following elective TKA and mortality over the year after TKA.

Our findings raise several important questions. The decision to undergo elective surgery is complex. Arguably, individuals who undergo elective knee replacement should be healthy enough to undergo, recover, and reap the benefits from a procedure that does not extend life. The distribution of preoperative CAN 1-year mortality scores for our study population was similar to that of the general VHA enrollee population with similar measured mortality rates (≤ 0.5% vs ≥ 1.7% in the low and high terciles, respectively).1 Further study comparing outcomes in matched cohorts who did and did not undergo joint replacement would be of interest. In lieu of this, though, the association of high but not extreme CAN scores with increased hospital LOS may potentially be used to guide allocation of resources to this group, obviating the increased cost and risk to which this group is exposed. And the additional insight afforded by CAN scores may enhance shared decision-making models by identifying patients at the very highest risk (eg, 1-year mortality CAN score ≥ 90), patients who conceivably might not survive long enough to recover from and enjoy their reconstructed knee, who might in the long run be harmed by undergoing the procedure.

Many total joint arthroplasties are performed in older patients, a population in which frailty is increasingly recognized as a significant risk factor for poor outcomes.20,21 CAN scores reliably identify high-risk patients and have been shown to correlate with frailty in this group.22 Multiple authors have reported improved outcomes with cost reductions after implementation of programs targeting modifiable risk factors in high-risk surgical candidates.23-25 A preoperative assessment that includes the CAN score may be valuable in identifying patients who would benefit most from prehabilitation programs or other interventions designed to blunt the impact of frailty. It is true that many elements used to calculate the CAN score would not be considered modifiable, especially in the short term. However, specific contributors to frailty, such as nutritional status and polypharmacy might be potential candidates. As with all multivariable risk prediction models, there are multiple paths to a high CAN score, and further research to identify clinically relevant subgroups may help inform efforts to improve perioperative care within this population.

Hospital LOS is of intense interest for many reasons, not least its utility as a surrogate for cost and increased risk for immediate perioperative adverse events, such as multidrug-resistant hospital acquired infections, need for postacute facility-based rehabilitation, and deconditioning that increase risks of falls and fractures in the older population.26-29 In addition, its importance is magnified due to the COVID-19 pandemic context in which restarting elective surgery programs has changed traditional criteria by which patients are scheduled for surgery.

We have shown that elevated CAN scores are able to identify patients at risk for extended hospital stays and, as such, may be useful additional data in allocating scarce operating room time and other resources for optimal patient and health care provider safety.30,31 Individual surgeons and hospital systems would, of course, decide which patients should be triaged to go first, based on local priorities; however, choosing lower risk patients with minimal risk of morbidity and mortality while pursuing prehabilitation for higher risk patients is a reasonable approach.

 

 

Limitations

Our study has several limitations. Only a single surgical procedure was included, albeit the most common one performed in the VHA. In addition, no information was available concerning the precise clinical course for these patients, such as the duration of surgery, anesthetic technique, and management of acute, perioperative course. Although the assumption was made that patients received standard care in a manner such that these factors would not significantly affect either their mortality or their LOS out of proportion to their preoperative clinical status, confounding cannot be excluded. Therefore, further study is necessary to determine whether CAN scores can accurately predict mortality and/or LOS for patients undergoing other procedures. Further, a clinical trial is required to assess whether systematic provision of the CAN score at the point of surgery would impact care and, more important, impact outcomes. In addition, multivariable analyses were not performed, including and excluding various components of the CAN score models. Currently, CAN scores could be made available to the surgical/anesthesia communities at minimal or no cost and are updated automatically. Model calibration and discrimination in this particular setting were not validated.

Because our interest is in leveraging an existing resource to a current clinical and operational problem rather than in creating or validating a new tool, we chose to test the simple bivariate relationship between preoperative CAN scores and outcomes. We chose the preoperative 1-year mortality CAN score from among the 4 options under the assumption that long-term survival is the most meaningful of the 4 candidate outcomes. Finally, while the CAN scores are currently only calculated and generated for patients cared for within the VHA, few data elements are unavailable to civilian health systems. The most problematic would be documentation of actual prescription filling, but this is a topic of increasing interest to the medical and academic communities and access to such information we hope will improve.32-34

Conclusions

Although designed for use by VHA primary care teams, CAN scores also may have value for perioperative clinicians, predicting mortality and prolonged hospital LOS in those with elevated 1-year mortality scores. Advantages of CAN scores relative to other perioperative risk calculators lies in their ability to predict long-term rather than 30-day survival and that they are automatically generated on a near-real-time basis for all patients who receive care in VHA ambulatory clinics. Further study is needed to determine practical utility in shared decision making, preoperative evaluation and optimization, and perioperative resource allocation.

Acknowledgments

This work was supported by the US Department of Veterans Affairs (VA) National Center for Patient Safety, Field Office 10A4E, through the Patient Safety Center of Inquiry at the Durham VA Medical Center in North Carolina. The study also received support from the Center of Innovation to Accelerate Discovery and Practice Transformation (CIN 13-410) at the Durham VA Health Care System.

References

1. McNair AGK, MacKichan F, Donovan JL, et al. What surgeons tell patients and what patients want to know before major cancer surgery: a qualitative study. BMC Cancer. 2016;16:258. doi:10.1186/s12885-016-2292-3

2. Grover FL, Hammermeister KE, Burchfiel C. Initial report of the Veterans Administration Preoperative Risk Assessment Study for Cardiac Surgery. Ann Thorac Surg. 1990;50(1):12-26; discussion 27-18. doi:10.1016/0003-4975(90)90073-f

3. Khuri SF, Daley J, Henderson W, et al. The National Veterans Administration Surgical Risk Study: risk adjustment for the comparative assessment of the quality of surgical care. J Am Coll Surg. 1995;180(5):519-531.

4. Glance LG, Lustik SJ, Hannan EL, et al. The Surgical Mortality Probability Model: derivation and validation of a simple simple risk prediction rule for noncardiac surgery. Ann Surg. 2012;255(4):696-702. doi:10.1097/SLA.0b013e31824b45af

5. Keller DS, Kroll D, Papaconstantinou HT, Ellis CN. Development and validation of a methodology to reduce mortality using the veterans affairs surgical quality improvement program risk calculator. J Am Coll Surg. 2017;224(4):602-607. doi:10.1016/j.jamcollsurg.2016.12.033

6. Bilimoria KY, Liu Y, Paruch JL, et al. Development and evaluation of the universal ACS NSQIP surgical risk calculator: a decision aid and informed consent tool for patients and surgeons. J Am Coll Surg. 2013;217(5):833-842.e831-833. doi:10.1016/j.jamcollsurg.2013.07.385

7. Ford MK, Beattie WS, Wijeysundera DN. Systematic review: prediction of perioperative cardiac complications and mortality by the revised cardiac risk index. Ann Intern Med. 2010;152(1):26-35. doi:10.7326/0003-4819-152-1-201001050-00007

8. Gupta PK, Gupta H, Sundaram A, et al. Development and validation of a risk calculator for prediction of cardiac risk after surgery. Circulation. 2011;124(4):381-387. doi:10.1161/CIRCULATIONAHA.110.015701

9. Lee TH, Marcantonio ER, Mangione CM, et al. Derivation and prospective validation of a simple index for prediction of cardiac risk of major noncardiac surgery. Circulation. 1999;100(10):1043-1049. doi:10.1161/01.cir.100.10.1043

10. Smith T, Li X, Nylander W, Gunnar W. Thirty-day postoperative mortality risk estimates and 1-year survival in Veterans Health Administration surgery patients. JAMA Surg. 2016;151(5):417-422. doi:10.1001/jamasurg.2015.4882

11. Damhuis RA, Wijnhoven BP, Plaisier PW, Kirkels WJ, Kranse R, van Lanschot JJ. Comparison of 30-day, 90- day and in-hospital postoperative mortality for eight different cancer types. Br J Surg. 2012;99(8):1149-1154. doi:10.1002/bjs.8813

12. Wang L, Porter B, Maynard C, et al. Predicting risk of hospitalization or death among patients receiving primary care in the Veterans Health Administration. Med Care. 2013;51(4):368-373. doi:10.1016/j.amjcard.2012.06.038

13. Fihn SD, Francis J, Clancy C, et al. Insights from advanced analytics at the Veterans Health Administration. Health Aff (Millwood). 2014;33(7):1203-1211. doi:10.1377/hlthaff.2014.0054

14. Noël PH, Copeland LA, Perrin RA, et al. VHA Corporate Data Warehouse height and weight data: opportunities and challenges for health services research. J Rehabil Res Dev. 2010;47(8):739-750. doi:10.1682/jrrd.2009.08.0110

15. Wong ES, Yoon J, Piegari RI, Rosland AM, Fihn SD, Chang ET. Identifying latent subgroups of high-risk patients using risk score trajectories. J Gen Intern Med. 2018;33(12):2120-2126. doi:10.1007/s11606-018-4653-x

16. Chen Q, Hsia HL, Overman R, et al. Impact of an opioid safety initiative on patients undergoing total knee arthroplasty: a time series analysis. Anesthesiology. 2019;131(2):369-380. doi:10.1097/ALN.0000000000002771

17. Hsia HL, Takemoto S, van de Ven T, et al. Acute pain is associated with chronic opioid use after total knee arthroplasty. Reg Anesth Pain Med. 2018;43(7):705-711. doi:10.1097/AAP.0000000000000831

18. Inacio MCS, Dillon MT, Miric A, Navarro RA, Paxton EW. Mortality after total knee and total hip arthroplasty in a large integrated health care system. Perm J. 2017;21:16-171. doi:10.7812/TPP/16-171

19. Lee QJ, Mak WP, Wong YC. Mortality following primary total knee replacement in public hospitals in Hong Kong. Hong Kong Med J. 2016;22(3):237-241. doi:10.12809/hkmj154712

20. Lin HS, Watts JN, Peel NM, Hubbard RE. Frailty and post-operative outcomes in older surgical patients: a systematic review. BMC Geriatr. 2016;16(1):157. doi:10.1186/s12877-016-0329-8

21. Shinall MC Jr, Arya S, Youk A, et al. Association of preoperative patient frailty and operative stress with postoperative mortality. JAMA Surg. 2019;155(1):e194620. doi:10.1001/jamasurg.2019.4620

22. Ruiz JG, Priyadarshni S, Rahaman Z, et al. Validation of an automatically generated screening score for frailty: the care assessment need (CAN) score. BMC Geriatr. 2018;18(1):106. doi:10.1186/s12877-018-0802-7

23. Bernstein DN, Liu TC, Winegar AL, et al. Evaluation of a preoperative optimization protocol for primary hip and knee arthroplasty patients. J Arthroplasty. 2018;33(12):3642- 3648. doi:10.1016/j.arth.2018.08.018

24. Sodhi N, Anis HK, Coste M, et al. A nationwide analysis of preoperative planning on operative times and postoperative complications in total knee arthroplasty. J Knee Surg. 2019;32(11):1040-1045. doi:10.1055/s-0039-1677790

25. Krause A, Sayeed Z, El-Othmani M, Pallekonda V, Mihalko W, Saleh KJ. Outpatient total knee arthroplasty: are we there yet? (part 1). Orthop Clin North Am. 2018;49(1):1-6. doi:10.1016/j.ocl.2017.08.002

26. Barrasa-Villar JI, Aibar-Remón C, Prieto-Andrés P, Mareca- Doñate R, Moliner-Lahoz J. Impact on morbidity, mortality, and length of stay of hospital-acquired infections by resistant microorganisms. Clin Infect Dis. 2017;65(4):644-652. doi:10.1093/cid/cix411

27. Nikkel LE, Kates SL, Schreck M, Maceroli M, Mahmood B, Elfar JC. Length of hospital stay after hip fracture and risk of early mortality after discharge in New York state: retrospective cohort study. BMJ. 2015;351:h6246. doi:10.1136/bmj.h6246

28. Marfil-Garza BA, Belaunzarán-Zamudio PF, Gulias-Herrero A, et al. Risk factors associated with prolonged hospital length-of-stay: 18-year retrospective study of hospitalizations in a tertiary healthcare center in Mexico. PLoS One. 2018;13(11):e0207203. doi:10.1371/journal.pone.0207203

29. Hirsch CH, Sommers L, Olsen A, Mullen L, Winograd CH. The natural history of functional morbidity in hospitalized older patients. J Am Geriatr Soc. 1990;38(12):1296-1303. doi:10.1111/j.1532-5415.1990.tb03451.x

30. Iyengar KP, Jain VK, Vaish A, Vaishya R, Maini L, Lal H. Post COVID-19: planning strategies to resume orthopaedic surgery -challenges and considerations. J Clin Orthop Trauma. 2020;11(suppl 3):S291-S295. doi:10.1016/j.jcot.2020.04.028

31. O’Connor CM, Anoushiravani AA, DiCaprio MR, Healy WL, Iorio R. Economic recovery after the COVID-19 pandemic: resuming elective orthopedic surgery and total joint arthroplasty. J Arthroplasty. 2020;35(suppl 7):S32-S36. doi:10.1016/j.arth.2020.04.038.

32. Mauseth SA, Skurtveit S, Skovlund E, Langhammer A, Spigset O. Medication use and association with urinary incontinence in women: data from the Norwegian Prescription Database and the HUNT study. Neurourol Urodyn. 2018;37(4):1448-1457. doi:10.1002/nau.23473

33. Sultan RS, Correll CU, Schoenbaum M, King M, Walkup JT, Olfson M. National patterns of commonly prescribed psychotropic medications to young people. J Child Adolesc Psychopharmacol. 2018;28(3):158-165. doi:10.1089/cap.2017.0077

34. McCoy RG, Dykhoff HJ, Sangaralingham L, et al. Adoption of new glucose-lowering medications in the U.S.-the case of SGLT2 inhibitors: nationwide cohort study. Diabetes Technol Ther. 2019;21(12):702-712. doi:10.1089/dia.2019.0213

References

1. McNair AGK, MacKichan F, Donovan JL, et al. What surgeons tell patients and what patients want to know before major cancer surgery: a qualitative study. BMC Cancer. 2016;16:258. doi:10.1186/s12885-016-2292-3

2. Grover FL, Hammermeister KE, Burchfiel C. Initial report of the Veterans Administration Preoperative Risk Assessment Study for Cardiac Surgery. Ann Thorac Surg. 1990;50(1):12-26; discussion 27-18. doi:10.1016/0003-4975(90)90073-f

3. Khuri SF, Daley J, Henderson W, et al. The National Veterans Administration Surgical Risk Study: risk adjustment for the comparative assessment of the quality of surgical care. J Am Coll Surg. 1995;180(5):519-531.

4. Glance LG, Lustik SJ, Hannan EL, et al. The Surgical Mortality Probability Model: derivation and validation of a simple simple risk prediction rule for noncardiac surgery. Ann Surg. 2012;255(4):696-702. doi:10.1097/SLA.0b013e31824b45af

5. Keller DS, Kroll D, Papaconstantinou HT, Ellis CN. Development and validation of a methodology to reduce mortality using the veterans affairs surgical quality improvement program risk calculator. J Am Coll Surg. 2017;224(4):602-607. doi:10.1016/j.jamcollsurg.2016.12.033

6. Bilimoria KY, Liu Y, Paruch JL, et al. Development and evaluation of the universal ACS NSQIP surgical risk calculator: a decision aid and informed consent tool for patients and surgeons. J Am Coll Surg. 2013;217(5):833-842.e831-833. doi:10.1016/j.jamcollsurg.2013.07.385

7. Ford MK, Beattie WS, Wijeysundera DN. Systematic review: prediction of perioperative cardiac complications and mortality by the revised cardiac risk index. Ann Intern Med. 2010;152(1):26-35. doi:10.7326/0003-4819-152-1-201001050-00007

8. Gupta PK, Gupta H, Sundaram A, et al. Development and validation of a risk calculator for prediction of cardiac risk after surgery. Circulation. 2011;124(4):381-387. doi:10.1161/CIRCULATIONAHA.110.015701

9. Lee TH, Marcantonio ER, Mangione CM, et al. Derivation and prospective validation of a simple index for prediction of cardiac risk of major noncardiac surgery. Circulation. 1999;100(10):1043-1049. doi:10.1161/01.cir.100.10.1043

10. Smith T, Li X, Nylander W, Gunnar W. Thirty-day postoperative mortality risk estimates and 1-year survival in Veterans Health Administration surgery patients. JAMA Surg. 2016;151(5):417-422. doi:10.1001/jamasurg.2015.4882

11. Damhuis RA, Wijnhoven BP, Plaisier PW, Kirkels WJ, Kranse R, van Lanschot JJ. Comparison of 30-day, 90- day and in-hospital postoperative mortality for eight different cancer types. Br J Surg. 2012;99(8):1149-1154. doi:10.1002/bjs.8813

12. Wang L, Porter B, Maynard C, et al. Predicting risk of hospitalization or death among patients receiving primary care in the Veterans Health Administration. Med Care. 2013;51(4):368-373. doi:10.1016/j.amjcard.2012.06.038

13. Fihn SD, Francis J, Clancy C, et al. Insights from advanced analytics at the Veterans Health Administration. Health Aff (Millwood). 2014;33(7):1203-1211. doi:10.1377/hlthaff.2014.0054

14. Noël PH, Copeland LA, Perrin RA, et al. VHA Corporate Data Warehouse height and weight data: opportunities and challenges for health services research. J Rehabil Res Dev. 2010;47(8):739-750. doi:10.1682/jrrd.2009.08.0110

15. Wong ES, Yoon J, Piegari RI, Rosland AM, Fihn SD, Chang ET. Identifying latent subgroups of high-risk patients using risk score trajectories. J Gen Intern Med. 2018;33(12):2120-2126. doi:10.1007/s11606-018-4653-x

16. Chen Q, Hsia HL, Overman R, et al. Impact of an opioid safety initiative on patients undergoing total knee arthroplasty: a time series analysis. Anesthesiology. 2019;131(2):369-380. doi:10.1097/ALN.0000000000002771

17. Hsia HL, Takemoto S, van de Ven T, et al. Acute pain is associated with chronic opioid use after total knee arthroplasty. Reg Anesth Pain Med. 2018;43(7):705-711. doi:10.1097/AAP.0000000000000831

18. Inacio MCS, Dillon MT, Miric A, Navarro RA, Paxton EW. Mortality after total knee and total hip arthroplasty in a large integrated health care system. Perm J. 2017;21:16-171. doi:10.7812/TPP/16-171

19. Lee QJ, Mak WP, Wong YC. Mortality following primary total knee replacement in public hospitals in Hong Kong. Hong Kong Med J. 2016;22(3):237-241. doi:10.12809/hkmj154712

20. Lin HS, Watts JN, Peel NM, Hubbard RE. Frailty and post-operative outcomes in older surgical patients: a systematic review. BMC Geriatr. 2016;16(1):157. doi:10.1186/s12877-016-0329-8

21. Shinall MC Jr, Arya S, Youk A, et al. Association of preoperative patient frailty and operative stress with postoperative mortality. JAMA Surg. 2019;155(1):e194620. doi:10.1001/jamasurg.2019.4620

22. Ruiz JG, Priyadarshni S, Rahaman Z, et al. Validation of an automatically generated screening score for frailty: the care assessment need (CAN) score. BMC Geriatr. 2018;18(1):106. doi:10.1186/s12877-018-0802-7

23. Bernstein DN, Liu TC, Winegar AL, et al. Evaluation of a preoperative optimization protocol for primary hip and knee arthroplasty patients. J Arthroplasty. 2018;33(12):3642- 3648. doi:10.1016/j.arth.2018.08.018

24. Sodhi N, Anis HK, Coste M, et al. A nationwide analysis of preoperative planning on operative times and postoperative complications in total knee arthroplasty. J Knee Surg. 2019;32(11):1040-1045. doi:10.1055/s-0039-1677790

25. Krause A, Sayeed Z, El-Othmani M, Pallekonda V, Mihalko W, Saleh KJ. Outpatient total knee arthroplasty: are we there yet? (part 1). Orthop Clin North Am. 2018;49(1):1-6. doi:10.1016/j.ocl.2017.08.002

26. Barrasa-Villar JI, Aibar-Remón C, Prieto-Andrés P, Mareca- Doñate R, Moliner-Lahoz J. Impact on morbidity, mortality, and length of stay of hospital-acquired infections by resistant microorganisms. Clin Infect Dis. 2017;65(4):644-652. doi:10.1093/cid/cix411

27. Nikkel LE, Kates SL, Schreck M, Maceroli M, Mahmood B, Elfar JC. Length of hospital stay after hip fracture and risk of early mortality after discharge in New York state: retrospective cohort study. BMJ. 2015;351:h6246. doi:10.1136/bmj.h6246

28. Marfil-Garza BA, Belaunzarán-Zamudio PF, Gulias-Herrero A, et al. Risk factors associated with prolonged hospital length-of-stay: 18-year retrospective study of hospitalizations in a tertiary healthcare center in Mexico. PLoS One. 2018;13(11):e0207203. doi:10.1371/journal.pone.0207203

29. Hirsch CH, Sommers L, Olsen A, Mullen L, Winograd CH. The natural history of functional morbidity in hospitalized older patients. J Am Geriatr Soc. 1990;38(12):1296-1303. doi:10.1111/j.1532-5415.1990.tb03451.x

30. Iyengar KP, Jain VK, Vaish A, Vaishya R, Maini L, Lal H. Post COVID-19: planning strategies to resume orthopaedic surgery -challenges and considerations. J Clin Orthop Trauma. 2020;11(suppl 3):S291-S295. doi:10.1016/j.jcot.2020.04.028

31. O’Connor CM, Anoushiravani AA, DiCaprio MR, Healy WL, Iorio R. Economic recovery after the COVID-19 pandemic: resuming elective orthopedic surgery and total joint arthroplasty. J Arthroplasty. 2020;35(suppl 7):S32-S36. doi:10.1016/j.arth.2020.04.038.

32. Mauseth SA, Skurtveit S, Skovlund E, Langhammer A, Spigset O. Medication use and association with urinary incontinence in women: data from the Norwegian Prescription Database and the HUNT study. Neurourol Urodyn. 2018;37(4):1448-1457. doi:10.1002/nau.23473

33. Sultan RS, Correll CU, Schoenbaum M, King M, Walkup JT, Olfson M. National patterns of commonly prescribed psychotropic medications to young people. J Child Adolesc Psychopharmacol. 2018;28(3):158-165. doi:10.1089/cap.2017.0077

34. McCoy RG, Dykhoff HJ, Sangaralingham L, et al. Adoption of new glucose-lowering medications in the U.S.-the case of SGLT2 inhibitors: nationwide cohort study. Diabetes Technol Ther. 2019;21(12):702-712. doi:10.1089/dia.2019.0213

Issue
Federal Practitioner - 38(7)a
Issue
Federal Practitioner - 38(7)a
Page Number
316-324
Page Number
316-324
Publications
Publications
Topics
Article Type
Sections
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media

The Hospital Readmissions Reduction Program: Inconvenient Observations

Article Type
Changed
Thu, 07/01/2021 - 11:26

Centers for Medicare and Medicaid Services (CMS)–promulgated quality metrics continue to attract critics. Physicians decry that many metrics are outside their control, while patient groups are frustrated that metrics lack meaning for beneficiaries. The Hospital Readmissions Reduction Program (HRRP) reduces payments for “excess” 30-day risk-standardized readmissions for six conditions and procedures, and may be less effective in reducing readmissions than previously reported due to intentional and increasing use of hospital observation stays.1

In this issue, Sheehy et al2 report that nearly one in five rehospitalizations were unrecognized because either the index hospitalization or the rehospitalization was an observation stay, highlighting yet another challenge with the HRRP. Limitations of their study include the use of a single year of claims data and the exclusion of Medicare Advantage claims data, as one might expect lower readmission rates in this capitated program. Opportunities for improving the HRRP could consist of updating the HRRP metric to include observation stays and, for surgical hospitalizations, extended-stay surgical recovery, wherein patients may be observed for up to 2 days following a procedure. Unfortunately, despite the HRRP missing nearly one in five readmissions, CMS would likely need additional statutory authority from Congress in order to reinterpret the definition of readmission3 to include observation stays.

Challenges with the HRRP metrics raise broader concerns about the program. For decades, administrators viewed readmissions as a utilization metric, only to have the Affordable Care Act re-designate and define all-cause readmissions as a quality metric. Yet hospitals and health systems control only some factors driving readmission. Readmissions occur for a variety of reasons, including not only poor quality of initial hospital care and inadequate care coordination, but also factors that are beyond the hospital’s purview, such as lack of access to ambulatory services, multiple and severe chronic conditions that progress or remain unresponsive to intervention,4 and demographic and social factors such as housing instability, health literacy, or residence in a food desert. These non-hospital factors reside within the domain of other market participants or local, state, and federal government agencies.

Challenges to the utility, validity, and appropriateness of HRRP metrics should remind policymakers of the dangers of over-legislating the details of healthcare policy and the statutory inflexibility that can ensue. Clinical care evolves, and artificial constructs—including payment categories such as observation status—may age poorly over time, exemplified best by the challenges of accessing post-acute care due to the 3-day rule.5 Introduced as a statutory requirement in 1967, when the average length of stay was 13.8 days and observation care did not exist as a payment category, the 3-day rule requires Medicare beneficiaries to spend 3 days admitted to the hospital in order to qualify for coverage of post-acute care, creating care gaps for observation stay patients.

Observation care itself is an artificial construct of CMS payment policy. In the Medicare program, observation care falls under Part B, exposing patients to both greater financial responsibility and billing complexity through the engagement of their supplemental insurance, even though those receiving observation care experience the same care as if hospitalized— routine monitoring, nursing care, blood draws, imaging, and diagnostic tests. While CMS requires notification of observation status and explanation of the difference in patient financial responsibility, in clinical practice, patient understanding is limited. Policymakers can support both Medicare beneficiaries and hospitals by reexamining observation care as a payment category.

Sheehy and colleagues’ work simultaneously challenges the face validity of the HRRP and the reasonableness of categorizing some inpatient stays as outpatient care in the hospital—issues that policymakers can and should address.

References

1. Sabbatini AK, Wright B. Excluding observation stays from readmission rates – what quality measures are missing. N Engl J Med. 2018;378(22):2062-2065. https://doi.org/10.1056/NEJMp1800732
2. Sheehy AM, Kaiksow F, Powell WR, et al. The hospital readmissions reduction program’s blind spot: observation hospitalizations. J Hosp Med. 2021;16(7):409-411. https://doi.org/10.12788/jhm.3634
3. The Patient Protection and Affordable Care Act, 42 USC 18001§3025 (2010).
4. Reuben DB, Tinetti ME. The hospital-dependent patient. N Engl J Med. 2014;370(8):694-697. https://doi.org/10.1056/NEJMp1315568
5. Patel N, Slota JM, Miller BJ. The continued conundrum of discharge to a skilled nursing facility after a medicare observation stay. JAMA Health Forum. 2020;1(5):e200577. https://doi.org/10.1001/jamahealthforum.2020.0577

Article PDF
Author and Disclosure Information

1Division of Hospital Medicine, Department of Medicine, The Johns Hopkins University School of Medicine, Baltimore, Maryland; 2Johns Hopkins Carey Business School, Baltimore, Maryland; 3Johns Hopkins University School of Nursing, Baltimore, Maryland.

Disclosures
Dr Miller formerly served as a Fellow at the Centers for Medicare & Medicaid Services. He reports serving as a member of the CMS Medicare Evidence Development and Coverage Advisory Committee, and receiving fees outside the related work from the Federal Trade Commission, the Health Resources and Services Administration, Radyus Research, Oxidien Pharmaceuticals, and the Heritage Foundation. Ms Deutschendorf and Dr Brotman report no relevant conflicts of interest.

Issue
Journal of Hospital Medicine 16(7)
Publications
Topics
Page Number
448
Sections
Author and Disclosure Information

1Division of Hospital Medicine, Department of Medicine, The Johns Hopkins University School of Medicine, Baltimore, Maryland; 2Johns Hopkins Carey Business School, Baltimore, Maryland; 3Johns Hopkins University School of Nursing, Baltimore, Maryland.

Disclosures
Dr Miller formerly served as a Fellow at the Centers for Medicare & Medicaid Services. He reports serving as a member of the CMS Medicare Evidence Development and Coverage Advisory Committee, and receiving fees outside the related work from the Federal Trade Commission, the Health Resources and Services Administration, Radyus Research, Oxidien Pharmaceuticals, and the Heritage Foundation. Ms Deutschendorf and Dr Brotman report no relevant conflicts of interest.

Author and Disclosure Information

1Division of Hospital Medicine, Department of Medicine, The Johns Hopkins University School of Medicine, Baltimore, Maryland; 2Johns Hopkins Carey Business School, Baltimore, Maryland; 3Johns Hopkins University School of Nursing, Baltimore, Maryland.

Disclosures
Dr Miller formerly served as a Fellow at the Centers for Medicare & Medicaid Services. He reports serving as a member of the CMS Medicare Evidence Development and Coverage Advisory Committee, and receiving fees outside the related work from the Federal Trade Commission, the Health Resources and Services Administration, Radyus Research, Oxidien Pharmaceuticals, and the Heritage Foundation. Ms Deutschendorf and Dr Brotman report no relevant conflicts of interest.

Article PDF
Article PDF
Related Articles

Centers for Medicare and Medicaid Services (CMS)–promulgated quality metrics continue to attract critics. Physicians decry that many metrics are outside their control, while patient groups are frustrated that metrics lack meaning for beneficiaries. The Hospital Readmissions Reduction Program (HRRP) reduces payments for “excess” 30-day risk-standardized readmissions for six conditions and procedures, and may be less effective in reducing readmissions than previously reported due to intentional and increasing use of hospital observation stays.1

In this issue, Sheehy et al2 report that nearly one in five rehospitalizations were unrecognized because either the index hospitalization or the rehospitalization was an observation stay, highlighting yet another challenge with the HRRP. Limitations of their study include the use of a single year of claims data and the exclusion of Medicare Advantage claims data, as one might expect lower readmission rates in this capitated program. Opportunities for improving the HRRP could consist of updating the HRRP metric to include observation stays and, for surgical hospitalizations, extended-stay surgical recovery, wherein patients may be observed for up to 2 days following a procedure. Unfortunately, despite the HRRP missing nearly one in five readmissions, CMS would likely need additional statutory authority from Congress in order to reinterpret the definition of readmission3 to include observation stays.

Challenges with the HRRP metrics raise broader concerns about the program. For decades, administrators viewed readmissions as a utilization metric, only to have the Affordable Care Act re-designate and define all-cause readmissions as a quality metric. Yet hospitals and health systems control only some factors driving readmission. Readmissions occur for a variety of reasons, including not only poor quality of initial hospital care and inadequate care coordination, but also factors that are beyond the hospital’s purview, such as lack of access to ambulatory services, multiple and severe chronic conditions that progress or remain unresponsive to intervention,4 and demographic and social factors such as housing instability, health literacy, or residence in a food desert. These non-hospital factors reside within the domain of other market participants or local, state, and federal government agencies.

Challenges to the utility, validity, and appropriateness of HRRP metrics should remind policymakers of the dangers of over-legislating the details of healthcare policy and the statutory inflexibility that can ensue. Clinical care evolves, and artificial constructs—including payment categories such as observation status—may age poorly over time, exemplified best by the challenges of accessing post-acute care due to the 3-day rule.5 Introduced as a statutory requirement in 1967, when the average length of stay was 13.8 days and observation care did not exist as a payment category, the 3-day rule requires Medicare beneficiaries to spend 3 days admitted to the hospital in order to qualify for coverage of post-acute care, creating care gaps for observation stay patients.

Observation care itself is an artificial construct of CMS payment policy. In the Medicare program, observation care falls under Part B, exposing patients to both greater financial responsibility and billing complexity through the engagement of their supplemental insurance, even though those receiving observation care experience the same care as if hospitalized— routine monitoring, nursing care, blood draws, imaging, and diagnostic tests. While CMS requires notification of observation status and explanation of the difference in patient financial responsibility, in clinical practice, patient understanding is limited. Policymakers can support both Medicare beneficiaries and hospitals by reexamining observation care as a payment category.

Sheehy and colleagues’ work simultaneously challenges the face validity of the HRRP and the reasonableness of categorizing some inpatient stays as outpatient care in the hospital—issues that policymakers can and should address.

Centers for Medicare and Medicaid Services (CMS)–promulgated quality metrics continue to attract critics. Physicians decry that many metrics are outside their control, while patient groups are frustrated that metrics lack meaning for beneficiaries. The Hospital Readmissions Reduction Program (HRRP) reduces payments for “excess” 30-day risk-standardized readmissions for six conditions and procedures, and may be less effective in reducing readmissions than previously reported due to intentional and increasing use of hospital observation stays.1

In this issue, Sheehy et al2 report that nearly one in five rehospitalizations were unrecognized because either the index hospitalization or the rehospitalization was an observation stay, highlighting yet another challenge with the HRRP. Limitations of their study include the use of a single year of claims data and the exclusion of Medicare Advantage claims data, as one might expect lower readmission rates in this capitated program. Opportunities for improving the HRRP could consist of updating the HRRP metric to include observation stays and, for surgical hospitalizations, extended-stay surgical recovery, wherein patients may be observed for up to 2 days following a procedure. Unfortunately, despite the HRRP missing nearly one in five readmissions, CMS would likely need additional statutory authority from Congress in order to reinterpret the definition of readmission3 to include observation stays.

Challenges with the HRRP metrics raise broader concerns about the program. For decades, administrators viewed readmissions as a utilization metric, only to have the Affordable Care Act re-designate and define all-cause readmissions as a quality metric. Yet hospitals and health systems control only some factors driving readmission. Readmissions occur for a variety of reasons, including not only poor quality of initial hospital care and inadequate care coordination, but also factors that are beyond the hospital’s purview, such as lack of access to ambulatory services, multiple and severe chronic conditions that progress or remain unresponsive to intervention,4 and demographic and social factors such as housing instability, health literacy, or residence in a food desert. These non-hospital factors reside within the domain of other market participants or local, state, and federal government agencies.

Challenges to the utility, validity, and appropriateness of HRRP metrics should remind policymakers of the dangers of over-legislating the details of healthcare policy and the statutory inflexibility that can ensue. Clinical care evolves, and artificial constructs—including payment categories such as observation status—may age poorly over time, exemplified best by the challenges of accessing post-acute care due to the 3-day rule.5 Introduced as a statutory requirement in 1967, when the average length of stay was 13.8 days and observation care did not exist as a payment category, the 3-day rule requires Medicare beneficiaries to spend 3 days admitted to the hospital in order to qualify for coverage of post-acute care, creating care gaps for observation stay patients.

Observation care itself is an artificial construct of CMS payment policy. In the Medicare program, observation care falls under Part B, exposing patients to both greater financial responsibility and billing complexity through the engagement of their supplemental insurance, even though those receiving observation care experience the same care as if hospitalized— routine monitoring, nursing care, blood draws, imaging, and diagnostic tests. While CMS requires notification of observation status and explanation of the difference in patient financial responsibility, in clinical practice, patient understanding is limited. Policymakers can support both Medicare beneficiaries and hospitals by reexamining observation care as a payment category.

Sheehy and colleagues’ work simultaneously challenges the face validity of the HRRP and the reasonableness of categorizing some inpatient stays as outpatient care in the hospital—issues that policymakers can and should address.

References

1. Sabbatini AK, Wright B. Excluding observation stays from readmission rates – what quality measures are missing. N Engl J Med. 2018;378(22):2062-2065. https://doi.org/10.1056/NEJMp1800732
2. Sheehy AM, Kaiksow F, Powell WR, et al. The hospital readmissions reduction program’s blind spot: observation hospitalizations. J Hosp Med. 2021;16(7):409-411. https://doi.org/10.12788/jhm.3634
3. The Patient Protection and Affordable Care Act, 42 USC 18001§3025 (2010).
4. Reuben DB, Tinetti ME. The hospital-dependent patient. N Engl J Med. 2014;370(8):694-697. https://doi.org/10.1056/NEJMp1315568
5. Patel N, Slota JM, Miller BJ. The continued conundrum of discharge to a skilled nursing facility after a medicare observation stay. JAMA Health Forum. 2020;1(5):e200577. https://doi.org/10.1001/jamahealthforum.2020.0577

References

1. Sabbatini AK, Wright B. Excluding observation stays from readmission rates – what quality measures are missing. N Engl J Med. 2018;378(22):2062-2065. https://doi.org/10.1056/NEJMp1800732
2. Sheehy AM, Kaiksow F, Powell WR, et al. The hospital readmissions reduction program’s blind spot: observation hospitalizations. J Hosp Med. 2021;16(7):409-411. https://doi.org/10.12788/jhm.3634
3. The Patient Protection and Affordable Care Act, 42 USC 18001§3025 (2010).
4. Reuben DB, Tinetti ME. The hospital-dependent patient. N Engl J Med. 2014;370(8):694-697. https://doi.org/10.1056/NEJMp1315568
5. Patel N, Slota JM, Miller BJ. The continued conundrum of discharge to a skilled nursing facility after a medicare observation stay. JAMA Health Forum. 2020;1(5):e200577. https://doi.org/10.1001/jamahealthforum.2020.0577

Issue
Journal of Hospital Medicine 16(7)
Issue
Journal of Hospital Medicine 16(7)
Page Number
448
Page Number
448
Publications
Publications
Topics
Article Type
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Brian J. Miller, MD, MBA, MPH; Email: [email protected]; Telephone: 410-614-4474; Twitter: 4_BetterHealth.
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Page Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media

Measuring Trainee Duty Hours: The Times They Are a-Changin’

Article Type
Changed
Thu, 07/01/2021 - 11:08
Display Headline
Measuring Trainee Duty Hours: The Times They Are a-Changin’

“If your time to you is worth savin’

Then you better start swimmin’ or you’ll sink like a stone

For the times they are a-changin’...”

–Bob Dylan

The Accreditation Council for Graduate Medical Education requires residency programs to limit and track trainee work hours to reduce the risk of fatigue, burnout, and medical errors. These hours are documented most often by self-report, at the cost of additional administrative burden for trainees and programs, dubious accuracy, and potentially incentivizing misrepresentation.1

Thus, the study by Soleimani and colleagues2 in this issue is a welcome addition to the literature on duty-hours tracking. Using timestamp data from the electronic health record (EHR), the authors developed and collected validity evidence for an automated computerized algorithm to measure how much time trainees were spending on clinical work. The study was conducted at a large academic internal medicine residency program and tracked 203 trainees working 14,610 days. The authors compared their results to trainee self-report data. Though the approach centered on EHR access logs, it accommodated common scenarios of time away from the computer while at the hospital (eg, during patient rounds). Crucially, the algorithm included EHR access while at home. The absolute discrepancy between the algorithm and self-report averaged 1.38 hours per day. Notably, EHR work at home accounted for about an extra hour per day. When considering in-hospital work alone, the authors found 3% to 13% of trainees exceeding 80-hour workweek limits, but when adding out-of-hospital work, this percentage rose to 10% to 21%.

The authors used inventive methods to improve accuracy. They prespecified EHR functions that constituted active clinical work, classifying reading without editing notes or placing orders simply as “educational study,” which they excluded from duty hours. They ensured that time spent off-site was included and that logins from personal devices while in-hospital were not double-counted. Caveats to the study include the limited generalizability for institutions without the computational resources to replicate the model. The authors acknowledged the inherent flaw in using trainee self-report as the “gold standard,” and potentially some subset of the results could have been corroborated with time-motion observation studies.3 The decision to exclude passive medical record review at home as work arguably discounts the integral value that the “chart biopsy” has on direct patient care; it probably led to systematic underestimation of duty hours for junior and senior residents, who may be most likely to contribute in this way. Similarly, not counting time spent with patients at the end of the day after sign-out risks undercounting hours as well. Nonetheless, this study represents a rigorously designed and scalable approach to meeting regulatory requirements that can potentially lighten the administrative task load for trainees, improve reporting accuracy, and facilitate research comparing work hours to other variables of interest (eg, efficiency). The model can be generalized to other specialties and could document workload for staff physicians as well.

Merits of the study aside, the algorithm underscores troubling realities about the practice of medicine in the 21st century. Do we now equate clinical work with time on the computer? Is our contribution as physicians defined primarily by our presence at the keyboard, rather than the bedside?4 Future research facilitated by automated hours tracking is likely to further elucidate a connection between time spent in the EHR with burnout4 and job dissatisfaction, and the premise of this study is emblematic of the erosion of clinical work-life boundaries that began even before the pandemic.5 While the “times they are a-changin’,” in this respect, it may not be for the better.

References

1. Grabski DF, Goudreau BJ, Gillen JR, et al. Compliance with the Accreditation Council for Graduate Medical Education duty hours in a general surgery residency program: challenges and solutions in a teaching hospital. Surgery. 2020;167(2):302-307. https://doi.org/10.1016/j.surg.2019.05.029
2. Soleimani H, Adler-Milstein J, Cucina RJ, Murray SG. Automating measurement of trainee work hours. J Hosp Med. 2021;16(7):404-408. https://doi.org/10.12788/jhm.3607
3. Tipping MD, Forth VE, O’Leary KJ, et al. Where did the day go?—a time-motion study of hospitalists. J Hosp Med. 2010;5(6):323-328. https://doi.org/10.1002/jhm.790
4. Gardner RL, Cooper E, Haskell J, et al. Physician stress and burnout: the impact of health information technology. J Am Med Inform Assoc. 2019;26(2):106-114. https://doi.org/10.1093/jamia/ocy145
5. Saag HS, Shah K, Jones SA, Testa PA, Horwitz LI. Pajama time: working after work in the electronic health record. J Gen Intern Med. 2019;34(9):1695-1696. https://doi.org/10.1007/s11606-019-05055-x

Article PDF
Author and Disclosure Information

1Section of Hospital Medicine, Department of Medicine, Beth Israel Deaconess Medical Center, Boston, Massachusetts; 2Harvard Medical School, Boston, Massachusetts.

Disclosures
The authors have no conflicts of interest to disclose.

Issue
Journal of Hospital Medicine 16(7)
Publications
Topics
Page Number
477
Sections
Author and Disclosure Information

1Section of Hospital Medicine, Department of Medicine, Beth Israel Deaconess Medical Center, Boston, Massachusetts; 2Harvard Medical School, Boston, Massachusetts.

Disclosures
The authors have no conflicts of interest to disclose.

Author and Disclosure Information

1Section of Hospital Medicine, Department of Medicine, Beth Israel Deaconess Medical Center, Boston, Massachusetts; 2Harvard Medical School, Boston, Massachusetts.

Disclosures
The authors have no conflicts of interest to disclose.

Article PDF
Article PDF
Related Articles

“If your time to you is worth savin’

Then you better start swimmin’ or you’ll sink like a stone

For the times they are a-changin’...”

–Bob Dylan

The Accreditation Council for Graduate Medical Education requires residency programs to limit and track trainee work hours to reduce the risk of fatigue, burnout, and medical errors. These hours are documented most often by self-report, at the cost of additional administrative burden for trainees and programs, dubious accuracy, and potentially incentivizing misrepresentation.1

Thus, the study by Soleimani and colleagues2 in this issue is a welcome addition to the literature on duty-hours tracking. Using timestamp data from the electronic health record (EHR), the authors developed and collected validity evidence for an automated computerized algorithm to measure how much time trainees were spending on clinical work. The study was conducted at a large academic internal medicine residency program and tracked 203 trainees working 14,610 days. The authors compared their results to trainee self-report data. Though the approach centered on EHR access logs, it accommodated common scenarios of time away from the computer while at the hospital (eg, during patient rounds). Crucially, the algorithm included EHR access while at home. The absolute discrepancy between the algorithm and self-report averaged 1.38 hours per day. Notably, EHR work at home accounted for about an extra hour per day. When considering in-hospital work alone, the authors found 3% to 13% of trainees exceeding 80-hour workweek limits, but when adding out-of-hospital work, this percentage rose to 10% to 21%.

The authors used inventive methods to improve accuracy. They prespecified EHR functions that constituted active clinical work, classifying reading without editing notes or placing orders simply as “educational study,” which they excluded from duty hours. They ensured that time spent off-site was included and that logins from personal devices while in-hospital were not double-counted. Caveats to the study include the limited generalizability for institutions without the computational resources to replicate the model. The authors acknowledged the inherent flaw in using trainee self-report as the “gold standard,” and potentially some subset of the results could have been corroborated with time-motion observation studies.3 The decision to exclude passive medical record review at home as work arguably discounts the integral value that the “chart biopsy” has on direct patient care; it probably led to systematic underestimation of duty hours for junior and senior residents, who may be most likely to contribute in this way. Similarly, not counting time spent with patients at the end of the day after sign-out risks undercounting hours as well. Nonetheless, this study represents a rigorously designed and scalable approach to meeting regulatory requirements that can potentially lighten the administrative task load for trainees, improve reporting accuracy, and facilitate research comparing work hours to other variables of interest (eg, efficiency). The model can be generalized to other specialties and could document workload for staff physicians as well.

Merits of the study aside, the algorithm underscores troubling realities about the practice of medicine in the 21st century. Do we now equate clinical work with time on the computer? Is our contribution as physicians defined primarily by our presence at the keyboard, rather than the bedside?4 Future research facilitated by automated hours tracking is likely to further elucidate a connection between time spent in the EHR with burnout4 and job dissatisfaction, and the premise of this study is emblematic of the erosion of clinical work-life boundaries that began even before the pandemic.5 While the “times they are a-changin’,” in this respect, it may not be for the better.

“If your time to you is worth savin’

Then you better start swimmin’ or you’ll sink like a stone

For the times they are a-changin’...”

–Bob Dylan

The Accreditation Council for Graduate Medical Education requires residency programs to limit and track trainee work hours to reduce the risk of fatigue, burnout, and medical errors. These hours are documented most often by self-report, at the cost of additional administrative burden for trainees and programs, dubious accuracy, and potentially incentivizing misrepresentation.1

Thus, the study by Soleimani and colleagues2 in this issue is a welcome addition to the literature on duty-hours tracking. Using timestamp data from the electronic health record (EHR), the authors developed and collected validity evidence for an automated computerized algorithm to measure how much time trainees were spending on clinical work. The study was conducted at a large academic internal medicine residency program and tracked 203 trainees working 14,610 days. The authors compared their results to trainee self-report data. Though the approach centered on EHR access logs, it accommodated common scenarios of time away from the computer while at the hospital (eg, during patient rounds). Crucially, the algorithm included EHR access while at home. The absolute discrepancy between the algorithm and self-report averaged 1.38 hours per day. Notably, EHR work at home accounted for about an extra hour per day. When considering in-hospital work alone, the authors found 3% to 13% of trainees exceeding 80-hour workweek limits, but when adding out-of-hospital work, this percentage rose to 10% to 21%.

The authors used inventive methods to improve accuracy. They prespecified EHR functions that constituted active clinical work, classifying reading without editing notes or placing orders simply as “educational study,” which they excluded from duty hours. They ensured that time spent off-site was included and that logins from personal devices while in-hospital were not double-counted. Caveats to the study include the limited generalizability for institutions without the computational resources to replicate the model. The authors acknowledged the inherent flaw in using trainee self-report as the “gold standard,” and potentially some subset of the results could have been corroborated with time-motion observation studies.3 The decision to exclude passive medical record review at home as work arguably discounts the integral value that the “chart biopsy” has on direct patient care; it probably led to systematic underestimation of duty hours for junior and senior residents, who may be most likely to contribute in this way. Similarly, not counting time spent with patients at the end of the day after sign-out risks undercounting hours as well. Nonetheless, this study represents a rigorously designed and scalable approach to meeting regulatory requirements that can potentially lighten the administrative task load for trainees, improve reporting accuracy, and facilitate research comparing work hours to other variables of interest (eg, efficiency). The model can be generalized to other specialties and could document workload for staff physicians as well.

Merits of the study aside, the algorithm underscores troubling realities about the practice of medicine in the 21st century. Do we now equate clinical work with time on the computer? Is our contribution as physicians defined primarily by our presence at the keyboard, rather than the bedside?4 Future research facilitated by automated hours tracking is likely to further elucidate a connection between time spent in the EHR with burnout4 and job dissatisfaction, and the premise of this study is emblematic of the erosion of clinical work-life boundaries that began even before the pandemic.5 While the “times they are a-changin’,” in this respect, it may not be for the better.

References

1. Grabski DF, Goudreau BJ, Gillen JR, et al. Compliance with the Accreditation Council for Graduate Medical Education duty hours in a general surgery residency program: challenges and solutions in a teaching hospital. Surgery. 2020;167(2):302-307. https://doi.org/10.1016/j.surg.2019.05.029
2. Soleimani H, Adler-Milstein J, Cucina RJ, Murray SG. Automating measurement of trainee work hours. J Hosp Med. 2021;16(7):404-408. https://doi.org/10.12788/jhm.3607
3. Tipping MD, Forth VE, O’Leary KJ, et al. Where did the day go?—a time-motion study of hospitalists. J Hosp Med. 2010;5(6):323-328. https://doi.org/10.1002/jhm.790
4. Gardner RL, Cooper E, Haskell J, et al. Physician stress and burnout: the impact of health information technology. J Am Med Inform Assoc. 2019;26(2):106-114. https://doi.org/10.1093/jamia/ocy145
5. Saag HS, Shah K, Jones SA, Testa PA, Horwitz LI. Pajama time: working after work in the electronic health record. J Gen Intern Med. 2019;34(9):1695-1696. https://doi.org/10.1007/s11606-019-05055-x

References

1. Grabski DF, Goudreau BJ, Gillen JR, et al. Compliance with the Accreditation Council for Graduate Medical Education duty hours in a general surgery residency program: challenges and solutions in a teaching hospital. Surgery. 2020;167(2):302-307. https://doi.org/10.1016/j.surg.2019.05.029
2. Soleimani H, Adler-Milstein J, Cucina RJ, Murray SG. Automating measurement of trainee work hours. J Hosp Med. 2021;16(7):404-408. https://doi.org/10.12788/jhm.3607
3. Tipping MD, Forth VE, O’Leary KJ, et al. Where did the day go?—a time-motion study of hospitalists. J Hosp Med. 2010;5(6):323-328. https://doi.org/10.1002/jhm.790
4. Gardner RL, Cooper E, Haskell J, et al. Physician stress and burnout: the impact of health information technology. J Am Med Inform Assoc. 2019;26(2):106-114. https://doi.org/10.1093/jamia/ocy145
5. Saag HS, Shah K, Jones SA, Testa PA, Horwitz LI. Pajama time: working after work in the electronic health record. J Gen Intern Med. 2019;34(9):1695-1696. https://doi.org/10.1007/s11606-019-05055-x

Issue
Journal of Hospital Medicine 16(7)
Issue
Journal of Hospital Medicine 16(7)
Page Number
477
Page Number
477
Publications
Publications
Topics
Article Type
Display Headline
Measuring Trainee Duty Hours: The Times They Are a-Changin’
Display Headline
Measuring Trainee Duty Hours: The Times They Are a-Changin’
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Grace C Huang, MD; Email: [email protected]; Telephone: 617-686-1428; Twitter: @GraceHuangMD.
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Page Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media

The Medical Liability Environment: Is It Really Any Worse for Hospitalists?

Article Type
Changed
Thu, 07/01/2021 - 11:04
Display Headline
The Medical Liability Environment: Is It Really Any Worse for Hospitalists?

Although malpractice “crises” come and go, liability fears persist near top of mind for most physicians.1 Liability insurance premiums have plateaued in recent years, but remain at high levels, and the prospect of being reported to the National Practitioner Data Bank (NPDB) or listed on a state medical board’s website for a paid liability claim is unsettling. The high-acuity setting and the absence of longitudinal patient relationships in hospital medicine may theoretically raise malpractice risk, yet hospitalists’ liability risk remains understudied.2

The contribution by Schaffer and colleagues3 in this issue of the Journal of Hospital Medicine is thus welcome and illuminating. The researchers examine the liability risk of hospitalists compared to that of other specialties by utilizing a large database of malpractice claims compiled from multiple insurers across a decade.3 In a field of research plagued by inadequate data, the Comparative Benchmarking System (CBS) built by CRICO/RMF is a treasure. Unlike the primary national database of malpractice claims, the NPDB, the CBS contains information on claims that did not result in a payment, as well as physicians’ specialty and detailed information on the allegations, injuries, and their causes. The CBS contains almost a third of all medical liability claims made in the United States during the study period, supporting generalizability.

Schaffer and colleagues1 found that hospitalists had a lower claims rate than physicians in emergency medicine or neurosurgery. The rate was on par with that for non-hospital general internists, even though hospitalists often care for higher-acuity patients. Although claims rates dropped over the study period for physicians in neurosurgery, emergency medicine, psychiatry, and internal medicine subspecialties, the rate for hospitalists did not change significantly. Further, the median payout on claims against hospitalists was the highest of all the specialties examined, except neurosurgery. This reflects higher injury severity in hospitalist cases: half the claims against hospitalists involved death and three-quarters were high severity.

The study is not without limitations. Due to missing data, only a fraction of the claims (8.2% to 11%) in the full dataset are used in the claims rate analysis. Regression models predicting a payment are based on a small number of payments for hospitalists (n = 363). Further, the authors advance, as a potential explanation for hospitalists’ higher liability risk, that hospitalists are disproportionately young compared to other specialists, but the dataset lacks age data. These limitations suggest caution in the authors’ overall conclusion that “the malpractice environment for hospitalists is becoming less favorable.”

Nevertheless, several important insights emerge from their analysis. The very existence of claims demonstrates that patient harm continues. The contributing factors and judgment errors found in these claims demonstrate that much of this harm is potentially preventable and a risk to patient safety. Whether or not the authors’ young-hospitalist hypothesis is ultimately proven, it is difficult to argue with more mentorship as a means to improve safety. Also, preventing or intercepting judgment errors remains a vexing challenge in medicine that undoubtedly calls for creative clinical decision support solutions. Schaffer and colleagues1 also note that hospitalists are increasingly co-managing patients with other specialties, such as orthopedic surgery. Whether this new practice model drives hospitalist liability risk because hospitalists are practicing in areas in which they have less experience (as the authors posit) or whether hospitalists are simply more likely to be named in a suit as part of a specialty team with higher liability risk remains unknown and merits further investigation.

Ultimately, regardless of whether the liability environment is worsening for hospitalists, the need to improve our liability system is clear. There is room to improve the system on a number of metrics, including properly compensating negligently harmed patients without unduly burdening providers. The system also induces defensive medicine and has not driven safety improvements as expected. The liability environment, as a result, remains challenging not just for hospitalists, but for all patients and physicians as well.

References

1. Sage WM, Boothman RC, Gallagher TH. Another medical malpractice crisis? Try something different. JAMA. 2020;324(14):1395-1396. https://doi.org/10.1001/jama.2020.16557
2. Schaffer AC, Puopolo AL, Raman S, Kachalia A. Liability impact of the hospitalist model of care. J Hosp Med. 2014;9(12):750-755. https://doi.org/10.1002/jhm.2244
3. Schaffer AC, Yu-Moe CW, Babayan A, Wachter RM, Einbinder JS. Rates and characteristics of medical malpractice claims against hospitalists. J Hosp Med. 2021;16(7):390-396. https://doi.org/10.12788/jhm.3557

Article PDF
Author and Disclosure Information

1Armstrong Institute for Patient Safety and Quality, and Department of Medicine, Johns Hopkins Medicine, Baltimore, Maryland; 2Stanford Law School, Stanford, California; 3Stanford Health Policy and Department of Medicine, Stanford University School of Medicine, Stanford, California; 4Freeman Spogli Institute for International Studies, Stanford, California.

Disclosures 
Drs Kachalia and Mello report receiving grant funding through the Massachusetts Alliance for Communication and Resolution following Medical Injury (MACRMI) for work on a project implementing and evaluating communication-and-resolution programs in Massachusetts hospitals; funding for that project came partially from CRICO, which employs authors of the study that the present commentary concerns.

Issue
Journal of Hospital Medicine 16(7)
Publications
Topics
Page Number
446
Sections
Author and Disclosure Information

1Armstrong Institute for Patient Safety and Quality, and Department of Medicine, Johns Hopkins Medicine, Baltimore, Maryland; 2Stanford Law School, Stanford, California; 3Stanford Health Policy and Department of Medicine, Stanford University School of Medicine, Stanford, California; 4Freeman Spogli Institute for International Studies, Stanford, California.

Disclosures 
Drs Kachalia and Mello report receiving grant funding through the Massachusetts Alliance for Communication and Resolution following Medical Injury (MACRMI) for work on a project implementing and evaluating communication-and-resolution programs in Massachusetts hospitals; funding for that project came partially from CRICO, which employs authors of the study that the present commentary concerns.

Author and Disclosure Information

1Armstrong Institute for Patient Safety and Quality, and Department of Medicine, Johns Hopkins Medicine, Baltimore, Maryland; 2Stanford Law School, Stanford, California; 3Stanford Health Policy and Department of Medicine, Stanford University School of Medicine, Stanford, California; 4Freeman Spogli Institute for International Studies, Stanford, California.

Disclosures 
Drs Kachalia and Mello report receiving grant funding through the Massachusetts Alliance for Communication and Resolution following Medical Injury (MACRMI) for work on a project implementing and evaluating communication-and-resolution programs in Massachusetts hospitals; funding for that project came partially from CRICO, which employs authors of the study that the present commentary concerns.

Article PDF
Article PDF
Related Articles

Although malpractice “crises” come and go, liability fears persist near top of mind for most physicians.1 Liability insurance premiums have plateaued in recent years, but remain at high levels, and the prospect of being reported to the National Practitioner Data Bank (NPDB) or listed on a state medical board’s website for a paid liability claim is unsettling. The high-acuity setting and the absence of longitudinal patient relationships in hospital medicine may theoretically raise malpractice risk, yet hospitalists’ liability risk remains understudied.2

The contribution by Schaffer and colleagues3 in this issue of the Journal of Hospital Medicine is thus welcome and illuminating. The researchers examine the liability risk of hospitalists compared to that of other specialties by utilizing a large database of malpractice claims compiled from multiple insurers across a decade.3 In a field of research plagued by inadequate data, the Comparative Benchmarking System (CBS) built by CRICO/RMF is a treasure. Unlike the primary national database of malpractice claims, the NPDB, the CBS contains information on claims that did not result in a payment, as well as physicians’ specialty and detailed information on the allegations, injuries, and their causes. The CBS contains almost a third of all medical liability claims made in the United States during the study period, supporting generalizability.

Schaffer and colleagues1 found that hospitalists had a lower claims rate than physicians in emergency medicine or neurosurgery. The rate was on par with that for non-hospital general internists, even though hospitalists often care for higher-acuity patients. Although claims rates dropped over the study period for physicians in neurosurgery, emergency medicine, psychiatry, and internal medicine subspecialties, the rate for hospitalists did not change significantly. Further, the median payout on claims against hospitalists was the highest of all the specialties examined, except neurosurgery. This reflects higher injury severity in hospitalist cases: half the claims against hospitalists involved death and three-quarters were high severity.

The study is not without limitations. Due to missing data, only a fraction of the claims (8.2% to 11%) in the full dataset are used in the claims rate analysis. Regression models predicting a payment are based on a small number of payments for hospitalists (n = 363). Further, the authors advance, as a potential explanation for hospitalists’ higher liability risk, that hospitalists are disproportionately young compared to other specialists, but the dataset lacks age data. These limitations suggest caution in the authors’ overall conclusion that “the malpractice environment for hospitalists is becoming less favorable.”

Nevertheless, several important insights emerge from their analysis. The very existence of claims demonstrates that patient harm continues. The contributing factors and judgment errors found in these claims demonstrate that much of this harm is potentially preventable and a risk to patient safety. Whether or not the authors’ young-hospitalist hypothesis is ultimately proven, it is difficult to argue with more mentorship as a means to improve safety. Also, preventing or intercepting judgment errors remains a vexing challenge in medicine that undoubtedly calls for creative clinical decision support solutions. Schaffer and colleagues1 also note that hospitalists are increasingly co-managing patients with other specialties, such as orthopedic surgery. Whether this new practice model drives hospitalist liability risk because hospitalists are practicing in areas in which they have less experience (as the authors posit) or whether hospitalists are simply more likely to be named in a suit as part of a specialty team with higher liability risk remains unknown and merits further investigation.

Ultimately, regardless of whether the liability environment is worsening for hospitalists, the need to improve our liability system is clear. There is room to improve the system on a number of metrics, including properly compensating negligently harmed patients without unduly burdening providers. The system also induces defensive medicine and has not driven safety improvements as expected. The liability environment, as a result, remains challenging not just for hospitalists, but for all patients and physicians as well.

Although malpractice “crises” come and go, liability fears persist near top of mind for most physicians.1 Liability insurance premiums have plateaued in recent years, but remain at high levels, and the prospect of being reported to the National Practitioner Data Bank (NPDB) or listed on a state medical board’s website for a paid liability claim is unsettling. The high-acuity setting and the absence of longitudinal patient relationships in hospital medicine may theoretically raise malpractice risk, yet hospitalists’ liability risk remains understudied.2

The contribution by Schaffer and colleagues3 in this issue of the Journal of Hospital Medicine is thus welcome and illuminating. The researchers examine the liability risk of hospitalists compared to that of other specialties by utilizing a large database of malpractice claims compiled from multiple insurers across a decade.3 In a field of research plagued by inadequate data, the Comparative Benchmarking System (CBS) built by CRICO/RMF is a treasure. Unlike the primary national database of malpractice claims, the NPDB, the CBS contains information on claims that did not result in a payment, as well as physicians’ specialty and detailed information on the allegations, injuries, and their causes. The CBS contains almost a third of all medical liability claims made in the United States during the study period, supporting generalizability.

Schaffer and colleagues1 found that hospitalists had a lower claims rate than physicians in emergency medicine or neurosurgery. The rate was on par with that for non-hospital general internists, even though hospitalists often care for higher-acuity patients. Although claims rates dropped over the study period for physicians in neurosurgery, emergency medicine, psychiatry, and internal medicine subspecialties, the rate for hospitalists did not change significantly. Further, the median payout on claims against hospitalists was the highest of all the specialties examined, except neurosurgery. This reflects higher injury severity in hospitalist cases: half the claims against hospitalists involved death and three-quarters were high severity.

The study is not without limitations. Due to missing data, only a fraction of the claims (8.2% to 11%) in the full dataset are used in the claims rate analysis. Regression models predicting a payment are based on a small number of payments for hospitalists (n = 363). Further, the authors advance, as a potential explanation for hospitalists’ higher liability risk, that hospitalists are disproportionately young compared to other specialists, but the dataset lacks age data. These limitations suggest caution in the authors’ overall conclusion that “the malpractice environment for hospitalists is becoming less favorable.”

Nevertheless, several important insights emerge from their analysis. The very existence of claims demonstrates that patient harm continues. The contributing factors and judgment errors found in these claims demonstrate that much of this harm is potentially preventable and a risk to patient safety. Whether or not the authors’ young-hospitalist hypothesis is ultimately proven, it is difficult to argue with more mentorship as a means to improve safety. Also, preventing or intercepting judgment errors remains a vexing challenge in medicine that undoubtedly calls for creative clinical decision support solutions. Schaffer and colleagues1 also note that hospitalists are increasingly co-managing patients with other specialties, such as orthopedic surgery. Whether this new practice model drives hospitalist liability risk because hospitalists are practicing in areas in which they have less experience (as the authors posit) or whether hospitalists are simply more likely to be named in a suit as part of a specialty team with higher liability risk remains unknown and merits further investigation.

Ultimately, regardless of whether the liability environment is worsening for hospitalists, the need to improve our liability system is clear. There is room to improve the system on a number of metrics, including properly compensating negligently harmed patients without unduly burdening providers. The system also induces defensive medicine and has not driven safety improvements as expected. The liability environment, as a result, remains challenging not just for hospitalists, but for all patients and physicians as well.

References

1. Sage WM, Boothman RC, Gallagher TH. Another medical malpractice crisis? Try something different. JAMA. 2020;324(14):1395-1396. https://doi.org/10.1001/jama.2020.16557
2. Schaffer AC, Puopolo AL, Raman S, Kachalia A. Liability impact of the hospitalist model of care. J Hosp Med. 2014;9(12):750-755. https://doi.org/10.1002/jhm.2244
3. Schaffer AC, Yu-Moe CW, Babayan A, Wachter RM, Einbinder JS. Rates and characteristics of medical malpractice claims against hospitalists. J Hosp Med. 2021;16(7):390-396. https://doi.org/10.12788/jhm.3557

References

1. Sage WM, Boothman RC, Gallagher TH. Another medical malpractice crisis? Try something different. JAMA. 2020;324(14):1395-1396. https://doi.org/10.1001/jama.2020.16557
2. Schaffer AC, Puopolo AL, Raman S, Kachalia A. Liability impact of the hospitalist model of care. J Hosp Med. 2014;9(12):750-755. https://doi.org/10.1002/jhm.2244
3. Schaffer AC, Yu-Moe CW, Babayan A, Wachter RM, Einbinder JS. Rates and characteristics of medical malpractice claims against hospitalists. J Hosp Med. 2021;16(7):390-396. https://doi.org/10.12788/jhm.3557

Issue
Journal of Hospital Medicine 16(7)
Issue
Journal of Hospital Medicine 16(7)
Page Number
446
Page Number
446
Publications
Publications
Topics
Article Type
Display Headline
The Medical Liability Environment: Is It Really Any Worse for Hospitalists?
Display Headline
The Medical Liability Environment: Is It Really Any Worse for Hospitalists?
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Allen Kachalia, MD, JD; Email: [email protected]; Twitter: @allenkachalia.
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Page Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media

Leadership & Professional Development: Cultivating Microcultures of Well-being

Article Type
Changed
Thu, 07/01/2021 - 10:53
Display Headline
Leadership & Professional Development: Cultivating Microcultures of Well-being

“As we work to create light for others, we naturally light our own way.”

– Mary Anne Radmacher

Perhaps unknowingly, hospitalists establish microcultures in their everyday work. Hospitalists’ interactions with colleagues often occur in the context of shared workspaces. The nature of these seemingly minor exchanges shapes the microculture, often described as the culture shared by a small group based on location within an organization. Hospitalists have an opportunity to cultivate well-being within these microcultures through gracious and thoughtful acknowledgments of their peers. Collegial support at the micro level influences wellness at the organizational level. A larger shared culture of wellness is necessary to nurture physicians’ personal fulfillment and professional development.1

We propose the CARE framework for cultivating well-being within the microcultures of hospital medicine shared workspaces. CARE consists of Capitalization, Active listening, Recognition, and Empathy. This framework is based on positive psychology research and inspired by lessons from The Happiness Advantage by Shawn Achor.2

Capitalization. Capitalization is defined as sharing upbeat news and receiving a positive reaction. Emotional support during good times, more so than during bad times, strengthens relationships. When a peer shares good news, show enthusiasm and counter with an active, constructive response to maximize the validation she perceives.2

For example, Alex sits at her desk and says to Kristen:

My workshop proposal was accepted for medical education day!”

Congratulations, Alex! Tell me more about the workshop.”

Active listening. Active listening requires concentration and observation of body language. Show engagement by maintaining an open posture, using positive facial expressions, and providing occasional cues that you’re paying attention. Paraphrasing and asking targeted questions to dive deeper demonstrates genuine interest. 

“Katie, I could use your advice. Do you have a minute?”

Katie turns to face John and smiles. “Of course. How can I help?”

My team seems drained after a code this morning. I planned a lecture for later, but I’m not sure this is the right time.”

Katie nods. “I think you’re right, John. How have you thought about handling the situation?”

Recognition. Acts of recognition and encouragement are catalysts for boosting morale. Even brief expressions of gratitude can have a significant emotional impact. Recognition is most meaningful when delivered deliberately and with warmth.

Kevin walks into the hospitalist workroom. “Diane, congratulations on your publication! I plan to make a medication interaction review part of my discharge workflow.”

Leah turns to Diane. “Diane, that’s great news! Can you send me the link to your article?”

Empathy. Burnout is prevalent in medicine, and our fellow hospitalists deserve empathy. Showing empathy reduces stress and promotes connectedness. Sense when your colleagues are in distress and take time to share in their feelings and emotions. Draw on your own clinical experience to find common ground and convey understanding.

“I transferred another patient with COVID-19 to the ICU. I spent the last hour talking to family.”

“Ashwin, you’ve had a tough week. I know how you must feel—I had to transfer a patient yesterday. Want to take a quick walk outside?”

Hospitalists are inherently busy while on service, but these four interventions are brief, requiring only several minutes. Each small investment of your time will pay significant emotional dividends. These practices will not only enhance your colleagues’ sense of well-being, but will also bolster your happiness and productivity. A positive mindset fosters creative thinking and enhances complex problem solving. Recharging the microcultures of hospitalist workspaces with positivity will spark a larger transformation at the organizational level. That’s because positive actions are contagious.2 One hospitalist’s commitment to CARE will encourage other hospitalists to adopt these behaviors, establishing a virtuous cycle that sustains an organization’s culture of wellness.

References

1. Bohman B, Dyrbye L, Sinsky CA, et al. Physician well-being: the reciprocity of practice efficiency, culture of wellness, and personal resilience. NEJM Catalyst. August 7, 2017. Accessed June 24, 2021. https://catalyst.nejm.org/doi/full/10.1056/CAT.17.0429
2. Achor S. The Happiness Advantage: How a Positive Brain Fuels Success in Work and Life. Currency; 2010.

Article PDF
Author and Disclosure Information

1Department of Medicine, Beth Israel Deaconess Medical Center, Boston, Massachusetts; 2Harvard Medical School, Boston, Massachusetts; 3Carl J Shapiro Institute for Education and Research, Boston, Massachusetts.

Disclosures
The authors have no conflicts to disclose.

Issue
Journal of Hospital Medicine 16(7)
Publications
Topics
Page Number
416
Sections
Author and Disclosure Information

1Department of Medicine, Beth Israel Deaconess Medical Center, Boston, Massachusetts; 2Harvard Medical School, Boston, Massachusetts; 3Carl J Shapiro Institute for Education and Research, Boston, Massachusetts.

Disclosures
The authors have no conflicts to disclose.

Author and Disclosure Information

1Department of Medicine, Beth Israel Deaconess Medical Center, Boston, Massachusetts; 2Harvard Medical School, Boston, Massachusetts; 3Carl J Shapiro Institute for Education and Research, Boston, Massachusetts.

Disclosures
The authors have no conflicts to disclose.

Article PDF
Article PDF
Related Articles

“As we work to create light for others, we naturally light our own way.”

– Mary Anne Radmacher

Perhaps unknowingly, hospitalists establish microcultures in their everyday work. Hospitalists’ interactions with colleagues often occur in the context of shared workspaces. The nature of these seemingly minor exchanges shapes the microculture, often described as the culture shared by a small group based on location within an organization. Hospitalists have an opportunity to cultivate well-being within these microcultures through gracious and thoughtful acknowledgments of their peers. Collegial support at the micro level influences wellness at the organizational level. A larger shared culture of wellness is necessary to nurture physicians’ personal fulfillment and professional development.1

We propose the CARE framework for cultivating well-being within the microcultures of hospital medicine shared workspaces. CARE consists of Capitalization, Active listening, Recognition, and Empathy. This framework is based on positive psychology research and inspired by lessons from The Happiness Advantage by Shawn Achor.2

Capitalization. Capitalization is defined as sharing upbeat news and receiving a positive reaction. Emotional support during good times, more so than during bad times, strengthens relationships. When a peer shares good news, show enthusiasm and counter with an active, constructive response to maximize the validation she perceives.2

For example, Alex sits at her desk and says to Kristen:

My workshop proposal was accepted for medical education day!”

Congratulations, Alex! Tell me more about the workshop.”

Active listening. Active listening requires concentration and observation of body language. Show engagement by maintaining an open posture, using positive facial expressions, and providing occasional cues that you’re paying attention. Paraphrasing and asking targeted questions to dive deeper demonstrates genuine interest. 

“Katie, I could use your advice. Do you have a minute?”

Katie turns to face John and smiles. “Of course. How can I help?”

My team seems drained after a code this morning. I planned a lecture for later, but I’m not sure this is the right time.”

Katie nods. “I think you’re right, John. How have you thought about handling the situation?”

Recognition. Acts of recognition and encouragement are catalysts for boosting morale. Even brief expressions of gratitude can have a significant emotional impact. Recognition is most meaningful when delivered deliberately and with warmth.

Kevin walks into the hospitalist workroom. “Diane, congratulations on your publication! I plan to make a medication interaction review part of my discharge workflow.”

Leah turns to Diane. “Diane, that’s great news! Can you send me the link to your article?”

Empathy. Burnout is prevalent in medicine, and our fellow hospitalists deserve empathy. Showing empathy reduces stress and promotes connectedness. Sense when your colleagues are in distress and take time to share in their feelings and emotions. Draw on your own clinical experience to find common ground and convey understanding.

“I transferred another patient with COVID-19 to the ICU. I spent the last hour talking to family.”

“Ashwin, you’ve had a tough week. I know how you must feel—I had to transfer a patient yesterday. Want to take a quick walk outside?”

Hospitalists are inherently busy while on service, but these four interventions are brief, requiring only several minutes. Each small investment of your time will pay significant emotional dividends. These practices will not only enhance your colleagues’ sense of well-being, but will also bolster your happiness and productivity. A positive mindset fosters creative thinking and enhances complex problem solving. Recharging the microcultures of hospitalist workspaces with positivity will spark a larger transformation at the organizational level. That’s because positive actions are contagious.2 One hospitalist’s commitment to CARE will encourage other hospitalists to adopt these behaviors, establishing a virtuous cycle that sustains an organization’s culture of wellness.

“As we work to create light for others, we naturally light our own way.”

– Mary Anne Radmacher

Perhaps unknowingly, hospitalists establish microcultures in their everyday work. Hospitalists’ interactions with colleagues often occur in the context of shared workspaces. The nature of these seemingly minor exchanges shapes the microculture, often described as the culture shared by a small group based on location within an organization. Hospitalists have an opportunity to cultivate well-being within these microcultures through gracious and thoughtful acknowledgments of their peers. Collegial support at the micro level influences wellness at the organizational level. A larger shared culture of wellness is necessary to nurture physicians’ personal fulfillment and professional development.1

We propose the CARE framework for cultivating well-being within the microcultures of hospital medicine shared workspaces. CARE consists of Capitalization, Active listening, Recognition, and Empathy. This framework is based on positive psychology research and inspired by lessons from The Happiness Advantage by Shawn Achor.2

Capitalization. Capitalization is defined as sharing upbeat news and receiving a positive reaction. Emotional support during good times, more so than during bad times, strengthens relationships. When a peer shares good news, show enthusiasm and counter with an active, constructive response to maximize the validation she perceives.2

For example, Alex sits at her desk and says to Kristen:

My workshop proposal was accepted for medical education day!”

Congratulations, Alex! Tell me more about the workshop.”

Active listening. Active listening requires concentration and observation of body language. Show engagement by maintaining an open posture, using positive facial expressions, and providing occasional cues that you’re paying attention. Paraphrasing and asking targeted questions to dive deeper demonstrates genuine interest. 

“Katie, I could use your advice. Do you have a minute?”

Katie turns to face John and smiles. “Of course. How can I help?”

My team seems drained after a code this morning. I planned a lecture for later, but I’m not sure this is the right time.”

Katie nods. “I think you’re right, John. How have you thought about handling the situation?”

Recognition. Acts of recognition and encouragement are catalysts for boosting morale. Even brief expressions of gratitude can have a significant emotional impact. Recognition is most meaningful when delivered deliberately and with warmth.

Kevin walks into the hospitalist workroom. “Diane, congratulations on your publication! I plan to make a medication interaction review part of my discharge workflow.”

Leah turns to Diane. “Diane, that’s great news! Can you send me the link to your article?”

Empathy. Burnout is prevalent in medicine, and our fellow hospitalists deserve empathy. Showing empathy reduces stress and promotes connectedness. Sense when your colleagues are in distress and take time to share in their feelings and emotions. Draw on your own clinical experience to find common ground and convey understanding.

“I transferred another patient with COVID-19 to the ICU. I spent the last hour talking to family.”

“Ashwin, you’ve had a tough week. I know how you must feel—I had to transfer a patient yesterday. Want to take a quick walk outside?”

Hospitalists are inherently busy while on service, but these four interventions are brief, requiring only several minutes. Each small investment of your time will pay significant emotional dividends. These practices will not only enhance your colleagues’ sense of well-being, but will also bolster your happiness and productivity. A positive mindset fosters creative thinking and enhances complex problem solving. Recharging the microcultures of hospitalist workspaces with positivity will spark a larger transformation at the organizational level. That’s because positive actions are contagious.2 One hospitalist’s commitment to CARE will encourage other hospitalists to adopt these behaviors, establishing a virtuous cycle that sustains an organization’s culture of wellness.

References

1. Bohman B, Dyrbye L, Sinsky CA, et al. Physician well-being: the reciprocity of practice efficiency, culture of wellness, and personal resilience. NEJM Catalyst. August 7, 2017. Accessed June 24, 2021. https://catalyst.nejm.org/doi/full/10.1056/CAT.17.0429
2. Achor S. The Happiness Advantage: How a Positive Brain Fuels Success in Work and Life. Currency; 2010.

References

1. Bohman B, Dyrbye L, Sinsky CA, et al. Physician well-being: the reciprocity of practice efficiency, culture of wellness, and personal resilience. NEJM Catalyst. August 7, 2017. Accessed June 24, 2021. https://catalyst.nejm.org/doi/full/10.1056/CAT.17.0429
2. Achor S. The Happiness Advantage: How a Positive Brain Fuels Success in Work and Life. Currency; 2010.

Issue
Journal of Hospital Medicine 16(7)
Issue
Journal of Hospital Medicine 16(7)
Page Number
416
Page Number
416
Publications
Publications
Topics
Article Type
Display Headline
Leadership & Professional Development: Cultivating Microcultures of Well-being
Display Headline
Leadership & Professional Development: Cultivating Microcultures of Well-being
Sections
Article Source

© 2021 Society of Hospital Medicine

Disallow All Ads
Correspondence Location
Ryan E Nelson, MD; Email: [email protected]; Telephone: 617-754-4677.
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Page Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media

Algorithms for Prediction of Clinical Deterioration on the General Wards: A Scoping Review

Article Type
Changed
Fri, 10/01/2021 - 07:51
Display Headline
Algorithms for Prediction of Clinical Deterioration on the General Wards: A Scoping Review

The early identification of clinical deterioration among adult hospitalized patients remains a challenge.1 Delayed identification is associated with increased morbidity and mortality, unplanned intensive care unit (ICU) admissions, prolonged hospitalization, and higher costs.2,3 Earlier detection of deterioration using predictive algorithms of vital sign monitoring might avoid these negative outcomes.4 In this scoping review, we summarize current algorithms and their evidence.

Vital signs provide the backbone for detecting clinical deterioration. Early warning scores (EWS) and outreach protocols were developed to bring structure to the assessment of vital signs. Most EWS claim to predict clinical end points such as unplanned ICU admission up to 24 hours in advance.5,6 Reviews of EWS showed a positive trend toward reduced length of stay and mortality. However, conclusions about general efficacy could not be generated because of case heterogeneity and methodologic shortcomings.4,7 Continuous automated vital sign monitoring of patients on the general ward can now be accomplished with wearable devices.8 The first reports on continuous monitoring showed earlier detection of deterioration but not improved clinical end points.4,9 Since then, different reports on continuous monitoring have shown positive effects but concluded that unprocessed monitoring data per se falls short of generating actionable alarms.4,10,11

Predictive algorithms, which often use artificial intelligence (AI), are increasingly employed to recognize complex patterns or abnormalities and support predictions of events in big data sets.12,13 Especially when combined with continuous vital sign monitoring, predictive algorithms have the potential to expedite detection of clinical deterioration and improve patient outcomes. Predictive algorithms using vital signs in the ICU have shown promising results.14 The impact of predictive algorithms on the general wards, however, is unclear.

The aims of our scoping review were to explore the extent and range of and evidence for predictive vital signs–based algorithms on the adult general ward; to describe the variety of these algorithms; and to categorize effects, facilitators, and barriers of their implementation.15

 

 

MATERIALS AND METHODS

We performed a scoping review to create a summary of the current state of research. We used the five-step method of Levac and followed the Preferred Reporting Items for Systematic reviews and Meta-Analyses Extension for Scoping Reviews guidelines (Appendix 1).16,17

PubMed, Embase, and CINAHL databases were searched for English-language articles written between January 1, 2010, and November 20, 2020. We developed the search queries with an experienced information scientist, and we used database-specific terms and strategies for input, clinical outcome, method, predictive capability, and population (Appendix 2). Additionally, we searched the references of the selected articles, as well as publications citing these articles.

All studies identified were screened by title and abstract by two researchers (RP and YE). The selected studies were read in their entirety and checked for eligibility using the following inclusion criteria: automated algorithm; vital signs-based; real-time prediction; of clinical deterioration; in an adult, general ward population. In cases where there were successive publications with the same algorithm and population, we selected the most recent study.

For screening and selection, we used the Rayyan QCRI online tool (Qatar Computing Research Institute) and Endnote X9 (Clarivate Analytics). We extracted information using a data extraction form and organized it into descriptive characteristics of the selected studies (Table 1): an input data table showing number of admissions, intermittent or continuous measurements, vital signs measured, laboratory results (Appendix Table 1), a table summarizing study designs and settings (Appendix Table 2), and a prediction performance table (Table 2). We report characteristics of the populations and algorithms, prediction specifications such as area under the receiver operating curve (AUROC), and predictive values. Predictive values are affected by prevalence, which may differ among populations. To compare the algorithms, we calculated an indexed positive predictive value (PPV) and a number needed to evaluate (NNE) using a weighted average prevalence of clinical deterioration of 3.0%.



We defined clinical deterioration as end points, including rapid response team activation, cardiopulmonary resuscitation, transfer to an ICU, or death. Real-time was defined by the ability to automatically update predictions as new measurements are added. Predictions were defined as data-derived warnings for events in the near future. Prediction horizon was defined as the period for which a prediction is made. Special interest was given to algorithms that involved AI, which we defined as any form of machine learning or other nonclassical statistical algorithm.

Effects, facilitators, and barriers were identified and categorized using ATLAS.ti 8 software (ATLAS.ti) and evaluated by three researchers (RP, MK, and THvdB). These were categorized using the adapted frameworks of Gagnon et al18 for the barriers and facilitators and of Donabedian19 for the effects (Appendix 3).

The Gagnon et al framework was adapted by changing two of four domains—that is, “Individual” was changed to “Professional” and “Human” to “Physiology.” The domains of “Technology” and “Organization” remained unchanged. The Donabedian domains of “Outcome,” “Process,” and “Structure” also remained unchanged (Table 3).



We divided the studies into two groups: studies on predictive algorithms with and without AI when reporting on characteristics and performance. For the secondary aim of exploring implementation impact, we reported facilitators and barriers in a narrative way, highlighting the most frequent and notable findings.

 

 

RESULTS

As shown in the Figure, we found 1741 publications, of which we read the full-text of 109. There were 1632 publications that did not meet the inclusion criteria. The publications by Churpek et al,20,21 Bartkiowak et al,22 Edelson et al,23 Escobar et al,24,25 and Kipnis et al26 reported on the same algorithms or databases but had significantly different approaches. For multiple publications using the same algorithm and population, the most recent was named with inclusion of the earlier findings.20,21,27-29 The resulting 21 papers are included in this review.

Descriptive characteristics of the studies are summarized in Table 1. Nineteen of the publications were full papers and two were conference abstracts. Most of the studies (n = 18) were from the United States; there was one study from South Korea,30 one study from Portugal,31 and one study from the United Kingdom.32 In 15 of the studies, there was a strict focus on general or specific wards; 6 studies also included the ICU and/or emergency departments.

Two of the studies were clinical trials, 2 were prospective observational studies, and 17 were retrospective studies. Five studies reported on an active predictive model during admission. Of these, 3 reported that the model was clinically implemented, using the predictions in their clinical workflow. None of the implemented studies used AI.

All input variables are presented in Appendix Table 1. In 10 of the studies, vital signs were combined with laboratory results; in 13 of the studies, vital signs were combined with patient characteristics. All of the studies used data derived from electronic medical records (EMRs), except for Bai et al33 and Hu et al,34 who used single-source waveforms directly from the bedside monitor. Three studies focused on continuous vital sign measurements.27,33,34Most authors reported an AUROC to describe the predictive value of their algorithms. As shown in Table 2, AUROCs varied from 0.65 to 0.95, with indexed PPVs between 0.24 and 0.75. Sensitivity ranged from 7.2% to 52.5% in non-AI models and up to 82.4% in AI models. Prediction definitions, horizons, and the reported metrics differed too much to directly compare studies.

The non-AI algorithm prediction horizons ranged from 4 to 24 hours, with a median of 24 hours (interquartile range [IQR], 12-24 hours). The AI algorithms ranged from 2 to 48 hours and had a median horizon of 14 hours (IQR, 12-24 hours).

We found three studies reporting patient outcomes. The most recent of these was a large multicenter implementation study by Escobar et al25 that included an extensive follow-up response. This study reported a significantly decreased 30-day mortality in the intervention cohort. A smaller randomized controlled trial reported no significant differences in patient outcomes with earlier warning alarms.27 A third study reported more appropriate rapid response team deployment and decreased mortality in a subgroup analysis.35

Effects, Facilitators, and Barriers

As shown in the Appendix Figure and further detailed in Table 3, the described effects were predominantly positive—57 positive effects vs 11 negative effects. These positive effects sorted primarily into the outcome and process domains.

 

 

All of the studies that compared their proposed model with one of various warning systems (eg, EWS, National Early Warning Score [NEWS], Modified Early Warning Score [MEWS]) showed superior performance (based on AUROC and reported predictive values). In 17 studies, the authors reported their model as more useful or superior to the EWS.20-23,26-28,34,36-41 Four studies reported real-time detection of deterioration before regular EWS,20,26,42 and three studies reported positive effects on patient-related outcomes.26,35 Four negative effects were noted on the controllability, validity, and potential limitations.27,42

There were 26 positive effects on the clinical process mentioned, 7 of which pointed out the effects of earlier, predictive alarming. Algorithms with higher PPVs reported greater rates of actionable alarms, less alarm fatigue, and improved workflow.21,22,24-27,30,32,33,35-38,40 Potential alarm fatigue was named as a barrier.27,42 Smoother scoring instead of binned categories was mentioned positively.24,26In the infrastructure domain, very few items were reported. The increased need for education on the used techniques was reported once as a negative effect.34 One of the positive infrastructural effects noted was more efficient planning and use of resources.24,37,40We identified 57 facilitators and 48 barriers for the clinical implementation and use of real-time predictive analytics (Appendix Figure). In the Technology domain, there were 18 facilitators and 20 barriers cited, and in the Organization domain, 25 and 14, respectively. They were equally present in the Professional and Physiology domains (6 vs 5, 8 vs 9).

Of the 38 remarks in the Technology domain, difficulty with implementation in daily practice was a commonly cited barrier.22,24,40,42 Difficulties included creating real-time data feeds out of the EMR, though there were mentions of some successful examples.25,27,36 Difficulty in the interpretability of AI was also considered a potential barrier.30,32,33,35,39,41 There were remarks as to the applicability of the prolonged prediction horizon because of the associated decoupling from the clinical view.39,42

Conservative attitudes toward new technologies and inadequate knowledge were mentioned as barriers.39 Repeated remarks were made on the difficulty of interpreting and responding to a predicted escalation, as the clinical pattern might not be recognizable at such an early stage. On the other hand, it is expected that less invasive countermeasures would be adequate to avert further escalation. Earlier recognition of possible escalations also raised potential ethical questions, such as when to discuss palliative care.24

The heterogeneity of the general ward population and the relatively low prevalence of deterioration were mentioned as barriers.24,30,38,41 There were also concerns that not all escalations are preventable and that some patient outcomes may not be modifiable.24,38

Many investigators expected reductions in false alarms and associated alarm fatigue (reflected as higher PPVs). Furthermore, they expected workflow to improve and workload to decrease.21,23,27,31,33,35,38,41 Despite the capacity of modern EMRs to store large amounts of patient data, some investigators felt improvements to real-time access, data quality and validity, and data density are needed to ensure valid associated predictions.21,22,24,32,37

DISCUSSION

As the complexity and comorbidity of hospitalized adults grow, predicting clinical deterioration is becoming more important. With an ever-increasing amount of available patient data, real-time algorithms can predict the patient’s clinical course with increasing accuracy, positively affecting outcomes.4,21,25,43 The studies identified in this scoping review, as measured by higher AUROC scores and improved PPVs, show that predictive algorithms can outperform more conventional EWS, enable earlier and more efficient alarming, and be successfully implemented on the general wards. However, formal meta-analysis was made infeasible by differences in populations, use of different endpoint definitions, cut-off points, time-horizons to prediction, and other methodologic heterogeneity.

 

 

There are several important limitations across these studies. In a clinical setting, these models would function as a screening test. Almost all studies report an AUROC; however, sensitivity and PPV or NNE (defined as 1/PPV) may be more useful than AUROC when predicting low-frequency events with high-potential clinical impact.44 Assessing the NNE is especially relevant because of its relation to alarm fatigue and responsiveness of clinicians.43 Alarm fatigue and lack of adequate response to alarms were repeatedly cited as potential barriers for application of automated scores. A more useful metric might be NNE over a certain timeframe and across a specified number of patients to more clearly reflect associated workload. Future studies should include these metrics as indicators of the usability and clinical impact of predictive models. This review could not assess PPV or NNE systematically due to inconsistencies in the reporting of these metrics.

Although the results of our scoping review are promising, there are limited data on clinical outcomes using these algorithms. Only three of five algorithms were used to guide clinical decision-making.25,27,35 Kollef et al27 showed shorter hospitalizations and Evans et al35 found decreased mortality rates in a multimorbid subgroup. Escobar et al25 found an overall and consistent decrease in mortality in a large, heterogenic population of inpatients across 21 hospitals. While Escobar et al’s findings provide strong evidence that predictive algorithms and structured follow-up on alarms can improve patient outcomes, it recognizes that not all facilities will have the resources to implement them.25 Dedicated round-the-clock follow-up of alarms has yet to be proven feasible for smaller institutions, and leaner solutions must be explored. The example set by Escobar et al25 should be translated into various settings to prove its reproducibility and to substantiate the clinical impact of predictive models and structured follow-up.

According to expert opinion, the use of high-frequency or continuous monitoring at low-acuity wards and AI algorithms to detect trends and patterns will reduce failure-to-rescue rates.4,9,43 However, most studies in our review focused on periodic spot-checked vital signs, and none of the AI algorithms were implemented in clinical care (Appendix Table 1). A significant barrier to implementation was uncertainty surrounding how to react to generated alarms.9,45 As algorithms become more complex and predict earlier, interpretability and causality in general can diminish, and the response to this type of alarm will be different from that of an acute warning from an EWS.

The assessment of predictive algorithm protocols must include their impact on clinical workflow, workload, and resource utilization. Earlier detection of deterioration can potentially allow coordinated alarm follow-up and lead to more efficient use of resources.20,21,31,43,46,47

Greater numbers of variables do not always improve the quality of monitoring. For example, in one study, an algorithm combining only heart rate, respiration rate, and age outperformed an EWS that tracked six vital sign measures.23 Algorithms using fewer variables may facilitate more frequent and less complex error-sensitive monitoring. Leaner measurements may also lead to higher patient and clinician acceptance.43,45The end goal of implementing predictive algorithms on the general ward is to provide timely, reliable, and actionable clinical decision support.43 As shown in a recent study by Blackwell et al,48 multiple prediction models for specific clinical events may increase interpretability and performance. Disease-specific algorithms may complement general algorithms for clinical deterioration and enhance overall performance.

 

 

STRENGTHS AND LIMITATIONS

We performed a comprehensive review of the current literature using a clear and reproducible methodology to minimize the risk of missing relevant publications. The identified research is mainly limited to large US centers and consists of mostly retrospective studies. Heterogeneity among inputs, endpoints, time horizons, and evaluation metrics make comparisons challenging. Comments on facilitators, barriers, and effects were limited. Positive publication bias may have led to overrepresentation of models showing clinical benefit.

RECOMMENDATIONS FOR FUTURE RESEARCH

Artificial intelligence and the use of continuous monitoring hold great promise in creating optimal predictive algorithms. Future studies should directly compare AI- and non-AI-based algorithms using continuous monitoring to determine predictive accuracy, feasibility, costs, and outcomes. A consensus on endpoint definitions, input variables, methodology, and reporting is needed to enhance reproducibility, comparability, and generalizability of future research. The current research is limited to a few research groups, predominantly in the United States. More international research could enhance validity and increase applicability across varying populations and settings. Greater collaboration would accelerate research and enhance short-cycled continuous improvements. Sharing databases with different populations, variables, and outcomes, such as the Medical Information Mart for Intensive Care database,49 could help develop, test, and compare models and contribute to consensus in data standardization and consistent reporting of results. Studies should be designed to determine clinical, societal, and economic effects in accordance with the Quadruple Aim principle.50 Successful implementation will depend not only on improved patient outcomes but also on cost-effectiveness, robust statistics, and end-user acceptance. Follow-up protocols and workflows also should be studied and optimized.

CONCLUSION

Predictive analytics based on vital sign monitoring can identify clinical deterioration at an earlier stage and can do so more accurately than conventional EWS. Implementation of such monitoring can simultaneously decrease alarm-related workload and enhance the efficiency of follow-up. While there is also emerging evidence of associated mortality reduction, it may be too soon to know how representative these findings are. The current literature is limited by heterogeneity across populations studied, monitoring frequency, definitions of deterioration, and clinical outcomes. Consensus is therefore needed to better compare tools and harmonize results. Artificial intelligence and continuous monitoring show great promise in advancing the field; however, additional studies to assess cost, actionability of results, and end-user acceptance are required.

Files
References
  1. van Galen LS, Struik PW, Driesen BEJM, et al. Delayed recognition of deterioration of patients in general wards is mostly caused by human related monitoring failures: a root cause analysis of unplanned ICU admissions. PLoS One. 2016;11(8):e0161393. https://doi.org/10.1371/journal. pone.0161393
  2. Mardini L, Lipes J, Jayaraman D. Adverse outcomes associated with delayed intensive care consultation in medical and surgical inpatients. J Crit Care. 2012;27(6):688-693. https://doi.org/10.1016/j.jcrc.2012.04.011
  3. Young MP, Gooder VJ, McBride K, James B, Fisher ES. Inpatient transfers to the intensive care unit: delays are associated with increased mortality and morbidity. J Gen Intern Med. 2003;18(2):77-83. https://doi.org/10.1046/ j.1525-1497.2003.20441.x
  4. Khanna AK, Hoppe P, Saugel B. Automated continuous noninvasive ward monitoring: future directions and challenges. Crit Care. 2019;23(1):194. https://doi.org/10.1186/s13054-019-2485-7
  5. Ludikhuize J, Hamming A, de Jonge E, Fikkers BG. Rapid response systems in The Netherlands. Jt Comm J Qual Patient Saf. 2011;37(3):138-197. https:// doi.org/10.1016/s1553-7250(11)37017-1
  6. Cuthbertson BH, Boroujerdi M, McKie L, Aucott L, Prescott G. Can physiological variables and early warning scoring systems allow early recognition of the deteriorating surgical patient? Crit Care Med. 2007;35(2):402-409. https://doi.org/10.1097/01.ccm.0000254826.10520.87
  7. Alam N, Hobbelink EL, van Tienhoven AJ, van de Ven PM, Jansma EP, Nanayakkara PWB. The impact of the use of the Early Warning Score (EWS) on patient outcomes: a systematic review. Resuscitation. 2014;85(5):587-594. https://doi.org/10.1016/j.resuscitation.2014.01.013
  8. Weenk M, Koeneman M, van de Belt TH, Engelen LJLPG, van Goor H, Bredie SJH. Wireless and continuous monitoring of vital signs in patients at the general ward. Resuscitation. 2019;136:47-53. https://doi.org/10.1016/j.resuscitation.2019.01.017
  9. Cardona-Morrell M, Prgomet M, Turner RM, Nicholson M, Hillman K. Effectiveness of continuous or intermittent vital signs monitoring in preventing adverse events on general wards: a systematic review and meta-analysis. Int J Clin Pract. 2016;70(10):806-824. https://doi.org/10.1111/ijcp.12846
  10. Brown H, Terrence J, Vasquez P, Bates DW, Zimlichman E. Continuous monitoring in an inpatient medical-surgical unit: a controlled clinical trial. Am J Med. 2014;127(3):226-232. https://doi.org/10.1016/j.amjmed.2013.12.004
  11. Mestrom E, De Bie A, van de Steeg M, Driessen M, Atallah L, Bezemer R. Implementation of an automated early warning scoring system in a E8 Journal of Hospital Medicine® Published Online June 2021 An Official Publication of the Society of Hospital Medicine Peelen et al | Predicting Deterioration: A Scoping Review surgical ward: practical use and effects on patient outcomes. PLoS One. 2019;14(5):e0213402. https://doi.org/10.1371/journal.pone.0213402
  12. Jiang F, Jiang Y, Zhi H, et al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol. 2017;2(4):230-243. https://doi.org/10.1136/ svn-2017-000101
  13. Iwashyna TJ, Liu V. What’s so different about big data? A primer for clinicians trained to think epidemiologically. Ann Am Thorac Soc. 2014;11(7):1130- 1135. https://doi.org/10.1513/annalsats.201405-185as
  14. Jalali A, Bender D, Rehman M, Nadkanri V, Nataraj C. Advanced analytics for outcome prediction in intensive care units. Conf Proc IEEE Eng Med Biol Soc. 2016;2016:2520-2524. https://doi.org/10.1109/embc.2016.7591243
  15. Munn Z, Peters MDJ, Stern C, Tufanaru C, McArthur A, Aromataris E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. 2018;18(1):143. https://doi.org/10.1186/s12874-018-0611-x
  16. Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19-32. https://doi.org/10.1080/13645 57032000119616
  17. Tricco AC, Lillie E, Zarin W, et al. PRISMA extension for scoping reviews (PRISMAScR): checklist and explanation. Ann Intern Med. 2018;169(7):467- 473. https://doi.org/10.7326/m18-0850
  18. Gagnon MP, Desmartis M, Gagnon J, et al. Framework for user involvement in health technology assessment at the local level: views of health managers, user representatives, and clinicians. Int J Technol Assess Health Care. 2015;31(1-2):68-77. https://doi.org/10.1017/s0266462315000070
  19. Donabedian A. The quality of care. How can it be assessed? JAMA. 1988;260(12):1743-1748. https://doi.org/10.1001/jama.260.12.1743
  20. Churpek MM, Yuen TC, Winslow C, et al. Multicenter development and validation of a risk stratification tool for ward patients. Am J Respir Crit Care Med. 2014;190(6):649-655. https://doi.org/10.1164/rccm.201406-1022oc
  21. Churpek MM, Yuen TC, Winslow C, Meltzer DO, Kattan MW, Edelson DP. Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards. Crit Care Med. 2016;44(2):368-374. https://doi.org/10.1097/ccm.0000000000001571
  22. Bartkowiak B, Snyder AM, Benjamin A, et al. Validating the electronic cardiac arrest risk triage (eCART) score for risk stratification of surgical inpatients in the postoperative setting: retrospective cohort study. Ann Surg. 2019;269(6):1059-1063. https://doi.org/10.1097/sla.0000000000002665
  23. Edelson DP, Carey K, Winslow CJ, Churpek MM. Less is more: detecting clinical deterioration in the hospital with machine learning using only age, heart rate and respiratory rate. Abstract presented at: American Thoracic Society International Conference; May 22, 2018; San Diego, California. Am J Resp Crit Care Med. 2018;197:A4444.
  24. Escobar GJ, LaGuardia JC, Turk BJ, Ragins A, Kipnis P, Draper D. Early detection of impending physiologic deterioration among patients who are not in intensive care: development of predictive models using data from an automated electronic medical record. J Hosp Med. 2012;7(5):388-395. https:// doi.org/10.1002/jhm.1929
  25. Escobar GJ, Liu VX, Schuler A, Lawson B, Greene JD, Kipnis P. Automated identification of adults at risk for in-hospital clinical deterioration. N Engl J Med. 2020;383(20):1951-1960. https://doi.org/10.1056/nejmsa2001090
  26. Kipnis P, Turk BJ, Wulf DA, et al. Development and validation of an electronic medical record-based alert score for detection of inpatient deterioration outside the ICU. J Biomed Inform. 2016;64:10-19. https://doi.org/10.1016/j. jbi.2016.09.013
  27. Kollef MH, Chen Y, Heard K, et al. A randomized trial of real-time automated clinical deterioration alerts sent to a rapid response team. J Hosp Med. 2014;9(7):424-429. https://doi.org/10.1002/jhm.2193
  28. Hackmann G, Chen M, Chipara O, et al. Toward a two-tier clinical warning system for hospitalized patients. AMIA Annu Symp Proc. 2011;2011:511-519.
  29. Bailey TC, Chen Y, Mao Y, Lu, C, Hackmann G, Micek ST. A trial of a real-time alert for clinical deterioration in patients hospitalized on general medical wards. J Hosp Med. 2013;8(5):236-242. https://doi.org/10.1002/jhm.2009
  30. Kwon JM, Lee Y, Lee Y, Lee S, Park J. An algorithm based on deep learning for predicting in-hospital cardiac arrest. J Am Heart Assoc. 2018;7(13):e008678. https://doi.org/10.1161/jaha.118.008678
  31. Correia S, Gomes A, Shahriari S, Almeida JP, Severo M, Azevedo A. Performance of the early warning system vital to predict unanticipated higher-level of care admission and in-hospital death of ward patients. Value Health. 2018;21(S3):S360. https://doi.org/10.1016/j.jval.2018.09.2152
  32. Shamout FE, Zhu T, Sharma P, Watkinson PJ, Clifton DA. Deep interpretable early warning system for the detection of clinical deterioration. IEEE J Biomed Health Inform. 2020;24(2):437-446. https://doi.org/10.1109/ jbhi.2019.2937803
  33. Bai Y, Do DH, Harris PRE, et al. Integrating monitor alarms with laboratory test results to enhance patient deterioration prediction. J Biomed Inform. 2015;53:81-92. https://doi.org/10.1016/j.jbi.2014.09.006
  34. Hu X, Sapo M, Nenov V, et al. Predictive combinations of monitor alarms preceding in-hospital code blue events. J Biomed Inform. 2012;45(5):913-921. https://doi.org/10.1016/j.jbi.2012.03.001
  35. Evans RS, Kuttler KG, Simpson KJ, et al. Automated detection of physiologic deterioration in hospitalized patients. J Am Med Inform Assoc. 2015;22(2):350-360. https://doi.org/10.1136/amiajnl-2014-002816
  36. Ghosh E, Eshelman L, Yang L, Carlson E, Lord B. Early deterioration indicator: data-driven approach to detecting deterioration in general ward. Resuscitation. 2018;122:99-105. https://doi.org/10.1016/j.resuscitation. 2017.10.026
  37. Kang MA, Churpek MM, Zadravecz FJ, Adhikari R, Twu NM, Edelson DP: Real-time risk prediction on the wards: a feasibility study. Crit Care Med. 2016;44(8):1468-1473. https://doi.org/10.1097/ccm.0000000000001716
  38. Hu SB, Wong DJL, Correa A, Li N, Deng JC. Prediction of clinical deterioration in hospitalized adult patients with hematologic malignancies using a neural network model. PLoS One. 2016;11(8):e0161401. https://doi. org/10.1371/journal.pone.0161401
  39. Rothman MJ, Rothman SI, Beals J 4th. Development and validation of a continuous measure of patient condition using the electronic medical record. J Biomed Inform. 2013;46(5):837-848. https://doi.org/10.1016/j. jbi.2013.06.011
  40. Alaa AM, Yoon J, Hu S, van der Schaar M. Personalized risk scoring for critical care prognosis using mixtures of Gaussian processes. IEEE Trans Biomed Eng. 2018;65(1):207-218. https://doi.org/10.1109/tbme.2017.2698602
  41. Mohamadlou H, Panchavati S, Calvert J, et al. Multicenter validation of a machine-learning algorithm for 48-h all-cause mortality prediction. Health Informatics J. 2020;26(3):1912-1925. https://doi.org/10.1177/1460458219894494
  42. Alvarez CA, Clark CA, Zhang S, et al. Predicting out of intensive care unit cardiopulmonary arrest or death using electronic medical record data. BMC Med Inform Decis Mak. 2013;13:28. https://doi.org/10.1186/1472-6947-13-28
  43. Vincent JL, Einav S, Pearse R, et al. Improving detection of patient deterioration in the general hospital ward environment. Eur J Anaesthesiol. 2018;35(5):325-333. https://doi.org/10.1097/eja.0000000000000798
  44. Romero-Brufau S, Huddleston JM, Escobar GJ, Liebow M. Why the C-statistic is not informative to evaluate early warning scores and what metrics to use. Crit Care. 2015;19(1):285. https://doi.org/10.1186/s13054-015-0999-1
  45. Weenk M, Bredie SJ, Koeneman M, Hesselink G, van Goor H, van de Belt TH. Continuous monitoring of the vital signs in the general ward using wearable devices: randomized controlled trial. J Med Internet Res. 2020;22(6):e15471. https://doi.org/10.2196/15471
  46. Wellner B, Grand J, Canzone E, et al. Predicting unplanned transfers to the intensive care unit: a machine learning approach leveraging diverse clinical elements. JMIR Med Inform. 2017;5(4):e45. https://doi.org/10.2196/medinform.8680
  47. Elliott M, Baird J. Pulse oximetry and the enduring neglect of respiratory rate assessment: a commentary on patient surveillance. Br J Nurs. 2019;28(19):1256-1259. https://doi.org/10.12968/bjon.2019.28.19.1256
  48. Blackwell JN, Keim-Malpass J, Clark MT, et al. Early detection of in-patient deterioration: one prediction model does not fit all. Crit Care Explor. 2020;2(5):e0116. https://doi.org/10.1097/cce.0000000000000116
  49. Johnson AEW, Pollard TJ, Shen L, et al. MIMIC-III, a freely accessible critical care database. Sci Data. 2016;3:160035. https://doi.org/10.1038/sdata.2016.35
  50. Bodenheimer T, Sinsky C. From triple to quadruple aim: care of the patient requires care of the provider. Ann Fam Med. 2014;12(6):573-576. https://doi. org/10.1370/afm.1713
  51.  Kirkland LL, Malinchoc M, O’Byrne M, et al. A clinical deterioration prediction tool for internal medicine patients. Am J Med Qual. 2013;28(2):135-142 https://doi.org/10.1177/1062860612450459
Article PDF
Author and Disclosure Information

1Radboud University Medical Center, Department of Internal Medicine, Nijmegen, The Netherlands; 2Radboud University Medical Center, Department of Surgery, Nijmegen, The Netherlands; 3Radboud University Medical Center, REshape and Innovation Center, Nijmegen, The Netherlands.

Disclosures

The authors have no conflicts of interest to disclose.

Issue
Journal of Hospital Medicine 16(10)
Publications
Topics
Page Number
612-619. Published Online First June 25, 2021
Sections
Files
Files
Author and Disclosure Information

1Radboud University Medical Center, Department of Internal Medicine, Nijmegen, The Netherlands; 2Radboud University Medical Center, Department of Surgery, Nijmegen, The Netherlands; 3Radboud University Medical Center, REshape and Innovation Center, Nijmegen, The Netherlands.

Disclosures

The authors have no conflicts of interest to disclose.

Author and Disclosure Information

1Radboud University Medical Center, Department of Internal Medicine, Nijmegen, The Netherlands; 2Radboud University Medical Center, Department of Surgery, Nijmegen, The Netherlands; 3Radboud University Medical Center, REshape and Innovation Center, Nijmegen, The Netherlands.

Disclosures

The authors have no conflicts of interest to disclose.

Article PDF
Article PDF
Related Articles

The early identification of clinical deterioration among adult hospitalized patients remains a challenge.1 Delayed identification is associated with increased morbidity and mortality, unplanned intensive care unit (ICU) admissions, prolonged hospitalization, and higher costs.2,3 Earlier detection of deterioration using predictive algorithms of vital sign monitoring might avoid these negative outcomes.4 In this scoping review, we summarize current algorithms and their evidence.

Vital signs provide the backbone for detecting clinical deterioration. Early warning scores (EWS) and outreach protocols were developed to bring structure to the assessment of vital signs. Most EWS claim to predict clinical end points such as unplanned ICU admission up to 24 hours in advance.5,6 Reviews of EWS showed a positive trend toward reduced length of stay and mortality. However, conclusions about general efficacy could not be generated because of case heterogeneity and methodologic shortcomings.4,7 Continuous automated vital sign monitoring of patients on the general ward can now be accomplished with wearable devices.8 The first reports on continuous monitoring showed earlier detection of deterioration but not improved clinical end points.4,9 Since then, different reports on continuous monitoring have shown positive effects but concluded that unprocessed monitoring data per se falls short of generating actionable alarms.4,10,11

Predictive algorithms, which often use artificial intelligence (AI), are increasingly employed to recognize complex patterns or abnormalities and support predictions of events in big data sets.12,13 Especially when combined with continuous vital sign monitoring, predictive algorithms have the potential to expedite detection of clinical deterioration and improve patient outcomes. Predictive algorithms using vital signs in the ICU have shown promising results.14 The impact of predictive algorithms on the general wards, however, is unclear.

The aims of our scoping review were to explore the extent and range of and evidence for predictive vital signs–based algorithms on the adult general ward; to describe the variety of these algorithms; and to categorize effects, facilitators, and barriers of their implementation.15

 

 

MATERIALS AND METHODS

We performed a scoping review to create a summary of the current state of research. We used the five-step method of Levac and followed the Preferred Reporting Items for Systematic reviews and Meta-Analyses Extension for Scoping Reviews guidelines (Appendix 1).16,17

PubMed, Embase, and CINAHL databases were searched for English-language articles written between January 1, 2010, and November 20, 2020. We developed the search queries with an experienced information scientist, and we used database-specific terms and strategies for input, clinical outcome, method, predictive capability, and population (Appendix 2). Additionally, we searched the references of the selected articles, as well as publications citing these articles.

All studies identified were screened by title and abstract by two researchers (RP and YE). The selected studies were read in their entirety and checked for eligibility using the following inclusion criteria: automated algorithm; vital signs-based; real-time prediction; of clinical deterioration; in an adult, general ward population. In cases where there were successive publications with the same algorithm and population, we selected the most recent study.

For screening and selection, we used the Rayyan QCRI online tool (Qatar Computing Research Institute) and Endnote X9 (Clarivate Analytics). We extracted information using a data extraction form and organized it into descriptive characteristics of the selected studies (Table 1): an input data table showing number of admissions, intermittent or continuous measurements, vital signs measured, laboratory results (Appendix Table 1), a table summarizing study designs and settings (Appendix Table 2), and a prediction performance table (Table 2). We report characteristics of the populations and algorithms, prediction specifications such as area under the receiver operating curve (AUROC), and predictive values. Predictive values are affected by prevalence, which may differ among populations. To compare the algorithms, we calculated an indexed positive predictive value (PPV) and a number needed to evaluate (NNE) using a weighted average prevalence of clinical deterioration of 3.0%.



We defined clinical deterioration as end points, including rapid response team activation, cardiopulmonary resuscitation, transfer to an ICU, or death. Real-time was defined by the ability to automatically update predictions as new measurements are added. Predictions were defined as data-derived warnings for events in the near future. Prediction horizon was defined as the period for which a prediction is made. Special interest was given to algorithms that involved AI, which we defined as any form of machine learning or other nonclassical statistical algorithm.

Effects, facilitators, and barriers were identified and categorized using ATLAS.ti 8 software (ATLAS.ti) and evaluated by three researchers (RP, MK, and THvdB). These were categorized using the adapted frameworks of Gagnon et al18 for the barriers and facilitators and of Donabedian19 for the effects (Appendix 3).

The Gagnon et al framework was adapted by changing two of four domains—that is, “Individual” was changed to “Professional” and “Human” to “Physiology.” The domains of “Technology” and “Organization” remained unchanged. The Donabedian domains of “Outcome,” “Process,” and “Structure” also remained unchanged (Table 3).



We divided the studies into two groups: studies on predictive algorithms with and without AI when reporting on characteristics and performance. For the secondary aim of exploring implementation impact, we reported facilitators and barriers in a narrative way, highlighting the most frequent and notable findings.

 

 

RESULTS

As shown in the Figure, we found 1741 publications, of which we read the full-text of 109. There were 1632 publications that did not meet the inclusion criteria. The publications by Churpek et al,20,21 Bartkiowak et al,22 Edelson et al,23 Escobar et al,24,25 and Kipnis et al26 reported on the same algorithms or databases but had significantly different approaches. For multiple publications using the same algorithm and population, the most recent was named with inclusion of the earlier findings.20,21,27-29 The resulting 21 papers are included in this review.

Descriptive characteristics of the studies are summarized in Table 1. Nineteen of the publications were full papers and two were conference abstracts. Most of the studies (n = 18) were from the United States; there was one study from South Korea,30 one study from Portugal,31 and one study from the United Kingdom.32 In 15 of the studies, there was a strict focus on general or specific wards; 6 studies also included the ICU and/or emergency departments.

Two of the studies were clinical trials, 2 were prospective observational studies, and 17 were retrospective studies. Five studies reported on an active predictive model during admission. Of these, 3 reported that the model was clinically implemented, using the predictions in their clinical workflow. None of the implemented studies used AI.

All input variables are presented in Appendix Table 1. In 10 of the studies, vital signs were combined with laboratory results; in 13 of the studies, vital signs were combined with patient characteristics. All of the studies used data derived from electronic medical records (EMRs), except for Bai et al33 and Hu et al,34 who used single-source waveforms directly from the bedside monitor. Three studies focused on continuous vital sign measurements.27,33,34Most authors reported an AUROC to describe the predictive value of their algorithms. As shown in Table 2, AUROCs varied from 0.65 to 0.95, with indexed PPVs between 0.24 and 0.75. Sensitivity ranged from 7.2% to 52.5% in non-AI models and up to 82.4% in AI models. Prediction definitions, horizons, and the reported metrics differed too much to directly compare studies.

The non-AI algorithm prediction horizons ranged from 4 to 24 hours, with a median of 24 hours (interquartile range [IQR], 12-24 hours). The AI algorithms ranged from 2 to 48 hours and had a median horizon of 14 hours (IQR, 12-24 hours).

We found three studies reporting patient outcomes. The most recent of these was a large multicenter implementation study by Escobar et al25 that included an extensive follow-up response. This study reported a significantly decreased 30-day mortality in the intervention cohort. A smaller randomized controlled trial reported no significant differences in patient outcomes with earlier warning alarms.27 A third study reported more appropriate rapid response team deployment and decreased mortality in a subgroup analysis.35

Effects, Facilitators, and Barriers

As shown in the Appendix Figure and further detailed in Table 3, the described effects were predominantly positive—57 positive effects vs 11 negative effects. These positive effects sorted primarily into the outcome and process domains.

 

 

All of the studies that compared their proposed model with one of various warning systems (eg, EWS, National Early Warning Score [NEWS], Modified Early Warning Score [MEWS]) showed superior performance (based on AUROC and reported predictive values). In 17 studies, the authors reported their model as more useful or superior to the EWS.20-23,26-28,34,36-41 Four studies reported real-time detection of deterioration before regular EWS,20,26,42 and three studies reported positive effects on patient-related outcomes.26,35 Four negative effects were noted on the controllability, validity, and potential limitations.27,42

There were 26 positive effects on the clinical process mentioned, 7 of which pointed out the effects of earlier, predictive alarming. Algorithms with higher PPVs reported greater rates of actionable alarms, less alarm fatigue, and improved workflow.21,22,24-27,30,32,33,35-38,40 Potential alarm fatigue was named as a barrier.27,42 Smoother scoring instead of binned categories was mentioned positively.24,26In the infrastructure domain, very few items were reported. The increased need for education on the used techniques was reported once as a negative effect.34 One of the positive infrastructural effects noted was more efficient planning and use of resources.24,37,40We identified 57 facilitators and 48 barriers for the clinical implementation and use of real-time predictive analytics (Appendix Figure). In the Technology domain, there were 18 facilitators and 20 barriers cited, and in the Organization domain, 25 and 14, respectively. They were equally present in the Professional and Physiology domains (6 vs 5, 8 vs 9).

Of the 38 remarks in the Technology domain, difficulty with implementation in daily practice was a commonly cited barrier.22,24,40,42 Difficulties included creating real-time data feeds out of the EMR, though there were mentions of some successful examples.25,27,36 Difficulty in the interpretability of AI was also considered a potential barrier.30,32,33,35,39,41 There were remarks as to the applicability of the prolonged prediction horizon because of the associated decoupling from the clinical view.39,42

Conservative attitudes toward new technologies and inadequate knowledge were mentioned as barriers.39 Repeated remarks were made on the difficulty of interpreting and responding to a predicted escalation, as the clinical pattern might not be recognizable at such an early stage. On the other hand, it is expected that less invasive countermeasures would be adequate to avert further escalation. Earlier recognition of possible escalations also raised potential ethical questions, such as when to discuss palliative care.24

The heterogeneity of the general ward population and the relatively low prevalence of deterioration were mentioned as barriers.24,30,38,41 There were also concerns that not all escalations are preventable and that some patient outcomes may not be modifiable.24,38

Many investigators expected reductions in false alarms and associated alarm fatigue (reflected as higher PPVs). Furthermore, they expected workflow to improve and workload to decrease.21,23,27,31,33,35,38,41 Despite the capacity of modern EMRs to store large amounts of patient data, some investigators felt improvements to real-time access, data quality and validity, and data density are needed to ensure valid associated predictions.21,22,24,32,37

DISCUSSION

As the complexity and comorbidity of hospitalized adults grow, predicting clinical deterioration is becoming more important. With an ever-increasing amount of available patient data, real-time algorithms can predict the patient’s clinical course with increasing accuracy, positively affecting outcomes.4,21,25,43 The studies identified in this scoping review, as measured by higher AUROC scores and improved PPVs, show that predictive algorithms can outperform more conventional EWS, enable earlier and more efficient alarming, and be successfully implemented on the general wards. However, formal meta-analysis was made infeasible by differences in populations, use of different endpoint definitions, cut-off points, time-horizons to prediction, and other methodologic heterogeneity.

 

 

There are several important limitations across these studies. In a clinical setting, these models would function as a screening test. Almost all studies report an AUROC; however, sensitivity and PPV or NNE (defined as 1/PPV) may be more useful than AUROC when predicting low-frequency events with high-potential clinical impact.44 Assessing the NNE is especially relevant because of its relation to alarm fatigue and responsiveness of clinicians.43 Alarm fatigue and lack of adequate response to alarms were repeatedly cited as potential barriers for application of automated scores. A more useful metric might be NNE over a certain timeframe and across a specified number of patients to more clearly reflect associated workload. Future studies should include these metrics as indicators of the usability and clinical impact of predictive models. This review could not assess PPV or NNE systematically due to inconsistencies in the reporting of these metrics.

Although the results of our scoping review are promising, there are limited data on clinical outcomes using these algorithms. Only three of five algorithms were used to guide clinical decision-making.25,27,35 Kollef et al27 showed shorter hospitalizations and Evans et al35 found decreased mortality rates in a multimorbid subgroup. Escobar et al25 found an overall and consistent decrease in mortality in a large, heterogenic population of inpatients across 21 hospitals. While Escobar et al’s findings provide strong evidence that predictive algorithms and structured follow-up on alarms can improve patient outcomes, it recognizes that not all facilities will have the resources to implement them.25 Dedicated round-the-clock follow-up of alarms has yet to be proven feasible for smaller institutions, and leaner solutions must be explored. The example set by Escobar et al25 should be translated into various settings to prove its reproducibility and to substantiate the clinical impact of predictive models and structured follow-up.

According to expert opinion, the use of high-frequency or continuous monitoring at low-acuity wards and AI algorithms to detect trends and patterns will reduce failure-to-rescue rates.4,9,43 However, most studies in our review focused on periodic spot-checked vital signs, and none of the AI algorithms were implemented in clinical care (Appendix Table 1). A significant barrier to implementation was uncertainty surrounding how to react to generated alarms.9,45 As algorithms become more complex and predict earlier, interpretability and causality in general can diminish, and the response to this type of alarm will be different from that of an acute warning from an EWS.

The assessment of predictive algorithm protocols must include their impact on clinical workflow, workload, and resource utilization. Earlier detection of deterioration can potentially allow coordinated alarm follow-up and lead to more efficient use of resources.20,21,31,43,46,47

Greater numbers of variables do not always improve the quality of monitoring. For example, in one study, an algorithm combining only heart rate, respiration rate, and age outperformed an EWS that tracked six vital sign measures.23 Algorithms using fewer variables may facilitate more frequent and less complex error-sensitive monitoring. Leaner measurements may also lead to higher patient and clinician acceptance.43,45The end goal of implementing predictive algorithms on the general ward is to provide timely, reliable, and actionable clinical decision support.43 As shown in a recent study by Blackwell et al,48 multiple prediction models for specific clinical events may increase interpretability and performance. Disease-specific algorithms may complement general algorithms for clinical deterioration and enhance overall performance.

 

 

STRENGTHS AND LIMITATIONS

We performed a comprehensive review of the current literature using a clear and reproducible methodology to minimize the risk of missing relevant publications. The identified research is mainly limited to large US centers and consists of mostly retrospective studies. Heterogeneity among inputs, endpoints, time horizons, and evaluation metrics make comparisons challenging. Comments on facilitators, barriers, and effects were limited. Positive publication bias may have led to overrepresentation of models showing clinical benefit.

RECOMMENDATIONS FOR FUTURE RESEARCH

Artificial intelligence and the use of continuous monitoring hold great promise in creating optimal predictive algorithms. Future studies should directly compare AI- and non-AI-based algorithms using continuous monitoring to determine predictive accuracy, feasibility, costs, and outcomes. A consensus on endpoint definitions, input variables, methodology, and reporting is needed to enhance reproducibility, comparability, and generalizability of future research. The current research is limited to a few research groups, predominantly in the United States. More international research could enhance validity and increase applicability across varying populations and settings. Greater collaboration would accelerate research and enhance short-cycled continuous improvements. Sharing databases with different populations, variables, and outcomes, such as the Medical Information Mart for Intensive Care database,49 could help develop, test, and compare models and contribute to consensus in data standardization and consistent reporting of results. Studies should be designed to determine clinical, societal, and economic effects in accordance with the Quadruple Aim principle.50 Successful implementation will depend not only on improved patient outcomes but also on cost-effectiveness, robust statistics, and end-user acceptance. Follow-up protocols and workflows also should be studied and optimized.

CONCLUSION

Predictive analytics based on vital sign monitoring can identify clinical deterioration at an earlier stage and can do so more accurately than conventional EWS. Implementation of such monitoring can simultaneously decrease alarm-related workload and enhance the efficiency of follow-up. While there is also emerging evidence of associated mortality reduction, it may be too soon to know how representative these findings are. The current literature is limited by heterogeneity across populations studied, monitoring frequency, definitions of deterioration, and clinical outcomes. Consensus is therefore needed to better compare tools and harmonize results. Artificial intelligence and continuous monitoring show great promise in advancing the field; however, additional studies to assess cost, actionability of results, and end-user acceptance are required.

The early identification of clinical deterioration among adult hospitalized patients remains a challenge.1 Delayed identification is associated with increased morbidity and mortality, unplanned intensive care unit (ICU) admissions, prolonged hospitalization, and higher costs.2,3 Earlier detection of deterioration using predictive algorithms of vital sign monitoring might avoid these negative outcomes.4 In this scoping review, we summarize current algorithms and their evidence.

Vital signs provide the backbone for detecting clinical deterioration. Early warning scores (EWS) and outreach protocols were developed to bring structure to the assessment of vital signs. Most EWS claim to predict clinical end points such as unplanned ICU admission up to 24 hours in advance.5,6 Reviews of EWS showed a positive trend toward reduced length of stay and mortality. However, conclusions about general efficacy could not be generated because of case heterogeneity and methodologic shortcomings.4,7 Continuous automated vital sign monitoring of patients on the general ward can now be accomplished with wearable devices.8 The first reports on continuous monitoring showed earlier detection of deterioration but not improved clinical end points.4,9 Since then, different reports on continuous monitoring have shown positive effects but concluded that unprocessed monitoring data per se falls short of generating actionable alarms.4,10,11

Predictive algorithms, which often use artificial intelligence (AI), are increasingly employed to recognize complex patterns or abnormalities and support predictions of events in big data sets.12,13 Especially when combined with continuous vital sign monitoring, predictive algorithms have the potential to expedite detection of clinical deterioration and improve patient outcomes. Predictive algorithms using vital signs in the ICU have shown promising results.14 The impact of predictive algorithms on the general wards, however, is unclear.

The aims of our scoping review were to explore the extent and range of and evidence for predictive vital signs–based algorithms on the adult general ward; to describe the variety of these algorithms; and to categorize effects, facilitators, and barriers of their implementation.15

 

 

MATERIALS AND METHODS

We performed a scoping review to create a summary of the current state of research. We used the five-step method of Levac and followed the Preferred Reporting Items for Systematic reviews and Meta-Analyses Extension for Scoping Reviews guidelines (Appendix 1).16,17

PubMed, Embase, and CINAHL databases were searched for English-language articles written between January 1, 2010, and November 20, 2020. We developed the search queries with an experienced information scientist, and we used database-specific terms and strategies for input, clinical outcome, method, predictive capability, and population (Appendix 2). Additionally, we searched the references of the selected articles, as well as publications citing these articles.

All studies identified were screened by title and abstract by two researchers (RP and YE). The selected studies were read in their entirety and checked for eligibility using the following inclusion criteria: automated algorithm; vital signs-based; real-time prediction; of clinical deterioration; in an adult, general ward population. In cases where there were successive publications with the same algorithm and population, we selected the most recent study.

For screening and selection, we used the Rayyan QCRI online tool (Qatar Computing Research Institute) and Endnote X9 (Clarivate Analytics). We extracted information using a data extraction form and organized it into descriptive characteristics of the selected studies (Table 1): an input data table showing number of admissions, intermittent or continuous measurements, vital signs measured, laboratory results (Appendix Table 1), a table summarizing study designs and settings (Appendix Table 2), and a prediction performance table (Table 2). We report characteristics of the populations and algorithms, prediction specifications such as area under the receiver operating curve (AUROC), and predictive values. Predictive values are affected by prevalence, which may differ among populations. To compare the algorithms, we calculated an indexed positive predictive value (PPV) and a number needed to evaluate (NNE) using a weighted average prevalence of clinical deterioration of 3.0%.



We defined clinical deterioration as end points, including rapid response team activation, cardiopulmonary resuscitation, transfer to an ICU, or death. Real-time was defined by the ability to automatically update predictions as new measurements are added. Predictions were defined as data-derived warnings for events in the near future. Prediction horizon was defined as the period for which a prediction is made. Special interest was given to algorithms that involved AI, which we defined as any form of machine learning or other nonclassical statistical algorithm.

Effects, facilitators, and barriers were identified and categorized using ATLAS.ti 8 software (ATLAS.ti) and evaluated by three researchers (RP, MK, and THvdB). These were categorized using the adapted frameworks of Gagnon et al18 for the barriers and facilitators and of Donabedian19 for the effects (Appendix 3).

The Gagnon et al framework was adapted by changing two of four domains—that is, “Individual” was changed to “Professional” and “Human” to “Physiology.” The domains of “Technology” and “Organization” remained unchanged. The Donabedian domains of “Outcome,” “Process,” and “Structure” also remained unchanged (Table 3).



We divided the studies into two groups: studies on predictive algorithms with and without AI when reporting on characteristics and performance. For the secondary aim of exploring implementation impact, we reported facilitators and barriers in a narrative way, highlighting the most frequent and notable findings.

 

 

RESULTS

As shown in the Figure, we found 1741 publications, of which we read the full-text of 109. There were 1632 publications that did not meet the inclusion criteria. The publications by Churpek et al,20,21 Bartkiowak et al,22 Edelson et al,23 Escobar et al,24,25 and Kipnis et al26 reported on the same algorithms or databases but had significantly different approaches. For multiple publications using the same algorithm and population, the most recent was named with inclusion of the earlier findings.20,21,27-29 The resulting 21 papers are included in this review.

Descriptive characteristics of the studies are summarized in Table 1. Nineteen of the publications were full papers and two were conference abstracts. Most of the studies (n = 18) were from the United States; there was one study from South Korea,30 one study from Portugal,31 and one study from the United Kingdom.32 In 15 of the studies, there was a strict focus on general or specific wards; 6 studies also included the ICU and/or emergency departments.

Two of the studies were clinical trials, 2 were prospective observational studies, and 17 were retrospective studies. Five studies reported on an active predictive model during admission. Of these, 3 reported that the model was clinically implemented, using the predictions in their clinical workflow. None of the implemented studies used AI.

All input variables are presented in Appendix Table 1. In 10 of the studies, vital signs were combined with laboratory results; in 13 of the studies, vital signs were combined with patient characteristics. All of the studies used data derived from electronic medical records (EMRs), except for Bai et al33 and Hu et al,34 who used single-source waveforms directly from the bedside monitor. Three studies focused on continuous vital sign measurements.27,33,34Most authors reported an AUROC to describe the predictive value of their algorithms. As shown in Table 2, AUROCs varied from 0.65 to 0.95, with indexed PPVs between 0.24 and 0.75. Sensitivity ranged from 7.2% to 52.5% in non-AI models and up to 82.4% in AI models. Prediction definitions, horizons, and the reported metrics differed too much to directly compare studies.

The non-AI algorithm prediction horizons ranged from 4 to 24 hours, with a median of 24 hours (interquartile range [IQR], 12-24 hours). The AI algorithms ranged from 2 to 48 hours and had a median horizon of 14 hours (IQR, 12-24 hours).

We found three studies reporting patient outcomes. The most recent of these was a large multicenter implementation study by Escobar et al25 that included an extensive follow-up response. This study reported a significantly decreased 30-day mortality in the intervention cohort. A smaller randomized controlled trial reported no significant differences in patient outcomes with earlier warning alarms.27 A third study reported more appropriate rapid response team deployment and decreased mortality in a subgroup analysis.35

Effects, Facilitators, and Barriers

As shown in the Appendix Figure and further detailed in Table 3, the described effects were predominantly positive—57 positive effects vs 11 negative effects. These positive effects sorted primarily into the outcome and process domains.

 

 

All of the studies that compared their proposed model with one of various warning systems (eg, EWS, National Early Warning Score [NEWS], Modified Early Warning Score [MEWS]) showed superior performance (based on AUROC and reported predictive values). In 17 studies, the authors reported their model as more useful or superior to the EWS.20-23,26-28,34,36-41 Four studies reported real-time detection of deterioration before regular EWS,20,26,42 and three studies reported positive effects on patient-related outcomes.26,35 Four negative effects were noted on the controllability, validity, and potential limitations.27,42

There were 26 positive effects on the clinical process mentioned, 7 of which pointed out the effects of earlier, predictive alarming. Algorithms with higher PPVs reported greater rates of actionable alarms, less alarm fatigue, and improved workflow.21,22,24-27,30,32,33,35-38,40 Potential alarm fatigue was named as a barrier.27,42 Smoother scoring instead of binned categories was mentioned positively.24,26In the infrastructure domain, very few items were reported. The increased need for education on the used techniques was reported once as a negative effect.34 One of the positive infrastructural effects noted was more efficient planning and use of resources.24,37,40We identified 57 facilitators and 48 barriers for the clinical implementation and use of real-time predictive analytics (Appendix Figure). In the Technology domain, there were 18 facilitators and 20 barriers cited, and in the Organization domain, 25 and 14, respectively. They were equally present in the Professional and Physiology domains (6 vs 5, 8 vs 9).

Of the 38 remarks in the Technology domain, difficulty with implementation in daily practice was a commonly cited barrier.22,24,40,42 Difficulties included creating real-time data feeds out of the EMR, though there were mentions of some successful examples.25,27,36 Difficulty in the interpretability of AI was also considered a potential barrier.30,32,33,35,39,41 There were remarks as to the applicability of the prolonged prediction horizon because of the associated decoupling from the clinical view.39,42

Conservative attitudes toward new technologies and inadequate knowledge were mentioned as barriers.39 Repeated remarks were made on the difficulty of interpreting and responding to a predicted escalation, as the clinical pattern might not be recognizable at such an early stage. On the other hand, it is expected that less invasive countermeasures would be adequate to avert further escalation. Earlier recognition of possible escalations also raised potential ethical questions, such as when to discuss palliative care.24

The heterogeneity of the general ward population and the relatively low prevalence of deterioration were mentioned as barriers.24,30,38,41 There were also concerns that not all escalations are preventable and that some patient outcomes may not be modifiable.24,38

Many investigators expected reductions in false alarms and associated alarm fatigue (reflected as higher PPVs). Furthermore, they expected workflow to improve and workload to decrease.21,23,27,31,33,35,38,41 Despite the capacity of modern EMRs to store large amounts of patient data, some investigators felt improvements to real-time access, data quality and validity, and data density are needed to ensure valid associated predictions.21,22,24,32,37

DISCUSSION

As the complexity and comorbidity of hospitalized adults grow, predicting clinical deterioration is becoming more important. With an ever-increasing amount of available patient data, real-time algorithms can predict the patient’s clinical course with increasing accuracy, positively affecting outcomes.4,21,25,43 The studies identified in this scoping review, as measured by higher AUROC scores and improved PPVs, show that predictive algorithms can outperform more conventional EWS, enable earlier and more efficient alarming, and be successfully implemented on the general wards. However, formal meta-analysis was made infeasible by differences in populations, use of different endpoint definitions, cut-off points, time-horizons to prediction, and other methodologic heterogeneity.

 

 

There are several important limitations across these studies. In a clinical setting, these models would function as a screening test. Almost all studies report an AUROC; however, sensitivity and PPV or NNE (defined as 1/PPV) may be more useful than AUROC when predicting low-frequency events with high-potential clinical impact.44 Assessing the NNE is especially relevant because of its relation to alarm fatigue and responsiveness of clinicians.43 Alarm fatigue and lack of adequate response to alarms were repeatedly cited as potential barriers for application of automated scores. A more useful metric might be NNE over a certain timeframe and across a specified number of patients to more clearly reflect associated workload. Future studies should include these metrics as indicators of the usability and clinical impact of predictive models. This review could not assess PPV or NNE systematically due to inconsistencies in the reporting of these metrics.

Although the results of our scoping review are promising, there are limited data on clinical outcomes using these algorithms. Only three of five algorithms were used to guide clinical decision-making.25,27,35 Kollef et al27 showed shorter hospitalizations and Evans et al35 found decreased mortality rates in a multimorbid subgroup. Escobar et al25 found an overall and consistent decrease in mortality in a large, heterogenic population of inpatients across 21 hospitals. While Escobar et al’s findings provide strong evidence that predictive algorithms and structured follow-up on alarms can improve patient outcomes, it recognizes that not all facilities will have the resources to implement them.25 Dedicated round-the-clock follow-up of alarms has yet to be proven feasible for smaller institutions, and leaner solutions must be explored. The example set by Escobar et al25 should be translated into various settings to prove its reproducibility and to substantiate the clinical impact of predictive models and structured follow-up.

According to expert opinion, the use of high-frequency or continuous monitoring at low-acuity wards and AI algorithms to detect trends and patterns will reduce failure-to-rescue rates.4,9,43 However, most studies in our review focused on periodic spot-checked vital signs, and none of the AI algorithms were implemented in clinical care (Appendix Table 1). A significant barrier to implementation was uncertainty surrounding how to react to generated alarms.9,45 As algorithms become more complex and predict earlier, interpretability and causality in general can diminish, and the response to this type of alarm will be different from that of an acute warning from an EWS.

The assessment of predictive algorithm protocols must include their impact on clinical workflow, workload, and resource utilization. Earlier detection of deterioration can potentially allow coordinated alarm follow-up and lead to more efficient use of resources.20,21,31,43,46,47

Greater numbers of variables do not always improve the quality of monitoring. For example, in one study, an algorithm combining only heart rate, respiration rate, and age outperformed an EWS that tracked six vital sign measures.23 Algorithms using fewer variables may facilitate more frequent and less complex error-sensitive monitoring. Leaner measurements may also lead to higher patient and clinician acceptance.43,45The end goal of implementing predictive algorithms on the general ward is to provide timely, reliable, and actionable clinical decision support.43 As shown in a recent study by Blackwell et al,48 multiple prediction models for specific clinical events may increase interpretability and performance. Disease-specific algorithms may complement general algorithms for clinical deterioration and enhance overall performance.

 

 

STRENGTHS AND LIMITATIONS

We performed a comprehensive review of the current literature using a clear and reproducible methodology to minimize the risk of missing relevant publications. The identified research is mainly limited to large US centers and consists of mostly retrospective studies. Heterogeneity among inputs, endpoints, time horizons, and evaluation metrics make comparisons challenging. Comments on facilitators, barriers, and effects were limited. Positive publication bias may have led to overrepresentation of models showing clinical benefit.

RECOMMENDATIONS FOR FUTURE RESEARCH

Artificial intelligence and the use of continuous monitoring hold great promise in creating optimal predictive algorithms. Future studies should directly compare AI- and non-AI-based algorithms using continuous monitoring to determine predictive accuracy, feasibility, costs, and outcomes. A consensus on endpoint definitions, input variables, methodology, and reporting is needed to enhance reproducibility, comparability, and generalizability of future research. The current research is limited to a few research groups, predominantly in the United States. More international research could enhance validity and increase applicability across varying populations and settings. Greater collaboration would accelerate research and enhance short-cycled continuous improvements. Sharing databases with different populations, variables, and outcomes, such as the Medical Information Mart for Intensive Care database,49 could help develop, test, and compare models and contribute to consensus in data standardization and consistent reporting of results. Studies should be designed to determine clinical, societal, and economic effects in accordance with the Quadruple Aim principle.50 Successful implementation will depend not only on improved patient outcomes but also on cost-effectiveness, robust statistics, and end-user acceptance. Follow-up protocols and workflows also should be studied and optimized.

CONCLUSION

Predictive analytics based on vital sign monitoring can identify clinical deterioration at an earlier stage and can do so more accurately than conventional EWS. Implementation of such monitoring can simultaneously decrease alarm-related workload and enhance the efficiency of follow-up. While there is also emerging evidence of associated mortality reduction, it may be too soon to know how representative these findings are. The current literature is limited by heterogeneity across populations studied, monitoring frequency, definitions of deterioration, and clinical outcomes. Consensus is therefore needed to better compare tools and harmonize results. Artificial intelligence and continuous monitoring show great promise in advancing the field; however, additional studies to assess cost, actionability of results, and end-user acceptance are required.

References
  1. van Galen LS, Struik PW, Driesen BEJM, et al. Delayed recognition of deterioration of patients in general wards is mostly caused by human related monitoring failures: a root cause analysis of unplanned ICU admissions. PLoS One. 2016;11(8):e0161393. https://doi.org/10.1371/journal. pone.0161393
  2. Mardini L, Lipes J, Jayaraman D. Adverse outcomes associated with delayed intensive care consultation in medical and surgical inpatients. J Crit Care. 2012;27(6):688-693. https://doi.org/10.1016/j.jcrc.2012.04.011
  3. Young MP, Gooder VJ, McBride K, James B, Fisher ES. Inpatient transfers to the intensive care unit: delays are associated with increased mortality and morbidity. J Gen Intern Med. 2003;18(2):77-83. https://doi.org/10.1046/ j.1525-1497.2003.20441.x
  4. Khanna AK, Hoppe P, Saugel B. Automated continuous noninvasive ward monitoring: future directions and challenges. Crit Care. 2019;23(1):194. https://doi.org/10.1186/s13054-019-2485-7
  5. Ludikhuize J, Hamming A, de Jonge E, Fikkers BG. Rapid response systems in The Netherlands. Jt Comm J Qual Patient Saf. 2011;37(3):138-197. https:// doi.org/10.1016/s1553-7250(11)37017-1
  6. Cuthbertson BH, Boroujerdi M, McKie L, Aucott L, Prescott G. Can physiological variables and early warning scoring systems allow early recognition of the deteriorating surgical patient? Crit Care Med. 2007;35(2):402-409. https://doi.org/10.1097/01.ccm.0000254826.10520.87
  7. Alam N, Hobbelink EL, van Tienhoven AJ, van de Ven PM, Jansma EP, Nanayakkara PWB. The impact of the use of the Early Warning Score (EWS) on patient outcomes: a systematic review. Resuscitation. 2014;85(5):587-594. https://doi.org/10.1016/j.resuscitation.2014.01.013
  8. Weenk M, Koeneman M, van de Belt TH, Engelen LJLPG, van Goor H, Bredie SJH. Wireless and continuous monitoring of vital signs in patients at the general ward. Resuscitation. 2019;136:47-53. https://doi.org/10.1016/j.resuscitation.2019.01.017
  9. Cardona-Morrell M, Prgomet M, Turner RM, Nicholson M, Hillman K. Effectiveness of continuous or intermittent vital signs monitoring in preventing adverse events on general wards: a systematic review and meta-analysis. Int J Clin Pract. 2016;70(10):806-824. https://doi.org/10.1111/ijcp.12846
  10. Brown H, Terrence J, Vasquez P, Bates DW, Zimlichman E. Continuous monitoring in an inpatient medical-surgical unit: a controlled clinical trial. Am J Med. 2014;127(3):226-232. https://doi.org/10.1016/j.amjmed.2013.12.004
  11. Mestrom E, De Bie A, van de Steeg M, Driessen M, Atallah L, Bezemer R. Implementation of an automated early warning scoring system in a E8 Journal of Hospital Medicine® Published Online June 2021 An Official Publication of the Society of Hospital Medicine Peelen et al | Predicting Deterioration: A Scoping Review surgical ward: practical use and effects on patient outcomes. PLoS One. 2019;14(5):e0213402. https://doi.org/10.1371/journal.pone.0213402
  12. Jiang F, Jiang Y, Zhi H, et al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol. 2017;2(4):230-243. https://doi.org/10.1136/ svn-2017-000101
  13. Iwashyna TJ, Liu V. What’s so different about big data? A primer for clinicians trained to think epidemiologically. Ann Am Thorac Soc. 2014;11(7):1130- 1135. https://doi.org/10.1513/annalsats.201405-185as
  14. Jalali A, Bender D, Rehman M, Nadkanri V, Nataraj C. Advanced analytics for outcome prediction in intensive care units. Conf Proc IEEE Eng Med Biol Soc. 2016;2016:2520-2524. https://doi.org/10.1109/embc.2016.7591243
  15. Munn Z, Peters MDJ, Stern C, Tufanaru C, McArthur A, Aromataris E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. 2018;18(1):143. https://doi.org/10.1186/s12874-018-0611-x
  16. Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19-32. https://doi.org/10.1080/13645 57032000119616
  17. Tricco AC, Lillie E, Zarin W, et al. PRISMA extension for scoping reviews (PRISMAScR): checklist and explanation. Ann Intern Med. 2018;169(7):467- 473. https://doi.org/10.7326/m18-0850
  18. Gagnon MP, Desmartis M, Gagnon J, et al. Framework for user involvement in health technology assessment at the local level: views of health managers, user representatives, and clinicians. Int J Technol Assess Health Care. 2015;31(1-2):68-77. https://doi.org/10.1017/s0266462315000070
  19. Donabedian A. The quality of care. How can it be assessed? JAMA. 1988;260(12):1743-1748. https://doi.org/10.1001/jama.260.12.1743
  20. Churpek MM, Yuen TC, Winslow C, et al. Multicenter development and validation of a risk stratification tool for ward patients. Am J Respir Crit Care Med. 2014;190(6):649-655. https://doi.org/10.1164/rccm.201406-1022oc
  21. Churpek MM, Yuen TC, Winslow C, Meltzer DO, Kattan MW, Edelson DP. Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards. Crit Care Med. 2016;44(2):368-374. https://doi.org/10.1097/ccm.0000000000001571
  22. Bartkowiak B, Snyder AM, Benjamin A, et al. Validating the electronic cardiac arrest risk triage (eCART) score for risk stratification of surgical inpatients in the postoperative setting: retrospective cohort study. Ann Surg. 2019;269(6):1059-1063. https://doi.org/10.1097/sla.0000000000002665
  23. Edelson DP, Carey K, Winslow CJ, Churpek MM. Less is more: detecting clinical deterioration in the hospital with machine learning using only age, heart rate and respiratory rate. Abstract presented at: American Thoracic Society International Conference; May 22, 2018; San Diego, California. Am J Resp Crit Care Med. 2018;197:A4444.
  24. Escobar GJ, LaGuardia JC, Turk BJ, Ragins A, Kipnis P, Draper D. Early detection of impending physiologic deterioration among patients who are not in intensive care: development of predictive models using data from an automated electronic medical record. J Hosp Med. 2012;7(5):388-395. https:// doi.org/10.1002/jhm.1929
  25. Escobar GJ, Liu VX, Schuler A, Lawson B, Greene JD, Kipnis P. Automated identification of adults at risk for in-hospital clinical deterioration. N Engl J Med. 2020;383(20):1951-1960. https://doi.org/10.1056/nejmsa2001090
  26. Kipnis P, Turk BJ, Wulf DA, et al. Development and validation of an electronic medical record-based alert score for detection of inpatient deterioration outside the ICU. J Biomed Inform. 2016;64:10-19. https://doi.org/10.1016/j. jbi.2016.09.013
  27. Kollef MH, Chen Y, Heard K, et al. A randomized trial of real-time automated clinical deterioration alerts sent to a rapid response team. J Hosp Med. 2014;9(7):424-429. https://doi.org/10.1002/jhm.2193
  28. Hackmann G, Chen M, Chipara O, et al. Toward a two-tier clinical warning system for hospitalized patients. AMIA Annu Symp Proc. 2011;2011:511-519.
  29. Bailey TC, Chen Y, Mao Y, Lu, C, Hackmann G, Micek ST. A trial of a real-time alert for clinical deterioration in patients hospitalized on general medical wards. J Hosp Med. 2013;8(5):236-242. https://doi.org/10.1002/jhm.2009
  30. Kwon JM, Lee Y, Lee Y, Lee S, Park J. An algorithm based on deep learning for predicting in-hospital cardiac arrest. J Am Heart Assoc. 2018;7(13):e008678. https://doi.org/10.1161/jaha.118.008678
  31. Correia S, Gomes A, Shahriari S, Almeida JP, Severo M, Azevedo A. Performance of the early warning system vital to predict unanticipated higher-level of care admission and in-hospital death of ward patients. Value Health. 2018;21(S3):S360. https://doi.org/10.1016/j.jval.2018.09.2152
  32. Shamout FE, Zhu T, Sharma P, Watkinson PJ, Clifton DA. Deep interpretable early warning system for the detection of clinical deterioration. IEEE J Biomed Health Inform. 2020;24(2):437-446. https://doi.org/10.1109/ jbhi.2019.2937803
  33. Bai Y, Do DH, Harris PRE, et al. Integrating monitor alarms with laboratory test results to enhance patient deterioration prediction. J Biomed Inform. 2015;53:81-92. https://doi.org/10.1016/j.jbi.2014.09.006
  34. Hu X, Sapo M, Nenov V, et al. Predictive combinations of monitor alarms preceding in-hospital code blue events. J Biomed Inform. 2012;45(5):913-921. https://doi.org/10.1016/j.jbi.2012.03.001
  35. Evans RS, Kuttler KG, Simpson KJ, et al. Automated detection of physiologic deterioration in hospitalized patients. J Am Med Inform Assoc. 2015;22(2):350-360. https://doi.org/10.1136/amiajnl-2014-002816
  36. Ghosh E, Eshelman L, Yang L, Carlson E, Lord B. Early deterioration indicator: data-driven approach to detecting deterioration in general ward. Resuscitation. 2018;122:99-105. https://doi.org/10.1016/j.resuscitation. 2017.10.026
  37. Kang MA, Churpek MM, Zadravecz FJ, Adhikari R, Twu NM, Edelson DP: Real-time risk prediction on the wards: a feasibility study. Crit Care Med. 2016;44(8):1468-1473. https://doi.org/10.1097/ccm.0000000000001716
  38. Hu SB, Wong DJL, Correa A, Li N, Deng JC. Prediction of clinical deterioration in hospitalized adult patients with hematologic malignancies using a neural network model. PLoS One. 2016;11(8):e0161401. https://doi. org/10.1371/journal.pone.0161401
  39. Rothman MJ, Rothman SI, Beals J 4th. Development and validation of a continuous measure of patient condition using the electronic medical record. J Biomed Inform. 2013;46(5):837-848. https://doi.org/10.1016/j. jbi.2013.06.011
  40. Alaa AM, Yoon J, Hu S, van der Schaar M. Personalized risk scoring for critical care prognosis using mixtures of Gaussian processes. IEEE Trans Biomed Eng. 2018;65(1):207-218. https://doi.org/10.1109/tbme.2017.2698602
  41. Mohamadlou H, Panchavati S, Calvert J, et al. Multicenter validation of a machine-learning algorithm for 48-h all-cause mortality prediction. Health Informatics J. 2020;26(3):1912-1925. https://doi.org/10.1177/1460458219894494
  42. Alvarez CA, Clark CA, Zhang S, et al. Predicting out of intensive care unit cardiopulmonary arrest or death using electronic medical record data. BMC Med Inform Decis Mak. 2013;13:28. https://doi.org/10.1186/1472-6947-13-28
  43. Vincent JL, Einav S, Pearse R, et al. Improving detection of patient deterioration in the general hospital ward environment. Eur J Anaesthesiol. 2018;35(5):325-333. https://doi.org/10.1097/eja.0000000000000798
  44. Romero-Brufau S, Huddleston JM, Escobar GJ, Liebow M. Why the C-statistic is not informative to evaluate early warning scores and what metrics to use. Crit Care. 2015;19(1):285. https://doi.org/10.1186/s13054-015-0999-1
  45. Weenk M, Bredie SJ, Koeneman M, Hesselink G, van Goor H, van de Belt TH. Continuous monitoring of the vital signs in the general ward using wearable devices: randomized controlled trial. J Med Internet Res. 2020;22(6):e15471. https://doi.org/10.2196/15471
  46. Wellner B, Grand J, Canzone E, et al. Predicting unplanned transfers to the intensive care unit: a machine learning approach leveraging diverse clinical elements. JMIR Med Inform. 2017;5(4):e45. https://doi.org/10.2196/medinform.8680
  47. Elliott M, Baird J. Pulse oximetry and the enduring neglect of respiratory rate assessment: a commentary on patient surveillance. Br J Nurs. 2019;28(19):1256-1259. https://doi.org/10.12968/bjon.2019.28.19.1256
  48. Blackwell JN, Keim-Malpass J, Clark MT, et al. Early detection of in-patient deterioration: one prediction model does not fit all. Crit Care Explor. 2020;2(5):e0116. https://doi.org/10.1097/cce.0000000000000116
  49. Johnson AEW, Pollard TJ, Shen L, et al. MIMIC-III, a freely accessible critical care database. Sci Data. 2016;3:160035. https://doi.org/10.1038/sdata.2016.35
  50. Bodenheimer T, Sinsky C. From triple to quadruple aim: care of the patient requires care of the provider. Ann Fam Med. 2014;12(6):573-576. https://doi. org/10.1370/afm.1713
  51.  Kirkland LL, Malinchoc M, O’Byrne M, et al. A clinical deterioration prediction tool for internal medicine patients. Am J Med Qual. 2013;28(2):135-142 https://doi.org/10.1177/1062860612450459
References
  1. van Galen LS, Struik PW, Driesen BEJM, et al. Delayed recognition of deterioration of patients in general wards is mostly caused by human related monitoring failures: a root cause analysis of unplanned ICU admissions. PLoS One. 2016;11(8):e0161393. https://doi.org/10.1371/journal. pone.0161393
  2. Mardini L, Lipes J, Jayaraman D. Adverse outcomes associated with delayed intensive care consultation in medical and surgical inpatients. J Crit Care. 2012;27(6):688-693. https://doi.org/10.1016/j.jcrc.2012.04.011
  3. Young MP, Gooder VJ, McBride K, James B, Fisher ES. Inpatient transfers to the intensive care unit: delays are associated with increased mortality and morbidity. J Gen Intern Med. 2003;18(2):77-83. https://doi.org/10.1046/ j.1525-1497.2003.20441.x
  4. Khanna AK, Hoppe P, Saugel B. Automated continuous noninvasive ward monitoring: future directions and challenges. Crit Care. 2019;23(1):194. https://doi.org/10.1186/s13054-019-2485-7
  5. Ludikhuize J, Hamming A, de Jonge E, Fikkers BG. Rapid response systems in The Netherlands. Jt Comm J Qual Patient Saf. 2011;37(3):138-197. https:// doi.org/10.1016/s1553-7250(11)37017-1
  6. Cuthbertson BH, Boroujerdi M, McKie L, Aucott L, Prescott G. Can physiological variables and early warning scoring systems allow early recognition of the deteriorating surgical patient? Crit Care Med. 2007;35(2):402-409. https://doi.org/10.1097/01.ccm.0000254826.10520.87
  7. Alam N, Hobbelink EL, van Tienhoven AJ, van de Ven PM, Jansma EP, Nanayakkara PWB. The impact of the use of the Early Warning Score (EWS) on patient outcomes: a systematic review. Resuscitation. 2014;85(5):587-594. https://doi.org/10.1016/j.resuscitation.2014.01.013
  8. Weenk M, Koeneman M, van de Belt TH, Engelen LJLPG, van Goor H, Bredie SJH. Wireless and continuous monitoring of vital signs in patients at the general ward. Resuscitation. 2019;136:47-53. https://doi.org/10.1016/j.resuscitation.2019.01.017
  9. Cardona-Morrell M, Prgomet M, Turner RM, Nicholson M, Hillman K. Effectiveness of continuous or intermittent vital signs monitoring in preventing adverse events on general wards: a systematic review and meta-analysis. Int J Clin Pract. 2016;70(10):806-824. https://doi.org/10.1111/ijcp.12846
  10. Brown H, Terrence J, Vasquez P, Bates DW, Zimlichman E. Continuous monitoring in an inpatient medical-surgical unit: a controlled clinical trial. Am J Med. 2014;127(3):226-232. https://doi.org/10.1016/j.amjmed.2013.12.004
  11. Mestrom E, De Bie A, van de Steeg M, Driessen M, Atallah L, Bezemer R. Implementation of an automated early warning scoring system in a E8 Journal of Hospital Medicine® Published Online June 2021 An Official Publication of the Society of Hospital Medicine Peelen et al | Predicting Deterioration: A Scoping Review surgical ward: practical use and effects on patient outcomes. PLoS One. 2019;14(5):e0213402. https://doi.org/10.1371/journal.pone.0213402
  12. Jiang F, Jiang Y, Zhi H, et al. Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol. 2017;2(4):230-243. https://doi.org/10.1136/ svn-2017-000101
  13. Iwashyna TJ, Liu V. What’s so different about big data? A primer for clinicians trained to think epidemiologically. Ann Am Thorac Soc. 2014;11(7):1130- 1135. https://doi.org/10.1513/annalsats.201405-185as
  14. Jalali A, Bender D, Rehman M, Nadkanri V, Nataraj C. Advanced analytics for outcome prediction in intensive care units. Conf Proc IEEE Eng Med Biol Soc. 2016;2016:2520-2524. https://doi.org/10.1109/embc.2016.7591243
  15. Munn Z, Peters MDJ, Stern C, Tufanaru C, McArthur A, Aromataris E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. 2018;18(1):143. https://doi.org/10.1186/s12874-018-0611-x
  16. Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19-32. https://doi.org/10.1080/13645 57032000119616
  17. Tricco AC, Lillie E, Zarin W, et al. PRISMA extension for scoping reviews (PRISMAScR): checklist and explanation. Ann Intern Med. 2018;169(7):467- 473. https://doi.org/10.7326/m18-0850
  18. Gagnon MP, Desmartis M, Gagnon J, et al. Framework for user involvement in health technology assessment at the local level: views of health managers, user representatives, and clinicians. Int J Technol Assess Health Care. 2015;31(1-2):68-77. https://doi.org/10.1017/s0266462315000070
  19. Donabedian A. The quality of care. How can it be assessed? JAMA. 1988;260(12):1743-1748. https://doi.org/10.1001/jama.260.12.1743
  20. Churpek MM, Yuen TC, Winslow C, et al. Multicenter development and validation of a risk stratification tool for ward patients. Am J Respir Crit Care Med. 2014;190(6):649-655. https://doi.org/10.1164/rccm.201406-1022oc
  21. Churpek MM, Yuen TC, Winslow C, Meltzer DO, Kattan MW, Edelson DP. Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards. Crit Care Med. 2016;44(2):368-374. https://doi.org/10.1097/ccm.0000000000001571
  22. Bartkowiak B, Snyder AM, Benjamin A, et al. Validating the electronic cardiac arrest risk triage (eCART) score for risk stratification of surgical inpatients in the postoperative setting: retrospective cohort study. Ann Surg. 2019;269(6):1059-1063. https://doi.org/10.1097/sla.0000000000002665
  23. Edelson DP, Carey K, Winslow CJ, Churpek MM. Less is more: detecting clinical deterioration in the hospital with machine learning using only age, heart rate and respiratory rate. Abstract presented at: American Thoracic Society International Conference; May 22, 2018; San Diego, California. Am J Resp Crit Care Med. 2018;197:A4444.
  24. Escobar GJ, LaGuardia JC, Turk BJ, Ragins A, Kipnis P, Draper D. Early detection of impending physiologic deterioration among patients who are not in intensive care: development of predictive models using data from an automated electronic medical record. J Hosp Med. 2012;7(5):388-395. https:// doi.org/10.1002/jhm.1929
  25. Escobar GJ, Liu VX, Schuler A, Lawson B, Greene JD, Kipnis P. Automated identification of adults at risk for in-hospital clinical deterioration. N Engl J Med. 2020;383(20):1951-1960. https://doi.org/10.1056/nejmsa2001090
  26. Kipnis P, Turk BJ, Wulf DA, et al. Development and validation of an electronic medical record-based alert score for detection of inpatient deterioration outside the ICU. J Biomed Inform. 2016;64:10-19. https://doi.org/10.1016/j. jbi.2016.09.013
  27. Kollef MH, Chen Y, Heard K, et al. A randomized trial of real-time automated clinical deterioration alerts sent to a rapid response team. J Hosp Med. 2014;9(7):424-429. https://doi.org/10.1002/jhm.2193
  28. Hackmann G, Chen M, Chipara O, et al. Toward a two-tier clinical warning system for hospitalized patients. AMIA Annu Symp Proc. 2011;2011:511-519.
  29. Bailey TC, Chen Y, Mao Y, Lu, C, Hackmann G, Micek ST. A trial of a real-time alert for clinical deterioration in patients hospitalized on general medical wards. J Hosp Med. 2013;8(5):236-242. https://doi.org/10.1002/jhm.2009
  30. Kwon JM, Lee Y, Lee Y, Lee S, Park J. An algorithm based on deep learning for predicting in-hospital cardiac arrest. J Am Heart Assoc. 2018;7(13):e008678. https://doi.org/10.1161/jaha.118.008678
  31. Correia S, Gomes A, Shahriari S, Almeida JP, Severo M, Azevedo A. Performance of the early warning system vital to predict unanticipated higher-level of care admission and in-hospital death of ward patients. Value Health. 2018;21(S3):S360. https://doi.org/10.1016/j.jval.2018.09.2152
  32. Shamout FE, Zhu T, Sharma P, Watkinson PJ, Clifton DA. Deep interpretable early warning system for the detection of clinical deterioration. IEEE J Biomed Health Inform. 2020;24(2):437-446. https://doi.org/10.1109/ jbhi.2019.2937803
  33. Bai Y, Do DH, Harris PRE, et al. Integrating monitor alarms with laboratory test results to enhance patient deterioration prediction. J Biomed Inform. 2015;53:81-92. https://doi.org/10.1016/j.jbi.2014.09.006
  34. Hu X, Sapo M, Nenov V, et al. Predictive combinations of monitor alarms preceding in-hospital code blue events. J Biomed Inform. 2012;45(5):913-921. https://doi.org/10.1016/j.jbi.2012.03.001
  35. Evans RS, Kuttler KG, Simpson KJ, et al. Automated detection of physiologic deterioration in hospitalized patients. J Am Med Inform Assoc. 2015;22(2):350-360. https://doi.org/10.1136/amiajnl-2014-002816
  36. Ghosh E, Eshelman L, Yang L, Carlson E, Lord B. Early deterioration indicator: data-driven approach to detecting deterioration in general ward. Resuscitation. 2018;122:99-105. https://doi.org/10.1016/j.resuscitation. 2017.10.026
  37. Kang MA, Churpek MM, Zadravecz FJ, Adhikari R, Twu NM, Edelson DP: Real-time risk prediction on the wards: a feasibility study. Crit Care Med. 2016;44(8):1468-1473. https://doi.org/10.1097/ccm.0000000000001716
  38. Hu SB, Wong DJL, Correa A, Li N, Deng JC. Prediction of clinical deterioration in hospitalized adult patients with hematologic malignancies using a neural network model. PLoS One. 2016;11(8):e0161401. https://doi. org/10.1371/journal.pone.0161401
  39. Rothman MJ, Rothman SI, Beals J 4th. Development and validation of a continuous measure of patient condition using the electronic medical record. J Biomed Inform. 2013;46(5):837-848. https://doi.org/10.1016/j. jbi.2013.06.011
  40. Alaa AM, Yoon J, Hu S, van der Schaar M. Personalized risk scoring for critical care prognosis using mixtures of Gaussian processes. IEEE Trans Biomed Eng. 2018;65(1):207-218. https://doi.org/10.1109/tbme.2017.2698602
  41. Mohamadlou H, Panchavati S, Calvert J, et al. Multicenter validation of a machine-learning algorithm for 48-h all-cause mortality prediction. Health Informatics J. 2020;26(3):1912-1925. https://doi.org/10.1177/1460458219894494
  42. Alvarez CA, Clark CA, Zhang S, et al. Predicting out of intensive care unit cardiopulmonary arrest or death using electronic medical record data. BMC Med Inform Decis Mak. 2013;13:28. https://doi.org/10.1186/1472-6947-13-28
  43. Vincent JL, Einav S, Pearse R, et al. Improving detection of patient deterioration in the general hospital ward environment. Eur J Anaesthesiol. 2018;35(5):325-333. https://doi.org/10.1097/eja.0000000000000798
  44. Romero-Brufau S, Huddleston JM, Escobar GJ, Liebow M. Why the C-statistic is not informative to evaluate early warning scores and what metrics to use. Crit Care. 2015;19(1):285. https://doi.org/10.1186/s13054-015-0999-1
  45. Weenk M, Bredie SJ, Koeneman M, Hesselink G, van Goor H, van de Belt TH. Continuous monitoring of the vital signs in the general ward using wearable devices: randomized controlled trial. J Med Internet Res. 2020;22(6):e15471. https://doi.org/10.2196/15471
  46. Wellner B, Grand J, Canzone E, et al. Predicting unplanned transfers to the intensive care unit: a machine learning approach leveraging diverse clinical elements. JMIR Med Inform. 2017;5(4):e45. https://doi.org/10.2196/medinform.8680
  47. Elliott M, Baird J. Pulse oximetry and the enduring neglect of respiratory rate assessment: a commentary on patient surveillance. Br J Nurs. 2019;28(19):1256-1259. https://doi.org/10.12968/bjon.2019.28.19.1256
  48. Blackwell JN, Keim-Malpass J, Clark MT, et al. Early detection of in-patient deterioration: one prediction model does not fit all. Crit Care Explor. 2020;2(5):e0116. https://doi.org/10.1097/cce.0000000000000116
  49. Johnson AEW, Pollard TJ, Shen L, et al. MIMIC-III, a freely accessible critical care database. Sci Data. 2016;3:160035. https://doi.org/10.1038/sdata.2016.35
  50. Bodenheimer T, Sinsky C. From triple to quadruple aim: care of the patient requires care of the provider. Ann Fam Med. 2014;12(6):573-576. https://doi. org/10.1370/afm.1713
  51.  Kirkland LL, Malinchoc M, O’Byrne M, et al. A clinical deterioration prediction tool for internal medicine patients. Am J Med Qual. 2013;28(2):135-142 https://doi.org/10.1177/1062860612450459
Issue
Journal of Hospital Medicine 16(10)
Issue
Journal of Hospital Medicine 16(10)
Page Number
612-619. Published Online First June 25, 2021
Page Number
612-619. Published Online First June 25, 2021
Publications
Publications
Topics
Article Type
Display Headline
Algorithms for Prediction of Clinical Deterioration on the General Wards: A Scoping Review
Display Headline
Algorithms for Prediction of Clinical Deterioration on the General Wards: A Scoping Review
Sections
Disallow All Ads
Correspondence Location
Email: [email protected]; Telephone: +31 6 3043 8698; Twitter: @PeelenRoel.
Content Gating
Gated (full article locked unless allowed per User)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Gating Strategy
First Peek Free
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media
Media Files