User login
The value of using ultrasound to rule out deep vein thrombosis in cases of cellulitis
The “Things We Do for No Reason” series reviews practices which have become common parts of hospital care but which may provide little value to our patients. Practices reviewed in the TWDFNR series do not represent “black and white” conclusions or clinical practice standards, but are meant as a starting place for research and active discussions among hospitalists and patients. We invite you to be part of that discussion. https://www.choosingwisely.org/
Because of overlapping clinical manifestations, clinicians often order ultrasound to rule out deep vein thrombosis (DVT) in cases of cellulitis. Ultrasound testing is performed for 16% to 73% of patients diagnosed with cellulitis. Although testing is common, the pooled incidence of DVT is low (3.1%). Few data elucidate which patients with cellulitis are more likely to have concurrent DVT and require further testing. The Wells clinical prediction rule with
CASE REPORT
A 50-year-old man presented to the emergency department with a 3-day-old cut on his anterior right shin. Associated redness, warmth, pain, and swelling had progressed. The patient had no history of prior DVT or pulmonary embolism (PE). His temperature was 38.5°C, and his white blood cell count of 18,000. On review of systems, he denied shortness of breath and chest pain. He was diagnosed with cellulitis and administered intravenous fluids and cefazolin. The clinician wondered whether to perform lower extremity ultrasound to rule out concurrent DVT.
WHY YOU MIGHT THINK ULTRASOUND IS HELPFUL IN RULING OUT DVT IN CELLULITIS
Lower extremity cellulitis, a common infection of the skin and subcutaneous tissues, is characterized by unilateral erythema, pain, warmth, and swelling. The infection usually follows a skin breach that allows bacteria to enter. DVT may present similarly, and symptoms can include mild leukocytosis and elevated temperature. Because of the clinical similarities, clinicians often order compression ultrasound of the extremity to rule out concurrent DVT in cellulitis. Further impetus for testing stems from fear of the potential complications of untreated DVT, including post-thrombotic syndrome, chronic venous insufficiency, and venous ulceration. A subsequent PE can be fatal, or can cause significant morbidity, including chronic VTE with associated pulmonary hypertension. An estimated quarter of all PEs present as sudden death.1
WHY ULTRASOUND IS NOT HELPFUL IN THIS SETTING
Studies have shown that ultrasound is ordered for 16% to 73% of patients with a cellulitis diagnosis.2,3 Although testing is commonly performed, a meta-analysis of 9 studies of cellulitis patients who underwent ultrasound testing for concurrent DVT revealed a low pooled incidence of total DVT (3.1%) and proximal DVT (2.1%).4 Maze et al.2 retrospectively reviewed 1515 cellulitis cases (identified by International Classification of Diseases, Ninth Revision codes) at a single center in New Zealand over 3 years. Of the 1515 patients, 240 (16%) had ultrasound performed, and only 3 (1.3%) were found to have DVT. Two of the 3 had active malignancy, and the third had injected battery acid into the area. In a 5-year retrospective cohort study at a Veterans Administration hospital in Connecticut, Gunderson and Chang3 reviewed the cases of 183 patients with cellulitis and found ultrasound testing commonly performed (73% of cases) to assess for DVT. Only 1 patient (<1%) was diagnosed with new DVT in the ipsilateral leg, and acute DVT was diagnosed in the contralateral leg of 2 other patients. Overall, these studies indicate the incidence of concurrent DVT in cellulitis is low, regardless of the frequency of ultrasound testing.
Although the cost of a single ultrasound test is not prohibitive, annual total costs hospital-wide and nationally are large. In the United States, the charge for a unilateral duplex ultrasound of the extremity ranges from $260 to $1300, and there is an additional charge for interpretation by a radiologist.5 In a retrospective study spanning 3.5 years and involving 2 community hospitals in Michigan, an estimated $290,000 was spent on ultrasound tests defined as unnecessary for patients with cellulitis.6 A limitation of the study was defining a test as unnecessary based on its result being negative.
DOES WELLS SCORE WITH D-DIMER HELP DEFINE A LOW-RISK POPULATION?
The Wells clinical prediction rule is commonly used to assess the pretest probability of DVT in patients presenting with unilateral leg symptoms. The Wells score is often combined with
WHEN MIGHT ULTRASOUND BE HELPFUL IN CELLULITIS?
Investigators have described possible DVT risk factors in patients with cellulitis, but definitive associations are lacking because of the insufficient number of patients studied.8,9 The most consistently identified DVT risk factor is history of previous thromboembolism. In a retrospective analysis of patients with cellulitis, Afzal et al.6 found that, of the 66.8% who underwent ultrasound testing, 5.5% were identified as having concurrent DVT. The authors performed univariate analyses of 15 potential risk factors, including active malignancy, oral contraceptive pill use, recent hospitalization, and surgery. A higher incidence of DVT was found for patients with history of VTE (odds ratio [OR], 5.7; 95% confidence interval [CI], 2.3-13.7), calf swelling (OR, 4.5; 95% CI, 1.3-15.8), CVA (OR, 3.5; 95% CI, 1.2-10.1), or hypertension (OR, 3.5; 95% CI, 0.98-12.2). Given the wide confidence intervals, paucity of studies, and lack of definitive data in the setting of cellulitis, clinicians may want to consider the risk factors established in larger trials in other settings, including known immobility (OR, <2); thrombophilia, CHF, and CVA with hemiparesis (OR, 2-9); and trauma and recent surgery (OR, >10).10
WHAT YOU SHOULD DO INSTEAD
As the incidence of concurrent VTE in patients with cellulitis is low, the essential step is to make a clear diagnosis of cellulitis based on its established signs and symptoms. A 2-center trial of 145 patients found that cellulitis was diagnosed accurately by general medicine and emergency medicine physicians 72% of the time, with evaluation by dermatologists and infectious disease specialists used as the gold standard. Only 5% of the misdiagnosed patients were diagnosed with DVT; stasis dermatitis was the most common alternative diagnosis. Taking a thorough history may elicit risk factors consistent with cellulitis, such as a recent injury with a break in the skin. On examination, cellulitis should be suspected for patients with fever and localized pain, redness, swelling, and warmth—the cardinal signs of dolor, rubor, tumor, and calor. An injury or entry site and leukocytosis also support the diagnosis of cellulitis. Distinct margins of erythema on the skin are highly suspicious for erysipelas.11 Other physical findings (eg, laceration, purulent drainage, lymphangitic spread, fluctuating mass) also are consistent with a diagnosis of cellulitis.
The patient’s history is also essential in determining whether any DVT risk factors are present. Past medical history of VTE or CVA, or recent history of surgery, immobility, or trauma, should alert the clinician to the possibility of DVT. Family history of VTE increases the likelihood of DVT. Acute shortness of breath or chest pain in the setting of concerning lower extremity findings for DVT should raise concern for DVT and concurrent PE.
If the classic features of cellulitis are present, empiric antibiotics should be initiated. Routine ultrasound testing for all patients with cellulitis is of low value. However, as the incidence of DVT in this population is not negligible, those with VTE risk factors should be targeted for testing. Studies in the setting of cellulitis provide little guidance regarding specific risk factors that can be used to determine who should undergo further testing. Given this limitation, we suggest that clinicians incorporate into their decision making the well-established VTE risk factors identified for large populations studied in other settings, such as the postoperative period. Specifically, clinicians should consider ultrasound testing for patients with cellulitis and prior history of VTE; immobility; thrombophilia, CHF, and CVA with hemiparesis; or trauma and recent surgery.10-12 Ultrasound should also be considered for patients with cellulitis that does not improve and for patients whose localized symptoms worsen despite use of antibiotics.
RECOMMENDATIONS
Do not routinely perform ultrasound to rule out concurrent DVT in cases of cellulitis.
Consider compression ultrasound if there is a history of VTE; immobility; thrombophilia, CHF, and CVA with hemiparesis; or trauma and recent surgery. Also consider it for patients who do not respond to antibiotics.
- In cases of cellulitis, avoid use of the Wells score alone or with
D -dimer testing, as it likely overestimates the DVT risk.
CONCLUSION
The current evidence shows that, for most patients with cellulitis, routine ultrasound testing for DVT is unnecessary. Ultrasound should be considered for patients with potent VTE risk factors. If symptoms do not improve, or if they worsen despite use of antibiotics, clinicians should be alert to potential anchoring bias and consider DVT. The Wells clinical prediction rule overestimates the incidence of DVT in cellulitis and has little value in this setting.
Disclosure
Nothing to report.
Do you think this is a low-value practice? Is this truly a “Thing We Do for No Reason”? Let us know what you do in your practice and propose ideas for other “Things We Do for No Reason” topics. Please join in the conversation online at Twitter (#TWDFNR)/Facebook and don’t forget to “Like It” on Facebook or retweet it on Twitter. We invite you to propose ideas for other “Things We Do for No Reason” topics by emailing [email protected].
1. Heit JA. The epidemiology of venous thromboembolism in the community: implications for prevention and management. J Thromb Thrombolysis. 2006;21(1):23-29. PubMed
2. Maze MJ, Pithie A, Dawes T, Chambers ST. An audit of venous duplex ultrasonography in patients with lower limb cellulitis. N Z Med J. 2011;124(1329):53-56. PubMed
3. Gunderson CG, Chang JJ. Overuse of compression ultrasound for patients with lower extremity cellulitis. Thromb Res. 2014;134(4):846-850. PubMed
4. Gunderson CG, Chang JJ. Risk of deep vein thrombosis in patients with cellulitis and erysipelas: a systematic review and meta-analysis. Thromb Res. 2013;132(3):336-340. PubMed
5. Extremity ultrasound (nonvascular) cost and procedure information. http://www.newchoicehealth.com/procedures/extremity-ultrasound-nonvascular. Accessed February 15, 2016.
6. Afzal MZ, Saleh MM, Razvi S, Hashmi H, Lampen R. Utility of lower extremity Doppler in patients with lower extremity cellulitis: a need to change the practice? South Med J. 2015;108(7):439-444. PubMed
7. Goodacre S, Sutton AJ, Sampson FC. Meta-analysis: the value of clinical assessment in the diagnosis of deep venous thrombosis. Ann Intern Med. 2005;143(2):129-139. PubMed
8. Maze MJ, Skea S, Pithie A, Metcalf S, Pearson JF, Chambers ST. Prevalence of concurrent deep vein thrombosis in patients with lower limb cellulitis: a prospective cohort study. BMC Infect Dis. 2013;13:141. PubMed
9. Bersier D, Bounameaux H. Cellulitis and deep vein thrombosis: a controversial association. J Thromb Haemost. 2003;1(4):867-868. PubMed
10. Anderson FA Jr, Spencer FA. Risk factors for venous thromboembolism. Circulation. 2003;107(23 suppl 1):I9-I16. PubMed
11. Rabuka CE, Azoulay LY, Kahn SR. Predictors of a positive duplex scan in patients with a clinical presentation compatible with deep vein thrombosis or cellulitis. Can J Infect Dis. 2003;14(4):210-214. PubMed
12. Samama MM. An epidemiologic study of risk factors for deep vein thrombosis in medical outpatients: the Sirius Study. Arch Intern Med. 2000;160(22):3415-3420. PubMed
The “Things We Do for No Reason” series reviews practices which have become common parts of hospital care but which may provide little value to our patients. Practices reviewed in the TWDFNR series do not represent “black and white” conclusions or clinical practice standards, but are meant as a starting place for research and active discussions among hospitalists and patients. We invite you to be part of that discussion. https://www.choosingwisely.org/
Because of overlapping clinical manifestations, clinicians often order ultrasound to rule out deep vein thrombosis (DVT) in cases of cellulitis. Ultrasound testing is performed for 16% to 73% of patients diagnosed with cellulitis. Although testing is common, the pooled incidence of DVT is low (3.1%). Few data elucidate which patients with cellulitis are more likely to have concurrent DVT and require further testing. The Wells clinical prediction rule with
CASE REPORT
A 50-year-old man presented to the emergency department with a 3-day-old cut on his anterior right shin. Associated redness, warmth, pain, and swelling had progressed. The patient had no history of prior DVT or pulmonary embolism (PE). His temperature was 38.5°C, and his white blood cell count of 18,000. On review of systems, he denied shortness of breath and chest pain. He was diagnosed with cellulitis and administered intravenous fluids and cefazolin. The clinician wondered whether to perform lower extremity ultrasound to rule out concurrent DVT.
WHY YOU MIGHT THINK ULTRASOUND IS HELPFUL IN RULING OUT DVT IN CELLULITIS
Lower extremity cellulitis, a common infection of the skin and subcutaneous tissues, is characterized by unilateral erythema, pain, warmth, and swelling. The infection usually follows a skin breach that allows bacteria to enter. DVT may present similarly, and symptoms can include mild leukocytosis and elevated temperature. Because of the clinical similarities, clinicians often order compression ultrasound of the extremity to rule out concurrent DVT in cellulitis. Further impetus for testing stems from fear of the potential complications of untreated DVT, including post-thrombotic syndrome, chronic venous insufficiency, and venous ulceration. A subsequent PE can be fatal, or can cause significant morbidity, including chronic VTE with associated pulmonary hypertension. An estimated quarter of all PEs present as sudden death.1
WHY ULTRASOUND IS NOT HELPFUL IN THIS SETTING
Studies have shown that ultrasound is ordered for 16% to 73% of patients with a cellulitis diagnosis.2,3 Although testing is commonly performed, a meta-analysis of 9 studies of cellulitis patients who underwent ultrasound testing for concurrent DVT revealed a low pooled incidence of total DVT (3.1%) and proximal DVT (2.1%).4 Maze et al.2 retrospectively reviewed 1515 cellulitis cases (identified by International Classification of Diseases, Ninth Revision codes) at a single center in New Zealand over 3 years. Of the 1515 patients, 240 (16%) had ultrasound performed, and only 3 (1.3%) were found to have DVT. Two of the 3 had active malignancy, and the third had injected battery acid into the area. In a 5-year retrospective cohort study at a Veterans Administration hospital in Connecticut, Gunderson and Chang3 reviewed the cases of 183 patients with cellulitis and found ultrasound testing commonly performed (73% of cases) to assess for DVT. Only 1 patient (<1%) was diagnosed with new DVT in the ipsilateral leg, and acute DVT was diagnosed in the contralateral leg of 2 other patients. Overall, these studies indicate the incidence of concurrent DVT in cellulitis is low, regardless of the frequency of ultrasound testing.
Although the cost of a single ultrasound test is not prohibitive, annual total costs hospital-wide and nationally are large. In the United States, the charge for a unilateral duplex ultrasound of the extremity ranges from $260 to $1300, and there is an additional charge for interpretation by a radiologist.5 In a retrospective study spanning 3.5 years and involving 2 community hospitals in Michigan, an estimated $290,000 was spent on ultrasound tests defined as unnecessary for patients with cellulitis.6 A limitation of the study was defining a test as unnecessary based on its result being negative.
DOES WELLS SCORE WITH D-DIMER HELP DEFINE A LOW-RISK POPULATION?
The Wells clinical prediction rule is commonly used to assess the pretest probability of DVT in patients presenting with unilateral leg symptoms. The Wells score is often combined with
WHEN MIGHT ULTRASOUND BE HELPFUL IN CELLULITIS?
Investigators have described possible DVT risk factors in patients with cellulitis, but definitive associations are lacking because of the insufficient number of patients studied.8,9 The most consistently identified DVT risk factor is history of previous thromboembolism. In a retrospective analysis of patients with cellulitis, Afzal et al.6 found that, of the 66.8% who underwent ultrasound testing, 5.5% were identified as having concurrent DVT. The authors performed univariate analyses of 15 potential risk factors, including active malignancy, oral contraceptive pill use, recent hospitalization, and surgery. A higher incidence of DVT was found for patients with history of VTE (odds ratio [OR], 5.7; 95% confidence interval [CI], 2.3-13.7), calf swelling (OR, 4.5; 95% CI, 1.3-15.8), CVA (OR, 3.5; 95% CI, 1.2-10.1), or hypertension (OR, 3.5; 95% CI, 0.98-12.2). Given the wide confidence intervals, paucity of studies, and lack of definitive data in the setting of cellulitis, clinicians may want to consider the risk factors established in larger trials in other settings, including known immobility (OR, <2); thrombophilia, CHF, and CVA with hemiparesis (OR, 2-9); and trauma and recent surgery (OR, >10).10
WHAT YOU SHOULD DO INSTEAD
As the incidence of concurrent VTE in patients with cellulitis is low, the essential step is to make a clear diagnosis of cellulitis based on its established signs and symptoms. A 2-center trial of 145 patients found that cellulitis was diagnosed accurately by general medicine and emergency medicine physicians 72% of the time, with evaluation by dermatologists and infectious disease specialists used as the gold standard. Only 5% of the misdiagnosed patients were diagnosed with DVT; stasis dermatitis was the most common alternative diagnosis. Taking a thorough history may elicit risk factors consistent with cellulitis, such as a recent injury with a break in the skin. On examination, cellulitis should be suspected for patients with fever and localized pain, redness, swelling, and warmth—the cardinal signs of dolor, rubor, tumor, and calor. An injury or entry site and leukocytosis also support the diagnosis of cellulitis. Distinct margins of erythema on the skin are highly suspicious for erysipelas.11 Other physical findings (eg, laceration, purulent drainage, lymphangitic spread, fluctuating mass) also are consistent with a diagnosis of cellulitis.
The patient’s history is also essential in determining whether any DVT risk factors are present. Past medical history of VTE or CVA, or recent history of surgery, immobility, or trauma, should alert the clinician to the possibility of DVT. Family history of VTE increases the likelihood of DVT. Acute shortness of breath or chest pain in the setting of concerning lower extremity findings for DVT should raise concern for DVT and concurrent PE.
If the classic features of cellulitis are present, empiric antibiotics should be initiated. Routine ultrasound testing for all patients with cellulitis is of low value. However, as the incidence of DVT in this population is not negligible, those with VTE risk factors should be targeted for testing. Studies in the setting of cellulitis provide little guidance regarding specific risk factors that can be used to determine who should undergo further testing. Given this limitation, we suggest that clinicians incorporate into their decision making the well-established VTE risk factors identified for large populations studied in other settings, such as the postoperative period. Specifically, clinicians should consider ultrasound testing for patients with cellulitis and prior history of VTE; immobility; thrombophilia, CHF, and CVA with hemiparesis; or trauma and recent surgery.10-12 Ultrasound should also be considered for patients with cellulitis that does not improve and for patients whose localized symptoms worsen despite use of antibiotics.
RECOMMENDATIONS
Do not routinely perform ultrasound to rule out concurrent DVT in cases of cellulitis.
Consider compression ultrasound if there is a history of VTE; immobility; thrombophilia, CHF, and CVA with hemiparesis; or trauma and recent surgery. Also consider it for patients who do not respond to antibiotics.
- In cases of cellulitis, avoid use of the Wells score alone or with
D -dimer testing, as it likely overestimates the DVT risk.
CONCLUSION
The current evidence shows that, for most patients with cellulitis, routine ultrasound testing for DVT is unnecessary. Ultrasound should be considered for patients with potent VTE risk factors. If symptoms do not improve, or if they worsen despite use of antibiotics, clinicians should be alert to potential anchoring bias and consider DVT. The Wells clinical prediction rule overestimates the incidence of DVT in cellulitis and has little value in this setting.
Disclosure
Nothing to report.
Do you think this is a low-value practice? Is this truly a “Thing We Do for No Reason”? Let us know what you do in your practice and propose ideas for other “Things We Do for No Reason” topics. Please join in the conversation online at Twitter (#TWDFNR)/Facebook and don’t forget to “Like It” on Facebook or retweet it on Twitter. We invite you to propose ideas for other “Things We Do for No Reason” topics by emailing [email protected].
The “Things We Do for No Reason” series reviews practices which have become common parts of hospital care but which may provide little value to our patients. Practices reviewed in the TWDFNR series do not represent “black and white” conclusions or clinical practice standards, but are meant as a starting place for research and active discussions among hospitalists and patients. We invite you to be part of that discussion. https://www.choosingwisely.org/
Because of overlapping clinical manifestations, clinicians often order ultrasound to rule out deep vein thrombosis (DVT) in cases of cellulitis. Ultrasound testing is performed for 16% to 73% of patients diagnosed with cellulitis. Although testing is common, the pooled incidence of DVT is low (3.1%). Few data elucidate which patients with cellulitis are more likely to have concurrent DVT and require further testing. The Wells clinical prediction rule with
CASE REPORT
A 50-year-old man presented to the emergency department with a 3-day-old cut on his anterior right shin. Associated redness, warmth, pain, and swelling had progressed. The patient had no history of prior DVT or pulmonary embolism (PE). His temperature was 38.5°C, and his white blood cell count of 18,000. On review of systems, he denied shortness of breath and chest pain. He was diagnosed with cellulitis and administered intravenous fluids and cefazolin. The clinician wondered whether to perform lower extremity ultrasound to rule out concurrent DVT.
WHY YOU MIGHT THINK ULTRASOUND IS HELPFUL IN RULING OUT DVT IN CELLULITIS
Lower extremity cellulitis, a common infection of the skin and subcutaneous tissues, is characterized by unilateral erythema, pain, warmth, and swelling. The infection usually follows a skin breach that allows bacteria to enter. DVT may present similarly, and symptoms can include mild leukocytosis and elevated temperature. Because of the clinical similarities, clinicians often order compression ultrasound of the extremity to rule out concurrent DVT in cellulitis. Further impetus for testing stems from fear of the potential complications of untreated DVT, including post-thrombotic syndrome, chronic venous insufficiency, and venous ulceration. A subsequent PE can be fatal, or can cause significant morbidity, including chronic VTE with associated pulmonary hypertension. An estimated quarter of all PEs present as sudden death.1
WHY ULTRASOUND IS NOT HELPFUL IN THIS SETTING
Studies have shown that ultrasound is ordered for 16% to 73% of patients with a cellulitis diagnosis.2,3 Although testing is commonly performed, a meta-analysis of 9 studies of cellulitis patients who underwent ultrasound testing for concurrent DVT revealed a low pooled incidence of total DVT (3.1%) and proximal DVT (2.1%).4 Maze et al.2 retrospectively reviewed 1515 cellulitis cases (identified by International Classification of Diseases, Ninth Revision codes) at a single center in New Zealand over 3 years. Of the 1515 patients, 240 (16%) had ultrasound performed, and only 3 (1.3%) were found to have DVT. Two of the 3 had active malignancy, and the third had injected battery acid into the area. In a 5-year retrospective cohort study at a Veterans Administration hospital in Connecticut, Gunderson and Chang3 reviewed the cases of 183 patients with cellulitis and found ultrasound testing commonly performed (73% of cases) to assess for DVT. Only 1 patient (<1%) was diagnosed with new DVT in the ipsilateral leg, and acute DVT was diagnosed in the contralateral leg of 2 other patients. Overall, these studies indicate the incidence of concurrent DVT in cellulitis is low, regardless of the frequency of ultrasound testing.
Although the cost of a single ultrasound test is not prohibitive, annual total costs hospital-wide and nationally are large. In the United States, the charge for a unilateral duplex ultrasound of the extremity ranges from $260 to $1300, and there is an additional charge for interpretation by a radiologist.5 In a retrospective study spanning 3.5 years and involving 2 community hospitals in Michigan, an estimated $290,000 was spent on ultrasound tests defined as unnecessary for patients with cellulitis.6 A limitation of the study was defining a test as unnecessary based on its result being negative.
DOES WELLS SCORE WITH D-DIMER HELP DEFINE A LOW-RISK POPULATION?
The Wells clinical prediction rule is commonly used to assess the pretest probability of DVT in patients presenting with unilateral leg symptoms. The Wells score is often combined with
WHEN MIGHT ULTRASOUND BE HELPFUL IN CELLULITIS?
Investigators have described possible DVT risk factors in patients with cellulitis, but definitive associations are lacking because of the insufficient number of patients studied.8,9 The most consistently identified DVT risk factor is history of previous thromboembolism. In a retrospective analysis of patients with cellulitis, Afzal et al.6 found that, of the 66.8% who underwent ultrasound testing, 5.5% were identified as having concurrent DVT. The authors performed univariate analyses of 15 potential risk factors, including active malignancy, oral contraceptive pill use, recent hospitalization, and surgery. A higher incidence of DVT was found for patients with history of VTE (odds ratio [OR], 5.7; 95% confidence interval [CI], 2.3-13.7), calf swelling (OR, 4.5; 95% CI, 1.3-15.8), CVA (OR, 3.5; 95% CI, 1.2-10.1), or hypertension (OR, 3.5; 95% CI, 0.98-12.2). Given the wide confidence intervals, paucity of studies, and lack of definitive data in the setting of cellulitis, clinicians may want to consider the risk factors established in larger trials in other settings, including known immobility (OR, <2); thrombophilia, CHF, and CVA with hemiparesis (OR, 2-9); and trauma and recent surgery (OR, >10).10
WHAT YOU SHOULD DO INSTEAD
As the incidence of concurrent VTE in patients with cellulitis is low, the essential step is to make a clear diagnosis of cellulitis based on its established signs and symptoms. A 2-center trial of 145 patients found that cellulitis was diagnosed accurately by general medicine and emergency medicine physicians 72% of the time, with evaluation by dermatologists and infectious disease specialists used as the gold standard. Only 5% of the misdiagnosed patients were diagnosed with DVT; stasis dermatitis was the most common alternative diagnosis. Taking a thorough history may elicit risk factors consistent with cellulitis, such as a recent injury with a break in the skin. On examination, cellulitis should be suspected for patients with fever and localized pain, redness, swelling, and warmth—the cardinal signs of dolor, rubor, tumor, and calor. An injury or entry site and leukocytosis also support the diagnosis of cellulitis. Distinct margins of erythema on the skin are highly suspicious for erysipelas.11 Other physical findings (eg, laceration, purulent drainage, lymphangitic spread, fluctuating mass) also are consistent with a diagnosis of cellulitis.
The patient’s history is also essential in determining whether any DVT risk factors are present. Past medical history of VTE or CVA, or recent history of surgery, immobility, or trauma, should alert the clinician to the possibility of DVT. Family history of VTE increases the likelihood of DVT. Acute shortness of breath or chest pain in the setting of concerning lower extremity findings for DVT should raise concern for DVT and concurrent PE.
If the classic features of cellulitis are present, empiric antibiotics should be initiated. Routine ultrasound testing for all patients with cellulitis is of low value. However, as the incidence of DVT in this population is not negligible, those with VTE risk factors should be targeted for testing. Studies in the setting of cellulitis provide little guidance regarding specific risk factors that can be used to determine who should undergo further testing. Given this limitation, we suggest that clinicians incorporate into their decision making the well-established VTE risk factors identified for large populations studied in other settings, such as the postoperative period. Specifically, clinicians should consider ultrasound testing for patients with cellulitis and prior history of VTE; immobility; thrombophilia, CHF, and CVA with hemiparesis; or trauma and recent surgery.10-12 Ultrasound should also be considered for patients with cellulitis that does not improve and for patients whose localized symptoms worsen despite use of antibiotics.
RECOMMENDATIONS
Do not routinely perform ultrasound to rule out concurrent DVT in cases of cellulitis.
Consider compression ultrasound if there is a history of VTE; immobility; thrombophilia, CHF, and CVA with hemiparesis; or trauma and recent surgery. Also consider it for patients who do not respond to antibiotics.
- In cases of cellulitis, avoid use of the Wells score alone or with
D -dimer testing, as it likely overestimates the DVT risk.
CONCLUSION
The current evidence shows that, for most patients with cellulitis, routine ultrasound testing for DVT is unnecessary. Ultrasound should be considered for patients with potent VTE risk factors. If symptoms do not improve, or if they worsen despite use of antibiotics, clinicians should be alert to potential anchoring bias and consider DVT. The Wells clinical prediction rule overestimates the incidence of DVT in cellulitis and has little value in this setting.
Disclosure
Nothing to report.
Do you think this is a low-value practice? Is this truly a “Thing We Do for No Reason”? Let us know what you do in your practice and propose ideas for other “Things We Do for No Reason” topics. Please join in the conversation online at Twitter (#TWDFNR)/Facebook and don’t forget to “Like It” on Facebook or retweet it on Twitter. We invite you to propose ideas for other “Things We Do for No Reason” topics by emailing [email protected].
1. Heit JA. The epidemiology of venous thromboembolism in the community: implications for prevention and management. J Thromb Thrombolysis. 2006;21(1):23-29. PubMed
2. Maze MJ, Pithie A, Dawes T, Chambers ST. An audit of venous duplex ultrasonography in patients with lower limb cellulitis. N Z Med J. 2011;124(1329):53-56. PubMed
3. Gunderson CG, Chang JJ. Overuse of compression ultrasound for patients with lower extremity cellulitis. Thromb Res. 2014;134(4):846-850. PubMed
4. Gunderson CG, Chang JJ. Risk of deep vein thrombosis in patients with cellulitis and erysipelas: a systematic review and meta-analysis. Thromb Res. 2013;132(3):336-340. PubMed
5. Extremity ultrasound (nonvascular) cost and procedure information. http://www.newchoicehealth.com/procedures/extremity-ultrasound-nonvascular. Accessed February 15, 2016.
6. Afzal MZ, Saleh MM, Razvi S, Hashmi H, Lampen R. Utility of lower extremity Doppler in patients with lower extremity cellulitis: a need to change the practice? South Med J. 2015;108(7):439-444. PubMed
7. Goodacre S, Sutton AJ, Sampson FC. Meta-analysis: the value of clinical assessment in the diagnosis of deep venous thrombosis. Ann Intern Med. 2005;143(2):129-139. PubMed
8. Maze MJ, Skea S, Pithie A, Metcalf S, Pearson JF, Chambers ST. Prevalence of concurrent deep vein thrombosis in patients with lower limb cellulitis: a prospective cohort study. BMC Infect Dis. 2013;13:141. PubMed
9. Bersier D, Bounameaux H. Cellulitis and deep vein thrombosis: a controversial association. J Thromb Haemost. 2003;1(4):867-868. PubMed
10. Anderson FA Jr, Spencer FA. Risk factors for venous thromboembolism. Circulation. 2003;107(23 suppl 1):I9-I16. PubMed
11. Rabuka CE, Azoulay LY, Kahn SR. Predictors of a positive duplex scan in patients with a clinical presentation compatible with deep vein thrombosis or cellulitis. Can J Infect Dis. 2003;14(4):210-214. PubMed
12. Samama MM. An epidemiologic study of risk factors for deep vein thrombosis in medical outpatients: the Sirius Study. Arch Intern Med. 2000;160(22):3415-3420. PubMed
1. Heit JA. The epidemiology of venous thromboembolism in the community: implications for prevention and management. J Thromb Thrombolysis. 2006;21(1):23-29. PubMed
2. Maze MJ, Pithie A, Dawes T, Chambers ST. An audit of venous duplex ultrasonography in patients with lower limb cellulitis. N Z Med J. 2011;124(1329):53-56. PubMed
3. Gunderson CG, Chang JJ. Overuse of compression ultrasound for patients with lower extremity cellulitis. Thromb Res. 2014;134(4):846-850. PubMed
4. Gunderson CG, Chang JJ. Risk of deep vein thrombosis in patients with cellulitis and erysipelas: a systematic review and meta-analysis. Thromb Res. 2013;132(3):336-340. PubMed
5. Extremity ultrasound (nonvascular) cost and procedure information. http://www.newchoicehealth.com/procedures/extremity-ultrasound-nonvascular. Accessed February 15, 2016.
6. Afzal MZ, Saleh MM, Razvi S, Hashmi H, Lampen R. Utility of lower extremity Doppler in patients with lower extremity cellulitis: a need to change the practice? South Med J. 2015;108(7):439-444. PubMed
7. Goodacre S, Sutton AJ, Sampson FC. Meta-analysis: the value of clinical assessment in the diagnosis of deep venous thrombosis. Ann Intern Med. 2005;143(2):129-139. PubMed
8. Maze MJ, Skea S, Pithie A, Metcalf S, Pearson JF, Chambers ST. Prevalence of concurrent deep vein thrombosis in patients with lower limb cellulitis: a prospective cohort study. BMC Infect Dis. 2013;13:141. PubMed
9. Bersier D, Bounameaux H. Cellulitis and deep vein thrombosis: a controversial association. J Thromb Haemost. 2003;1(4):867-868. PubMed
10. Anderson FA Jr, Spencer FA. Risk factors for venous thromboembolism. Circulation. 2003;107(23 suppl 1):I9-I16. PubMed
11. Rabuka CE, Azoulay LY, Kahn SR. Predictors of a positive duplex scan in patients with a clinical presentation compatible with deep vein thrombosis or cellulitis. Can J Infect Dis. 2003;14(4):210-214. PubMed
12. Samama MM. An epidemiologic study of risk factors for deep vein thrombosis in medical outpatients: the Sirius Study. Arch Intern Med. 2000;160(22):3415-3420. PubMed
© 2017 Society of Hospital Medicine
What are the chances?
The approach to clinical conundrums by an expert clinician is revealed through the presentation of an actual patient’s case in an approach typical of a morning report. Similarly to patient care, sequential pieces of information are provided to the clinician, who is unfamiliar with the case. The focus is on the thought processes of both the clinical team caring for the patient and the discussant. The bolded text represents the patient’s case. Each paragraph that follows represents the discussant’s thoughts.
Two weeks after undergoing a below-knee amputation (BKA) and 10 days after being discharged to a skilled nursing facility (SNF), an 87-year-old man returned to the emergency department (ED) for evaluation of somnolence and altered mental state. In the ED, he was disoriented and unable to provide a detailed history.
The differential diagnosis for acute confusion and altered consciousness is broad. Initial possibilities include toxic-metabolic abnormalities, medication side effects, and infections. Urinary tract infection, pneumonia, and surgical-site infection should be assessed for first, as they are common causes of postoperative altered mentation. Next to be considered are subclinical seizure, ischemic stroke, and infectious encephalitis or meningitis, along with hemorrhagic stroke and subdural hematoma.
During initial assessment, the clinician should ascertain baseline mental state, the timeline of the change in mental status, recent medication changes, history of substance abuse, and concern about any recent trauma, such as a fall. Performing the physical examination, the clinician should assess vital signs and then focus on identifying localizing neurologic deficits.
First steps in the work-up include a complete metabolic panel, complete blood cell count, urinalysis with culture, and a urine toxicology screen. If the patient has a “toxic” appearance, blood cultures should be obtained. An electrocardiogram should be used to screen for drug toxicity or evidence of cardiac ischemia. If laboratory test results do not reveal an obvious infectious or metabolic cause, a noncontrast computed tomography (CT) of the head should be obtained. In terms of early interventions, a low glucose level should be treated with thiamine and then glucose, and naloxone should be given if there is any suspicion of narcotic overdose.
More history was obtained from the patient’s records. The BKA was performed to address a nonhealing transmetatarsal amputation. Two months earlier, the transmetatarsal amputation had been performed as treatment for a diabetic forefoot ulcer with chronic osteomyelitis. The patient’s post-BKA course was uncomplicated. He was started on intravenous (IV) ertapenem on postoperative day 1, and on postoperative day 4 was discharged to the SNF to complete a 6-week course of antibiotics for osteomyelitis. Past medical history included paroxysmal atrial fibrillation, coronary artery disease, congestive heart failure (ejection fraction 40%), and type 2 diabetes mellitus. Medications given at the SNF were oxycodone, acetaminophen, cholecalciferol, melatonin, digoxin, ondansetron, furosemide, gabapentin, correctional insulin, tamsulosin, senna, docusate, warfarin, and metoprolol. While there, the patient’s family expressed concern about his diminishing “mental ability.” They reported he had been fully alert and oriented on arrival at the SNF, and living independently with his wife before the BKA. Then, a week before the ED presentation, he started becoming more somnolent and forgetful. The gabapentin and oxycodone dosages were reduced to minimize their sedative effects, but he showed no improvement. At the SNF, a somnolence work-up was not performed.
Several of the patient’s medications can contribute to altered mental state. Ertapenem can cause seizures as well as profound mental status changes, though these are more likely in the setting of poor renal function. The mental status changes were noticed about a week into the patient’s course of antibiotics, which suggests a possible temporal correlation with the initiation of ertapenem. An electroencephalogram is required to diagnose nonconvulsive seizure activity. Narcotic overdose should still be considered, despite the recent reduction in oxycodone dosage. Digoxin toxicity, though less likely when the dose is stable and there are no changes in renal function, can cause a confused state. Concurrent use of furosemide could potentiate the toxic effects of digoxin.
Non-medication-related concerns include hypoglycemia, hyperglycemia, and, given his history of atrial fibrillation, cardioembolic stroke. Although generalized confusion is not a common manifestation of stroke, a thalamic stroke can alter mental state but be easily missed if not specifically considered. Additional lab work-up should include a digoxin level and, since he is taking warfarin, a prothrombin time/international normalized ratio (PT/INR). If the initial laboratory studies and head CT do not explain the altered mental state, magnetic resonance imaging (MRI) of the brain should be performed to further assess for stroke.
On physical examination in the ED, the patient was resting comfortably with eyes closed, and arousing to voice. He obeyed commands and participated in the examination. His Glasgow Coma Scale score was 13; temperature, 36.8°C, heart rate, 80 beats per minute; respiratory rate, 16 breaths per minute; blood pressure, 90/57 mm Hg; and 100% peripheral capillary oxygen saturation while breathing ambient air. He appeared well developed. His heart rhythm was irregularly irregular, without murmurs, rubs, or gallops. Respiratory and abdominal examination findings were normal. The left BKA incision was well approximated, with no drainage, dehiscence, fluctuance, or erythema. On neurologic examination, the patient was intermittently oriented only to self. Pupils were equal, round, and reactive to light; extraocular movements were intact; face was symmetric; tongue was midline; sensation on face was equal bilaterally; and shoulder shrug was intact. Strength was 5/5 and symmetric in the elbow and hip and 5/5 in the right knee and ankle (not tested on left because of BKA). Deep tendon reflexes were 3+ and symmetrical at the biceps, brachioradialis, and triceps tendons and 3+ in the right patellar and Achilles tendons. Sensation was intact and symmetrical in the upper and lower extremities. The patient’s speech was slow and slurred, and his answers were unrelated to the questions being asked.
The patient’s mental state is best described as lethargic. As he is only intermittently oriented, he meets the criteria for delirium. He is not obtunded or comatose, and his pupils are at least reactive, not pinpoint, so narcotic overdose is less likely. Thalamic stroke remains in the differential diagnosis; despite the seemingly symmetrical sensation examination, hemisensory deficits cannot be definitively ruled out given the patient’s mental state. A rare entity such as carcinomatosis meningitis or another diffuse, infiltrative neoplastic process could be causing his condition. However, because focal deficits other than abnormal speech and diffuse hyperreflexia are absent, toxic, infectious, or metabolic causes are more likely than structural abnormalities. Still possible is a medication toxicity, such as ertapenem toxicity or, less likely, digoxin toxicity. In terms of infectious possibilities, urinary tract infection could certainly present in this fashion, especially if the patient had a somewhat low neurologic reserve at baseline, and hypotension could be secondary to sepsis. Encephalitis or meningitis remains in the differential diagnosis, though the patient appears nontoxic, and therefore a bacterial etiology is very unlikely.
The patient’s hyperreflexia may be an important clue. Although the strength of his reflexes at baseline is unknown, seizures can cause transiently increased reflexes as well as a confused, lethargic mental state. Reflexes can also be increased by a drug overdose that has caused serotonin syndrome. Of the patient’s medications, only ondansetron can cause this reaction. Hyperthyroidism can cause brisk reflexes and confusion, though more typically it causes agitated confusion. A thyroid-stimulating hormone level should be added to the initial laboratory panel.
A complete blood count revealed white blood cell count 11.86 K/uL with neutrophilic predominance and immature granulocytes, hemoglobin 11.5 g/dL, and platelet count 323 K/uL. Serum sodium was 141 mEq/L, potassium 4.2 mEq/L, chloride 103 mEq/L, bicarbonate 30 mEq/L, creatinine 1.14 mg/dL (prior baseline of 0.8-1.0 mg/dL), blood urea nitrogen 26 mg/dL, blood glucose 159 mg/dL, and calcium 9.1 mg/dL. His digoxin level was 1.3 ng/mL (reference range 0.5-1.9 mg/mL) and troponin was undetectable. INR was 2.7 and partial thromboplastin time (PTT) 60 seconds. Vitamin B12 level was 674 pg/mL (reference range >180). A urinalysis had 1+ hyaline casts and was negative for nitrites, leukocyte esterase, blood, and bacteria. An ECG revealed atrial fibrillation with a ventricular rate of 80 beats per minute. A chest radiograph showed clear lung fields. A CT of the head without IV contrast had no evidence of an acute intracranial abnormality. In the ED, 1 liter of IV normal saline was given and blood pressure improved to 127/72 mm Hg.
The head CT does not show intracranial bleeding, and, though it is reassuring that INR is in the therapeutic range, ischemic stroke must remain in the differential diagnosis. Sepsis is less likely given that the criteria for systemic inflammatory response syndrome are not met, and hypotension was rapidly corrected with administration of IV fluids. Urinary tract infection was ruled out with the negative urinalysis. Subclinical seizures remain possible, as does medication-related or other toxicity. A medication overdose, intentional or otherwise, should also be considered.
The patient was admitted to the hospital. On reassessment by the inpatient team, he was oriented only to self, frequently falling asleep, and not recalling earlier conversations when aroused. His speech remained slurred and difficult to understand. Neurologic examination findings were unchanged since the ED examination. On additional cerebellar examination, he had dysmetria with finger-to-nose testing bilaterally and dysdiadochokinesia (impaired rapid alternating movements) of the left hand.
His handedness is not mentioned; the dysdiadochokinesia of the left hand may reflect the patient’s being right-handed, or may signify a focal cerebellar lesion. The cerebellum is also implicated by the bilateral dysmetria. Persistent somnolence in the absence of CT findings suggests a metabolic or infectious process. Metabolic processes that can cause bilateral cerebellar ataxia and somnolence include overdose of a drug or medication. Use of alcohol or a medication such as phenytoin, valproic acid, or a benzodiazepine can cause the symptoms in this case, but was not reported by the family, and there was no documentation of it in the SNF records. Wernicke encephalopathy is rare and is not well supported by the patient’s presentation but should be considered, as it can be easily treated with thiamine. Meningoencephalitis affecting the cerebellum remains possible, but infection is less likely. Both electroencephalogram and brain MRI should be performed, with a specific interest in possible cerebellar lesions. If the MRI is unremarkable, a lumbar puncture should be performed to assess opening pressure and investigate for infectious etiologies.
MRI of the brain showed age-related volume loss and nonspecific white matter disease without acute changes. Lack of a clear explanation for the neurologic findings led to suspicion of a medication side effect. Ertapenem was stopped on admission because it has been reported to rarely cause altered mental status. IV moxifloxacin was started for the osteomyelitis. Over the next 2 days, symptoms began resolving; within 24 hours of ertapenem discontinuation, the patient was awake, alert, and talkative. On examination, he remained dysarthric but was no longer dysmetric. Within 48 hours, the dysarthria was completely resolved, and he was returned to the SNF to complete a course of IV moxifloxacin.
DISCUSSION
Among elderly patients presenting to the ED, altered mental status is a common complaint, accounting for 10% to 30% of visits.1 Medications are a common cause of altered mental status among the elderly and are responsible for 40% of delirium cases.1 The risk of adverse drug events (ADEs) rises with the number of medications prescribed.1-3 Among patients older than 60 years, the incidence of polypharmacy (defined as taking >5 prescription medications) increased from roughly 20% in 1999 to 40% in 2012.4,5 The most common ADEs in the ambulatory setting (25%) are central nervous system (CNS) symptoms, including dizziness, sleep disturbances, and mood changes.6 A medication effect should be suspected in any elderly patient presenting with altered mental state.
The present patient developed a constellation of neurologic symptoms after starting ertapenem, one of the carbapenem antibiotics, which is a class of medications that can cause CNS ADEs. Carbapenems are renally cleared, and adjustments must be made for acute or chronic changes in kidney function. Carbapenems are associated with increased risk of seizure; the incidence of seizure with ertapenem is 0.2%.7,8 Food and Drug Administration postmarketing reports have noted ertapenem can cause somnolence and dyskinesia,9 and several case reports have described ertapenem-associated CNS side effects, including psychosis and encephalopathy.10-13 Symptoms and examination findings can include confusion, disorientation, garbled speech, dysphagia, hallucinations, miosis, myoclonus, tremor, and agitation.10-13 Although reports of dysmetria and dysdiadochokinesia are lacking, suspicion of an ADE in this case was heightened by the timing of the exposure and the absence of alternative infectious, metabolic, and vascular explanations for bilateral cerebellar dysfunction.
The Naranjo Adverse Drug Reaction (ADR) scale may help clinicians differentiate ADEs from other etiologies of symptoms. It uses 10 weighted questions (Table) to estimate the probability that an adverse clinical event is caused by a drug reaction.14 The present case was assigned 1 point for prior reports of neurologic ADEs associated with ertapenem, 2 for the temporal association, 1 for resolution after medication withdrawal, 2 for lack of alternative causes, and 1 for objective evidence of neurologic dysfunction—for a total of 7 points, indicating ertapenem was probably the cause of the patient’s neurologic symptoms. Of 4 prior cases in which carbapenem toxicity was suspected and the Naranjo scale was used, 3 found a probable relationship, and the fourth a highly probable one.10,12 Confusion, disorientation, hallucinations, tangential thoughts, and garbled speech were reported in the 3 probable cases of ADEs. In the highly probable case, tangential thoughts, garbled speech, and miosis were noted on examination, and these findings returned after re-exposure to ertapenem. Of note, these ADEs occurred in patients with normal and abnormal renal function, and in middle-aged and elderly patients.10,11,13
Most medications have a long list of low-frequency and rarely reported adverse effects. The present case reminds clinicians to consider rare adverse effects, or variants of previously reported adverse effects, in a patient with unexplained symptoms. To estimate the probability that a drug is causing harm to a patient, using a validated tool such as the Naranjo scale helps answer the question, What are the chances?
KEY TEACHING POINTS
Clinicians should include rare adverse effects of common medications in the differential diagnosis.
The Naranjo score is a validated tool that can be used to systematically assess the probability of an adverse drug effect at the bedside.
- The presentation of ertapenem-associated neurotoxicity may include features of bilateral cerebellar dysfunction.
Disclosure
Nothing to report.
1. Inouye SK, Fearing MA, Marcantonio ER. Delirium. In: Halter JB, Ouslander JG, Tinetti ME, Studenski S, High KP, Asthana S, eds. Hazzard’s Geriatric Medicine and Gerontology. 6th ed. New York, NY: McGraw-Hill; 2009.
2. Sarkar U, López A, Maselli JH, Gonzales R. Adverse drug events in U.S. adult ambulatory medical care. Health Serv Res. 2011;46(5):1517-1533. PubMed
3. Chrischilles E, Rubenstein L, Van Gilder R, Voelker M, Wright K, Wallace R. Risk factors for adverse drug events in older adults with mobility limitations in the community setting. J Am Geriatr Soc. 2007;55(1):29-34. PubMed
4. Kaufman DW, Kelly JP, Rosenberg L, Anderson TE, Mitchell AA. Recent patterns of medication use in the ambulatory adult population of the United States: the Slone survey. JAMA. 2002;287(3):337-344. PubMed
5. Kantor ED, Rehm CD, Haas JS, Chan AT, Giovannucci EL. Trends in prescription drug use among adults in the United States from 1999-2012. JAMA. 2015;314(17):1818-1831. PubMed
6. Thomsen LA, Winterstein AG, Søndergaard B, Haugbølle LS, Melander A. Systematic review of the incidence and characteristics of preventable adverse drug events in ambulatory care. Ann Pharmacother. 2007;41(9):1411-1426. PubMed
7. Zhanel GG, Wiebe R, Dilay L, et al. Comparative review of the carbapenems. Drugs. 2007;67(7):1027-1052. PubMed
8. Cannon JP, Lee TA, Clark NM, Setlak P, Grim SA. The risk of seizures among the carbapenems: a meta-analysis. J Antimicrob Chemother. 2014;69(8):2043-2055. PubMed
9. US Food and Drug Administration. Invanz (ertapenem) injection [safety information]. http://www.fda.gov/Safety/MedWatch/SafetyInformation/ucm196605.htm. Published July 2013. Accessed July 6, 2015.
10. Oo Y, Packham D, Yau W, Munckhof WJ. Ertapenem-associated psychosis and encephalopathy. Intern Med J. 2014;44(8):817-819. PubMed
11. Wen MJ, Sung CC, Chau T, Lin SH. Acute prolonged neurotoxicity associated with recommended doses of ertapenem in 2 patients with advanced renal failure. Clin Nephrol. 2013;80(6):474-478. PubMed
12. Duquaine S, Kitchell E, Tate T, Tannen RC, Wickremasinghe IM. Central nervous system toxicity associated with ertapenem use. Ann Pharmacother. 2011;45(1):e6. PubMed
13. Kong V, Beckert L, Awunor-Renner C. A case of beta lactam-induced visual hallucination. N Z Med J. 2009;122(1298):76-77. PubMed
14. Naranjo CA, Busto U, Sellers EM, et al. A method for estimating the probability of adverse drug reactions. Clin Pharmacol Ther. 1981;30(2):239-245. PubMed
The approach to clinical conundrums by an expert clinician is revealed through the presentation of an actual patient’s case in an approach typical of a morning report. Similarly to patient care, sequential pieces of information are provided to the clinician, who is unfamiliar with the case. The focus is on the thought processes of both the clinical team caring for the patient and the discussant. The bolded text represents the patient’s case. Each paragraph that follows represents the discussant’s thoughts.
Two weeks after undergoing a below-knee amputation (BKA) and 10 days after being discharged to a skilled nursing facility (SNF), an 87-year-old man returned to the emergency department (ED) for evaluation of somnolence and altered mental state. In the ED, he was disoriented and unable to provide a detailed history.
The differential diagnosis for acute confusion and altered consciousness is broad. Initial possibilities include toxic-metabolic abnormalities, medication side effects, and infections. Urinary tract infection, pneumonia, and surgical-site infection should be assessed for first, as they are common causes of postoperative altered mentation. Next to be considered are subclinical seizure, ischemic stroke, and infectious encephalitis or meningitis, along with hemorrhagic stroke and subdural hematoma.
During initial assessment, the clinician should ascertain baseline mental state, the timeline of the change in mental status, recent medication changes, history of substance abuse, and concern about any recent trauma, such as a fall. Performing the physical examination, the clinician should assess vital signs and then focus on identifying localizing neurologic deficits.
First steps in the work-up include a complete metabolic panel, complete blood cell count, urinalysis with culture, and a urine toxicology screen. If the patient has a “toxic” appearance, blood cultures should be obtained. An electrocardiogram should be used to screen for drug toxicity or evidence of cardiac ischemia. If laboratory test results do not reveal an obvious infectious or metabolic cause, a noncontrast computed tomography (CT) of the head should be obtained. In terms of early interventions, a low glucose level should be treated with thiamine and then glucose, and naloxone should be given if there is any suspicion of narcotic overdose.
More history was obtained from the patient’s records. The BKA was performed to address a nonhealing transmetatarsal amputation. Two months earlier, the transmetatarsal amputation had been performed as treatment for a diabetic forefoot ulcer with chronic osteomyelitis. The patient’s post-BKA course was uncomplicated. He was started on intravenous (IV) ertapenem on postoperative day 1, and on postoperative day 4 was discharged to the SNF to complete a 6-week course of antibiotics for osteomyelitis. Past medical history included paroxysmal atrial fibrillation, coronary artery disease, congestive heart failure (ejection fraction 40%), and type 2 diabetes mellitus. Medications given at the SNF were oxycodone, acetaminophen, cholecalciferol, melatonin, digoxin, ondansetron, furosemide, gabapentin, correctional insulin, tamsulosin, senna, docusate, warfarin, and metoprolol. While there, the patient’s family expressed concern about his diminishing “mental ability.” They reported he had been fully alert and oriented on arrival at the SNF, and living independently with his wife before the BKA. Then, a week before the ED presentation, he started becoming more somnolent and forgetful. The gabapentin and oxycodone dosages were reduced to minimize their sedative effects, but he showed no improvement. At the SNF, a somnolence work-up was not performed.
Several of the patient’s medications can contribute to altered mental state. Ertapenem can cause seizures as well as profound mental status changes, though these are more likely in the setting of poor renal function. The mental status changes were noticed about a week into the patient’s course of antibiotics, which suggests a possible temporal correlation with the initiation of ertapenem. An electroencephalogram is required to diagnose nonconvulsive seizure activity. Narcotic overdose should still be considered, despite the recent reduction in oxycodone dosage. Digoxin toxicity, though less likely when the dose is stable and there are no changes in renal function, can cause a confused state. Concurrent use of furosemide could potentiate the toxic effects of digoxin.
Non-medication-related concerns include hypoglycemia, hyperglycemia, and, given his history of atrial fibrillation, cardioembolic stroke. Although generalized confusion is not a common manifestation of stroke, a thalamic stroke can alter mental state but be easily missed if not specifically considered. Additional lab work-up should include a digoxin level and, since he is taking warfarin, a prothrombin time/international normalized ratio (PT/INR). If the initial laboratory studies and head CT do not explain the altered mental state, magnetic resonance imaging (MRI) of the brain should be performed to further assess for stroke.
On physical examination in the ED, the patient was resting comfortably with eyes closed, and arousing to voice. He obeyed commands and participated in the examination. His Glasgow Coma Scale score was 13; temperature, 36.8°C, heart rate, 80 beats per minute; respiratory rate, 16 breaths per minute; blood pressure, 90/57 mm Hg; and 100% peripheral capillary oxygen saturation while breathing ambient air. He appeared well developed. His heart rhythm was irregularly irregular, without murmurs, rubs, or gallops. Respiratory and abdominal examination findings were normal. The left BKA incision was well approximated, with no drainage, dehiscence, fluctuance, or erythema. On neurologic examination, the patient was intermittently oriented only to self. Pupils were equal, round, and reactive to light; extraocular movements were intact; face was symmetric; tongue was midline; sensation on face was equal bilaterally; and shoulder shrug was intact. Strength was 5/5 and symmetric in the elbow and hip and 5/5 in the right knee and ankle (not tested on left because of BKA). Deep tendon reflexes were 3+ and symmetrical at the biceps, brachioradialis, and triceps tendons and 3+ in the right patellar and Achilles tendons. Sensation was intact and symmetrical in the upper and lower extremities. The patient’s speech was slow and slurred, and his answers were unrelated to the questions being asked.
The patient’s mental state is best described as lethargic. As he is only intermittently oriented, he meets the criteria for delirium. He is not obtunded or comatose, and his pupils are at least reactive, not pinpoint, so narcotic overdose is less likely. Thalamic stroke remains in the differential diagnosis; despite the seemingly symmetrical sensation examination, hemisensory deficits cannot be definitively ruled out given the patient’s mental state. A rare entity such as carcinomatosis meningitis or another diffuse, infiltrative neoplastic process could be causing his condition. However, because focal deficits other than abnormal speech and diffuse hyperreflexia are absent, toxic, infectious, or metabolic causes are more likely than structural abnormalities. Still possible is a medication toxicity, such as ertapenem toxicity or, less likely, digoxin toxicity. In terms of infectious possibilities, urinary tract infection could certainly present in this fashion, especially if the patient had a somewhat low neurologic reserve at baseline, and hypotension could be secondary to sepsis. Encephalitis or meningitis remains in the differential diagnosis, though the patient appears nontoxic, and therefore a bacterial etiology is very unlikely.
The patient’s hyperreflexia may be an important clue. Although the strength of his reflexes at baseline is unknown, seizures can cause transiently increased reflexes as well as a confused, lethargic mental state. Reflexes can also be increased by a drug overdose that has caused serotonin syndrome. Of the patient’s medications, only ondansetron can cause this reaction. Hyperthyroidism can cause brisk reflexes and confusion, though more typically it causes agitated confusion. A thyroid-stimulating hormone level should be added to the initial laboratory panel.
A complete blood count revealed white blood cell count 11.86 K/uL with neutrophilic predominance and immature granulocytes, hemoglobin 11.5 g/dL, and platelet count 323 K/uL. Serum sodium was 141 mEq/L, potassium 4.2 mEq/L, chloride 103 mEq/L, bicarbonate 30 mEq/L, creatinine 1.14 mg/dL (prior baseline of 0.8-1.0 mg/dL), blood urea nitrogen 26 mg/dL, blood glucose 159 mg/dL, and calcium 9.1 mg/dL. His digoxin level was 1.3 ng/mL (reference range 0.5-1.9 mg/mL) and troponin was undetectable. INR was 2.7 and partial thromboplastin time (PTT) 60 seconds. Vitamin B12 level was 674 pg/mL (reference range >180). A urinalysis had 1+ hyaline casts and was negative for nitrites, leukocyte esterase, blood, and bacteria. An ECG revealed atrial fibrillation with a ventricular rate of 80 beats per minute. A chest radiograph showed clear lung fields. A CT of the head without IV contrast had no evidence of an acute intracranial abnormality. In the ED, 1 liter of IV normal saline was given and blood pressure improved to 127/72 mm Hg.
The head CT does not show intracranial bleeding, and, though it is reassuring that INR is in the therapeutic range, ischemic stroke must remain in the differential diagnosis. Sepsis is less likely given that the criteria for systemic inflammatory response syndrome are not met, and hypotension was rapidly corrected with administration of IV fluids. Urinary tract infection was ruled out with the negative urinalysis. Subclinical seizures remain possible, as does medication-related or other toxicity. A medication overdose, intentional or otherwise, should also be considered.
The patient was admitted to the hospital. On reassessment by the inpatient team, he was oriented only to self, frequently falling asleep, and not recalling earlier conversations when aroused. His speech remained slurred and difficult to understand. Neurologic examination findings were unchanged since the ED examination. On additional cerebellar examination, he had dysmetria with finger-to-nose testing bilaterally and dysdiadochokinesia (impaired rapid alternating movements) of the left hand.
His handedness is not mentioned; the dysdiadochokinesia of the left hand may reflect the patient’s being right-handed, or may signify a focal cerebellar lesion. The cerebellum is also implicated by the bilateral dysmetria. Persistent somnolence in the absence of CT findings suggests a metabolic or infectious process. Metabolic processes that can cause bilateral cerebellar ataxia and somnolence include overdose of a drug or medication. Use of alcohol or a medication such as phenytoin, valproic acid, or a benzodiazepine can cause the symptoms in this case, but was not reported by the family, and there was no documentation of it in the SNF records. Wernicke encephalopathy is rare and is not well supported by the patient’s presentation but should be considered, as it can be easily treated with thiamine. Meningoencephalitis affecting the cerebellum remains possible, but infection is less likely. Both electroencephalogram and brain MRI should be performed, with a specific interest in possible cerebellar lesions. If the MRI is unremarkable, a lumbar puncture should be performed to assess opening pressure and investigate for infectious etiologies.
MRI of the brain showed age-related volume loss and nonspecific white matter disease without acute changes. Lack of a clear explanation for the neurologic findings led to suspicion of a medication side effect. Ertapenem was stopped on admission because it has been reported to rarely cause altered mental status. IV moxifloxacin was started for the osteomyelitis. Over the next 2 days, symptoms began resolving; within 24 hours of ertapenem discontinuation, the patient was awake, alert, and talkative. On examination, he remained dysarthric but was no longer dysmetric. Within 48 hours, the dysarthria was completely resolved, and he was returned to the SNF to complete a course of IV moxifloxacin.
DISCUSSION
Among elderly patients presenting to the ED, altered mental status is a common complaint, accounting for 10% to 30% of visits.1 Medications are a common cause of altered mental status among the elderly and are responsible for 40% of delirium cases.1 The risk of adverse drug events (ADEs) rises with the number of medications prescribed.1-3 Among patients older than 60 years, the incidence of polypharmacy (defined as taking >5 prescription medications) increased from roughly 20% in 1999 to 40% in 2012.4,5 The most common ADEs in the ambulatory setting (25%) are central nervous system (CNS) symptoms, including dizziness, sleep disturbances, and mood changes.6 A medication effect should be suspected in any elderly patient presenting with altered mental state.
The present patient developed a constellation of neurologic symptoms after starting ertapenem, one of the carbapenem antibiotics, which is a class of medications that can cause CNS ADEs. Carbapenems are renally cleared, and adjustments must be made for acute or chronic changes in kidney function. Carbapenems are associated with increased risk of seizure; the incidence of seizure with ertapenem is 0.2%.7,8 Food and Drug Administration postmarketing reports have noted ertapenem can cause somnolence and dyskinesia,9 and several case reports have described ertapenem-associated CNS side effects, including psychosis and encephalopathy.10-13 Symptoms and examination findings can include confusion, disorientation, garbled speech, dysphagia, hallucinations, miosis, myoclonus, tremor, and agitation.10-13 Although reports of dysmetria and dysdiadochokinesia are lacking, suspicion of an ADE in this case was heightened by the timing of the exposure and the absence of alternative infectious, metabolic, and vascular explanations for bilateral cerebellar dysfunction.
The Naranjo Adverse Drug Reaction (ADR) scale may help clinicians differentiate ADEs from other etiologies of symptoms. It uses 10 weighted questions (Table) to estimate the probability that an adverse clinical event is caused by a drug reaction.14 The present case was assigned 1 point for prior reports of neurologic ADEs associated with ertapenem, 2 for the temporal association, 1 for resolution after medication withdrawal, 2 for lack of alternative causes, and 1 for objective evidence of neurologic dysfunction—for a total of 7 points, indicating ertapenem was probably the cause of the patient’s neurologic symptoms. Of 4 prior cases in which carbapenem toxicity was suspected and the Naranjo scale was used, 3 found a probable relationship, and the fourth a highly probable one.10,12 Confusion, disorientation, hallucinations, tangential thoughts, and garbled speech were reported in the 3 probable cases of ADEs. In the highly probable case, tangential thoughts, garbled speech, and miosis were noted on examination, and these findings returned after re-exposure to ertapenem. Of note, these ADEs occurred in patients with normal and abnormal renal function, and in middle-aged and elderly patients.10,11,13
Most medications have a long list of low-frequency and rarely reported adverse effects. The present case reminds clinicians to consider rare adverse effects, or variants of previously reported adverse effects, in a patient with unexplained symptoms. To estimate the probability that a drug is causing harm to a patient, using a validated tool such as the Naranjo scale helps answer the question, What are the chances?
KEY TEACHING POINTS
Clinicians should include rare adverse effects of common medications in the differential diagnosis.
The Naranjo score is a validated tool that can be used to systematically assess the probability of an adverse drug effect at the bedside.
- The presentation of ertapenem-associated neurotoxicity may include features of bilateral cerebellar dysfunction.
Disclosure
Nothing to report.
The approach to clinical conundrums by an expert clinician is revealed through the presentation of an actual patient’s case in an approach typical of a morning report. Similarly to patient care, sequential pieces of information are provided to the clinician, who is unfamiliar with the case. The focus is on the thought processes of both the clinical team caring for the patient and the discussant. The bolded text represents the patient’s case. Each paragraph that follows represents the discussant’s thoughts.
Two weeks after undergoing a below-knee amputation (BKA) and 10 days after being discharged to a skilled nursing facility (SNF), an 87-year-old man returned to the emergency department (ED) for evaluation of somnolence and altered mental state. In the ED, he was disoriented and unable to provide a detailed history.
The differential diagnosis for acute confusion and altered consciousness is broad. Initial possibilities include toxic-metabolic abnormalities, medication side effects, and infections. Urinary tract infection, pneumonia, and surgical-site infection should be assessed for first, as they are common causes of postoperative altered mentation. Next to be considered are subclinical seizure, ischemic stroke, and infectious encephalitis or meningitis, along with hemorrhagic stroke and subdural hematoma.
During initial assessment, the clinician should ascertain baseline mental state, the timeline of the change in mental status, recent medication changes, history of substance abuse, and concern about any recent trauma, such as a fall. Performing the physical examination, the clinician should assess vital signs and then focus on identifying localizing neurologic deficits.
First steps in the work-up include a complete metabolic panel, complete blood cell count, urinalysis with culture, and a urine toxicology screen. If the patient has a “toxic” appearance, blood cultures should be obtained. An electrocardiogram should be used to screen for drug toxicity or evidence of cardiac ischemia. If laboratory test results do not reveal an obvious infectious or metabolic cause, a noncontrast computed tomography (CT) of the head should be obtained. In terms of early interventions, a low glucose level should be treated with thiamine and then glucose, and naloxone should be given if there is any suspicion of narcotic overdose.
More history was obtained from the patient’s records. The BKA was performed to address a nonhealing transmetatarsal amputation. Two months earlier, the transmetatarsal amputation had been performed as treatment for a diabetic forefoot ulcer with chronic osteomyelitis. The patient’s post-BKA course was uncomplicated. He was started on intravenous (IV) ertapenem on postoperative day 1, and on postoperative day 4 was discharged to the SNF to complete a 6-week course of antibiotics for osteomyelitis. Past medical history included paroxysmal atrial fibrillation, coronary artery disease, congestive heart failure (ejection fraction 40%), and type 2 diabetes mellitus. Medications given at the SNF were oxycodone, acetaminophen, cholecalciferol, melatonin, digoxin, ondansetron, furosemide, gabapentin, correctional insulin, tamsulosin, senna, docusate, warfarin, and metoprolol. While there, the patient’s family expressed concern about his diminishing “mental ability.” They reported he had been fully alert and oriented on arrival at the SNF, and living independently with his wife before the BKA. Then, a week before the ED presentation, he started becoming more somnolent and forgetful. The gabapentin and oxycodone dosages were reduced to minimize their sedative effects, but he showed no improvement. At the SNF, a somnolence work-up was not performed.
Several of the patient’s medications can contribute to altered mental state. Ertapenem can cause seizures as well as profound mental status changes, though these are more likely in the setting of poor renal function. The mental status changes were noticed about a week into the patient’s course of antibiotics, which suggests a possible temporal correlation with the initiation of ertapenem. An electroencephalogram is required to diagnose nonconvulsive seizure activity. Narcotic overdose should still be considered, despite the recent reduction in oxycodone dosage. Digoxin toxicity, though less likely when the dose is stable and there are no changes in renal function, can cause a confused state. Concurrent use of furosemide could potentiate the toxic effects of digoxin.
Non-medication-related concerns include hypoglycemia, hyperglycemia, and, given his history of atrial fibrillation, cardioembolic stroke. Although generalized confusion is not a common manifestation of stroke, a thalamic stroke can alter mental state but be easily missed if not specifically considered. Additional lab work-up should include a digoxin level and, since he is taking warfarin, a prothrombin time/international normalized ratio (PT/INR). If the initial laboratory studies and head CT do not explain the altered mental state, magnetic resonance imaging (MRI) of the brain should be performed to further assess for stroke.
On physical examination in the ED, the patient was resting comfortably with eyes closed, and arousing to voice. He obeyed commands and participated in the examination. His Glasgow Coma Scale score was 13; temperature, 36.8°C, heart rate, 80 beats per minute; respiratory rate, 16 breaths per minute; blood pressure, 90/57 mm Hg; and 100% peripheral capillary oxygen saturation while breathing ambient air. He appeared well developed. His heart rhythm was irregularly irregular, without murmurs, rubs, or gallops. Respiratory and abdominal examination findings were normal. The left BKA incision was well approximated, with no drainage, dehiscence, fluctuance, or erythema. On neurologic examination, the patient was intermittently oriented only to self. Pupils were equal, round, and reactive to light; extraocular movements were intact; face was symmetric; tongue was midline; sensation on face was equal bilaterally; and shoulder shrug was intact. Strength was 5/5 and symmetric in the elbow and hip and 5/5 in the right knee and ankle (not tested on left because of BKA). Deep tendon reflexes were 3+ and symmetrical at the biceps, brachioradialis, and triceps tendons and 3+ in the right patellar and Achilles tendons. Sensation was intact and symmetrical in the upper and lower extremities. The patient’s speech was slow and slurred, and his answers were unrelated to the questions being asked.
The patient’s mental state is best described as lethargic. As he is only intermittently oriented, he meets the criteria for delirium. He is not obtunded or comatose, and his pupils are at least reactive, not pinpoint, so narcotic overdose is less likely. Thalamic stroke remains in the differential diagnosis; despite the seemingly symmetrical sensation examination, hemisensory deficits cannot be definitively ruled out given the patient’s mental state. A rare entity such as carcinomatosis meningitis or another diffuse, infiltrative neoplastic process could be causing his condition. However, because focal deficits other than abnormal speech and diffuse hyperreflexia are absent, toxic, infectious, or metabolic causes are more likely than structural abnormalities. Still possible is a medication toxicity, such as ertapenem toxicity or, less likely, digoxin toxicity. In terms of infectious possibilities, urinary tract infection could certainly present in this fashion, especially if the patient had a somewhat low neurologic reserve at baseline, and hypotension could be secondary to sepsis. Encephalitis or meningitis remains in the differential diagnosis, though the patient appears nontoxic, and therefore a bacterial etiology is very unlikely.
The patient’s hyperreflexia may be an important clue. Although the strength of his reflexes at baseline is unknown, seizures can cause transiently increased reflexes as well as a confused, lethargic mental state. Reflexes can also be increased by a drug overdose that has caused serotonin syndrome. Of the patient’s medications, only ondansetron can cause this reaction. Hyperthyroidism can cause brisk reflexes and confusion, though more typically it causes agitated confusion. A thyroid-stimulating hormone level should be added to the initial laboratory panel.
A complete blood count revealed white blood cell count 11.86 K/uL with neutrophilic predominance and immature granulocytes, hemoglobin 11.5 g/dL, and platelet count 323 K/uL. Serum sodium was 141 mEq/L, potassium 4.2 mEq/L, chloride 103 mEq/L, bicarbonate 30 mEq/L, creatinine 1.14 mg/dL (prior baseline of 0.8-1.0 mg/dL), blood urea nitrogen 26 mg/dL, blood glucose 159 mg/dL, and calcium 9.1 mg/dL. His digoxin level was 1.3 ng/mL (reference range 0.5-1.9 mg/mL) and troponin was undetectable. INR was 2.7 and partial thromboplastin time (PTT) 60 seconds. Vitamin B12 level was 674 pg/mL (reference range >180). A urinalysis had 1+ hyaline casts and was negative for nitrites, leukocyte esterase, blood, and bacteria. An ECG revealed atrial fibrillation with a ventricular rate of 80 beats per minute. A chest radiograph showed clear lung fields. A CT of the head without IV contrast had no evidence of an acute intracranial abnormality. In the ED, 1 liter of IV normal saline was given and blood pressure improved to 127/72 mm Hg.
The head CT does not show intracranial bleeding, and, though it is reassuring that INR is in the therapeutic range, ischemic stroke must remain in the differential diagnosis. Sepsis is less likely given that the criteria for systemic inflammatory response syndrome are not met, and hypotension was rapidly corrected with administration of IV fluids. Urinary tract infection was ruled out with the negative urinalysis. Subclinical seizures remain possible, as does medication-related or other toxicity. A medication overdose, intentional or otherwise, should also be considered.
The patient was admitted to the hospital. On reassessment by the inpatient team, he was oriented only to self, frequently falling asleep, and not recalling earlier conversations when aroused. His speech remained slurred and difficult to understand. Neurologic examination findings were unchanged since the ED examination. On additional cerebellar examination, he had dysmetria with finger-to-nose testing bilaterally and dysdiadochokinesia (impaired rapid alternating movements) of the left hand.
His handedness is not mentioned; the dysdiadochokinesia of the left hand may reflect the patient’s being right-handed, or may signify a focal cerebellar lesion. The cerebellum is also implicated by the bilateral dysmetria. Persistent somnolence in the absence of CT findings suggests a metabolic or infectious process. Metabolic processes that can cause bilateral cerebellar ataxia and somnolence include overdose of a drug or medication. Use of alcohol or a medication such as phenytoin, valproic acid, or a benzodiazepine can cause the symptoms in this case, but was not reported by the family, and there was no documentation of it in the SNF records. Wernicke encephalopathy is rare and is not well supported by the patient’s presentation but should be considered, as it can be easily treated with thiamine. Meningoencephalitis affecting the cerebellum remains possible, but infection is less likely. Both electroencephalogram and brain MRI should be performed, with a specific interest in possible cerebellar lesions. If the MRI is unremarkable, a lumbar puncture should be performed to assess opening pressure and investigate for infectious etiologies.
MRI of the brain showed age-related volume loss and nonspecific white matter disease without acute changes. Lack of a clear explanation for the neurologic findings led to suspicion of a medication side effect. Ertapenem was stopped on admission because it has been reported to rarely cause altered mental status. IV moxifloxacin was started for the osteomyelitis. Over the next 2 days, symptoms began resolving; within 24 hours of ertapenem discontinuation, the patient was awake, alert, and talkative. On examination, he remained dysarthric but was no longer dysmetric. Within 48 hours, the dysarthria was completely resolved, and he was returned to the SNF to complete a course of IV moxifloxacin.
DISCUSSION
Among elderly patients presenting to the ED, altered mental status is a common complaint, accounting for 10% to 30% of visits.1 Medications are a common cause of altered mental status among the elderly and are responsible for 40% of delirium cases.1 The risk of adverse drug events (ADEs) rises with the number of medications prescribed.1-3 Among patients older than 60 years, the incidence of polypharmacy (defined as taking >5 prescription medications) increased from roughly 20% in 1999 to 40% in 2012.4,5 The most common ADEs in the ambulatory setting (25%) are central nervous system (CNS) symptoms, including dizziness, sleep disturbances, and mood changes.6 A medication effect should be suspected in any elderly patient presenting with altered mental state.
The present patient developed a constellation of neurologic symptoms after starting ertapenem, one of the carbapenem antibiotics, which is a class of medications that can cause CNS ADEs. Carbapenems are renally cleared, and adjustments must be made for acute or chronic changes in kidney function. Carbapenems are associated with increased risk of seizure; the incidence of seizure with ertapenem is 0.2%.7,8 Food and Drug Administration postmarketing reports have noted ertapenem can cause somnolence and dyskinesia,9 and several case reports have described ertapenem-associated CNS side effects, including psychosis and encephalopathy.10-13 Symptoms and examination findings can include confusion, disorientation, garbled speech, dysphagia, hallucinations, miosis, myoclonus, tremor, and agitation.10-13 Although reports of dysmetria and dysdiadochokinesia are lacking, suspicion of an ADE in this case was heightened by the timing of the exposure and the absence of alternative infectious, metabolic, and vascular explanations for bilateral cerebellar dysfunction.
The Naranjo Adverse Drug Reaction (ADR) scale may help clinicians differentiate ADEs from other etiologies of symptoms. It uses 10 weighted questions (Table) to estimate the probability that an adverse clinical event is caused by a drug reaction.14 The present case was assigned 1 point for prior reports of neurologic ADEs associated with ertapenem, 2 for the temporal association, 1 for resolution after medication withdrawal, 2 for lack of alternative causes, and 1 for objective evidence of neurologic dysfunction—for a total of 7 points, indicating ertapenem was probably the cause of the patient’s neurologic symptoms. Of 4 prior cases in which carbapenem toxicity was suspected and the Naranjo scale was used, 3 found a probable relationship, and the fourth a highly probable one.10,12 Confusion, disorientation, hallucinations, tangential thoughts, and garbled speech were reported in the 3 probable cases of ADEs. In the highly probable case, tangential thoughts, garbled speech, and miosis were noted on examination, and these findings returned after re-exposure to ertapenem. Of note, these ADEs occurred in patients with normal and abnormal renal function, and in middle-aged and elderly patients.10,11,13
Most medications have a long list of low-frequency and rarely reported adverse effects. The present case reminds clinicians to consider rare adverse effects, or variants of previously reported adverse effects, in a patient with unexplained symptoms. To estimate the probability that a drug is causing harm to a patient, using a validated tool such as the Naranjo scale helps answer the question, What are the chances?
KEY TEACHING POINTS
Clinicians should include rare adverse effects of common medications in the differential diagnosis.
The Naranjo score is a validated tool that can be used to systematically assess the probability of an adverse drug effect at the bedside.
- The presentation of ertapenem-associated neurotoxicity may include features of bilateral cerebellar dysfunction.
Disclosure
Nothing to report.
1. Inouye SK, Fearing MA, Marcantonio ER. Delirium. In: Halter JB, Ouslander JG, Tinetti ME, Studenski S, High KP, Asthana S, eds. Hazzard’s Geriatric Medicine and Gerontology. 6th ed. New York, NY: McGraw-Hill; 2009.
2. Sarkar U, López A, Maselli JH, Gonzales R. Adverse drug events in U.S. adult ambulatory medical care. Health Serv Res. 2011;46(5):1517-1533. PubMed
3. Chrischilles E, Rubenstein L, Van Gilder R, Voelker M, Wright K, Wallace R. Risk factors for adverse drug events in older adults with mobility limitations in the community setting. J Am Geriatr Soc. 2007;55(1):29-34. PubMed
4. Kaufman DW, Kelly JP, Rosenberg L, Anderson TE, Mitchell AA. Recent patterns of medication use in the ambulatory adult population of the United States: the Slone survey. JAMA. 2002;287(3):337-344. PubMed
5. Kantor ED, Rehm CD, Haas JS, Chan AT, Giovannucci EL. Trends in prescription drug use among adults in the United States from 1999-2012. JAMA. 2015;314(17):1818-1831. PubMed
6. Thomsen LA, Winterstein AG, Søndergaard B, Haugbølle LS, Melander A. Systematic review of the incidence and characteristics of preventable adverse drug events in ambulatory care. Ann Pharmacother. 2007;41(9):1411-1426. PubMed
7. Zhanel GG, Wiebe R, Dilay L, et al. Comparative review of the carbapenems. Drugs. 2007;67(7):1027-1052. PubMed
8. Cannon JP, Lee TA, Clark NM, Setlak P, Grim SA. The risk of seizures among the carbapenems: a meta-analysis. J Antimicrob Chemother. 2014;69(8):2043-2055. PubMed
9. US Food and Drug Administration. Invanz (ertapenem) injection [safety information]. http://www.fda.gov/Safety/MedWatch/SafetyInformation/ucm196605.htm. Published July 2013. Accessed July 6, 2015.
10. Oo Y, Packham D, Yau W, Munckhof WJ. Ertapenem-associated psychosis and encephalopathy. Intern Med J. 2014;44(8):817-819. PubMed
11. Wen MJ, Sung CC, Chau T, Lin SH. Acute prolonged neurotoxicity associated with recommended doses of ertapenem in 2 patients with advanced renal failure. Clin Nephrol. 2013;80(6):474-478. PubMed
12. Duquaine S, Kitchell E, Tate T, Tannen RC, Wickremasinghe IM. Central nervous system toxicity associated with ertapenem use. Ann Pharmacother. 2011;45(1):e6. PubMed
13. Kong V, Beckert L, Awunor-Renner C. A case of beta lactam-induced visual hallucination. N Z Med J. 2009;122(1298):76-77. PubMed
14. Naranjo CA, Busto U, Sellers EM, et al. A method for estimating the probability of adverse drug reactions. Clin Pharmacol Ther. 1981;30(2):239-245. PubMed
1. Inouye SK, Fearing MA, Marcantonio ER. Delirium. In: Halter JB, Ouslander JG, Tinetti ME, Studenski S, High KP, Asthana S, eds. Hazzard’s Geriatric Medicine and Gerontology. 6th ed. New York, NY: McGraw-Hill; 2009.
2. Sarkar U, López A, Maselli JH, Gonzales R. Adverse drug events in U.S. adult ambulatory medical care. Health Serv Res. 2011;46(5):1517-1533. PubMed
3. Chrischilles E, Rubenstein L, Van Gilder R, Voelker M, Wright K, Wallace R. Risk factors for adverse drug events in older adults with mobility limitations in the community setting. J Am Geriatr Soc. 2007;55(1):29-34. PubMed
4. Kaufman DW, Kelly JP, Rosenberg L, Anderson TE, Mitchell AA. Recent patterns of medication use in the ambulatory adult population of the United States: the Slone survey. JAMA. 2002;287(3):337-344. PubMed
5. Kantor ED, Rehm CD, Haas JS, Chan AT, Giovannucci EL. Trends in prescription drug use among adults in the United States from 1999-2012. JAMA. 2015;314(17):1818-1831. PubMed
6. Thomsen LA, Winterstein AG, Søndergaard B, Haugbølle LS, Melander A. Systematic review of the incidence and characteristics of preventable adverse drug events in ambulatory care. Ann Pharmacother. 2007;41(9):1411-1426. PubMed
7. Zhanel GG, Wiebe R, Dilay L, et al. Comparative review of the carbapenems. Drugs. 2007;67(7):1027-1052. PubMed
8. Cannon JP, Lee TA, Clark NM, Setlak P, Grim SA. The risk of seizures among the carbapenems: a meta-analysis. J Antimicrob Chemother. 2014;69(8):2043-2055. PubMed
9. US Food and Drug Administration. Invanz (ertapenem) injection [safety information]. http://www.fda.gov/Safety/MedWatch/SafetyInformation/ucm196605.htm. Published July 2013. Accessed July 6, 2015.
10. Oo Y, Packham D, Yau W, Munckhof WJ. Ertapenem-associated psychosis and encephalopathy. Intern Med J. 2014;44(8):817-819. PubMed
11. Wen MJ, Sung CC, Chau T, Lin SH. Acute prolonged neurotoxicity associated with recommended doses of ertapenem in 2 patients with advanced renal failure. Clin Nephrol. 2013;80(6):474-478. PubMed
12. Duquaine S, Kitchell E, Tate T, Tannen RC, Wickremasinghe IM. Central nervous system toxicity associated with ertapenem use. Ann Pharmacother. 2011;45(1):e6. PubMed
13. Kong V, Beckert L, Awunor-Renner C. A case of beta lactam-induced visual hallucination. N Z Med J. 2009;122(1298):76-77. PubMed
14. Naranjo CA, Busto U, Sellers EM, et al. A method for estimating the probability of adverse drug reactions. Clin Pharmacol Ther. 1981;30(2):239-245. PubMed
© 2017 Society of Hospital Medicine
Safe and effective bedside thoracentesis: A review of the evidence for practicing clinicians
Pleural effusion can occur in myriad conditions including infection, heart failure, liver disease, and cancer.1 Consequently, physicians from many disciplines routinely encounter both inpatients and outpatients with this diagnosis. Often, evaluation and treatment require thoracentesis to obtain fluid for analysis or symptom relief.
Although historically performed at the bedside without imaging guidance or intraprocedural monitoring, thoracentesis performed in this fashion carries considerable risk of complications. In fact, it has 1 of the highest rates of iatrogenic pneumothorax among bedside procedures.2 However, recent advances in practice and adoption of newer technologies have helped to mitigate risks associated with this procedure. These advances are relevant because approximately 50% of thoracenteses are still performed at the bedside.3 In this review, we aim to identify the most recent key practices that enhance the safety and the effectiveness of thoracentesis for practicing clinicians.
METHODS
Information Sources and Search Strategy
With the assistance of a research librarian, we performed a systematic search of PubMed-indexed articles from January 1, 2000 to September 30, 2015. Articles were identified using search terms such as thoracentesis, pleural effusion, safety, medical error, adverse event, and ultrasound in combination with Boolean operators. Of note, as thoracentesis is indexed as a subgroup of paracentesis in PubMed, this term was also included to increase the sensitivity of the search. The full search strategy is available in the Appendix. Any references cited in this review outside of the date range of our search are provided only to give relevant background information or establish the origin of commonly performed practices.
Study Eligibility and Selection Criteria
Studies were included if they reported clinical aspects related to thoracentesis. We defined clinical aspects as those strategies that focused on operator training, procedural techniques, technology, management, or prevention of complications. Non-English language articles, animal studies, case reports, conference proceedings, and abstracts were excluded. As our intention was to focus on the contemporary advances related to thoracentesis performance, (eg, ultrasound [US]), our search was limited to studies published after the year 2000. Two authors, Drs. Schildhouse and Lai independently screened studies to determine inclusion, excluding studies with weak methodology, very small sample sizes, and those only tangentially related to our aim. Disagreements regarding study inclusion were resolved by consensus. Drs. Lai, Barsuk, and Mourad identified additional studies by hand review of reference lists and content experts (Figure 1).
Conceptual Framework
All selected articles were categorized by temporal relationship to thoracentesis as pre-, intra-, or postprocedure. Pre-procedural topics were those outcomes that had been identified and addressed before attempting thoracentesis, such as physician training or perceived risks of harm. Intraprocedural considerations included aspects such as use of bedside US, pleural manometry, and large-volume drainage. Finally, postprocedural factors were those related to evaluation after thoracentesis, such as follow-up imaging. This conceptual framework is outlined in Figure 2.
RESULTS
The PubMed search returned a total of 1170 manuscripts, of which 56 articles met inclusion criteria. Four additional articles were identified by experts and included in the study.4-7 Therefore, 60 articles were identified and included in this review. Study designs included cohort studies, case control studies, systematic reviews, meta-analyses, narrative reviews, consensus guidelines, and randomized controlled trials. A summary of all included articles by topic can be found in the Table.
PRE-PROCEDURAL CONSIDERATIONS
Physician Training
Studies indicate that graduate medical education may not adequately prepare clinicians to perform thoracentesis.8 In fact, residents have the least exposure and confidence in performing thoracentesis when compared to other bedside procedures.9,10 In 1 survey, 69% of medical trainees desired more exposure to procedures, and 98% felt that procedural skills were important to master.11 Not surprisingly, then, graduating internal medicine residents perform poorly when assessed on a thoracentesis simulator.12
Supplemental training outside of residency is useful to develop and maintain skills for thoracentesis, such as simulation with direct observation in a zero-risk environment. In 1 study, “simulation-based mastery learning” combined an educational video presentation with repeated, deliberate practice on a simulator until procedural competence was acquired, over two 2-hour sessions. In this study, 40 third-year medicine residents demonstrated a 71% improvement in clinical skills performance after course completion, with 93% achieving a passing score. The remaining 7% also achieved passing scores with extra practice time.12 Others have built upon the concept of simulation-based training. For instance, 2 studies suggest that use of a simulation-based curriculum improved both thoracentesis knowledge and performance skills in a 3-hour session.13,14 Similarly, 1 prospective study reported that a half-day thoracentesis workshop using simulation and 1:1 direct observation successfully lowered pneumothorax rates from 8.6% to 1.8% in a group of practicing clinicians. Notably, additional interventions including use of bedside US, limiting operators to a focused group, and standardization of equipment were also a part of this quality improvement initiative.7 Although repetition is required to gain proficiency when using a simulator, performance and confidence appear to plateau with only 4 simulator trials. In medical students, improvements derived through simulator-based teaching were sustained when retested 6 months following training.15
An instrument to ensure competency is necessary, given variability in procedural experience among both new graduates and practicing physicians,. Our search did not identify any clinically validated tools that adequately assessed thoracentesis performance. However, some have been proposed16 and 1 validated in a simulation environment.12 Regarding the incorporation of US for effusion markup, 1 validated tool used an 11-domain assessment covering knowledge of US machine manipulation, recognition of images with common pleural effusion characteristics, and performance of thoracic US with puncture-site marking on a simulator. When used on 22 participants, scores with the tool could reliably differentiate between novice, intermediate, and advanced groups (P < 0.0001).17
Patient Selection
Coagulopathies and Anticoagulation. Historically, the accepted cutoff for performing thoracentesis is an international normalized ratio (INR) less than 1.5 and a platelet count greater than 50,000/µL. McVay et al.18 first showed in 1991 that use of these cutoffs was associated with low rates of periprocedural bleeding, leading to endorsement in the British Thoracic Society (BTS) Pleural Disease Guideline 2010.19 Other recommendations include the 2012 Society for Interventional Radiology guidelines that endorse correction of an INR greater than 2, or platelets less than 50,000/µL, based almost exclusively on expert opinion.5
However, data suggest that thoracentesis may be safely performed outside these parameters. For instance, a prospective study of approximately 9000 thoracenteses over 12 years found that patients with an INR of 1.5-2.9 or platelets of 20,000 - 49,000/µL experienced rates of bleeding complications similar to those with normal values.20 Similarly, a 2014 review21 found that the overall risk of hemorrhage during thoracentesis in the setting of moderate coagulopathy (defined as an INR of 1.5 - 3 or platelets of 25,000-50,000/µL), was not increased. In 1 retrospective study of more than 1000 procedures, no differences in hemorrhagic events were noted in patients with bleeding diatheses that received prophylactic fresh frozen plasma or platelets vs. those who did not.22 Of note, included studies used a variety of criteria to define a hemorrhagic complication, which included: an isolated 2 g/dL or more decrement in hemoglobin, presence of bloody fluid on repeat tap with associated hemoglobin decrement, rapid re-accumulation of fluid with a hemoglobin decrement, or transfusion of 2 units or more of whole blood.
Whether it is safe to perform thoracentesis on patients taking antiplatelet therapy is less well understood. Although data are limited, a few small-scale studies23,24 suggest that hemorrhagic complications following thoracentesis in patients receiving clopidogrel are comparable to the general population. We found no compelling data regarding the safety of thoracentesis in the setting of direct oral anticoagulants, heparin, low-molecular weight heparin, or intravenous direct thrombin inhibitors. Current practice is to generally avoid thoracentesis while these therapeutic anticoagulants are used.
Invasive mechanical ventilation. Pleural effusion is common in patients in the intensive care unit, including those requiring mechanical ventilation.25 Thoracentesis in this population is clinically important: fluid analysis in 1 study was shown to aid the diagnosis in 45% of cases and changes in treatment in 33%.26 However, clinicians may be reluctant to perform thoracentesis on patients who require mechanical ventilation, given the perception of a greater risk of pneumothorax from positive pressure ventilation.
Despite this concern, a 2011 meta-analysis including 19 studies and more than 1100 patients revealed rates of pneumothorax and hemothorax comparable to nonventilated patients.25 Furthermore, a 2015 prospective study that examined thoracentesis in 1377 mechanically ventilated patients revealed no difference in complication rates as well.20 Therefore, evidence suggests that performance of thoracentesis in mechanically ventilated patients is not contraindicated.
Skin Disinfection and Antisepsis Precautions
The 2010 BTS guidelines list empyema and wound infection as possible complications of thoracentesis.19 However, no data regarding incidence are provided. Additionally, an alcohol-based skin cleanser (such as 2% chlorhexidine gluconate/70% isopropyl alcohol), along with sterile gloves, field, and dressing are suggested as precautionary measures.19 In 1 single-center registry of 2489 thoracenteses performed using alcohol or iodine-based antiseptic and sterile drapes, no postprocedure infections were identified.27 Of note, we did not find other studies (including case reports) that reported either incidence or rate of infectious complications such as wound infection and empyema. In an era of modern skin antiseptics that have effectively reduced complications such as catheter-related bloodstream infection,28 the incidence of this event is thus likely to be low.
INTRAPROCEDURAL CONSIDERATIONS
Use of Bedside Ultrasound
Portable US has particular advantages for evaluation of pleural effusion vs other imaging modalities. Compared with computerized tomography (CT), bedside US offers similar performance but is less costly, avoids both radiation exposure and need for patient transportation, and provides results instantaneously.29,30 Compared to chest x-ray (CXR), US is more sensitive at detecting the presence, volume, and characteristics of pleural fluid30,31 and can be up to 100% sensitive for effusions greater than 100 mL.29 Furthermore, whereas CXR typically requires 200 mL of fluid to be present for detection of an effusion, US can reliably detect as little as 20 mL of fluid.29 When US was used to confirm thoracentesis puncture sites in a study involving 30 physicians of varying experience and 67 consecutive patients, 15% of sites found by clinical exam were inaccurate (less than 10 mm fluid present), 10% were at high risk for organ puncture, and a suitable fluid pocket was found 54% of times when exam could not.4
A 2010 meta-analysis of 24 studies and 6605 thoracenteses estimated the overall rate of pneumothorax at 6%; however, procedures performed with US guidance were associated with a 70% reduced risk of this event (odds ratio, 0.30; 95% confidence interval, 0.20 - 0.70).32 In a 2014 randomized control trial of 160 patients that compared thoracentesis with US guidance for site marking vs no US use, 10 pneumothoraces occurred in the control group vs 1 in the US group (12.5% vs 1.25%, P = 0.009).33 Similarly, another retrospective review of 445 consecutive patients with malignant effusions revealed a pneumothorax rate of 0.97% using US in real time during needle insertion compared to 8.89% for unguided thoracenteses (P < 0.0001).34 Several other studies using US guidance for either site markup or in real time reported similar pneumothorax rates, ranging from 1.1% - 4.8%.35-37 However, it is unclear if real-time US specifically provides an additive effect vs site marking alone, as no studies directly comparing the 2 methods were found.
Benefits of US also include a higher rate of procedural success, with 1 study demonstrating a 99% success rate when using US vs. 90% without (P = 0.030).33 A larger volume of fluid removed has been observed with US use as well, and methods have been described using fluid-pocket depth to guide puncture site localization and maximize drainage.38 Finally, US use for thoracentesis has been associated with lower costs and length of stay.39,40
Intercostal Artery Localization
Although rare (incidence, 0.18%-2%20,21,39), the occurrence of hemothorax following thoracentesis is potentially catastrophic. This serious complication is often caused by laceration of the intercostal artery (ICA) or 1 of its branches during needle insertion.41
While risk of injury is theoretically reduced by needle insertion superior to the rib, studies using cadaver dissection and 3D angiography show significant tortuosity of the ICA.6,41-43 The degree of tortuosity is increased within 6 cm of the midline, in more cephalad rib spaces, and in the elderly (older than 60 years).41-43 Furthermore, 1 cadaveric study also demonstrated the presence of arterial collaterals branching off the ICA at multiple intercostal spaces, ranging between 8 cm and 11 cm from the midline.41 This anatomic variability may explain why some have observed low complication and hemothorax rates with an extreme lateral approach.35 Bedside US with color flow Doppler imaging has been used to identify the ICA, with 88% sensitivity compared to CT imaging while adding little to exam time.44,45 Of note, a 37% drop in the rate of hemothorax was observed in 1 study with routine US guidance alone.39
Pleural Pressure Monitoring and Large-Volume Thoracentesis
While normal intrapleural pressures are approximately -5 to -10 cm H2O,46 the presence of a pleural effusion creates a complex interaction between fluid, compressed lung, and chest wall that can increase these pressures.47 During drainage of an effusion, pleural pressures may rapidly drop, provoking re-expansion pulmonary edema (REPE). While rare (0 -1%), clinically-diagnosed REPE is a serious complication that can lead to rapid respiratory failure and death.20,48 REPE is postulated to be caused by increased capillary permeability resulting from inflammation, driven by rapid re-inflation of the lung when exposed to highly negative intrapleural pressures.47,49
Measurement of intrapleural pressure using a water manometer during thoracentesis may minimize REPE by terminating fluid drainage when intrapleural pressure begins to drop rapidly.50,51 A cutoff of -20 cm H2O has been cited repeatedly as safe since being suggested by Light in 1980, but this is based on animal models.50,52 In 1 prospective study of 185 thoracenteses in which manometry was performed, 15% of patients had intrapleural pressure drop to less than -20 cm H2O (at which point the procedure was terminated) but suffered no REPE.50
Manometry is valuable in the identification of an unexpandable or trapped lung when pleural pressures drop rapidly with only minimal fluid volume removal.47,53 Other findings correlated with an unexpandable lung include a negative opening pressure47 and large fluctuations in pressure during the respiratory cycle.54
While development of symptoms (eg, chest pain, cough, or dyspnea) is often used as a surrogate, the correlation between intrapleural pressure and patient symptoms is inconsistent and not a reliable proxy.55 One study found that 22% of patients with chest pain during thoracentesis had intrapleural pressures lower than -20 cm H2O compared with 8.6% of asymptomatic patients,56 but it is unclear if the association is causal.
Thoracentesis is often performed for symptomatic relief and removal of large fluid volume. However, it remains common to halt fluid removal after 1.5 L, a threshold endorsed by BTS.19 While some investigators have suggested that removal of 2 L or more of pleural fluid does not compromise safety,57,58 a 4- to 5-fold rise in the risk of pneumothorax was noted in 2 studies.20,59 when more than 1.5 L of fluid was removed. The majority of these may be related to pneumothorax ex vacuo, a condition in which fluid is drained from the chest, but the lung is unable to expand and fill the space (eg, “trapped lung”), resulting in a persistent pneumothorax. This condition generally does not require treatment.60 When manometry is employed at 200-mL intervals with termination at an intrapleural pressure of less than 20 mm H2O, drainage of 3 L or more has been reported with low rates of pneumothorax and very low rates of REPE.50,51 However, whether this is cause and effect is unknown because REPE is rare, and more work is needed to determine the role of manometry for its prevention.
POSTPROCEDURAL CONSIDERATIONS
Postprocedure Imaging
Performing an upright CXR following thoracentesis is a practice that remains routinely done by many practitioners to monitor for complications. Such imaging was also endorsed by the American Thoracic Society guidelines.61 However, more recent data question the utility of this practice. Multiple studies have confirmed that post-thoracentesis CXR is unnecessary unless clinical suspicion for pneumothorax or REPE is present.36,58,62,63 The BTS guidelines also advocate this approach.19 Interestingly, a potentially more effective way to screen for postprocedure complications is through bedside US, which has been shown to be more sensitive than CXR in detecting pneumothorax.64 In 1 study of 185 patients, bedside US demonstrated a sensitivity of 88% and a specificity of 97% for diagnosing pneumothorax in patients with adequate quality scans, with positive and negative likelihood ratios of 55 and 0.17, respectively.65
DISCUSSION
Thoracentesis remains a core procedural skill for hospitalists, critical care physicians, and emergency physicians. It is the foundational component when investigating and treating pleural effusions. When the most current training, techniques, and technology are used, data suggest this procedure is safe to perform at the bedside. Our review highlights these strategies and evaluates which aspects might be most applicable to clinical practice.
Our findings have several implications for those who perform this procedure. First, appropriate training is central to procedural safety, and both simulation and direct observation by procedural experts have been shown by multiple investigators to improve knowledge and skill. This training should integrate the use of US in performing a focused thoracic exam.
Second, recommendations regarding coagulopathy and a “safe cutoff” of an INR less than 1.5 or platelets greater than 50,000/µL had limited evidentiary support. Rather, multiple studies suggest no difference in bleeding risk following thoracentesis with an INR as high as 3.0 and platelets greater than 25,000/µL. Furthermore, prophylactic transfusion with fresh frozen plasma or platelets before thoracentesis did not alter bleeding risk and exposes patients to transfusion complications. Thus, routine use of this practice can no longer be recommended. Third, further research is needed to understand the bleeding risk for patients on antiplatelet medications, heparin products, and also direct oral anticoagulants, given the growing popularity in their use and the potential consequences of even temporary cessation. Regarding patients on mechanical ventilation, thoracentesis demonstrated no difference in complication rates vs. the general population, and its performance in this population is encouraged when clinically indicated.
Intraprocedural considerations include the use of bedside US. Due to multiple benefits including effusion characterization, puncture site localization, and significantly lower rates of pneumothorax, the standard of care should be to perform thoracentesis with US guidance. Both use of US to mark an effusion immediately prior to puncture or in real time during needle insertion demonstrated benefit; however, it is unclear if 1 method is superior because no direct comparison studies were found. Further work is needed to investigate this potential.
Our review suggests that the location and course of the ICA is variable, especially near the midline, in the elderly, and in higher intercostal spaces, leaving it vulnerable to laceration. We recommend physicians only attempt thoracentesis at least 6 cm lateral to the midline due to ICA tortuosity and, ideally, 12 cm lateral, to avoid the presence of collaterals. Although only 2 small-scale studies were found pertaining to the use of US in identifying the ICA, we encourage physicians to consider learning how to screen for its presence as a part of their routine thoracic US exam in the area underlying the planned puncture site.
Manometry is beneficial because it can diagnose a nonexpandable lung and allows for pleural pressure monitoring.52,53 A simple U-shaped manometer can be constructed from intravenous tubing included in most thoracentesis kits, which adds little to overall procedure time. While low rates of REPE have been observed when terminating thoracentesis if pressures drop below -20 cm H2O or chest pain develops, neither measure appears to have reliable predictive value, limiting clinical utility. Further work is required to determine if a “safe pressure cutoff” exists. In general, we recommend the use of manometry when a nonexpandable (trapped) lung is suspected, because large drops in intrapleural pressure, a negative opening pressure, and respiratory variation can help confirm the diagnosis and avoid pneumothorax ex vacuo or unnecessary procedures in the future. As this condition appears to be more common in the setting of larger effusions, use of manometry when large-volume thoracenteses are planned is also reasonable.
Postprocedurally, routine imaging after thoracentesis is not recommended unless there is objective concern for complication. When indicated, bedside US is better positioned for this role compared with CXR, because it is more sensitive in detecting pneumothorax, provides instantaneous results, and avoids radiation exposure.
Our review has limitations. First, we searched only for articles between defined time periods, restricted our search to a single database, and excluded non-English articles. This has the potential to introduce selection bias, as nonprimary articles that fall within our time restrictions may cite older studies that are outside our search range. To minimize this effect, we performed a critical review of all included studies, especially nonprimary articles. Second, despite the focus of our search strategy to identify any articles related to patient safety and adverse events, we cannot guarantee that all relevant articles for any particular complication or risk factor were captured given the lack of more specific search terms. Third, although we performed a systematic search of the literature, we did not perform a formal systematic review or formally grade included studies. As the goal of our review was to categorize and operationalize clinical aspects, this approach was necessary, and we acknowledge that the quality of studies is variable. Lastly, we aimed to generate clinical recommendations for physicians performing thoracentesis at the bedside; others reviewing this literature may find or emphasize different aspects relevant to practice outside this setting.
In conclusion, evaluation and treatment of pleural effusions with bedside thoracentesis is an important skill for physicians of many disciplines. The evidence presented in this review will help inform the process and ensure patient safety. Physicians should consider incorporating these recommendations into their practice.
Acknowledgments
The authors thank Whitney Townsend, MLIS, health sciences informationist, for assistance with serial literature searches.
Disclosure
Nothing to report.
1. Kasper DL. Harrison's Principles of Internal Medicine. 19th ed. New York, NY: McGraw Hill Education; 2015.
2. Celik B, Sahin E, Nadir A, Kaptanoglu M. Iatrogenic pneumothorax: etiology, incidence and risk factors. Thorac Cardiovasc Surg. 2009;57(5):286-290. PubMed
3. Hooper CE, Welham SA, Maskell NA, Soc BT. Pleural procedures and patient safety: a national BTS audit of practice. Thorax. 2015;70(2):189-191. PubMed
4. Diacon AH, Brutsche MH, Soler M. Accuracy of pleural puncture sites: a prospective comparison of clinical examination with ultrasound. Chest. 2003;123(2):436-441. PubMed
5. Patel IJ, Davidson JC, Nikolic B, et al. Consensus guidelines for periprocedural management of coagulation status and hemostasis risk in percutaneous image-guided interventions. J Vasc Interv Radiol. 2012;23(6):727-736. PubMed
6. Wraight WM, Tweedie DJ, Parkin IG. Neurovascular anatomy and variation in the fourth, fifth, and sixth intercostal spaces in the mid-axillary line: a cadaveric study in respect of chest drain insertion. Clin Anat. 2005;18(5):346-349. PubMed
7. Duncan DR, Morgenthaler TI, Ryu JH, Daniels CE. Reducing iatrogenic risk in thoracentesis: establishing best practice via experiential training in a zero-risk environment. Chest. 2009;135(5):1315-1320. PubMed
8. Grover S, Currier PF, Elinoff JM, Mouchantaf KJ, Katz JT, McMahon GT. Development of a test to evaluate residents' knowledge of medical procedures. J Hosp Med. 2009;4(7):430-432. PubMed
9. Promes SB, Chudgar SM, Grochowski CO, et al. Gaps in procedural experience and competency in medical school graduates. Acad Emerg Med. 2009;16 Suppl 2:S58-62. PubMed
10. Huang GC, Smith CC, Gordon CE, et al. Beyond the comfort zone: residents assess their comfort performing inpatient medical procedures. Am J Med. 2006;119(1):71 e17-24. PubMed
11. Lagan J, Cutts L, Zaidi S, Benton I, Rylance J. Are we failing our trainees in providing opportunities to attain procedural confidence? Br J Hosp Med (Lond). 2015;76(2):105-108. PubMed
12. Wayne DB, Barsuk JH, O'Leary KJ, Fudala MJ, McGaghie WC. Mastery learning of thoracentesis skills by internal medicine residents using simulation technology and deliberate practice. J Hosp Med. 2008;3(1):48-54. PubMed
13. Lenchus JD. End of the "see one, do one, teach one" era: the next generation of invasive bedside procedural instruction. J Am Osteopath Assoc. 2010;110(6):340-346. PubMed
14. Lenchus J, Issenberg SB, Murphy D, et al. A blended approach to invasive bedside procedural instruction. Med Teach. 2011;33(2):116-123. PubMed
15. Jiang G, Chen H, Wang S, et al. Learning curves and long-term outcome of simulation-based thoracentesis training for medical students. BMC Med Educ. 2011;11:39. PubMed
16. Berg D, Berg K, Riesenberg LA, et al. The development of a validated checklist for thoracentesis: preliminary results. Am J Med Qual. 2013;28(3):220-226. PubMed
17. Salamonsen M, McGrath D, Steiler G, Ware R, Colt H, Fielding D. A new instrument to assess physician skill at thoracic ultrasound, including pleural effusion markup. Chest. 2013;144(3):930-934. PubMed
18. McVay PA, Toy PT. Lack of increased bleeding after paracentesis and thoracentesis in patients with mild coagulation abnormalities. Transfusion. 1991;31(2):164-171. PubMed
19. Havelock T, Teoh R, Laws D, Gleeson F, Group BTSPDG. Pleural procedures and thoracic ultrasound: British Thoracic Society Pleural Disease Guideline 2010. Thorax. 2010;65 Suppl 2:ii61-76. PubMed
20. Ault MJ, Rosen BT, Scher J, Feinglass J, Barsuk JH. Thoracentesis outcomes: a 12-year experience. Thorax. 2015;70(2):127-132. PubMed
21. Puchalski J. Thoracentesis and the risks for bleeding: a new era. Curr Opin Pulm Med. 2014;20(4):377-384. PubMed
22. Hibbert RM, Atwell TD, Lekah A, et al. Safety of ultrasound-guided thoracentesis in patients with abnormal preprocedural coagulation parameters. Chest. 2013;144(2):456-463. PubMed
23. Zalt MB, Bechara RI, Parks C, Berkowitz DM. Effect of routine clopidogrel use on bleeding complications after ultrasound-guided thoracentesis. J Bronchology Interv Pulmonol. 2012;19(4):284-287. PubMed
24. Mahmood K, Shofer SL, Moser BK, Argento AC, Smathers EC, Wahidi MM. Hemorrhagic complications of thoracentesis and small-bore chest tube placement in patients taking clopidogrel. Ann Am Thorac Soc. 2014;11(1):73-79. PubMed
25. Goligher EC, Leis JA, Fowler RA, Pinto R, Adhikari NK, Ferguson ND. Utility and safety of draining pleural effusions in mechanically ventilated patients: a systematic review and meta-analysis. Crit Care. 2011;15(1):R46. PubMed
26. Fartoukh M, Azoulay E, Galliot R, et al. Clinically documented pleural effusions in medical ICU patients: how useful is routine thoracentesis? Chest. 2002;121(1):178-184. PubMed
27. Cervini P, Hesley GK, Thompson RL, Sampathkumar P, Knudsen JM. Incidence of infectious complications after an ultrasound-guided intervention. AJR Am J Roentgenol. 2010;195(4):846-850. PubMed
28. Mimoz O, Chopra V, Timsit JF. What's new in catheter-related infection: skin cleansing and skin antisepsis. Intensive Care Med. 2016;42(11):1784-1786. PubMed
29. Soni NJ, Franco R, Velez MI, et al. Ultrasound in the diagnosis and management of pleural effusions. J Hosp Med. 2015;10(12):811-816. PubMed
30. Feller-Kopman D. Ultrasound-guided thoracentesis. Chest. 2006;129(6):1709-1714. PubMed
31. Shojaee S, Argento AC. Ultrasound-guided pleural access. Semin Respir Crit Care Med. 2014;35(6):693-705. PubMed
32. Gordon CE, Feller-Kopman D, Balk EM, Smetana GW. Pneumothorax following thoracentesis: a systematic review and meta-analysis. Arch Intern Med. 2010;170(4):332-339. PubMed
33. Perazzo A, Gatto P, Barlascini C, Ferrari-Bravo M, Nicolini A. Can ultrasound guidance reduce the risk of pneumothorax following thoracentesis? J Bras Pneumol. 2014;40(1):6-12. PubMed
34. Cavanna L, Mordenti P, Berte R, et al. Ultrasound guidance reduces pneumothorax rate and improves safety of thoracentesis in malignant pleural effusion: report on 445 consecutive patients with advanced cancer. World J Surg Oncol. 2014;12:139. PubMed
35. Soldati G, Smargiassi A, Inchingolo R, Sher S, Valente S, Corbo GM. Ultrasound-guided pleural puncture in supine or recumbent lateral position - feasibility study. Multidiscip Respir Med. 2013;8(1):18. PubMed
36. Pihlajamaa K, Bode MK, Puumalainen T, Lehtimaki A, Marjelund S, Tikkakoski T. Pneumothorax and the value of chest radiography after ultrasound-guided thoracocentesis. Acta Radiol. 2004;45(8):828-832. PubMed
37. Barnes TW, Morgenthaler TI, Olson EJ, Hesley GK, Decker PA, Ryu JH. Sonographically guided thoracentesis and rate of pneumothorax. J Clin Ultrasound. 2005;33(9):442-446. PubMed
38. Zanforlin A, Gavelli G, Oboldi D, Galletti S. Ultrasound-guided thoracenthesis: the V-point as a site for optimal drainage positioning. Eur Rev Med Pharmacol Sci. 2013;17(1):25-28. PubMed
39. Patel PA, Ernst FR, Gunnarsson CL. Ultrasonography guidance reduces complications and costs associated with thoracentesis procedures. J Clin Ultrasound. 2012;40(3):135-141. PubMed
40. Mercaldi CJ, Lanes SF. Ultrasound guidance decreases complications and improves the cost of care among patients undergoing thoracentesis and paracentesis. Chest. 2013;143(2):532-538. PubMed
41. Shurtleff E, Olinger A. Posterior intercostal artery tortuosity and collateral branch points: a cadaveric study. Folia Morphol (Warsz). 2012;71(4):245-251. PubMed
42. Helm EJ, Rahman NM, Talakoub O, Fox DL, Gleeson FV. Course and variation of the intercostal artery by CT scan. Chest. 2013;143(3):634-639. PubMed
43. Yoneyama H, Arahata M, Temaru R, Ishizaka S, Minami S. Evaluation of the risk of intercostal artery laceration during thoracentesis in elderly patients by using 3D-CT angiography. Intern Med. 2010;49(4):289-292. PubMed
44. Salamonsen M, Ellis S, Paul E, Steinke K, Fielding D. Thoracic ultrasound demonstrates variable location of the intercostal artery. Respiration. 2012;83(4):323-329. PubMed
45. Salamonsen M, Dobeli K, McGrath D, et al. Physician-performed ultrasound can accurately screen for a vulnerable intercostal artery prior to chest drainage procedures. Respirology. 2013;18(6):942-947. PubMed
46. Grippi MA. Fishman's pulmonary diseases and disorders. Fifth edition. ed. New York: McGraw-Hill Education; 2015.
47. Huggins JT, Doelken P. Pleural manometry. Clin Chest Med. 2006;27(2):229-240. PubMed
48. Echevarria C, Twomey D, Dunning J, Chanda B. Does re-expansion pulmonary oedema exist? Interact Cardiovasc Thorac Surg. 2008;7(3):485-489. PubMed
49. Sue RD, Matthay MA, Ware LB. Hydrostatic mechanisms may contribute to the pathogenesis of human re-expansion pulmonary edema. Intensive Care Med. 2004;30(10):1921-1926. PubMed
50. Feller-Kopman D, Berkowitz D, Boiselle P, Ernst A. Large-volume thoracentesis and the risk of reexpansion pulmonary edema. Ann Thorac Surg. 2007;84(5):1656-1661. PubMed
51. Villena V, Lopez-Encuentra A, Pozo F, De-Pablo A, Martin-Escribano P. Measurement of pleural pressure during therapeutic thoracentesis. Am J Respir Crit Care Med. 2000;162(4 Pt 1):1534-1538. PubMed
52. Doelken P, Huggins JT, Pastis NJ, Sahn SA. Pleural manometry: technique and clinical implications. Chest. 2004;126(6):1764-1769. PubMed
53. Feller-Kopman D. Therapeutic thoracentesis: the role of ultrasound and pleural manometry. Curr Opin Pulm Med. 2007;13(4):312-318. PubMed
54. Boshuizen RC, Sinaasappel M, Vincent AD, Goldfinger V, Farag S, van den Heuvel MM. Pleural pressure swing and lung expansion after malignant pleural effusion drainage: the benefits of high-temporal resolution pleural manometry. J Bronchology Interv Pulmonol. 2013;20(3):200-205. PubMed
55. Pannu J, DePew ZS, Mullon JJ, Daniels CE, Hagen CE, Maldonado F. Impact of pleural manometry on the development of chest discomfort during thoracentesis: a symptom-based study. J Bronchology Interv Pulmonol. 2014;21(4):306-313. PubMed
56. Feller-Kopman D, Walkey A, Berkowitz D, Ernst A. The relationship of pleural pressure to symptom development during therapeutic thoracentesis. Chest. 2006;129(6):1556-1560. PubMed
57. Abunasser J, Brown R. Safety of large-volume thoracentesis. Conn Med. 2010;74(1):23-26. PubMed
58. Mynarek G, Brabrand K, Jakobsen JA, Kolbenstvedt A. Complications following ultrasound-guided thoracocentesis. Acta Radiol. 2004;45(5):519-522. PubMed
59. Josephson T, Nordenskjold CA, Larsson J, Rosenberg LU, Kaijser M. Amount drained at ultrasound-guided thoracentesis and risk of pneumothorax. Acta Radiol. 2009;50(1):42-47. PubMed
60. Heidecker J, Huggins JT, Sahn SA, Doelken P. Pathophysiology of pneumothorax following ultrasound-guided thoracentesis. Chest. 2006;130(4):1173-1184. PubMed
61. Sokolowski JW Jr, Burgher LW, Jones FL Jr, Patterson JR, Selecky PA. Guidelines for thoracentesis and needle biopsy of the pleura. This position paper of the American Thoracic Society was adopted by the ATS Board of Directors, June 1988. Am Rev Respir Dis. 1989;140(1):257-258. PubMed
62. Jones PW, Moyers JP, Rogers JT, Rodriguez RM, Lee YC, Light RW. Ultrasound-guided thoracentesis: is it a safer method? Chest. 2003;123(2):418-423. PubMed
63. Petersen WG, Zimmerman R. Limited utility of chest radiograph after thoracentesis. Chest. 2000;117(4):1038-1042. PubMed
64. Sachdeva A, Shepherd RW, Lee HJ. Thoracentesis and thoracic ultrasound: state of the art in 2013. Clin Chest Med. 2013;34(1):1-9. PubMed
65. Shostak E, Brylka D, Krepp J, Pua B, Sanders A. Bedside sonography for detection of postprocedure pneumothorax. J Ultrasound Med. 2013;32(6):1003-1009. PubMed
Pleural effusion can occur in myriad conditions including infection, heart failure, liver disease, and cancer.1 Consequently, physicians from many disciplines routinely encounter both inpatients and outpatients with this diagnosis. Often, evaluation and treatment require thoracentesis to obtain fluid for analysis or symptom relief.
Although historically performed at the bedside without imaging guidance or intraprocedural monitoring, thoracentesis performed in this fashion carries considerable risk of complications. In fact, it has 1 of the highest rates of iatrogenic pneumothorax among bedside procedures.2 However, recent advances in practice and adoption of newer technologies have helped to mitigate risks associated with this procedure. These advances are relevant because approximately 50% of thoracenteses are still performed at the bedside.3 In this review, we aim to identify the most recent key practices that enhance the safety and the effectiveness of thoracentesis for practicing clinicians.
METHODS
Information Sources and Search Strategy
With the assistance of a research librarian, we performed a systematic search of PubMed-indexed articles from January 1, 2000 to September 30, 2015. Articles were identified using search terms such as thoracentesis, pleural effusion, safety, medical error, adverse event, and ultrasound in combination with Boolean operators. Of note, as thoracentesis is indexed as a subgroup of paracentesis in PubMed, this term was also included to increase the sensitivity of the search. The full search strategy is available in the Appendix. Any references cited in this review outside of the date range of our search are provided only to give relevant background information or establish the origin of commonly performed practices.
Study Eligibility and Selection Criteria
Studies were included if they reported clinical aspects related to thoracentesis. We defined clinical aspects as those strategies that focused on operator training, procedural techniques, technology, management, or prevention of complications. Non-English language articles, animal studies, case reports, conference proceedings, and abstracts were excluded. As our intention was to focus on the contemporary advances related to thoracentesis performance, (eg, ultrasound [US]), our search was limited to studies published after the year 2000. Two authors, Drs. Schildhouse and Lai independently screened studies to determine inclusion, excluding studies with weak methodology, very small sample sizes, and those only tangentially related to our aim. Disagreements regarding study inclusion were resolved by consensus. Drs. Lai, Barsuk, and Mourad identified additional studies by hand review of reference lists and content experts (Figure 1).
Conceptual Framework
All selected articles were categorized by temporal relationship to thoracentesis as pre-, intra-, or postprocedure. Pre-procedural topics were those outcomes that had been identified and addressed before attempting thoracentesis, such as physician training or perceived risks of harm. Intraprocedural considerations included aspects such as use of bedside US, pleural manometry, and large-volume drainage. Finally, postprocedural factors were those related to evaluation after thoracentesis, such as follow-up imaging. This conceptual framework is outlined in Figure 2.
RESULTS
The PubMed search returned a total of 1170 manuscripts, of which 56 articles met inclusion criteria. Four additional articles were identified by experts and included in the study.4-7 Therefore, 60 articles were identified and included in this review. Study designs included cohort studies, case control studies, systematic reviews, meta-analyses, narrative reviews, consensus guidelines, and randomized controlled trials. A summary of all included articles by topic can be found in the Table.
PRE-PROCEDURAL CONSIDERATIONS
Physician Training
Studies indicate that graduate medical education may not adequately prepare clinicians to perform thoracentesis.8 In fact, residents have the least exposure and confidence in performing thoracentesis when compared to other bedside procedures.9,10 In 1 survey, 69% of medical trainees desired more exposure to procedures, and 98% felt that procedural skills were important to master.11 Not surprisingly, then, graduating internal medicine residents perform poorly when assessed on a thoracentesis simulator.12
Supplemental training outside of residency is useful to develop and maintain skills for thoracentesis, such as simulation with direct observation in a zero-risk environment. In 1 study, “simulation-based mastery learning” combined an educational video presentation with repeated, deliberate practice on a simulator until procedural competence was acquired, over two 2-hour sessions. In this study, 40 third-year medicine residents demonstrated a 71% improvement in clinical skills performance after course completion, with 93% achieving a passing score. The remaining 7% also achieved passing scores with extra practice time.12 Others have built upon the concept of simulation-based training. For instance, 2 studies suggest that use of a simulation-based curriculum improved both thoracentesis knowledge and performance skills in a 3-hour session.13,14 Similarly, 1 prospective study reported that a half-day thoracentesis workshop using simulation and 1:1 direct observation successfully lowered pneumothorax rates from 8.6% to 1.8% in a group of practicing clinicians. Notably, additional interventions including use of bedside US, limiting operators to a focused group, and standardization of equipment were also a part of this quality improvement initiative.7 Although repetition is required to gain proficiency when using a simulator, performance and confidence appear to plateau with only 4 simulator trials. In medical students, improvements derived through simulator-based teaching were sustained when retested 6 months following training.15
An instrument to ensure competency is necessary, given variability in procedural experience among both new graduates and practicing physicians,. Our search did not identify any clinically validated tools that adequately assessed thoracentesis performance. However, some have been proposed16 and 1 validated in a simulation environment.12 Regarding the incorporation of US for effusion markup, 1 validated tool used an 11-domain assessment covering knowledge of US machine manipulation, recognition of images with common pleural effusion characteristics, and performance of thoracic US with puncture-site marking on a simulator. When used on 22 participants, scores with the tool could reliably differentiate between novice, intermediate, and advanced groups (P < 0.0001).17
Patient Selection
Coagulopathies and Anticoagulation. Historically, the accepted cutoff for performing thoracentesis is an international normalized ratio (INR) less than 1.5 and a platelet count greater than 50,000/µL. McVay et al.18 first showed in 1991 that use of these cutoffs was associated with low rates of periprocedural bleeding, leading to endorsement in the British Thoracic Society (BTS) Pleural Disease Guideline 2010.19 Other recommendations include the 2012 Society for Interventional Radiology guidelines that endorse correction of an INR greater than 2, or platelets less than 50,000/µL, based almost exclusively on expert opinion.5
However, data suggest that thoracentesis may be safely performed outside these parameters. For instance, a prospective study of approximately 9000 thoracenteses over 12 years found that patients with an INR of 1.5-2.9 or platelets of 20,000 - 49,000/µL experienced rates of bleeding complications similar to those with normal values.20 Similarly, a 2014 review21 found that the overall risk of hemorrhage during thoracentesis in the setting of moderate coagulopathy (defined as an INR of 1.5 - 3 or platelets of 25,000-50,000/µL), was not increased. In 1 retrospective study of more than 1000 procedures, no differences in hemorrhagic events were noted in patients with bleeding diatheses that received prophylactic fresh frozen plasma or platelets vs. those who did not.22 Of note, included studies used a variety of criteria to define a hemorrhagic complication, which included: an isolated 2 g/dL or more decrement in hemoglobin, presence of bloody fluid on repeat tap with associated hemoglobin decrement, rapid re-accumulation of fluid with a hemoglobin decrement, or transfusion of 2 units or more of whole blood.
Whether it is safe to perform thoracentesis on patients taking antiplatelet therapy is less well understood. Although data are limited, a few small-scale studies23,24 suggest that hemorrhagic complications following thoracentesis in patients receiving clopidogrel are comparable to the general population. We found no compelling data regarding the safety of thoracentesis in the setting of direct oral anticoagulants, heparin, low-molecular weight heparin, or intravenous direct thrombin inhibitors. Current practice is to generally avoid thoracentesis while these therapeutic anticoagulants are used.
Invasive mechanical ventilation. Pleural effusion is common in patients in the intensive care unit, including those requiring mechanical ventilation.25 Thoracentesis in this population is clinically important: fluid analysis in 1 study was shown to aid the diagnosis in 45% of cases and changes in treatment in 33%.26 However, clinicians may be reluctant to perform thoracentesis on patients who require mechanical ventilation, given the perception of a greater risk of pneumothorax from positive pressure ventilation.
Despite this concern, a 2011 meta-analysis including 19 studies and more than 1100 patients revealed rates of pneumothorax and hemothorax comparable to nonventilated patients.25 Furthermore, a 2015 prospective study that examined thoracentesis in 1377 mechanically ventilated patients revealed no difference in complication rates as well.20 Therefore, evidence suggests that performance of thoracentesis in mechanically ventilated patients is not contraindicated.
Skin Disinfection and Antisepsis Precautions
The 2010 BTS guidelines list empyema and wound infection as possible complications of thoracentesis.19 However, no data regarding incidence are provided. Additionally, an alcohol-based skin cleanser (such as 2% chlorhexidine gluconate/70% isopropyl alcohol), along with sterile gloves, field, and dressing are suggested as precautionary measures.19 In 1 single-center registry of 2489 thoracenteses performed using alcohol or iodine-based antiseptic and sterile drapes, no postprocedure infections were identified.27 Of note, we did not find other studies (including case reports) that reported either incidence or rate of infectious complications such as wound infection and empyema. In an era of modern skin antiseptics that have effectively reduced complications such as catheter-related bloodstream infection,28 the incidence of this event is thus likely to be low.
INTRAPROCEDURAL CONSIDERATIONS
Use of Bedside Ultrasound
Portable US has particular advantages for evaluation of pleural effusion vs other imaging modalities. Compared with computerized tomography (CT), bedside US offers similar performance but is less costly, avoids both radiation exposure and need for patient transportation, and provides results instantaneously.29,30 Compared to chest x-ray (CXR), US is more sensitive at detecting the presence, volume, and characteristics of pleural fluid30,31 and can be up to 100% sensitive for effusions greater than 100 mL.29 Furthermore, whereas CXR typically requires 200 mL of fluid to be present for detection of an effusion, US can reliably detect as little as 20 mL of fluid.29 When US was used to confirm thoracentesis puncture sites in a study involving 30 physicians of varying experience and 67 consecutive patients, 15% of sites found by clinical exam were inaccurate (less than 10 mm fluid present), 10% were at high risk for organ puncture, and a suitable fluid pocket was found 54% of times when exam could not.4
A 2010 meta-analysis of 24 studies and 6605 thoracenteses estimated the overall rate of pneumothorax at 6%; however, procedures performed with US guidance were associated with a 70% reduced risk of this event (odds ratio, 0.30; 95% confidence interval, 0.20 - 0.70).32 In a 2014 randomized control trial of 160 patients that compared thoracentesis with US guidance for site marking vs no US use, 10 pneumothoraces occurred in the control group vs 1 in the US group (12.5% vs 1.25%, P = 0.009).33 Similarly, another retrospective review of 445 consecutive patients with malignant effusions revealed a pneumothorax rate of 0.97% using US in real time during needle insertion compared to 8.89% for unguided thoracenteses (P < 0.0001).34 Several other studies using US guidance for either site markup or in real time reported similar pneumothorax rates, ranging from 1.1% - 4.8%.35-37 However, it is unclear if real-time US specifically provides an additive effect vs site marking alone, as no studies directly comparing the 2 methods were found.
Benefits of US also include a higher rate of procedural success, with 1 study demonstrating a 99% success rate when using US vs. 90% without (P = 0.030).33 A larger volume of fluid removed has been observed with US use as well, and methods have been described using fluid-pocket depth to guide puncture site localization and maximize drainage.38 Finally, US use for thoracentesis has been associated with lower costs and length of stay.39,40
Intercostal Artery Localization
Although rare (incidence, 0.18%-2%20,21,39), the occurrence of hemothorax following thoracentesis is potentially catastrophic. This serious complication is often caused by laceration of the intercostal artery (ICA) or 1 of its branches during needle insertion.41
While risk of injury is theoretically reduced by needle insertion superior to the rib, studies using cadaver dissection and 3D angiography show significant tortuosity of the ICA.6,41-43 The degree of tortuosity is increased within 6 cm of the midline, in more cephalad rib spaces, and in the elderly (older than 60 years).41-43 Furthermore, 1 cadaveric study also demonstrated the presence of arterial collaterals branching off the ICA at multiple intercostal spaces, ranging between 8 cm and 11 cm from the midline.41 This anatomic variability may explain why some have observed low complication and hemothorax rates with an extreme lateral approach.35 Bedside US with color flow Doppler imaging has been used to identify the ICA, with 88% sensitivity compared to CT imaging while adding little to exam time.44,45 Of note, a 37% drop in the rate of hemothorax was observed in 1 study with routine US guidance alone.39
Pleural Pressure Monitoring and Large-Volume Thoracentesis
While normal intrapleural pressures are approximately -5 to -10 cm H2O,46 the presence of a pleural effusion creates a complex interaction between fluid, compressed lung, and chest wall that can increase these pressures.47 During drainage of an effusion, pleural pressures may rapidly drop, provoking re-expansion pulmonary edema (REPE). While rare (0 -1%), clinically-diagnosed REPE is a serious complication that can lead to rapid respiratory failure and death.20,48 REPE is postulated to be caused by increased capillary permeability resulting from inflammation, driven by rapid re-inflation of the lung when exposed to highly negative intrapleural pressures.47,49
Measurement of intrapleural pressure using a water manometer during thoracentesis may minimize REPE by terminating fluid drainage when intrapleural pressure begins to drop rapidly.50,51 A cutoff of -20 cm H2O has been cited repeatedly as safe since being suggested by Light in 1980, but this is based on animal models.50,52 In 1 prospective study of 185 thoracenteses in which manometry was performed, 15% of patients had intrapleural pressure drop to less than -20 cm H2O (at which point the procedure was terminated) but suffered no REPE.50
Manometry is valuable in the identification of an unexpandable or trapped lung when pleural pressures drop rapidly with only minimal fluid volume removal.47,53 Other findings correlated with an unexpandable lung include a negative opening pressure47 and large fluctuations in pressure during the respiratory cycle.54
While development of symptoms (eg, chest pain, cough, or dyspnea) is often used as a surrogate, the correlation between intrapleural pressure and patient symptoms is inconsistent and not a reliable proxy.55 One study found that 22% of patients with chest pain during thoracentesis had intrapleural pressures lower than -20 cm H2O compared with 8.6% of asymptomatic patients,56 but it is unclear if the association is causal.
Thoracentesis is often performed for symptomatic relief and removal of large fluid volume. However, it remains common to halt fluid removal after 1.5 L, a threshold endorsed by BTS.19 While some investigators have suggested that removal of 2 L or more of pleural fluid does not compromise safety,57,58 a 4- to 5-fold rise in the risk of pneumothorax was noted in 2 studies.20,59 when more than 1.5 L of fluid was removed. The majority of these may be related to pneumothorax ex vacuo, a condition in which fluid is drained from the chest, but the lung is unable to expand and fill the space (eg, “trapped lung”), resulting in a persistent pneumothorax. This condition generally does not require treatment.60 When manometry is employed at 200-mL intervals with termination at an intrapleural pressure of less than 20 mm H2O, drainage of 3 L or more has been reported with low rates of pneumothorax and very low rates of REPE.50,51 However, whether this is cause and effect is unknown because REPE is rare, and more work is needed to determine the role of manometry for its prevention.
POSTPROCEDURAL CONSIDERATIONS
Postprocedure Imaging
Performing an upright CXR following thoracentesis is a practice that remains routinely done by many practitioners to monitor for complications. Such imaging was also endorsed by the American Thoracic Society guidelines.61 However, more recent data question the utility of this practice. Multiple studies have confirmed that post-thoracentesis CXR is unnecessary unless clinical suspicion for pneumothorax or REPE is present.36,58,62,63 The BTS guidelines also advocate this approach.19 Interestingly, a potentially more effective way to screen for postprocedure complications is through bedside US, which has been shown to be more sensitive than CXR in detecting pneumothorax.64 In 1 study of 185 patients, bedside US demonstrated a sensitivity of 88% and a specificity of 97% for diagnosing pneumothorax in patients with adequate quality scans, with positive and negative likelihood ratios of 55 and 0.17, respectively.65
DISCUSSION
Thoracentesis remains a core procedural skill for hospitalists, critical care physicians, and emergency physicians. It is the foundational component when investigating and treating pleural effusions. When the most current training, techniques, and technology are used, data suggest this procedure is safe to perform at the bedside. Our review highlights these strategies and evaluates which aspects might be most applicable to clinical practice.
Our findings have several implications for those who perform this procedure. First, appropriate training is central to procedural safety, and both simulation and direct observation by procedural experts have been shown by multiple investigators to improve knowledge and skill. This training should integrate the use of US in performing a focused thoracic exam.
Second, recommendations regarding coagulopathy and a “safe cutoff” of an INR less than 1.5 or platelets greater than 50,000/µL had limited evidentiary support. Rather, multiple studies suggest no difference in bleeding risk following thoracentesis with an INR as high as 3.0 and platelets greater than 25,000/µL. Furthermore, prophylactic transfusion with fresh frozen plasma or platelets before thoracentesis did not alter bleeding risk and exposes patients to transfusion complications. Thus, routine use of this practice can no longer be recommended. Third, further research is needed to understand the bleeding risk for patients on antiplatelet medications, heparin products, and also direct oral anticoagulants, given the growing popularity in their use and the potential consequences of even temporary cessation. Regarding patients on mechanical ventilation, thoracentesis demonstrated no difference in complication rates vs. the general population, and its performance in this population is encouraged when clinically indicated.
Intraprocedural considerations include the use of bedside US. Due to multiple benefits including effusion characterization, puncture site localization, and significantly lower rates of pneumothorax, the standard of care should be to perform thoracentesis with US guidance. Both use of US to mark an effusion immediately prior to puncture or in real time during needle insertion demonstrated benefit; however, it is unclear if 1 method is superior because no direct comparison studies were found. Further work is needed to investigate this potential.
Our review suggests that the location and course of the ICA is variable, especially near the midline, in the elderly, and in higher intercostal spaces, leaving it vulnerable to laceration. We recommend physicians only attempt thoracentesis at least 6 cm lateral to the midline due to ICA tortuosity and, ideally, 12 cm lateral, to avoid the presence of collaterals. Although only 2 small-scale studies were found pertaining to the use of US in identifying the ICA, we encourage physicians to consider learning how to screen for its presence as a part of their routine thoracic US exam in the area underlying the planned puncture site.
Manometry is beneficial because it can diagnose a nonexpandable lung and allows for pleural pressure monitoring.52,53 A simple U-shaped manometer can be constructed from intravenous tubing included in most thoracentesis kits, which adds little to overall procedure time. While low rates of REPE have been observed when terminating thoracentesis if pressures drop below -20 cm H2O or chest pain develops, neither measure appears to have reliable predictive value, limiting clinical utility. Further work is required to determine if a “safe pressure cutoff” exists. In general, we recommend the use of manometry when a nonexpandable (trapped) lung is suspected, because large drops in intrapleural pressure, a negative opening pressure, and respiratory variation can help confirm the diagnosis and avoid pneumothorax ex vacuo or unnecessary procedures in the future. As this condition appears to be more common in the setting of larger effusions, use of manometry when large-volume thoracenteses are planned is also reasonable.
Postprocedurally, routine imaging after thoracentesis is not recommended unless there is objective concern for complication. When indicated, bedside US is better positioned for this role compared with CXR, because it is more sensitive in detecting pneumothorax, provides instantaneous results, and avoids radiation exposure.
Our review has limitations. First, we searched only for articles between defined time periods, restricted our search to a single database, and excluded non-English articles. This has the potential to introduce selection bias, as nonprimary articles that fall within our time restrictions may cite older studies that are outside our search range. To minimize this effect, we performed a critical review of all included studies, especially nonprimary articles. Second, despite the focus of our search strategy to identify any articles related to patient safety and adverse events, we cannot guarantee that all relevant articles for any particular complication or risk factor were captured given the lack of more specific search terms. Third, although we performed a systematic search of the literature, we did not perform a formal systematic review or formally grade included studies. As the goal of our review was to categorize and operationalize clinical aspects, this approach was necessary, and we acknowledge that the quality of studies is variable. Lastly, we aimed to generate clinical recommendations for physicians performing thoracentesis at the bedside; others reviewing this literature may find or emphasize different aspects relevant to practice outside this setting.
In conclusion, evaluation and treatment of pleural effusions with bedside thoracentesis is an important skill for physicians of many disciplines. The evidence presented in this review will help inform the process and ensure patient safety. Physicians should consider incorporating these recommendations into their practice.
Acknowledgments
The authors thank Whitney Townsend, MLIS, health sciences informationist, for assistance with serial literature searches.
Disclosure
Nothing to report.
Pleural effusion can occur in myriad conditions including infection, heart failure, liver disease, and cancer.1 Consequently, physicians from many disciplines routinely encounter both inpatients and outpatients with this diagnosis. Often, evaluation and treatment require thoracentesis to obtain fluid for analysis or symptom relief.
Although historically performed at the bedside without imaging guidance or intraprocedural monitoring, thoracentesis performed in this fashion carries considerable risk of complications. In fact, it has 1 of the highest rates of iatrogenic pneumothorax among bedside procedures.2 However, recent advances in practice and adoption of newer technologies have helped to mitigate risks associated with this procedure. These advances are relevant because approximately 50% of thoracenteses are still performed at the bedside.3 In this review, we aim to identify the most recent key practices that enhance the safety and the effectiveness of thoracentesis for practicing clinicians.
METHODS
Information Sources and Search Strategy
With the assistance of a research librarian, we performed a systematic search of PubMed-indexed articles from January 1, 2000 to September 30, 2015. Articles were identified using search terms such as thoracentesis, pleural effusion, safety, medical error, adverse event, and ultrasound in combination with Boolean operators. Of note, as thoracentesis is indexed as a subgroup of paracentesis in PubMed, this term was also included to increase the sensitivity of the search. The full search strategy is available in the Appendix. Any references cited in this review outside of the date range of our search are provided only to give relevant background information or establish the origin of commonly performed practices.
Study Eligibility and Selection Criteria
Studies were included if they reported clinical aspects related to thoracentesis. We defined clinical aspects as those strategies that focused on operator training, procedural techniques, technology, management, or prevention of complications. Non-English language articles, animal studies, case reports, conference proceedings, and abstracts were excluded. As our intention was to focus on the contemporary advances related to thoracentesis performance, (eg, ultrasound [US]), our search was limited to studies published after the year 2000. Two authors, Drs. Schildhouse and Lai independently screened studies to determine inclusion, excluding studies with weak methodology, very small sample sizes, and those only tangentially related to our aim. Disagreements regarding study inclusion were resolved by consensus. Drs. Lai, Barsuk, and Mourad identified additional studies by hand review of reference lists and content experts (Figure 1).
Conceptual Framework
All selected articles were categorized by temporal relationship to thoracentesis as pre-, intra-, or postprocedure. Pre-procedural topics were those outcomes that had been identified and addressed before attempting thoracentesis, such as physician training or perceived risks of harm. Intraprocedural considerations included aspects such as use of bedside US, pleural manometry, and large-volume drainage. Finally, postprocedural factors were those related to evaluation after thoracentesis, such as follow-up imaging. This conceptual framework is outlined in Figure 2.
RESULTS
The PubMed search returned a total of 1170 manuscripts, of which 56 articles met inclusion criteria. Four additional articles were identified by experts and included in the study.4-7 Therefore, 60 articles were identified and included in this review. Study designs included cohort studies, case control studies, systematic reviews, meta-analyses, narrative reviews, consensus guidelines, and randomized controlled trials. A summary of all included articles by topic can be found in the Table.
PRE-PROCEDURAL CONSIDERATIONS
Physician Training
Studies indicate that graduate medical education may not adequately prepare clinicians to perform thoracentesis.8 In fact, residents have the least exposure and confidence in performing thoracentesis when compared to other bedside procedures.9,10 In 1 survey, 69% of medical trainees desired more exposure to procedures, and 98% felt that procedural skills were important to master.11 Not surprisingly, then, graduating internal medicine residents perform poorly when assessed on a thoracentesis simulator.12
Supplemental training outside of residency is useful to develop and maintain skills for thoracentesis, such as simulation with direct observation in a zero-risk environment. In 1 study, “simulation-based mastery learning” combined an educational video presentation with repeated, deliberate practice on a simulator until procedural competence was acquired, over two 2-hour sessions. In this study, 40 third-year medicine residents demonstrated a 71% improvement in clinical skills performance after course completion, with 93% achieving a passing score. The remaining 7% also achieved passing scores with extra practice time.12 Others have built upon the concept of simulation-based training. For instance, 2 studies suggest that use of a simulation-based curriculum improved both thoracentesis knowledge and performance skills in a 3-hour session.13,14 Similarly, 1 prospective study reported that a half-day thoracentesis workshop using simulation and 1:1 direct observation successfully lowered pneumothorax rates from 8.6% to 1.8% in a group of practicing clinicians. Notably, additional interventions including use of bedside US, limiting operators to a focused group, and standardization of equipment were also a part of this quality improvement initiative.7 Although repetition is required to gain proficiency when using a simulator, performance and confidence appear to plateau with only 4 simulator trials. In medical students, improvements derived through simulator-based teaching were sustained when retested 6 months following training.15
An instrument to ensure competency is necessary, given variability in procedural experience among both new graduates and practicing physicians,. Our search did not identify any clinically validated tools that adequately assessed thoracentesis performance. However, some have been proposed16 and 1 validated in a simulation environment.12 Regarding the incorporation of US for effusion markup, 1 validated tool used an 11-domain assessment covering knowledge of US machine manipulation, recognition of images with common pleural effusion characteristics, and performance of thoracic US with puncture-site marking on a simulator. When used on 22 participants, scores with the tool could reliably differentiate between novice, intermediate, and advanced groups (P < 0.0001).17
Patient Selection
Coagulopathies and Anticoagulation. Historically, the accepted cutoff for performing thoracentesis is an international normalized ratio (INR) less than 1.5 and a platelet count greater than 50,000/µL. McVay et al.18 first showed in 1991 that use of these cutoffs was associated with low rates of periprocedural bleeding, leading to endorsement in the British Thoracic Society (BTS) Pleural Disease Guideline 2010.19 Other recommendations include the 2012 Society for Interventional Radiology guidelines that endorse correction of an INR greater than 2, or platelets less than 50,000/µL, based almost exclusively on expert opinion.5
However, data suggest that thoracentesis may be safely performed outside these parameters. For instance, a prospective study of approximately 9000 thoracenteses over 12 years found that patients with an INR of 1.5-2.9 or platelets of 20,000 - 49,000/µL experienced rates of bleeding complications similar to those with normal values.20 Similarly, a 2014 review21 found that the overall risk of hemorrhage during thoracentesis in the setting of moderate coagulopathy (defined as an INR of 1.5 - 3 or platelets of 25,000-50,000/µL), was not increased. In 1 retrospective study of more than 1000 procedures, no differences in hemorrhagic events were noted in patients with bleeding diatheses that received prophylactic fresh frozen plasma or platelets vs. those who did not.22 Of note, included studies used a variety of criteria to define a hemorrhagic complication, which included: an isolated 2 g/dL or more decrement in hemoglobin, presence of bloody fluid on repeat tap with associated hemoglobin decrement, rapid re-accumulation of fluid with a hemoglobin decrement, or transfusion of 2 units or more of whole blood.
Whether it is safe to perform thoracentesis on patients taking antiplatelet therapy is less well understood. Although data are limited, a few small-scale studies23,24 suggest that hemorrhagic complications following thoracentesis in patients receiving clopidogrel are comparable to the general population. We found no compelling data regarding the safety of thoracentesis in the setting of direct oral anticoagulants, heparin, low-molecular weight heparin, or intravenous direct thrombin inhibitors. Current practice is to generally avoid thoracentesis while these therapeutic anticoagulants are used.
Invasive mechanical ventilation. Pleural effusion is common in patients in the intensive care unit, including those requiring mechanical ventilation.25 Thoracentesis in this population is clinically important: fluid analysis in 1 study was shown to aid the diagnosis in 45% of cases and changes in treatment in 33%.26 However, clinicians may be reluctant to perform thoracentesis on patients who require mechanical ventilation, given the perception of a greater risk of pneumothorax from positive pressure ventilation.
Despite this concern, a 2011 meta-analysis including 19 studies and more than 1100 patients revealed rates of pneumothorax and hemothorax comparable to nonventilated patients.25 Furthermore, a 2015 prospective study that examined thoracentesis in 1377 mechanically ventilated patients revealed no difference in complication rates as well.20 Therefore, evidence suggests that performance of thoracentesis in mechanically ventilated patients is not contraindicated.
Skin Disinfection and Antisepsis Precautions
The 2010 BTS guidelines list empyema and wound infection as possible complications of thoracentesis.19 However, no data regarding incidence are provided. Additionally, an alcohol-based skin cleanser (such as 2% chlorhexidine gluconate/70% isopropyl alcohol), along with sterile gloves, field, and dressing are suggested as precautionary measures.19 In 1 single-center registry of 2489 thoracenteses performed using alcohol or iodine-based antiseptic and sterile drapes, no postprocedure infections were identified.27 Of note, we did not find other studies (including case reports) that reported either incidence or rate of infectious complications such as wound infection and empyema. In an era of modern skin antiseptics that have effectively reduced complications such as catheter-related bloodstream infection,28 the incidence of this event is thus likely to be low.
INTRAPROCEDURAL CONSIDERATIONS
Use of Bedside Ultrasound
Portable US has particular advantages for evaluation of pleural effusion vs other imaging modalities. Compared with computerized tomography (CT), bedside US offers similar performance but is less costly, avoids both radiation exposure and need for patient transportation, and provides results instantaneously.29,30 Compared to chest x-ray (CXR), US is more sensitive at detecting the presence, volume, and characteristics of pleural fluid30,31 and can be up to 100% sensitive for effusions greater than 100 mL.29 Furthermore, whereas CXR typically requires 200 mL of fluid to be present for detection of an effusion, US can reliably detect as little as 20 mL of fluid.29 When US was used to confirm thoracentesis puncture sites in a study involving 30 physicians of varying experience and 67 consecutive patients, 15% of sites found by clinical exam were inaccurate (less than 10 mm fluid present), 10% were at high risk for organ puncture, and a suitable fluid pocket was found 54% of times when exam could not.4
A 2010 meta-analysis of 24 studies and 6605 thoracenteses estimated the overall rate of pneumothorax at 6%; however, procedures performed with US guidance were associated with a 70% reduced risk of this event (odds ratio, 0.30; 95% confidence interval, 0.20 - 0.70).32 In a 2014 randomized control trial of 160 patients that compared thoracentesis with US guidance for site marking vs no US use, 10 pneumothoraces occurred in the control group vs 1 in the US group (12.5% vs 1.25%, P = 0.009).33 Similarly, another retrospective review of 445 consecutive patients with malignant effusions revealed a pneumothorax rate of 0.97% using US in real time during needle insertion compared to 8.89% for unguided thoracenteses (P < 0.0001).34 Several other studies using US guidance for either site markup or in real time reported similar pneumothorax rates, ranging from 1.1% - 4.8%.35-37 However, it is unclear if real-time US specifically provides an additive effect vs site marking alone, as no studies directly comparing the 2 methods were found.
Benefits of US also include a higher rate of procedural success, with 1 study demonstrating a 99% success rate when using US vs. 90% without (P = 0.030).33 A larger volume of fluid removed has been observed with US use as well, and methods have been described using fluid-pocket depth to guide puncture site localization and maximize drainage.38 Finally, US use for thoracentesis has been associated with lower costs and length of stay.39,40
Intercostal Artery Localization
Although rare (incidence, 0.18%-2%20,21,39), the occurrence of hemothorax following thoracentesis is potentially catastrophic. This serious complication is often caused by laceration of the intercostal artery (ICA) or 1 of its branches during needle insertion.41
While risk of injury is theoretically reduced by needle insertion superior to the rib, studies using cadaver dissection and 3D angiography show significant tortuosity of the ICA.6,41-43 The degree of tortuosity is increased within 6 cm of the midline, in more cephalad rib spaces, and in the elderly (older than 60 years).41-43 Furthermore, 1 cadaveric study also demonstrated the presence of arterial collaterals branching off the ICA at multiple intercostal spaces, ranging between 8 cm and 11 cm from the midline.41 This anatomic variability may explain why some have observed low complication and hemothorax rates with an extreme lateral approach.35 Bedside US with color flow Doppler imaging has been used to identify the ICA, with 88% sensitivity compared to CT imaging while adding little to exam time.44,45 Of note, a 37% drop in the rate of hemothorax was observed in 1 study with routine US guidance alone.39
Pleural Pressure Monitoring and Large-Volume Thoracentesis
While normal intrapleural pressures are approximately -5 to -10 cm H2O,46 the presence of a pleural effusion creates a complex interaction between fluid, compressed lung, and chest wall that can increase these pressures.47 During drainage of an effusion, pleural pressures may rapidly drop, provoking re-expansion pulmonary edema (REPE). While rare (0 -1%), clinically-diagnosed REPE is a serious complication that can lead to rapid respiratory failure and death.20,48 REPE is postulated to be caused by increased capillary permeability resulting from inflammation, driven by rapid re-inflation of the lung when exposed to highly negative intrapleural pressures.47,49
Measurement of intrapleural pressure using a water manometer during thoracentesis may minimize REPE by terminating fluid drainage when intrapleural pressure begins to drop rapidly.50,51 A cutoff of -20 cm H2O has been cited repeatedly as safe since being suggested by Light in 1980, but this is based on animal models.50,52 In 1 prospective study of 185 thoracenteses in which manometry was performed, 15% of patients had intrapleural pressure drop to less than -20 cm H2O (at which point the procedure was terminated) but suffered no REPE.50
Manometry is valuable in the identification of an unexpandable or trapped lung when pleural pressures drop rapidly with only minimal fluid volume removal.47,53 Other findings correlated with an unexpandable lung include a negative opening pressure47 and large fluctuations in pressure during the respiratory cycle.54
While development of symptoms (eg, chest pain, cough, or dyspnea) is often used as a surrogate, the correlation between intrapleural pressure and patient symptoms is inconsistent and not a reliable proxy.55 One study found that 22% of patients with chest pain during thoracentesis had intrapleural pressures lower than -20 cm H2O compared with 8.6% of asymptomatic patients,56 but it is unclear if the association is causal.
Thoracentesis is often performed for symptomatic relief and removal of large fluid volume. However, it remains common to halt fluid removal after 1.5 L, a threshold endorsed by BTS.19 While some investigators have suggested that removal of 2 L or more of pleural fluid does not compromise safety,57,58 a 4- to 5-fold rise in the risk of pneumothorax was noted in 2 studies.20,59 when more than 1.5 L of fluid was removed. The majority of these may be related to pneumothorax ex vacuo, a condition in which fluid is drained from the chest, but the lung is unable to expand and fill the space (eg, “trapped lung”), resulting in a persistent pneumothorax. This condition generally does not require treatment.60 When manometry is employed at 200-mL intervals with termination at an intrapleural pressure of less than 20 mm H2O, drainage of 3 L or more has been reported with low rates of pneumothorax and very low rates of REPE.50,51 However, whether this is cause and effect is unknown because REPE is rare, and more work is needed to determine the role of manometry for its prevention.
POSTPROCEDURAL CONSIDERATIONS
Postprocedure Imaging
Performing an upright CXR following thoracentesis is a practice that remains routinely done by many practitioners to monitor for complications. Such imaging was also endorsed by the American Thoracic Society guidelines.61 However, more recent data question the utility of this practice. Multiple studies have confirmed that post-thoracentesis CXR is unnecessary unless clinical suspicion for pneumothorax or REPE is present.36,58,62,63 The BTS guidelines also advocate this approach.19 Interestingly, a potentially more effective way to screen for postprocedure complications is through bedside US, which has been shown to be more sensitive than CXR in detecting pneumothorax.64 In 1 study of 185 patients, bedside US demonstrated a sensitivity of 88% and a specificity of 97% for diagnosing pneumothorax in patients with adequate quality scans, with positive and negative likelihood ratios of 55 and 0.17, respectively.65
DISCUSSION
Thoracentesis remains a core procedural skill for hospitalists, critical care physicians, and emergency physicians. It is the foundational component when investigating and treating pleural effusions. When the most current training, techniques, and technology are used, data suggest this procedure is safe to perform at the bedside. Our review highlights these strategies and evaluates which aspects might be most applicable to clinical practice.
Our findings have several implications for those who perform this procedure. First, appropriate training is central to procedural safety, and both simulation and direct observation by procedural experts have been shown by multiple investigators to improve knowledge and skill. This training should integrate the use of US in performing a focused thoracic exam.
Second, recommendations regarding coagulopathy and a “safe cutoff” of an INR less than 1.5 or platelets greater than 50,000/µL had limited evidentiary support. Rather, multiple studies suggest no difference in bleeding risk following thoracentesis with an INR as high as 3.0 and platelets greater than 25,000/µL. Furthermore, prophylactic transfusion with fresh frozen plasma or platelets before thoracentesis did not alter bleeding risk and exposes patients to transfusion complications. Thus, routine use of this practice can no longer be recommended. Third, further research is needed to understand the bleeding risk for patients on antiplatelet medications, heparin products, and also direct oral anticoagulants, given the growing popularity in their use and the potential consequences of even temporary cessation. Regarding patients on mechanical ventilation, thoracentesis demonstrated no difference in complication rates vs. the general population, and its performance in this population is encouraged when clinically indicated.
Intraprocedural considerations include the use of bedside US. Due to multiple benefits including effusion characterization, puncture site localization, and significantly lower rates of pneumothorax, the standard of care should be to perform thoracentesis with US guidance. Both use of US to mark an effusion immediately prior to puncture or in real time during needle insertion demonstrated benefit; however, it is unclear if 1 method is superior because no direct comparison studies were found. Further work is needed to investigate this potential.
Our review suggests that the location and course of the ICA is variable, especially near the midline, in the elderly, and in higher intercostal spaces, leaving it vulnerable to laceration. We recommend physicians only attempt thoracentesis at least 6 cm lateral to the midline due to ICA tortuosity and, ideally, 12 cm lateral, to avoid the presence of collaterals. Although only 2 small-scale studies were found pertaining to the use of US in identifying the ICA, we encourage physicians to consider learning how to screen for its presence as a part of their routine thoracic US exam in the area underlying the planned puncture site.
Manometry is beneficial because it can diagnose a nonexpandable lung and allows for pleural pressure monitoring.52,53 A simple U-shaped manometer can be constructed from intravenous tubing included in most thoracentesis kits, which adds little to overall procedure time. While low rates of REPE have been observed when terminating thoracentesis if pressures drop below -20 cm H2O or chest pain develops, neither measure appears to have reliable predictive value, limiting clinical utility. Further work is required to determine if a “safe pressure cutoff” exists. In general, we recommend the use of manometry when a nonexpandable (trapped) lung is suspected, because large drops in intrapleural pressure, a negative opening pressure, and respiratory variation can help confirm the diagnosis and avoid pneumothorax ex vacuo or unnecessary procedures in the future. As this condition appears to be more common in the setting of larger effusions, use of manometry when large-volume thoracenteses are planned is also reasonable.
Postprocedurally, routine imaging after thoracentesis is not recommended unless there is objective concern for complication. When indicated, bedside US is better positioned for this role compared with CXR, because it is more sensitive in detecting pneumothorax, provides instantaneous results, and avoids radiation exposure.
Our review has limitations. First, we searched only for articles between defined time periods, restricted our search to a single database, and excluded non-English articles. This has the potential to introduce selection bias, as nonprimary articles that fall within our time restrictions may cite older studies that are outside our search range. To minimize this effect, we performed a critical review of all included studies, especially nonprimary articles. Second, despite the focus of our search strategy to identify any articles related to patient safety and adverse events, we cannot guarantee that all relevant articles for any particular complication or risk factor were captured given the lack of more specific search terms. Third, although we performed a systematic search of the literature, we did not perform a formal systematic review or formally grade included studies. As the goal of our review was to categorize and operationalize clinical aspects, this approach was necessary, and we acknowledge that the quality of studies is variable. Lastly, we aimed to generate clinical recommendations for physicians performing thoracentesis at the bedside; others reviewing this literature may find or emphasize different aspects relevant to practice outside this setting.
In conclusion, evaluation and treatment of pleural effusions with bedside thoracentesis is an important skill for physicians of many disciplines. The evidence presented in this review will help inform the process and ensure patient safety. Physicians should consider incorporating these recommendations into their practice.
Acknowledgments
The authors thank Whitney Townsend, MLIS, health sciences informationist, for assistance with serial literature searches.
Disclosure
Nothing to report.
1. Kasper DL. Harrison's Principles of Internal Medicine. 19th ed. New York, NY: McGraw Hill Education; 2015.
2. Celik B, Sahin E, Nadir A, Kaptanoglu M. Iatrogenic pneumothorax: etiology, incidence and risk factors. Thorac Cardiovasc Surg. 2009;57(5):286-290. PubMed
3. Hooper CE, Welham SA, Maskell NA, Soc BT. Pleural procedures and patient safety: a national BTS audit of practice. Thorax. 2015;70(2):189-191. PubMed
4. Diacon AH, Brutsche MH, Soler M. Accuracy of pleural puncture sites: a prospective comparison of clinical examination with ultrasound. Chest. 2003;123(2):436-441. PubMed
5. Patel IJ, Davidson JC, Nikolic B, et al. Consensus guidelines for periprocedural management of coagulation status and hemostasis risk in percutaneous image-guided interventions. J Vasc Interv Radiol. 2012;23(6):727-736. PubMed
6. Wraight WM, Tweedie DJ, Parkin IG. Neurovascular anatomy and variation in the fourth, fifth, and sixth intercostal spaces in the mid-axillary line: a cadaveric study in respect of chest drain insertion. Clin Anat. 2005;18(5):346-349. PubMed
7. Duncan DR, Morgenthaler TI, Ryu JH, Daniels CE. Reducing iatrogenic risk in thoracentesis: establishing best practice via experiential training in a zero-risk environment. Chest. 2009;135(5):1315-1320. PubMed
8. Grover S, Currier PF, Elinoff JM, Mouchantaf KJ, Katz JT, McMahon GT. Development of a test to evaluate residents' knowledge of medical procedures. J Hosp Med. 2009;4(7):430-432. PubMed
9. Promes SB, Chudgar SM, Grochowski CO, et al. Gaps in procedural experience and competency in medical school graduates. Acad Emerg Med. 2009;16 Suppl 2:S58-62. PubMed
10. Huang GC, Smith CC, Gordon CE, et al. Beyond the comfort zone: residents assess their comfort performing inpatient medical procedures. Am J Med. 2006;119(1):71 e17-24. PubMed
11. Lagan J, Cutts L, Zaidi S, Benton I, Rylance J. Are we failing our trainees in providing opportunities to attain procedural confidence? Br J Hosp Med (Lond). 2015;76(2):105-108. PubMed
12. Wayne DB, Barsuk JH, O'Leary KJ, Fudala MJ, McGaghie WC. Mastery learning of thoracentesis skills by internal medicine residents using simulation technology and deliberate practice. J Hosp Med. 2008;3(1):48-54. PubMed
13. Lenchus JD. End of the "see one, do one, teach one" era: the next generation of invasive bedside procedural instruction. J Am Osteopath Assoc. 2010;110(6):340-346. PubMed
14. Lenchus J, Issenberg SB, Murphy D, et al. A blended approach to invasive bedside procedural instruction. Med Teach. 2011;33(2):116-123. PubMed
15. Jiang G, Chen H, Wang S, et al. Learning curves and long-term outcome of simulation-based thoracentesis training for medical students. BMC Med Educ. 2011;11:39. PubMed
16. Berg D, Berg K, Riesenberg LA, et al. The development of a validated checklist for thoracentesis: preliminary results. Am J Med Qual. 2013;28(3):220-226. PubMed
17. Salamonsen M, McGrath D, Steiler G, Ware R, Colt H, Fielding D. A new instrument to assess physician skill at thoracic ultrasound, including pleural effusion markup. Chest. 2013;144(3):930-934. PubMed
18. McVay PA, Toy PT. Lack of increased bleeding after paracentesis and thoracentesis in patients with mild coagulation abnormalities. Transfusion. 1991;31(2):164-171. PubMed
19. Havelock T, Teoh R, Laws D, Gleeson F, Group BTSPDG. Pleural procedures and thoracic ultrasound: British Thoracic Society Pleural Disease Guideline 2010. Thorax. 2010;65 Suppl 2:ii61-76. PubMed
20. Ault MJ, Rosen BT, Scher J, Feinglass J, Barsuk JH. Thoracentesis outcomes: a 12-year experience. Thorax. 2015;70(2):127-132. PubMed
21. Puchalski J. Thoracentesis and the risks for bleeding: a new era. Curr Opin Pulm Med. 2014;20(4):377-384. PubMed
22. Hibbert RM, Atwell TD, Lekah A, et al. Safety of ultrasound-guided thoracentesis in patients with abnormal preprocedural coagulation parameters. Chest. 2013;144(2):456-463. PubMed
23. Zalt MB, Bechara RI, Parks C, Berkowitz DM. Effect of routine clopidogrel use on bleeding complications after ultrasound-guided thoracentesis. J Bronchology Interv Pulmonol. 2012;19(4):284-287. PubMed
24. Mahmood K, Shofer SL, Moser BK, Argento AC, Smathers EC, Wahidi MM. Hemorrhagic complications of thoracentesis and small-bore chest tube placement in patients taking clopidogrel. Ann Am Thorac Soc. 2014;11(1):73-79. PubMed
25. Goligher EC, Leis JA, Fowler RA, Pinto R, Adhikari NK, Ferguson ND. Utility and safety of draining pleural effusions in mechanically ventilated patients: a systematic review and meta-analysis. Crit Care. 2011;15(1):R46. PubMed
26. Fartoukh M, Azoulay E, Galliot R, et al. Clinically documented pleural effusions in medical ICU patients: how useful is routine thoracentesis? Chest. 2002;121(1):178-184. PubMed
27. Cervini P, Hesley GK, Thompson RL, Sampathkumar P, Knudsen JM. Incidence of infectious complications after an ultrasound-guided intervention. AJR Am J Roentgenol. 2010;195(4):846-850. PubMed
28. Mimoz O, Chopra V, Timsit JF. What's new in catheter-related infection: skin cleansing and skin antisepsis. Intensive Care Med. 2016;42(11):1784-1786. PubMed
29. Soni NJ, Franco R, Velez MI, et al. Ultrasound in the diagnosis and management of pleural effusions. J Hosp Med. 2015;10(12):811-816. PubMed
30. Feller-Kopman D. Ultrasound-guided thoracentesis. Chest. 2006;129(6):1709-1714. PubMed
31. Shojaee S, Argento AC. Ultrasound-guided pleural access. Semin Respir Crit Care Med. 2014;35(6):693-705. PubMed
32. Gordon CE, Feller-Kopman D, Balk EM, Smetana GW. Pneumothorax following thoracentesis: a systematic review and meta-analysis. Arch Intern Med. 2010;170(4):332-339. PubMed
33. Perazzo A, Gatto P, Barlascini C, Ferrari-Bravo M, Nicolini A. Can ultrasound guidance reduce the risk of pneumothorax following thoracentesis? J Bras Pneumol. 2014;40(1):6-12. PubMed
34. Cavanna L, Mordenti P, Berte R, et al. Ultrasound guidance reduces pneumothorax rate and improves safety of thoracentesis in malignant pleural effusion: report on 445 consecutive patients with advanced cancer. World J Surg Oncol. 2014;12:139. PubMed
35. Soldati G, Smargiassi A, Inchingolo R, Sher S, Valente S, Corbo GM. Ultrasound-guided pleural puncture in supine or recumbent lateral position - feasibility study. Multidiscip Respir Med. 2013;8(1):18. PubMed
36. Pihlajamaa K, Bode MK, Puumalainen T, Lehtimaki A, Marjelund S, Tikkakoski T. Pneumothorax and the value of chest radiography after ultrasound-guided thoracocentesis. Acta Radiol. 2004;45(8):828-832. PubMed
37. Barnes TW, Morgenthaler TI, Olson EJ, Hesley GK, Decker PA, Ryu JH. Sonographically guided thoracentesis and rate of pneumothorax. J Clin Ultrasound. 2005;33(9):442-446. PubMed
38. Zanforlin A, Gavelli G, Oboldi D, Galletti S. Ultrasound-guided thoracenthesis: the V-point as a site for optimal drainage positioning. Eur Rev Med Pharmacol Sci. 2013;17(1):25-28. PubMed
39. Patel PA, Ernst FR, Gunnarsson CL. Ultrasonography guidance reduces complications and costs associated with thoracentesis procedures. J Clin Ultrasound. 2012;40(3):135-141. PubMed
40. Mercaldi CJ, Lanes SF. Ultrasound guidance decreases complications and improves the cost of care among patients undergoing thoracentesis and paracentesis. Chest. 2013;143(2):532-538. PubMed
41. Shurtleff E, Olinger A. Posterior intercostal artery tortuosity and collateral branch points: a cadaveric study. Folia Morphol (Warsz). 2012;71(4):245-251. PubMed
42. Helm EJ, Rahman NM, Talakoub O, Fox DL, Gleeson FV. Course and variation of the intercostal artery by CT scan. Chest. 2013;143(3):634-639. PubMed
43. Yoneyama H, Arahata M, Temaru R, Ishizaka S, Minami S. Evaluation of the risk of intercostal artery laceration during thoracentesis in elderly patients by using 3D-CT angiography. Intern Med. 2010;49(4):289-292. PubMed
44. Salamonsen M, Ellis S, Paul E, Steinke K, Fielding D. Thoracic ultrasound demonstrates variable location of the intercostal artery. Respiration. 2012;83(4):323-329. PubMed
45. Salamonsen M, Dobeli K, McGrath D, et al. Physician-performed ultrasound can accurately screen for a vulnerable intercostal artery prior to chest drainage procedures. Respirology. 2013;18(6):942-947. PubMed
46. Grippi MA. Fishman's pulmonary diseases and disorders. Fifth edition. ed. New York: McGraw-Hill Education; 2015.
47. Huggins JT, Doelken P. Pleural manometry. Clin Chest Med. 2006;27(2):229-240. PubMed
48. Echevarria C, Twomey D, Dunning J, Chanda B. Does re-expansion pulmonary oedema exist? Interact Cardiovasc Thorac Surg. 2008;7(3):485-489. PubMed
49. Sue RD, Matthay MA, Ware LB. Hydrostatic mechanisms may contribute to the pathogenesis of human re-expansion pulmonary edema. Intensive Care Med. 2004;30(10):1921-1926. PubMed
50. Feller-Kopman D, Berkowitz D, Boiselle P, Ernst A. Large-volume thoracentesis and the risk of reexpansion pulmonary edema. Ann Thorac Surg. 2007;84(5):1656-1661. PubMed
51. Villena V, Lopez-Encuentra A, Pozo F, De-Pablo A, Martin-Escribano P. Measurement of pleural pressure during therapeutic thoracentesis. Am J Respir Crit Care Med. 2000;162(4 Pt 1):1534-1538. PubMed
52. Doelken P, Huggins JT, Pastis NJ, Sahn SA. Pleural manometry: technique and clinical implications. Chest. 2004;126(6):1764-1769. PubMed
53. Feller-Kopman D. Therapeutic thoracentesis: the role of ultrasound and pleural manometry. Curr Opin Pulm Med. 2007;13(4):312-318. PubMed
54. Boshuizen RC, Sinaasappel M, Vincent AD, Goldfinger V, Farag S, van den Heuvel MM. Pleural pressure swing and lung expansion after malignant pleural effusion drainage: the benefits of high-temporal resolution pleural manometry. J Bronchology Interv Pulmonol. 2013;20(3):200-205. PubMed
55. Pannu J, DePew ZS, Mullon JJ, Daniels CE, Hagen CE, Maldonado F. Impact of pleural manometry on the development of chest discomfort during thoracentesis: a symptom-based study. J Bronchology Interv Pulmonol. 2014;21(4):306-313. PubMed
56. Feller-Kopman D, Walkey A, Berkowitz D, Ernst A. The relationship of pleural pressure to symptom development during therapeutic thoracentesis. Chest. 2006;129(6):1556-1560. PubMed
57. Abunasser J, Brown R. Safety of large-volume thoracentesis. Conn Med. 2010;74(1):23-26. PubMed
58. Mynarek G, Brabrand K, Jakobsen JA, Kolbenstvedt A. Complications following ultrasound-guided thoracocentesis. Acta Radiol. 2004;45(5):519-522. PubMed
59. Josephson T, Nordenskjold CA, Larsson J, Rosenberg LU, Kaijser M. Amount drained at ultrasound-guided thoracentesis and risk of pneumothorax. Acta Radiol. 2009;50(1):42-47. PubMed
60. Heidecker J, Huggins JT, Sahn SA, Doelken P. Pathophysiology of pneumothorax following ultrasound-guided thoracentesis. Chest. 2006;130(4):1173-1184. PubMed
61. Sokolowski JW Jr, Burgher LW, Jones FL Jr, Patterson JR, Selecky PA. Guidelines for thoracentesis and needle biopsy of the pleura. This position paper of the American Thoracic Society was adopted by the ATS Board of Directors, June 1988. Am Rev Respir Dis. 1989;140(1):257-258. PubMed
62. Jones PW, Moyers JP, Rogers JT, Rodriguez RM, Lee YC, Light RW. Ultrasound-guided thoracentesis: is it a safer method? Chest. 2003;123(2):418-423. PubMed
63. Petersen WG, Zimmerman R. Limited utility of chest radiograph after thoracentesis. Chest. 2000;117(4):1038-1042. PubMed
64. Sachdeva A, Shepherd RW, Lee HJ. Thoracentesis and thoracic ultrasound: state of the art in 2013. Clin Chest Med. 2013;34(1):1-9. PubMed
65. Shostak E, Brylka D, Krepp J, Pua B, Sanders A. Bedside sonography for detection of postprocedure pneumothorax. J Ultrasound Med. 2013;32(6):1003-1009. PubMed
1. Kasper DL. Harrison's Principles of Internal Medicine. 19th ed. New York, NY: McGraw Hill Education; 2015.
2. Celik B, Sahin E, Nadir A, Kaptanoglu M. Iatrogenic pneumothorax: etiology, incidence and risk factors. Thorac Cardiovasc Surg. 2009;57(5):286-290. PubMed
3. Hooper CE, Welham SA, Maskell NA, Soc BT. Pleural procedures and patient safety: a national BTS audit of practice. Thorax. 2015;70(2):189-191. PubMed
4. Diacon AH, Brutsche MH, Soler M. Accuracy of pleural puncture sites: a prospective comparison of clinical examination with ultrasound. Chest. 2003;123(2):436-441. PubMed
5. Patel IJ, Davidson JC, Nikolic B, et al. Consensus guidelines for periprocedural management of coagulation status and hemostasis risk in percutaneous image-guided interventions. J Vasc Interv Radiol. 2012;23(6):727-736. PubMed
6. Wraight WM, Tweedie DJ, Parkin IG. Neurovascular anatomy and variation in the fourth, fifth, and sixth intercostal spaces in the mid-axillary line: a cadaveric study in respect of chest drain insertion. Clin Anat. 2005;18(5):346-349. PubMed
7. Duncan DR, Morgenthaler TI, Ryu JH, Daniels CE. Reducing iatrogenic risk in thoracentesis: establishing best practice via experiential training in a zero-risk environment. Chest. 2009;135(5):1315-1320. PubMed
8. Grover S, Currier PF, Elinoff JM, Mouchantaf KJ, Katz JT, McMahon GT. Development of a test to evaluate residents' knowledge of medical procedures. J Hosp Med. 2009;4(7):430-432. PubMed
9. Promes SB, Chudgar SM, Grochowski CO, et al. Gaps in procedural experience and competency in medical school graduates. Acad Emerg Med. 2009;16 Suppl 2:S58-62. PubMed
10. Huang GC, Smith CC, Gordon CE, et al. Beyond the comfort zone: residents assess their comfort performing inpatient medical procedures. Am J Med. 2006;119(1):71 e17-24. PubMed
11. Lagan J, Cutts L, Zaidi S, Benton I, Rylance J. Are we failing our trainees in providing opportunities to attain procedural confidence? Br J Hosp Med (Lond). 2015;76(2):105-108. PubMed
12. Wayne DB, Barsuk JH, O'Leary KJ, Fudala MJ, McGaghie WC. Mastery learning of thoracentesis skills by internal medicine residents using simulation technology and deliberate practice. J Hosp Med. 2008;3(1):48-54. PubMed
13. Lenchus JD. End of the "see one, do one, teach one" era: the next generation of invasive bedside procedural instruction. J Am Osteopath Assoc. 2010;110(6):340-346. PubMed
14. Lenchus J, Issenberg SB, Murphy D, et al. A blended approach to invasive bedside procedural instruction. Med Teach. 2011;33(2):116-123. PubMed
15. Jiang G, Chen H, Wang S, et al. Learning curves and long-term outcome of simulation-based thoracentesis training for medical students. BMC Med Educ. 2011;11:39. PubMed
16. Berg D, Berg K, Riesenberg LA, et al. The development of a validated checklist for thoracentesis: preliminary results. Am J Med Qual. 2013;28(3):220-226. PubMed
17. Salamonsen M, McGrath D, Steiler G, Ware R, Colt H, Fielding D. A new instrument to assess physician skill at thoracic ultrasound, including pleural effusion markup. Chest. 2013;144(3):930-934. PubMed
18. McVay PA, Toy PT. Lack of increased bleeding after paracentesis and thoracentesis in patients with mild coagulation abnormalities. Transfusion. 1991;31(2):164-171. PubMed
19. Havelock T, Teoh R, Laws D, Gleeson F, Group BTSPDG. Pleural procedures and thoracic ultrasound: British Thoracic Society Pleural Disease Guideline 2010. Thorax. 2010;65 Suppl 2:ii61-76. PubMed
20. Ault MJ, Rosen BT, Scher J, Feinglass J, Barsuk JH. Thoracentesis outcomes: a 12-year experience. Thorax. 2015;70(2):127-132. PubMed
21. Puchalski J. Thoracentesis and the risks for bleeding: a new era. Curr Opin Pulm Med. 2014;20(4):377-384. PubMed
22. Hibbert RM, Atwell TD, Lekah A, et al. Safety of ultrasound-guided thoracentesis in patients with abnormal preprocedural coagulation parameters. Chest. 2013;144(2):456-463. PubMed
23. Zalt MB, Bechara RI, Parks C, Berkowitz DM. Effect of routine clopidogrel use on bleeding complications after ultrasound-guided thoracentesis. J Bronchology Interv Pulmonol. 2012;19(4):284-287. PubMed
24. Mahmood K, Shofer SL, Moser BK, Argento AC, Smathers EC, Wahidi MM. Hemorrhagic complications of thoracentesis and small-bore chest tube placement in patients taking clopidogrel. Ann Am Thorac Soc. 2014;11(1):73-79. PubMed
25. Goligher EC, Leis JA, Fowler RA, Pinto R, Adhikari NK, Ferguson ND. Utility and safety of draining pleural effusions in mechanically ventilated patients: a systematic review and meta-analysis. Crit Care. 2011;15(1):R46. PubMed
26. Fartoukh M, Azoulay E, Galliot R, et al. Clinically documented pleural effusions in medical ICU patients: how useful is routine thoracentesis? Chest. 2002;121(1):178-184. PubMed
27. Cervini P, Hesley GK, Thompson RL, Sampathkumar P, Knudsen JM. Incidence of infectious complications after an ultrasound-guided intervention. AJR Am J Roentgenol. 2010;195(4):846-850. PubMed
28. Mimoz O, Chopra V, Timsit JF. What's new in catheter-related infection: skin cleansing and skin antisepsis. Intensive Care Med. 2016;42(11):1784-1786. PubMed
29. Soni NJ, Franco R, Velez MI, et al. Ultrasound in the diagnosis and management of pleural effusions. J Hosp Med. 2015;10(12):811-816. PubMed
30. Feller-Kopman D. Ultrasound-guided thoracentesis. Chest. 2006;129(6):1709-1714. PubMed
31. Shojaee S, Argento AC. Ultrasound-guided pleural access. Semin Respir Crit Care Med. 2014;35(6):693-705. PubMed
32. Gordon CE, Feller-Kopman D, Balk EM, Smetana GW. Pneumothorax following thoracentesis: a systematic review and meta-analysis. Arch Intern Med. 2010;170(4):332-339. PubMed
33. Perazzo A, Gatto P, Barlascini C, Ferrari-Bravo M, Nicolini A. Can ultrasound guidance reduce the risk of pneumothorax following thoracentesis? J Bras Pneumol. 2014;40(1):6-12. PubMed
34. Cavanna L, Mordenti P, Berte R, et al. Ultrasound guidance reduces pneumothorax rate and improves safety of thoracentesis in malignant pleural effusion: report on 445 consecutive patients with advanced cancer. World J Surg Oncol. 2014;12:139. PubMed
35. Soldati G, Smargiassi A, Inchingolo R, Sher S, Valente S, Corbo GM. Ultrasound-guided pleural puncture in supine or recumbent lateral position - feasibility study. Multidiscip Respir Med. 2013;8(1):18. PubMed
36. Pihlajamaa K, Bode MK, Puumalainen T, Lehtimaki A, Marjelund S, Tikkakoski T. Pneumothorax and the value of chest radiography after ultrasound-guided thoracocentesis. Acta Radiol. 2004;45(8):828-832. PubMed
37. Barnes TW, Morgenthaler TI, Olson EJ, Hesley GK, Decker PA, Ryu JH. Sonographically guided thoracentesis and rate of pneumothorax. J Clin Ultrasound. 2005;33(9):442-446. PubMed
38. Zanforlin A, Gavelli G, Oboldi D, Galletti S. Ultrasound-guided thoracenthesis: the V-point as a site for optimal drainage positioning. Eur Rev Med Pharmacol Sci. 2013;17(1):25-28. PubMed
39. Patel PA, Ernst FR, Gunnarsson CL. Ultrasonography guidance reduces complications and costs associated with thoracentesis procedures. J Clin Ultrasound. 2012;40(3):135-141. PubMed
40. Mercaldi CJ, Lanes SF. Ultrasound guidance decreases complications and improves the cost of care among patients undergoing thoracentesis and paracentesis. Chest. 2013;143(2):532-538. PubMed
41. Shurtleff E, Olinger A. Posterior intercostal artery tortuosity and collateral branch points: a cadaveric study. Folia Morphol (Warsz). 2012;71(4):245-251. PubMed
42. Helm EJ, Rahman NM, Talakoub O, Fox DL, Gleeson FV. Course and variation of the intercostal artery by CT scan. Chest. 2013;143(3):634-639. PubMed
43. Yoneyama H, Arahata M, Temaru R, Ishizaka S, Minami S. Evaluation of the risk of intercostal artery laceration during thoracentesis in elderly patients by using 3D-CT angiography. Intern Med. 2010;49(4):289-292. PubMed
44. Salamonsen M, Ellis S, Paul E, Steinke K, Fielding D. Thoracic ultrasound demonstrates variable location of the intercostal artery. Respiration. 2012;83(4):323-329. PubMed
45. Salamonsen M, Dobeli K, McGrath D, et al. Physician-performed ultrasound can accurately screen for a vulnerable intercostal artery prior to chest drainage procedures. Respirology. 2013;18(6):942-947. PubMed
46. Grippi MA. Fishman's pulmonary diseases and disorders. Fifth edition. ed. New York: McGraw-Hill Education; 2015.
47. Huggins JT, Doelken P. Pleural manometry. Clin Chest Med. 2006;27(2):229-240. PubMed
48. Echevarria C, Twomey D, Dunning J, Chanda B. Does re-expansion pulmonary oedema exist? Interact Cardiovasc Thorac Surg. 2008;7(3):485-489. PubMed
49. Sue RD, Matthay MA, Ware LB. Hydrostatic mechanisms may contribute to the pathogenesis of human re-expansion pulmonary edema. Intensive Care Med. 2004;30(10):1921-1926. PubMed
50. Feller-Kopman D, Berkowitz D, Boiselle P, Ernst A. Large-volume thoracentesis and the risk of reexpansion pulmonary edema. Ann Thorac Surg. 2007;84(5):1656-1661. PubMed
51. Villena V, Lopez-Encuentra A, Pozo F, De-Pablo A, Martin-Escribano P. Measurement of pleural pressure during therapeutic thoracentesis. Am J Respir Crit Care Med. 2000;162(4 Pt 1):1534-1538. PubMed
52. Doelken P, Huggins JT, Pastis NJ, Sahn SA. Pleural manometry: technique and clinical implications. Chest. 2004;126(6):1764-1769. PubMed
53. Feller-Kopman D. Therapeutic thoracentesis: the role of ultrasound and pleural manometry. Curr Opin Pulm Med. 2007;13(4):312-318. PubMed
54. Boshuizen RC, Sinaasappel M, Vincent AD, Goldfinger V, Farag S, van den Heuvel MM. Pleural pressure swing and lung expansion after malignant pleural effusion drainage: the benefits of high-temporal resolution pleural manometry. J Bronchology Interv Pulmonol. 2013;20(3):200-205. PubMed
55. Pannu J, DePew ZS, Mullon JJ, Daniels CE, Hagen CE, Maldonado F. Impact of pleural manometry on the development of chest discomfort during thoracentesis: a symptom-based study. J Bronchology Interv Pulmonol. 2014;21(4):306-313. PubMed
56. Feller-Kopman D, Walkey A, Berkowitz D, Ernst A. The relationship of pleural pressure to symptom development during therapeutic thoracentesis. Chest. 2006;129(6):1556-1560. PubMed
57. Abunasser J, Brown R. Safety of large-volume thoracentesis. Conn Med. 2010;74(1):23-26. PubMed
58. Mynarek G, Brabrand K, Jakobsen JA, Kolbenstvedt A. Complications following ultrasound-guided thoracocentesis. Acta Radiol. 2004;45(5):519-522. PubMed
59. Josephson T, Nordenskjold CA, Larsson J, Rosenberg LU, Kaijser M. Amount drained at ultrasound-guided thoracentesis and risk of pneumothorax. Acta Radiol. 2009;50(1):42-47. PubMed
60. Heidecker J, Huggins JT, Sahn SA, Doelken P. Pathophysiology of pneumothorax following ultrasound-guided thoracentesis. Chest. 2006;130(4):1173-1184. PubMed
61. Sokolowski JW Jr, Burgher LW, Jones FL Jr, Patterson JR, Selecky PA. Guidelines for thoracentesis and needle biopsy of the pleura. This position paper of the American Thoracic Society was adopted by the ATS Board of Directors, June 1988. Am Rev Respir Dis. 1989;140(1):257-258. PubMed
62. Jones PW, Moyers JP, Rogers JT, Rodriguez RM, Lee YC, Light RW. Ultrasound-guided thoracentesis: is it a safer method? Chest. 2003;123(2):418-423. PubMed
63. Petersen WG, Zimmerman R. Limited utility of chest radiograph after thoracentesis. Chest. 2000;117(4):1038-1042. PubMed
64. Sachdeva A, Shepherd RW, Lee HJ. Thoracentesis and thoracic ultrasound: state of the art in 2013. Clin Chest Med. 2013;34(1):1-9. PubMed
65. Shostak E, Brylka D, Krepp J, Pua B, Sanders A. Bedside sonography for detection of postprocedure pneumothorax. J Ultrasound Med. 2013;32(6):1003-1009. PubMed
© 2017 Society of Hospital Medicine
Hospital medicine and perioperative care: A framework for high-quality, high-value collaborative care
Of the 36 million US hospitalizations each year, 22% are surgical.1 Although less frequent than medical hospitalizations, surgical hospitalizations are more than twice as costly.2 Additionally, surgical hospitalizations are on average longer than medical hospitalizations.2 Given the increased scrutiny on cost and efficiency of care, attention has turned to optimizing perioperative care. Hospitalists are well positioned to provide specific expertise in the complex interdisciplinary medical management of surgical patients.
In recent decades, multiple models of hospitalist involvement in perioperative care have evolved across the United States.3-19 To consolidate knowledge and experience and to develop a framework for providing the best care for surgical patients, the Society of Hospital Medicine organized the Perioperative Care Work Group in 2015. This framework was designed for interdisciplinary collaboration in building and strengthening perioperative care programs.
METHODS
The Society of Hospital Medicine recognized hospital medicine programs’ need for guidance in developing collaborative care in perioperative medicine and appointed the Perioperative Care Work Group in May 2015. Work group members are perioperative medicine experts from US medical centers. They have extensive knowledge of the literature as well as administrative and clinical experience in a variety of perioperative care models.
Topic Development. Initial work was focused on reviewing and discussing multiple models of perioperative care and exploring the roles that hospital medicine physicians have within these models. Useful information was summarized to guide hospitals and physicians in designing, implementing, and expanding patient-centric perioperative medicine services with a focus on preoperative and postoperative care. A final document was created; it outlines system-level issues in perioperative care, organized by perioperative phases.
Initial Framework. Group members submitted written descriptions of key issues in each of 4 phases: (1) preoperative, (2) day of surgery, (3) postoperative inpatient, and (4) postdischarge. These descriptions were merged and reviewed by the content experts. Editing and discussion from the entire group were incorporated into the final matrix, which highlighted (1) perioperative phase definitions, (2) requirements for patients to move to next phase, (3) elements of care coordination typically provided by surgery, anesthesiology, and medicine disciplines, (4) concerns and risks particular to each phase, (5) unique considerations for each phase, (6) suggested metrics of success, and (7) key questions for determining the effectiveness of perioperative care in an institution. All members provided final evaluation and editing.
Final Approval. The Perioperative Care Matrix for Inpatient Surgeries (PCMIS) was presented to the board of the Society of Hospital Medicine in fall 2015 and was approved for use in centering and directing discussions regarding perioperative care.
Models of Care. The Perioperative Care Work Group surveyed examples of hospitalist engagement in perioperative care and synthesized these into synopses of existing models of care for the preoperative, day-of-surgery, postoperative-inpatient, and postdischarge phases.
RESULTS
Defining Key Concepts and Issues
Hospitalists have participated in a variety of perioperative roles for more than a decade. Roles include performing in-depth preoperative assessments, providing oversight to presurgical advanced practice provider assessments, providing inpatient comanagement and consultation both before and after surgery, and providing postdischarge follow-up within the surgical period for medical comorbidities.
Although a comprehensive look at the entire perioperative period is important, 4 specific phases were defined to guide this work (Figure). The phases identified were based on time relative to surgery, with unique considerations as to the overall perioperative period. Concerns and potential risks specific to each phase were considered (Table 1).
The PCMIS was constructed to provide a single coherent vision of key concepts in perioperative care (Table 2). Also identified were several key questions for determining the effectiveness of perioperative care within an institution (Table 3).
Models of Care
Multiple examples of hospitalist involvement were collected to inform the program development guidelines. The specifics noted among the reviewed practice models are described here.
Preoperative. In some centers, all patients scheduled for surgery are required to undergo evaluation at the institution’s preoperative clinic. At most others, referral to the preoperative clinic is at the discretion of the surgical specialists, who have been informed of the clinic’s available resources. Factors determining whether a patient has an in-person clinic visit, undergoes a telephone-based medical evaluation, or has a referral deferred to the primary care physician (PCP) include patient complexity and surgery-specific risk. Patients who have major medical comorbidities (eg, chronic lung or heart disease) or are undergoing higher risk procedures (eg, those lasting >1 hour, laparotomy) most often undergo a formal clinic evaluation. Often, even for a patient whose preoperative evaluation is completed by a PCP, the preoperative nursing staff will call before surgery to provide instructions and to confirm that preoperative planning is complete. Confirmation includes ensuring that the surgery consent and preoperative history and physical examination documents are in the medical record, and that all recommended tests have been performed. If deficiencies are found, surgical and preoperative clinic staff are notified.
During a typical preoperative clinic visit, nursing staff complete necessary regulatory documentation requirements and ensure that all items on the preoperative checklist are completed before day of surgery. Nurses or pharmacists perform complete medication reconciliation. For medical evaluation at institutions with a multidisciplinary preoperative clinic, patients are triaged according to comorbidity and procedure. These clinics often have anesthesiology and hospital medicine clinicians collaborating with interdisciplinary colleagues and with patients’ longitudinal care providers (eg, PCP, cardiologist). Hospitalists evaluate patients with comorbid medical diseases and address uncontrolled conditions and newly identified symptomatology. Additional testing is determined by evidence- and guideline-based standards. Patients receive preoperative education, including simple template-based medication management instructions. Perioperative clinicians follow up on test results, adjust therapy, and counsel patients to optimize health in preparation for surgery.
Patients who present to the hospital and require urgent surgical intervention are most often admitted to the surgical service, and hospital medicine provides timely consultation for preoperative recommendations. At some institutions, protocols may dictate that certain surgical patients (eg, elderly with hip fracture) are admitted to the hospital medicine service. In these scenarios, the hospitalist serves as the primary inpatient care provider and ensures preoperative medical optimization and coordination with the surgical service to expedite plans for surgery.
Day of Surgery. On the day of surgery, the surgical team verifies all patient demographic and clinical information, confirms that all necessary documentation is complete (eg, consents, history, physical examination), and marks the surgical site. The anesthesia team performs a focused review and examination while explaining the perioperative care plan to the patient. Most often, the preoperative history and physical examination, completed by a preoperative clinic provider or the patient’s PCP, is used by the anesthesiologist as the basis for clinical assessment. However, when information is incomplete or contradictory, surgery may be delayed for further record review and consultation.
Hospital medicine teams may be called to the pre-anesthesia holding area to evaluate acute medical problems (eg, hypertension, hyperglycemia, new-onset arrhythmia) or to give a second opinion in cases in which the anesthesiologist disagrees with the recommendations made by the provider who completed the preoperative evaluation. In either scenario, hospitalists must provide rapid service in close collaboration with anesthesiologists and surgeons. If a patient is found to be sufficiently optimized for surgery, the hospitalist clearly documents the evaluation and recommendation in the medical record. For a patient who requires further medical intervention before surgery, the hospitalist often coordinates the immediate disposition (eg, hospital admission or discharge home) and plans for optimization in the timeliest manner possible.
Occasionally, hospitalists are called to evaluate a patient in the postanesthesia care unit (PACU) for a new or chronic medical problem before the patient is transitioned to the next level of care. At most institutions, all PACU care is provided under the direction of anesthesiology, so it is imperative to collaborate with the patient’s anesthesiologist for all recommendations. When a patient is to be discharged home, the hospitalist coordinates outpatient follow-up plans for any medical issues to be addressed postoperatively. Hospitalists also apply their knowledge of the limitations of non–intensive care unit hospital care to decisions regarding appropriate triage of patients being admitted after surgery.
Postoperative Inpatient. Hospitalists provide a 24/7 model of care that deploys a staff physician for prompt assessment and management of medical problems in surgical patients. This care can be provided as part of the duties of a standard hospital medicine team or can be delivered by a dedicated perioperative medical consultation and comanagement service. In either situation, the type of medical care, comanagement or consultation, is determined at the outset. As consultants, hospitalists provide recommendations for medical care but do not write orders or take primary responsibility for management. Comanagement agreements are common, especially for orthopedic surgery and neurosurgery; these agreements delineate the specific circumstances and responsibilities of the hospitalist and surgical teams. Indications for comanagement, which may be identified during preoperative clinic evaluation or on admission, include uncontrolled or multiple medical comorbidities or the development of nonsurgical complications in the perioperative period. In the comanagement model, care of most medical issues is provided at the discretion of the hospitalist. Although this care includes order-writing privileges, management of analgesics, wounds, blood products, and antithrombotics is usually reserved for the surgical team, with the hospitalist only providing recommendations. In some circumstances, hospitalists may determine that the patient’s care requires consultation with other specialists. Although it is useful for the hospitalist to speak directly with other consultants and coordinate their recommendations, the surgical service should agree to the involvement of other services.
In addition to providing medical care throughout a patient’s hospitalization, the hospitalist consultant is crucial in the discharge process. During the admission, ideally in collaboration with a pharmacist, the hospitalist reviews the home medications and may change chronic medications. The hospitalist may also identify specific postdischarge needs of which the surgical team is not fully aware. These medical plans are incorporated through shared responsibility for discharge orders or through a reliable mechanism for ensuring the surgical team assumes responsibility. Final medication reconciliation at discharge, and a plan for prior and new medications, can be formulated with pharmacy assistance. Finally, the hospitalist is responsible for coordinating medically related hospital follow-up and handover back to the patient’s longitudinal care providers. The latter occurs through inclusion of medical care plans in the discharge summary completed by the surgical service and, in complex cases, through direct communication with the patient’s outpatient providers.
For some patients, medical problems eclipse surgical care as the primary focus of management. Collaborative discussion between the medical and surgical teams helps determine if it is more appropriate for the medical team to become the primary service, with the surgical team consulting. Such triage decisions should be jointly made by the attending physicians of the services rather than by intermediaries.
Postdischarge. Similar to their being used for medical problems after hospitalization, hospitalist-led postdischarge and extensivist clinics may be used for rapid follow-up of medical concerns in patients discharged after surgical admissions. A key benefit of this model is increased availability over what primary care clinics may be able to provide on short notice, particularly for patients who previously did not have a PCP. Additionally, the handover of specific follow-up items is more streamlined because the transition of care is between hospitalists from the same institution. Through the postdischarge clinic, hospitalists can provide care through either clinic visits or telephone-based follow-up. Once a patient’s immediate postoperative medical issues are fully stabilized, the patient can be transitioned to long-term primary care follow-up.
DISCUSSION
The United States is focused on sensible, high-value care. Perioperative care is burgeoning with opportunities for improvement, including reducing avoidable complications, developing systems for early recognition and treatment of complications, and streamlining processes to shorten length of stay and improve patient experience. The PCMIS provides the needed platform to catalyze detailed collaborative work between disciplines engaged in perioperative care.
As average age and level of medical comorbidity increase among surgical patients, hospitalists will increasingly be called on to assist in perioperative care. Hospitalists have long been involved in caring for medically complex surgical patients, through comanagement, consultation, and preoperative evaluations. As a provider group, hospitalists have comprehensive skills in quality and systems improvement, and in program development across hospital systems nationwide. Hospitalists have demonstrated their value by focusing on improving patient outcomes and enhancing patient engagement and experiences. Additionally, the perioperative period is fraught with multiple and complicated handoffs, a problem area for which hospital medicine has pioneered solutions and developed unique expertise. Hospital medicine is well prepared to provide skilled and proven leadership in the timely development, improvement, and expansion of perioperative care for this increasingly older and chronically ill population.
Hospitalists are established in multiple perioperative roles for high-risk surgical patients and have the opportunity to expand optimal patient-centric perioperative care systems working in close concert with surgeons and anesthesiologists. The basics of developing these systems include (1) assessing risk for medical complications, (2) planning for perioperative care, (3) developing programs aimed at risk reduction for preventable complications and early identification and intervention for unavoidable complications, and (4) guiding quality improvement efforts, including planning for frequent handoffs and transitions.
As a key partner in developing comprehensive programs in perioperative care, hospital medicine will continue to shape the future of hospital care for all patients. The PCMIS, as developed with support from the Society of Hospital Medicine, will aid efforts to achieve the best perioperative care models for our surgical patients.
Disclosures
Financial activities outside the submitted work: Drs. Pfeifer and Jaffer report payment for development of educational presentations; Dr. Grant reports payment for expert testimony pertaining to hospital medicine; Drs. Grant and Jaffer report royalties from publishing; Drs. Thompson, Pfiefer, Grant, Slawski, and Jaffer report travel expenses for speaking and serving on national committees; and Drs. Slawski and Jaffer serve on the board of the Society of Perioperative Assessment and Quality Improvement. The other authors have nothing to report.
1. Colby SL, Ortman JM. Projections of the Size and Composition of the U.S. Population: 2014 to 2060 (Current Population Reports, P25-1143). Washington, DC: US Census Bureau; 2014. https://www.census.gov/content/dam/Census/library/publications/2015/demo/p25-1143.pdf. Published March 2015. Accessed May 26, 2016.
2. Steiner C, Andrews R, Barrett M, Weiss A. HCUP Projections: Cost of Inpatient Discharges 2003 to 2013 (Rep 2013-01). Rockville, MD: US Dept of Health and Human Services, Agency for Healthcare Research and Quality; 2013. http://www.hcup-us.ahrq.gov/reports/projections/2013-01.pdf. Published December 11, 2013. Accessed May 26, 2016.
3. Auerbach AD, Wachter RM, Cheng HQ, et al. Comanagement of surgical patients between neurosurgeons and hospitalists. Arch Intern Med. 2010;170(22):2004-2010. PubMed
4. Batsis JA, Phy MP, Melton LJ 3rd, et al. Effects of a hospitalist care model on mortality of elderly patients with hip fractures. J Hosp Med. 2007;2(4):219-225. PubMed
5. Carr AM, Irigoyen M, Wimmer RS, Arbeter AM. A pediatric residency experience with surgical co-management. Hosp Pediatr. 2013;3(2):144-148. PubMed
6. Della Rocca GJ, Moylan KC, Crist BD, Volgas DA, Stannard JP, Mehr DR. Comanagement of geriatric patients with hip fractures: a retrospective, controlled, cohort study. Geriatr Orthop Surg Rehabil. 2013;4(1):10-15. PubMed
7. Fisher AA, Davis MW, Rubenach SE, Sivakumaran S, Smith PN, Budge MM. Outcomes for older patients with hip fractures: the impact of orthopedic and geriatric medicine cocare. J Orthop Trauma. 2006;20(3):172-178. PubMed
8. Friedman SM, Mendelson DA, Kates SL, McCann RM. Geriatric co-management of proximal femur fractures: total quality management and protocol-driven care result in better outcomes for a frail patient population. J Am Geriatr Soc. 2008;56(7):1349-1356. PubMed
9. Huddleston JM, Long KH, Naessens JM, et al; Hospitalist-Orthopedic Team Trial Investigators. Medical and surgical comanagement after elective hip and knee arthroplasty: a randomized, controlled trial. Ann Intern Med. 2004;141(1):28-38. PubMed
10. Mendelson DA, Friedman SM. Principles of comanagement and the geriatric fracture center. Clin Geriatr Med. 2014;30(2):183-189. PubMed
11. Merli GJ. The hospitalist joins the surgical team. Ann Intern Med. 2004;141(1):67-69. PubMed
12. Phy MP, Vanness DJ, Melton LJ 3rd, et al. Effects of a hospitalist model on elderly patients with hip fracture. Arch Intern Med. 2005;165(7):796-801. PubMed
13. Pinzur MS, Gurza E, Kristopaitis T, et al. Hospitalist-orthopedic co-management of high-risk patients undergoing lower extremity reconstruction surgery. Orthopedics. 2009;32(7):495. PubMed
14. Rappaport DI, Adelizzi-Delany J, Rogers KJ, et al. Outcomes and costs associated with hospitalist comanagement of medically complex children undergoing spinal fusion surgery. Hosp Pediatr. 2013;3(3):233-241. PubMed
15. Rappaport DI, Cerra S, Hossain J, Sharif I, Pressel DM. Pediatric hospitalist preoperative evaluation of children with neuromuscular scoliosis. J Hosp Med. 2013;8(12):684-688. PubMed
16. Roy A, Heckman MG, Roy V. Associations between the hospitalist model of care and quality-of-care-related outcomes in patients undergoing hip fracture surgery. Mayo Clin Proc. 2006;81(1):28-31. PubMed
17. Sharma G, Kuo YF, Freeman J, Zhang DD, Goodwin JS. Comanagement of hospitalized surgical patients by medicine physicians in the United States. Arch Intern Med. 2010;170(4):363-368. PubMed
18. Simon TD, Eilert R, Dickinson LM, Kempe A, Benefield E, Berman S. Pediatric hospitalist comanagement of spinal fusion surgery patients. J Hosp Med. 2007;2(1):23-30. PubMed
19. Whinney C, Michota F. Surgical comanagement: a natural evolution of hospitalist practice. J Hosp Med. 2008;3(5):394-397. PubMed
Of the 36 million US hospitalizations each year, 22% are surgical.1 Although less frequent than medical hospitalizations, surgical hospitalizations are more than twice as costly.2 Additionally, surgical hospitalizations are on average longer than medical hospitalizations.2 Given the increased scrutiny on cost and efficiency of care, attention has turned to optimizing perioperative care. Hospitalists are well positioned to provide specific expertise in the complex interdisciplinary medical management of surgical patients.
In recent decades, multiple models of hospitalist involvement in perioperative care have evolved across the United States.3-19 To consolidate knowledge and experience and to develop a framework for providing the best care for surgical patients, the Society of Hospital Medicine organized the Perioperative Care Work Group in 2015. This framework was designed for interdisciplinary collaboration in building and strengthening perioperative care programs.
METHODS
The Society of Hospital Medicine recognized hospital medicine programs’ need for guidance in developing collaborative care in perioperative medicine and appointed the Perioperative Care Work Group in May 2015. Work group members are perioperative medicine experts from US medical centers. They have extensive knowledge of the literature as well as administrative and clinical experience in a variety of perioperative care models.
Topic Development. Initial work was focused on reviewing and discussing multiple models of perioperative care and exploring the roles that hospital medicine physicians have within these models. Useful information was summarized to guide hospitals and physicians in designing, implementing, and expanding patient-centric perioperative medicine services with a focus on preoperative and postoperative care. A final document was created; it outlines system-level issues in perioperative care, organized by perioperative phases.
Initial Framework. Group members submitted written descriptions of key issues in each of 4 phases: (1) preoperative, (2) day of surgery, (3) postoperative inpatient, and (4) postdischarge. These descriptions were merged and reviewed by the content experts. Editing and discussion from the entire group were incorporated into the final matrix, which highlighted (1) perioperative phase definitions, (2) requirements for patients to move to next phase, (3) elements of care coordination typically provided by surgery, anesthesiology, and medicine disciplines, (4) concerns and risks particular to each phase, (5) unique considerations for each phase, (6) suggested metrics of success, and (7) key questions for determining the effectiveness of perioperative care in an institution. All members provided final evaluation and editing.
Final Approval. The Perioperative Care Matrix for Inpatient Surgeries (PCMIS) was presented to the board of the Society of Hospital Medicine in fall 2015 and was approved for use in centering and directing discussions regarding perioperative care.
Models of Care. The Perioperative Care Work Group surveyed examples of hospitalist engagement in perioperative care and synthesized these into synopses of existing models of care for the preoperative, day-of-surgery, postoperative-inpatient, and postdischarge phases.
RESULTS
Defining Key Concepts and Issues
Hospitalists have participated in a variety of perioperative roles for more than a decade. Roles include performing in-depth preoperative assessments, providing oversight to presurgical advanced practice provider assessments, providing inpatient comanagement and consultation both before and after surgery, and providing postdischarge follow-up within the surgical period for medical comorbidities.
Although a comprehensive look at the entire perioperative period is important, 4 specific phases were defined to guide this work (Figure). The phases identified were based on time relative to surgery, with unique considerations as to the overall perioperative period. Concerns and potential risks specific to each phase were considered (Table 1).
The PCMIS was constructed to provide a single coherent vision of key concepts in perioperative care (Table 2). Also identified were several key questions for determining the effectiveness of perioperative care within an institution (Table 3).
Models of Care
Multiple examples of hospitalist involvement were collected to inform the program development guidelines. The specifics noted among the reviewed practice models are described here.
Preoperative. In some centers, all patients scheduled for surgery are required to undergo evaluation at the institution’s preoperative clinic. At most others, referral to the preoperative clinic is at the discretion of the surgical specialists, who have been informed of the clinic’s available resources. Factors determining whether a patient has an in-person clinic visit, undergoes a telephone-based medical evaluation, or has a referral deferred to the primary care physician (PCP) include patient complexity and surgery-specific risk. Patients who have major medical comorbidities (eg, chronic lung or heart disease) or are undergoing higher risk procedures (eg, those lasting >1 hour, laparotomy) most often undergo a formal clinic evaluation. Often, even for a patient whose preoperative evaluation is completed by a PCP, the preoperative nursing staff will call before surgery to provide instructions and to confirm that preoperative planning is complete. Confirmation includes ensuring that the surgery consent and preoperative history and physical examination documents are in the medical record, and that all recommended tests have been performed. If deficiencies are found, surgical and preoperative clinic staff are notified.
During a typical preoperative clinic visit, nursing staff complete necessary regulatory documentation requirements and ensure that all items on the preoperative checklist are completed before day of surgery. Nurses or pharmacists perform complete medication reconciliation. For medical evaluation at institutions with a multidisciplinary preoperative clinic, patients are triaged according to comorbidity and procedure. These clinics often have anesthesiology and hospital medicine clinicians collaborating with interdisciplinary colleagues and with patients’ longitudinal care providers (eg, PCP, cardiologist). Hospitalists evaluate patients with comorbid medical diseases and address uncontrolled conditions and newly identified symptomatology. Additional testing is determined by evidence- and guideline-based standards. Patients receive preoperative education, including simple template-based medication management instructions. Perioperative clinicians follow up on test results, adjust therapy, and counsel patients to optimize health in preparation for surgery.
Patients who present to the hospital and require urgent surgical intervention are most often admitted to the surgical service, and hospital medicine provides timely consultation for preoperative recommendations. At some institutions, protocols may dictate that certain surgical patients (eg, elderly with hip fracture) are admitted to the hospital medicine service. In these scenarios, the hospitalist serves as the primary inpatient care provider and ensures preoperative medical optimization and coordination with the surgical service to expedite plans for surgery.
Day of Surgery. On the day of surgery, the surgical team verifies all patient demographic and clinical information, confirms that all necessary documentation is complete (eg, consents, history, physical examination), and marks the surgical site. The anesthesia team performs a focused review and examination while explaining the perioperative care plan to the patient. Most often, the preoperative history and physical examination, completed by a preoperative clinic provider or the patient’s PCP, is used by the anesthesiologist as the basis for clinical assessment. However, when information is incomplete or contradictory, surgery may be delayed for further record review and consultation.
Hospital medicine teams may be called to the pre-anesthesia holding area to evaluate acute medical problems (eg, hypertension, hyperglycemia, new-onset arrhythmia) or to give a second opinion in cases in which the anesthesiologist disagrees with the recommendations made by the provider who completed the preoperative evaluation. In either scenario, hospitalists must provide rapid service in close collaboration with anesthesiologists and surgeons. If a patient is found to be sufficiently optimized for surgery, the hospitalist clearly documents the evaluation and recommendation in the medical record. For a patient who requires further medical intervention before surgery, the hospitalist often coordinates the immediate disposition (eg, hospital admission or discharge home) and plans for optimization in the timeliest manner possible.
Occasionally, hospitalists are called to evaluate a patient in the postanesthesia care unit (PACU) for a new or chronic medical problem before the patient is transitioned to the next level of care. At most institutions, all PACU care is provided under the direction of anesthesiology, so it is imperative to collaborate with the patient’s anesthesiologist for all recommendations. When a patient is to be discharged home, the hospitalist coordinates outpatient follow-up plans for any medical issues to be addressed postoperatively. Hospitalists also apply their knowledge of the limitations of non–intensive care unit hospital care to decisions regarding appropriate triage of patients being admitted after surgery.
Postoperative Inpatient. Hospitalists provide a 24/7 model of care that deploys a staff physician for prompt assessment and management of medical problems in surgical patients. This care can be provided as part of the duties of a standard hospital medicine team or can be delivered by a dedicated perioperative medical consultation and comanagement service. In either situation, the type of medical care, comanagement or consultation, is determined at the outset. As consultants, hospitalists provide recommendations for medical care but do not write orders or take primary responsibility for management. Comanagement agreements are common, especially for orthopedic surgery and neurosurgery; these agreements delineate the specific circumstances and responsibilities of the hospitalist and surgical teams. Indications for comanagement, which may be identified during preoperative clinic evaluation or on admission, include uncontrolled or multiple medical comorbidities or the development of nonsurgical complications in the perioperative period. In the comanagement model, care of most medical issues is provided at the discretion of the hospitalist. Although this care includes order-writing privileges, management of analgesics, wounds, blood products, and antithrombotics is usually reserved for the surgical team, with the hospitalist only providing recommendations. In some circumstances, hospitalists may determine that the patient’s care requires consultation with other specialists. Although it is useful for the hospitalist to speak directly with other consultants and coordinate their recommendations, the surgical service should agree to the involvement of other services.
In addition to providing medical care throughout a patient’s hospitalization, the hospitalist consultant is crucial in the discharge process. During the admission, ideally in collaboration with a pharmacist, the hospitalist reviews the home medications and may change chronic medications. The hospitalist may also identify specific postdischarge needs of which the surgical team is not fully aware. These medical plans are incorporated through shared responsibility for discharge orders or through a reliable mechanism for ensuring the surgical team assumes responsibility. Final medication reconciliation at discharge, and a plan for prior and new medications, can be formulated with pharmacy assistance. Finally, the hospitalist is responsible for coordinating medically related hospital follow-up and handover back to the patient’s longitudinal care providers. The latter occurs through inclusion of medical care plans in the discharge summary completed by the surgical service and, in complex cases, through direct communication with the patient’s outpatient providers.
For some patients, medical problems eclipse surgical care as the primary focus of management. Collaborative discussion between the medical and surgical teams helps determine if it is more appropriate for the medical team to become the primary service, with the surgical team consulting. Such triage decisions should be jointly made by the attending physicians of the services rather than by intermediaries.
Postdischarge. Similar to their being used for medical problems after hospitalization, hospitalist-led postdischarge and extensivist clinics may be used for rapid follow-up of medical concerns in patients discharged after surgical admissions. A key benefit of this model is increased availability over what primary care clinics may be able to provide on short notice, particularly for patients who previously did not have a PCP. Additionally, the handover of specific follow-up items is more streamlined because the transition of care is between hospitalists from the same institution. Through the postdischarge clinic, hospitalists can provide care through either clinic visits or telephone-based follow-up. Once a patient’s immediate postoperative medical issues are fully stabilized, the patient can be transitioned to long-term primary care follow-up.
DISCUSSION
The United States is focused on sensible, high-value care. Perioperative care is burgeoning with opportunities for improvement, including reducing avoidable complications, developing systems for early recognition and treatment of complications, and streamlining processes to shorten length of stay and improve patient experience. The PCMIS provides the needed platform to catalyze detailed collaborative work between disciplines engaged in perioperative care.
As average age and level of medical comorbidity increase among surgical patients, hospitalists will increasingly be called on to assist in perioperative care. Hospitalists have long been involved in caring for medically complex surgical patients, through comanagement, consultation, and preoperative evaluations. As a provider group, hospitalists have comprehensive skills in quality and systems improvement, and in program development across hospital systems nationwide. Hospitalists have demonstrated their value by focusing on improving patient outcomes and enhancing patient engagement and experiences. Additionally, the perioperative period is fraught with multiple and complicated handoffs, a problem area for which hospital medicine has pioneered solutions and developed unique expertise. Hospital medicine is well prepared to provide skilled and proven leadership in the timely development, improvement, and expansion of perioperative care for this increasingly older and chronically ill population.
Hospitalists are established in multiple perioperative roles for high-risk surgical patients and have the opportunity to expand optimal patient-centric perioperative care systems working in close concert with surgeons and anesthesiologists. The basics of developing these systems include (1) assessing risk for medical complications, (2) planning for perioperative care, (3) developing programs aimed at risk reduction for preventable complications and early identification and intervention for unavoidable complications, and (4) guiding quality improvement efforts, including planning for frequent handoffs and transitions.
As a key partner in developing comprehensive programs in perioperative care, hospital medicine will continue to shape the future of hospital care for all patients. The PCMIS, as developed with support from the Society of Hospital Medicine, will aid efforts to achieve the best perioperative care models for our surgical patients.
Disclosures
Financial activities outside the submitted work: Drs. Pfeifer and Jaffer report payment for development of educational presentations; Dr. Grant reports payment for expert testimony pertaining to hospital medicine; Drs. Grant and Jaffer report royalties from publishing; Drs. Thompson, Pfiefer, Grant, Slawski, and Jaffer report travel expenses for speaking and serving on national committees; and Drs. Slawski and Jaffer serve on the board of the Society of Perioperative Assessment and Quality Improvement. The other authors have nothing to report.
Of the 36 million US hospitalizations each year, 22% are surgical.1 Although less frequent than medical hospitalizations, surgical hospitalizations are more than twice as costly.2 Additionally, surgical hospitalizations are on average longer than medical hospitalizations.2 Given the increased scrutiny on cost and efficiency of care, attention has turned to optimizing perioperative care. Hospitalists are well positioned to provide specific expertise in the complex interdisciplinary medical management of surgical patients.
In recent decades, multiple models of hospitalist involvement in perioperative care have evolved across the United States.3-19 To consolidate knowledge and experience and to develop a framework for providing the best care for surgical patients, the Society of Hospital Medicine organized the Perioperative Care Work Group in 2015. This framework was designed for interdisciplinary collaboration in building and strengthening perioperative care programs.
METHODS
The Society of Hospital Medicine recognized hospital medicine programs’ need for guidance in developing collaborative care in perioperative medicine and appointed the Perioperative Care Work Group in May 2015. Work group members are perioperative medicine experts from US medical centers. They have extensive knowledge of the literature as well as administrative and clinical experience in a variety of perioperative care models.
Topic Development. Initial work was focused on reviewing and discussing multiple models of perioperative care and exploring the roles that hospital medicine physicians have within these models. Useful information was summarized to guide hospitals and physicians in designing, implementing, and expanding patient-centric perioperative medicine services with a focus on preoperative and postoperative care. A final document was created; it outlines system-level issues in perioperative care, organized by perioperative phases.
Initial Framework. Group members submitted written descriptions of key issues in each of 4 phases: (1) preoperative, (2) day of surgery, (3) postoperative inpatient, and (4) postdischarge. These descriptions were merged and reviewed by the content experts. Editing and discussion from the entire group were incorporated into the final matrix, which highlighted (1) perioperative phase definitions, (2) requirements for patients to move to next phase, (3) elements of care coordination typically provided by surgery, anesthesiology, and medicine disciplines, (4) concerns and risks particular to each phase, (5) unique considerations for each phase, (6) suggested metrics of success, and (7) key questions for determining the effectiveness of perioperative care in an institution. All members provided final evaluation and editing.
Final Approval. The Perioperative Care Matrix for Inpatient Surgeries (PCMIS) was presented to the board of the Society of Hospital Medicine in fall 2015 and was approved for use in centering and directing discussions regarding perioperative care.
Models of Care. The Perioperative Care Work Group surveyed examples of hospitalist engagement in perioperative care and synthesized these into synopses of existing models of care for the preoperative, day-of-surgery, postoperative-inpatient, and postdischarge phases.
RESULTS
Defining Key Concepts and Issues
Hospitalists have participated in a variety of perioperative roles for more than a decade. Roles include performing in-depth preoperative assessments, providing oversight to presurgical advanced practice provider assessments, providing inpatient comanagement and consultation both before and after surgery, and providing postdischarge follow-up within the surgical period for medical comorbidities.
Although a comprehensive look at the entire perioperative period is important, 4 specific phases were defined to guide this work (Figure). The phases identified were based on time relative to surgery, with unique considerations as to the overall perioperative period. Concerns and potential risks specific to each phase were considered (Table 1).
The PCMIS was constructed to provide a single coherent vision of key concepts in perioperative care (Table 2). Also identified were several key questions for determining the effectiveness of perioperative care within an institution (Table 3).
Models of Care
Multiple examples of hospitalist involvement were collected to inform the program development guidelines. The specifics noted among the reviewed practice models are described here.
Preoperative. In some centers, all patients scheduled for surgery are required to undergo evaluation at the institution’s preoperative clinic. At most others, referral to the preoperative clinic is at the discretion of the surgical specialists, who have been informed of the clinic’s available resources. Factors determining whether a patient has an in-person clinic visit, undergoes a telephone-based medical evaluation, or has a referral deferred to the primary care physician (PCP) include patient complexity and surgery-specific risk. Patients who have major medical comorbidities (eg, chronic lung or heart disease) or are undergoing higher risk procedures (eg, those lasting >1 hour, laparotomy) most often undergo a formal clinic evaluation. Often, even for a patient whose preoperative evaluation is completed by a PCP, the preoperative nursing staff will call before surgery to provide instructions and to confirm that preoperative planning is complete. Confirmation includes ensuring that the surgery consent and preoperative history and physical examination documents are in the medical record, and that all recommended tests have been performed. If deficiencies are found, surgical and preoperative clinic staff are notified.
During a typical preoperative clinic visit, nursing staff complete necessary regulatory documentation requirements and ensure that all items on the preoperative checklist are completed before day of surgery. Nurses or pharmacists perform complete medication reconciliation. For medical evaluation at institutions with a multidisciplinary preoperative clinic, patients are triaged according to comorbidity and procedure. These clinics often have anesthesiology and hospital medicine clinicians collaborating with interdisciplinary colleagues and with patients’ longitudinal care providers (eg, PCP, cardiologist). Hospitalists evaluate patients with comorbid medical diseases and address uncontrolled conditions and newly identified symptomatology. Additional testing is determined by evidence- and guideline-based standards. Patients receive preoperative education, including simple template-based medication management instructions. Perioperative clinicians follow up on test results, adjust therapy, and counsel patients to optimize health in preparation for surgery.
Patients who present to the hospital and require urgent surgical intervention are most often admitted to the surgical service, and hospital medicine provides timely consultation for preoperative recommendations. At some institutions, protocols may dictate that certain surgical patients (eg, elderly with hip fracture) are admitted to the hospital medicine service. In these scenarios, the hospitalist serves as the primary inpatient care provider and ensures preoperative medical optimization and coordination with the surgical service to expedite plans for surgery.
Day of Surgery. On the day of surgery, the surgical team verifies all patient demographic and clinical information, confirms that all necessary documentation is complete (eg, consents, history, physical examination), and marks the surgical site. The anesthesia team performs a focused review and examination while explaining the perioperative care plan to the patient. Most often, the preoperative history and physical examination, completed by a preoperative clinic provider or the patient’s PCP, is used by the anesthesiologist as the basis for clinical assessment. However, when information is incomplete or contradictory, surgery may be delayed for further record review and consultation.
Hospital medicine teams may be called to the pre-anesthesia holding area to evaluate acute medical problems (eg, hypertension, hyperglycemia, new-onset arrhythmia) or to give a second opinion in cases in which the anesthesiologist disagrees with the recommendations made by the provider who completed the preoperative evaluation. In either scenario, hospitalists must provide rapid service in close collaboration with anesthesiologists and surgeons. If a patient is found to be sufficiently optimized for surgery, the hospitalist clearly documents the evaluation and recommendation in the medical record. For a patient who requires further medical intervention before surgery, the hospitalist often coordinates the immediate disposition (eg, hospital admission or discharge home) and plans for optimization in the timeliest manner possible.
Occasionally, hospitalists are called to evaluate a patient in the postanesthesia care unit (PACU) for a new or chronic medical problem before the patient is transitioned to the next level of care. At most institutions, all PACU care is provided under the direction of anesthesiology, so it is imperative to collaborate with the patient’s anesthesiologist for all recommendations. When a patient is to be discharged home, the hospitalist coordinates outpatient follow-up plans for any medical issues to be addressed postoperatively. Hospitalists also apply their knowledge of the limitations of non–intensive care unit hospital care to decisions regarding appropriate triage of patients being admitted after surgery.
Postoperative Inpatient. Hospitalists provide a 24/7 model of care that deploys a staff physician for prompt assessment and management of medical problems in surgical patients. This care can be provided as part of the duties of a standard hospital medicine team or can be delivered by a dedicated perioperative medical consultation and comanagement service. In either situation, the type of medical care, comanagement or consultation, is determined at the outset. As consultants, hospitalists provide recommendations for medical care but do not write orders or take primary responsibility for management. Comanagement agreements are common, especially for orthopedic surgery and neurosurgery; these agreements delineate the specific circumstances and responsibilities of the hospitalist and surgical teams. Indications for comanagement, which may be identified during preoperative clinic evaluation or on admission, include uncontrolled or multiple medical comorbidities or the development of nonsurgical complications in the perioperative period. In the comanagement model, care of most medical issues is provided at the discretion of the hospitalist. Although this care includes order-writing privileges, management of analgesics, wounds, blood products, and antithrombotics is usually reserved for the surgical team, with the hospitalist only providing recommendations. In some circumstances, hospitalists may determine that the patient’s care requires consultation with other specialists. Although it is useful for the hospitalist to speak directly with other consultants and coordinate their recommendations, the surgical service should agree to the involvement of other services.
In addition to providing medical care throughout a patient’s hospitalization, the hospitalist consultant is crucial in the discharge process. During the admission, ideally in collaboration with a pharmacist, the hospitalist reviews the home medications and may change chronic medications. The hospitalist may also identify specific postdischarge needs of which the surgical team is not fully aware. These medical plans are incorporated through shared responsibility for discharge orders or through a reliable mechanism for ensuring the surgical team assumes responsibility. Final medication reconciliation at discharge, and a plan for prior and new medications, can be formulated with pharmacy assistance. Finally, the hospitalist is responsible for coordinating medically related hospital follow-up and handover back to the patient’s longitudinal care providers. The latter occurs through inclusion of medical care plans in the discharge summary completed by the surgical service and, in complex cases, through direct communication with the patient’s outpatient providers.
For some patients, medical problems eclipse surgical care as the primary focus of management. Collaborative discussion between the medical and surgical teams helps determine if it is more appropriate for the medical team to become the primary service, with the surgical team consulting. Such triage decisions should be jointly made by the attending physicians of the services rather than by intermediaries.
Postdischarge. Similar to their being used for medical problems after hospitalization, hospitalist-led postdischarge and extensivist clinics may be used for rapid follow-up of medical concerns in patients discharged after surgical admissions. A key benefit of this model is increased availability over what primary care clinics may be able to provide on short notice, particularly for patients who previously did not have a PCP. Additionally, the handover of specific follow-up items is more streamlined because the transition of care is between hospitalists from the same institution. Through the postdischarge clinic, hospitalists can provide care through either clinic visits or telephone-based follow-up. Once a patient’s immediate postoperative medical issues are fully stabilized, the patient can be transitioned to long-term primary care follow-up.
DISCUSSION
The United States is focused on sensible, high-value care. Perioperative care is burgeoning with opportunities for improvement, including reducing avoidable complications, developing systems for early recognition and treatment of complications, and streamlining processes to shorten length of stay and improve patient experience. The PCMIS provides the needed platform to catalyze detailed collaborative work between disciplines engaged in perioperative care.
As average age and level of medical comorbidity increase among surgical patients, hospitalists will increasingly be called on to assist in perioperative care. Hospitalists have long been involved in caring for medically complex surgical patients, through comanagement, consultation, and preoperative evaluations. As a provider group, hospitalists have comprehensive skills in quality and systems improvement, and in program development across hospital systems nationwide. Hospitalists have demonstrated their value by focusing on improving patient outcomes and enhancing patient engagement and experiences. Additionally, the perioperative period is fraught with multiple and complicated handoffs, a problem area for which hospital medicine has pioneered solutions and developed unique expertise. Hospital medicine is well prepared to provide skilled and proven leadership in the timely development, improvement, and expansion of perioperative care for this increasingly older and chronically ill population.
Hospitalists are established in multiple perioperative roles for high-risk surgical patients and have the opportunity to expand optimal patient-centric perioperative care systems working in close concert with surgeons and anesthesiologists. The basics of developing these systems include (1) assessing risk for medical complications, (2) planning for perioperative care, (3) developing programs aimed at risk reduction for preventable complications and early identification and intervention for unavoidable complications, and (4) guiding quality improvement efforts, including planning for frequent handoffs and transitions.
As a key partner in developing comprehensive programs in perioperative care, hospital medicine will continue to shape the future of hospital care for all patients. The PCMIS, as developed with support from the Society of Hospital Medicine, will aid efforts to achieve the best perioperative care models for our surgical patients.
Disclosures
Financial activities outside the submitted work: Drs. Pfeifer and Jaffer report payment for development of educational presentations; Dr. Grant reports payment for expert testimony pertaining to hospital medicine; Drs. Grant and Jaffer report royalties from publishing; Drs. Thompson, Pfiefer, Grant, Slawski, and Jaffer report travel expenses for speaking and serving on national committees; and Drs. Slawski and Jaffer serve on the board of the Society of Perioperative Assessment and Quality Improvement. The other authors have nothing to report.
1. Colby SL, Ortman JM. Projections of the Size and Composition of the U.S. Population: 2014 to 2060 (Current Population Reports, P25-1143). Washington, DC: US Census Bureau; 2014. https://www.census.gov/content/dam/Census/library/publications/2015/demo/p25-1143.pdf. Published March 2015. Accessed May 26, 2016.
2. Steiner C, Andrews R, Barrett M, Weiss A. HCUP Projections: Cost of Inpatient Discharges 2003 to 2013 (Rep 2013-01). Rockville, MD: US Dept of Health and Human Services, Agency for Healthcare Research and Quality; 2013. http://www.hcup-us.ahrq.gov/reports/projections/2013-01.pdf. Published December 11, 2013. Accessed May 26, 2016.
3. Auerbach AD, Wachter RM, Cheng HQ, et al. Comanagement of surgical patients between neurosurgeons and hospitalists. Arch Intern Med. 2010;170(22):2004-2010. PubMed
4. Batsis JA, Phy MP, Melton LJ 3rd, et al. Effects of a hospitalist care model on mortality of elderly patients with hip fractures. J Hosp Med. 2007;2(4):219-225. PubMed
5. Carr AM, Irigoyen M, Wimmer RS, Arbeter AM. A pediatric residency experience with surgical co-management. Hosp Pediatr. 2013;3(2):144-148. PubMed
6. Della Rocca GJ, Moylan KC, Crist BD, Volgas DA, Stannard JP, Mehr DR. Comanagement of geriatric patients with hip fractures: a retrospective, controlled, cohort study. Geriatr Orthop Surg Rehabil. 2013;4(1):10-15. PubMed
7. Fisher AA, Davis MW, Rubenach SE, Sivakumaran S, Smith PN, Budge MM. Outcomes for older patients with hip fractures: the impact of orthopedic and geriatric medicine cocare. J Orthop Trauma. 2006;20(3):172-178. PubMed
8. Friedman SM, Mendelson DA, Kates SL, McCann RM. Geriatric co-management of proximal femur fractures: total quality management and protocol-driven care result in better outcomes for a frail patient population. J Am Geriatr Soc. 2008;56(7):1349-1356. PubMed
9. Huddleston JM, Long KH, Naessens JM, et al; Hospitalist-Orthopedic Team Trial Investigators. Medical and surgical comanagement after elective hip and knee arthroplasty: a randomized, controlled trial. Ann Intern Med. 2004;141(1):28-38. PubMed
10. Mendelson DA, Friedman SM. Principles of comanagement and the geriatric fracture center. Clin Geriatr Med. 2014;30(2):183-189. PubMed
11. Merli GJ. The hospitalist joins the surgical team. Ann Intern Med. 2004;141(1):67-69. PubMed
12. Phy MP, Vanness DJ, Melton LJ 3rd, et al. Effects of a hospitalist model on elderly patients with hip fracture. Arch Intern Med. 2005;165(7):796-801. PubMed
13. Pinzur MS, Gurza E, Kristopaitis T, et al. Hospitalist-orthopedic co-management of high-risk patients undergoing lower extremity reconstruction surgery. Orthopedics. 2009;32(7):495. PubMed
14. Rappaport DI, Adelizzi-Delany J, Rogers KJ, et al. Outcomes and costs associated with hospitalist comanagement of medically complex children undergoing spinal fusion surgery. Hosp Pediatr. 2013;3(3):233-241. PubMed
15. Rappaport DI, Cerra S, Hossain J, Sharif I, Pressel DM. Pediatric hospitalist preoperative evaluation of children with neuromuscular scoliosis. J Hosp Med. 2013;8(12):684-688. PubMed
16. Roy A, Heckman MG, Roy V. Associations between the hospitalist model of care and quality-of-care-related outcomes in patients undergoing hip fracture surgery. Mayo Clin Proc. 2006;81(1):28-31. PubMed
17. Sharma G, Kuo YF, Freeman J, Zhang DD, Goodwin JS. Comanagement of hospitalized surgical patients by medicine physicians in the United States. Arch Intern Med. 2010;170(4):363-368. PubMed
18. Simon TD, Eilert R, Dickinson LM, Kempe A, Benefield E, Berman S. Pediatric hospitalist comanagement of spinal fusion surgery patients. J Hosp Med. 2007;2(1):23-30. PubMed
19. Whinney C, Michota F. Surgical comanagement: a natural evolution of hospitalist practice. J Hosp Med. 2008;3(5):394-397. PubMed
1. Colby SL, Ortman JM. Projections of the Size and Composition of the U.S. Population: 2014 to 2060 (Current Population Reports, P25-1143). Washington, DC: US Census Bureau; 2014. https://www.census.gov/content/dam/Census/library/publications/2015/demo/p25-1143.pdf. Published March 2015. Accessed May 26, 2016.
2. Steiner C, Andrews R, Barrett M, Weiss A. HCUP Projections: Cost of Inpatient Discharges 2003 to 2013 (Rep 2013-01). Rockville, MD: US Dept of Health and Human Services, Agency for Healthcare Research and Quality; 2013. http://www.hcup-us.ahrq.gov/reports/projections/2013-01.pdf. Published December 11, 2013. Accessed May 26, 2016.
3. Auerbach AD, Wachter RM, Cheng HQ, et al. Comanagement of surgical patients between neurosurgeons and hospitalists. Arch Intern Med. 2010;170(22):2004-2010. PubMed
4. Batsis JA, Phy MP, Melton LJ 3rd, et al. Effects of a hospitalist care model on mortality of elderly patients with hip fractures. J Hosp Med. 2007;2(4):219-225. PubMed
5. Carr AM, Irigoyen M, Wimmer RS, Arbeter AM. A pediatric residency experience with surgical co-management. Hosp Pediatr. 2013;3(2):144-148. PubMed
6. Della Rocca GJ, Moylan KC, Crist BD, Volgas DA, Stannard JP, Mehr DR. Comanagement of geriatric patients with hip fractures: a retrospective, controlled, cohort study. Geriatr Orthop Surg Rehabil. 2013;4(1):10-15. PubMed
7. Fisher AA, Davis MW, Rubenach SE, Sivakumaran S, Smith PN, Budge MM. Outcomes for older patients with hip fractures: the impact of orthopedic and geriatric medicine cocare. J Orthop Trauma. 2006;20(3):172-178. PubMed
8. Friedman SM, Mendelson DA, Kates SL, McCann RM. Geriatric co-management of proximal femur fractures: total quality management and protocol-driven care result in better outcomes for a frail patient population. J Am Geriatr Soc. 2008;56(7):1349-1356. PubMed
9. Huddleston JM, Long KH, Naessens JM, et al; Hospitalist-Orthopedic Team Trial Investigators. Medical and surgical comanagement after elective hip and knee arthroplasty: a randomized, controlled trial. Ann Intern Med. 2004;141(1):28-38. PubMed
10. Mendelson DA, Friedman SM. Principles of comanagement and the geriatric fracture center. Clin Geriatr Med. 2014;30(2):183-189. PubMed
11. Merli GJ. The hospitalist joins the surgical team. Ann Intern Med. 2004;141(1):67-69. PubMed
12. Phy MP, Vanness DJ, Melton LJ 3rd, et al. Effects of a hospitalist model on elderly patients with hip fracture. Arch Intern Med. 2005;165(7):796-801. PubMed
13. Pinzur MS, Gurza E, Kristopaitis T, et al. Hospitalist-orthopedic co-management of high-risk patients undergoing lower extremity reconstruction surgery. Orthopedics. 2009;32(7):495. PubMed
14. Rappaport DI, Adelizzi-Delany J, Rogers KJ, et al. Outcomes and costs associated with hospitalist comanagement of medically complex children undergoing spinal fusion surgery. Hosp Pediatr. 2013;3(3):233-241. PubMed
15. Rappaport DI, Cerra S, Hossain J, Sharif I, Pressel DM. Pediatric hospitalist preoperative evaluation of children with neuromuscular scoliosis. J Hosp Med. 2013;8(12):684-688. PubMed
16. Roy A, Heckman MG, Roy V. Associations between the hospitalist model of care and quality-of-care-related outcomes in patients undergoing hip fracture surgery. Mayo Clin Proc. 2006;81(1):28-31. PubMed
17. Sharma G, Kuo YF, Freeman J, Zhang DD, Goodwin JS. Comanagement of hospitalized surgical patients by medicine physicians in the United States. Arch Intern Med. 2010;170(4):363-368. PubMed
18. Simon TD, Eilert R, Dickinson LM, Kempe A, Benefield E, Berman S. Pediatric hospitalist comanagement of spinal fusion surgery patients. J Hosp Med. 2007;2(1):23-30. PubMed
19. Whinney C, Michota F. Surgical comanagement: a natural evolution of hospitalist practice. J Hosp Med. 2008;3(5):394-397. PubMed
© 2017 Society of Hospital Medicine
Focus on Reducing the Discomfort, Not the Fever
A child who has a cold, flu, or other acute illness may be what parents often call “fussy”: irritable, teary, and clingy. Such changes in behavior and mood, or “sickness behavior (SB),” are usually thought to be linked to fever. Actually, those symptoms are the immune system’s reactions to invasion by a pathogen, say French researchers—and they may be present whether the child has fever or not. The researchers’ say their multicenter study is the first to show dissociation between SB and the severity of the fever.
The researchers evaluated 6 parameters over the 2 hours preceding consultations with the parents of 200 children with and 200 without fever. Children with particularly painful illnesses and chronic diseases were excluded from the study. Parents used rating scales to report degrees of change in the time the child spent playing, the distance covered (ie. how far from the parent the child roamed), time the child spent seeking comfort, time spent whining or crying, time spent in a state of irritation or anger, most distorted facial expression (on a chart). The researchers also assessed time spent sleeping and appetite in the 24 hours before the consultation. Sickness behavior can’t be reduced to the observation of those 8 behavioral parameters, the researchers note, but they were easy for parents to use and assess.
The mean values of the 8 parameters differed significantly between the 2 groups but were independent of the height of fever in the febrile children. That independence suggests that SB and fever are expressions of 2 autonomous metabolic pathways that are activated simultaneously in febrile conditions, the researchers say, which is in accordance with current pathophysiologic knowledge.
Their findings are in harmony with current treatment recommendations, the researchers say. Because it’s hard to know when behavior changes are due to SB, pain, fatigue, or something else in a febrile child—especially one who is too young to talk about it—it’s more important to focus on relieving the discomfort than in reducing the fever.
Source:
Corrard F, Copin C, Wollner A, et al. PLoS One. 2017;12(3): e0171670.
doi: 10.1371/journal.pone.0171670.
A child who has a cold, flu, or other acute illness may be what parents often call “fussy”: irritable, teary, and clingy. Such changes in behavior and mood, or “sickness behavior (SB),” are usually thought to be linked to fever. Actually, those symptoms are the immune system’s reactions to invasion by a pathogen, say French researchers—and they may be present whether the child has fever or not. The researchers’ say their multicenter study is the first to show dissociation between SB and the severity of the fever.
The researchers evaluated 6 parameters over the 2 hours preceding consultations with the parents of 200 children with and 200 without fever. Children with particularly painful illnesses and chronic diseases were excluded from the study. Parents used rating scales to report degrees of change in the time the child spent playing, the distance covered (ie. how far from the parent the child roamed), time the child spent seeking comfort, time spent whining or crying, time spent in a state of irritation or anger, most distorted facial expression (on a chart). The researchers also assessed time spent sleeping and appetite in the 24 hours before the consultation. Sickness behavior can’t be reduced to the observation of those 8 behavioral parameters, the researchers note, but they were easy for parents to use and assess.
The mean values of the 8 parameters differed significantly between the 2 groups but were independent of the height of fever in the febrile children. That independence suggests that SB and fever are expressions of 2 autonomous metabolic pathways that are activated simultaneously in febrile conditions, the researchers say, which is in accordance with current pathophysiologic knowledge.
Their findings are in harmony with current treatment recommendations, the researchers say. Because it’s hard to know when behavior changes are due to SB, pain, fatigue, or something else in a febrile child—especially one who is too young to talk about it—it’s more important to focus on relieving the discomfort than in reducing the fever.
Source:
Corrard F, Copin C, Wollner A, et al. PLoS One. 2017;12(3): e0171670.
doi: 10.1371/journal.pone.0171670.
A child who has a cold, flu, or other acute illness may be what parents often call “fussy”: irritable, teary, and clingy. Such changes in behavior and mood, or “sickness behavior (SB),” are usually thought to be linked to fever. Actually, those symptoms are the immune system’s reactions to invasion by a pathogen, say French researchers—and they may be present whether the child has fever or not. The researchers’ say their multicenter study is the first to show dissociation between SB and the severity of the fever.
The researchers evaluated 6 parameters over the 2 hours preceding consultations with the parents of 200 children with and 200 without fever. Children with particularly painful illnesses and chronic diseases were excluded from the study. Parents used rating scales to report degrees of change in the time the child spent playing, the distance covered (ie. how far from the parent the child roamed), time the child spent seeking comfort, time spent whining or crying, time spent in a state of irritation or anger, most distorted facial expression (on a chart). The researchers also assessed time spent sleeping and appetite in the 24 hours before the consultation. Sickness behavior can’t be reduced to the observation of those 8 behavioral parameters, the researchers note, but they were easy for parents to use and assess.
The mean values of the 8 parameters differed significantly between the 2 groups but were independent of the height of fever in the febrile children. That independence suggests that SB and fever are expressions of 2 autonomous metabolic pathways that are activated simultaneously in febrile conditions, the researchers say, which is in accordance with current pathophysiologic knowledge.
Their findings are in harmony with current treatment recommendations, the researchers say. Because it’s hard to know when behavior changes are due to SB, pain, fatigue, or something else in a febrile child—especially one who is too young to talk about it—it’s more important to focus on relieving the discomfort than in reducing the fever.
Source:
Corrard F, Copin C, Wollner A, et al. PLoS One. 2017;12(3): e0171670.
doi: 10.1371/journal.pone.0171670.
Prognosticating with the hospitalized-patient one-year mortality risk score using information abstracted from the medical record
A patient’s prognosis can strongly influence their medical care. Decisions about diagnostic modalities, treatment options, and the use of preventive therapies can all be affected by the likelihood of a patient’s death in the near future. For example, patients with severely limited survival might forego prophylactic therapy, avoid interventions for asymptomatic issues, and cease screening interventions. Knowing survival probability would also be very helpful as a controlling variable in research analyses whenever death risk might be a possible confounder.
Sixteen indices that aim to predict patient death risk have been described by Yourman et al.1 They were all created from secondary analyses of clinical and administrative datasets, were applicable to patients in a variety of settings (including the community, nursing home, or hospital), and predicted survival probabilities in time horizons ranging from 6 months to 5 years. Prognostic factors that were most commonly included in these indices were comorbidity and functional status. In validation populations, the discrimination of these indices for 1-year survival in hospitalized patients was moderate (with C statistics that ranged from 0.64 to 0.79) with good calibration for broad prognostic ranges.
In 2014, we published the Hospitalized-patient One-year Mortality Risk (HOMR) score.2 This study used health administrative data for all adult Ontarians admitted in 2011 to hospital under nonpsychiatric services (n = 640,022) to estimate the probability of dying within 1 year of admission to hospital (which happened in 11.7% of people). The HOMR score included 12 patient and hospitalization factors (Table 1). It was highly discriminative (C statistic, 0.923; [0.922-0.924]) and well calibrated (the mean relative difference between observed and expected death risk was 2.0% [range, 0.0% to 7.0%]). It was externally validated in more than 3 million adults from Ontario, Alberta, and Boston in whom the C statistic ranged from 0.89 to 0.92 and calibration was excellent.3 We concluded from these studies that the HOMR score is excellent for prognosticating a diverse group of patients using health administrative data.
However, we do not know whether the HOMR score can be applied to patients using primary data (ie, those taken directly from the chart). This question is important for 2 reasons. First, if HOMR accurately predicts death risk using data abstracted from the medical record, it could be used in the clinical setting to assist in clinical decision-making. Second, HOMR uses multiple administrative datasets that are difficult to access and use by most clinical researchers; it is, therefore, important to determine if HOMR is accurate for clinical research based on primary medical record review. The primary objective of this study was to determine the accuracy of the HOMR score when calculated using data abstracted from clinical notes that were available when patients were admitted to hospital. Secondary objectives included determining whether functional measures abstracted were significantly associated with death risk beyond the HOMR score and whether HOMR scores calculated from chart review deviated from those calculated from administrative data.
METHODS
Study Cohort
The study, which was approved by our local research ethics board, took place at the Ottawa Hospital, a 1000-bed teaching hospital that is the primary referral center in our region. We used the hospital admission registry to identify all people 18 years or older who were admitted to a nonpsychiatric service at our hospital between January 1, 2011 and December 31, 2011 (this time frame corresponds with the year used to derive the HOMR score). We excluded overnight patients in the same-day surgery or the bone-marrow transplant units (since they would not have been included in the original study) and those without a valid health card number (which was required to link to provincial data to identify outcomes). From this list, we randomly selected 5000 patients.
Primary Data Collection
For each patient, we retrieved all data required to calculate the HOMR score from the medical record (Table 1). Patient registration information in our electronic medical record was used to identify patient age, sex, admitting service, number of emergency department (ED) visits in the previous year, number of admissions in the previous year (the nursing triage note was reviewed for each admission to determine if it was by ambulance), and whether or not the patient had been discharged from hospital in the previous 30 days. The admitting service consult note was used to determine the admitting diagnosis and whether or not the patient was admitted directly to the intensive care unit. If they were present, the emergency nursing triage note, the ED record of treatment, the admission consult note, the pre-operative consult note, and consult notes were all used to determine the patient’s comorbidities, living status, and home oxygen status. Admission urgency was determined using information from the patient registration information and the ED nursing triage note. All data were abstracted from information that had been registered prior to when the patient was physically transferred to their hospital bed. This ensured that we used only data available at the start of the admission.
Patient functional status has been shown to be strongly associated with survival4 but HOMR only indirectly captures functional information (through the patient’s living status). We, therefore, collected more detailed functional information from the medical record by determining if the patient was dependent for any activities of daily living (ADL) from the emergency nursing triage note, the ED record of treatment, the admission consult note, and the pre-operative consultation. We also collected information that might indicate frailty, which we defined per Clegg et al.5 as “a state of increased vulnerability to poor resolution of homeostasis following a stress.” This information included: delirium or more than 1 fall recorded on the emergency nursing triage note, the ED record of treatment, or the admission consultation note; or whether a geriatric nursing specialist assessment occurred in the ED in the previous 6 months. Finally, we recorded possible indicators of limited social support (no fixed address [from patient registration and nursing triage note], primary contact is not a family member [from the emergency notes, consult, and patient registration], and no religion noted in system [from patient registration]). Patients for whom religion status was missing were classified as having “no religion.”
Analysis
These data were encrypted and linked anonymously to population-based databases to determine whether patients died within 1 year of admission to hospital. We calculated the chart-HOMR score using information from the chart review and determined its association with the outcome using bivariate logistic regression. We compared observed and expected risk of death within 1 year of admission to hospital for each chart-HOMR score value, with expected risks determined from the external validation study.3 We regressed observed death risks on expected death risks for chart-HOMR scores (clustered into 22 groups to ensure adequate numbers in each group); and we gauged overall deviations from expected risk and the relationship between the observed and expected death risk (based on the chart-HOMR score) using the line’s intercept and slope, respectively.6 Next, we replicated methods from our studies2,3 to calculate the administrative-HOMR score in our study cohort using administrative databases. We compared these chart-HOMR and administrative-HOMR scores (and scores for each of its components). Finally, we determined which of the socio-functional factors were associated with 1-year death risk independent of the chart-HOMR score. We used the likelihood ratio test to determine whether these additional socio-functional factors significantly improved the model beyond the chart-HOMR score.7 This test subtracted the -2 logL value of the full model from that containing the chart-HOMR score alone, comparing its value to the χ2 distribution (with degrees of freedom equivalent to the number of additional parameters in the nested model) to determine statistical significance. All analyses were completed using SAS v9.4 (SAS Institute Inc., Cary, North Carolina).
RESULTS
There were 43,883 overnight hospitalizations at our hospital in 2011, and 38,886 hospitalizations were excluded: 1883 hospitalizations were in the same-day surgery or the bone-marrow transplant unit; 2485 did not have a valid health card number; 34,515 were not randomly selected; the records of 3 randomly selected patients had been blocked by our hospital’s privacy department; and 1 patient could not be linked with the population-based administrative datasets.
The 4996 study patients were middle-aged and predominantly female (Table 2). The extensive majority of patients was admitted from the community, was independent for ADL, had a family member as the principal contact, and had no admissions by ambulance in the previous year. Most people had no significant comorbidities or ED visits in the year prior to their admission. The mean chart-HOMR score was 22 (standard deviation [SD], 12), which is associated with a 1.2% expected risk of death within 1 year of hospital admission (Appendix 1).3
A total of 563 patients (11.3%) died within 1 year of admission to hospital (Table 2). In the study cohort, each chart-HOMR component was associated with death status. People who died were older, more likely to be male, had a greater number of important comorbidities, had more ED visits and admissions by ambulance in the previous year, and were more likely to have been discharged in the previous 30 days, and were admitted urgently, directly to the intensive care unit, or with complicated diagnoses. The mean chart-HOMR score differed extensively by survival status (37.4 [SD, 7.5] in those who died vs. 19.9 [SD, 12.2] in those who survived). Three of the socio-functional variables (delirium and falls noted on admission documents, and dependent for any ADL) also varied with death status.
The chart-HOMR score was strongly associated with the likelihood of death within 1 year of admission. When included in a logistic regression model having 1-year death as the outcome, a 1-point increase in the chart-HOMR score was associated with a 19% increase in the odds of death (P < 0.0001). This model (with only the chart-HOMR score) was highly discriminative (C statistic, 0.888) and well calibrated (Hosmer-Lemeshow test, 12.9 [8 df, P = 0.11]).
Observed and expected death risks by chart-HOMR score were similar (Figure 1). The observed total number of deaths (n = 563; 11.3%) exceeded the expected number of deaths (n = 437, 8.7%). When we regressed observed death risks on expected death risks for chart-HOMR scores (clustered into 22 groups), the Hosmer-Lemeshow test was significant, indicating that differences between observed and expected risks were beyond that expected by chance (Hosmer-Lemeshow test, 141.9, 21 df, P < 0.0001). The intercept of this model (0.035; 95% CI, 0.01-0.06) was statistically significant (P = 0.01), indicating that the observed number of cases significantly exceeded the expected; however, its calibration slope (1.02; 95% CI, 0.89-1.16) did not deviate significantly from unity, indicating that the relationship between the observed and expected death risk (based on the chart-HOMR score) remained intact (Figure 1).
The deviations between observed and expected death risks reflected deviations between the c chart-HOMR score and the administrative-HOMR score, with the former being significantly lower than the latter (Figure 2). Overall, the chart-HOMR score was 0.96 points lower (95% CI, 0.81-1.12) than the administrative-HOMR score. The HOMR score components that were notably underestimated using chart data included those for the age-Charlson Comorbidity Index interaction, living status, and admit points. Points for only 2 components (admitting service and admission urgency) were higher when calculated using chart data.
Four additional socio-functional variables collected from medical record review were significantly associated with 1-year death risk independent of the chart-HOMR score (Table 3). Admission documentation noting either delirium or falls were both associated with a significantly increased death risk (adjusted odds ratio [OR], 1.92 [95% CI, 1.24-2.96] and OR 1.96 [95% CI, 1.29-2.99], respectively). An independently increased death risk was also noted in patients who were dependent for any ADL (adjusted OR, 1.99 [95% CI, 1.24-3.19]). The presence of an ED geriatrics consultation within the previous 6 months was associated with a significantly decreased death risk of 60% (adjusted OR, 0.40 [95% CI, 0.20-0.81]). Adding these covariates to the logistic model with the chart-HOMR score significantly improved predictions (likelihood ratio statistic = 33.569, 4df, P < 0.00001).
DISCUSSION
In a large random sample of patients from our hospital, we found that the HOMR score using data abstracted from the medical record was significantly associated with 1-year death risk. The expected death risk based on the chart-HOMR score underestimated observed death risk but the relationship between the chart-HOMR score and death risk was similar to that in studies using administrative data. The HOMR score calculated using data from the chart was lower than that calculated using data from population-based administrative datasets; additional variables indicating patient frailty were significantly associated with 1-year death risk independent of the chart-HOMR score. Since the HOMR score was derived and initially validated using health administrative data, this study using data abstracted from the health record shows that the HOMR score has methodological generalizability.8
We think that our study has several notable findings. First, we found that data abstracted from the medical record can be used to calculate the HOMR score to accurately predict individual death risk. The chart-HOMR score discriminated very well between patients who did and did not die (C statistic, 0.88), which extensively exceeds the discrimination of published death risk indices (whose C statistics range between 0.69 and 0.82). It is also possible that chart abstraction for the HOMR score—without functional status—is simpler than other indices since its components are primarily very objective. (Other indices for hospital-based patients required factors that could be difficult to abstract reliably from the medical record including meeting more than 1 guideline for noncancer hospice care9; ambulation difficulties10; scales such as the Exton-Smith Scale or the Short Portable Mental Status Questionnaire11; weight loss12; functional status4; and pressure sore risk.13) Although expected risks for the chart-HOMR consistently underestimated observed risks (Figure 1), the mean deviation was small (with an absolute difference of 3.5% that can be used as a correction factor when determining expected risks with HOMR scores calculated from chart review), but it was an association between the chart-HOMR score and death risk that remained consistent through the cohort. Second, we found a small but significant decrease in the chart-HOMR score vs. the administrative-HOMR score (Figure 2). Some of these underestimates such as those for the number of ED visits or admissions by ambulance were expected since population-based health administrative databases would best capture such data. However, we were surprised that the comorbidity score was less when calculated using chart vs. database data (Figure 2). This finding is distinct from studies finding that particular comorbidities are documented in the chart are sometimes not coded.14,15 However, we identified comorbidities in the administrative databases using a 1-year ‘look-back’ period so that diagnostic codes from multiple hospitalizations (and from multiple hospitals) could be used to calculate the Charlson Comorbidity Index for a particular patient; this has been shown to increase the capture of comorbidities.16 Third, we found that variables from the chart review indicating frailty were predictive of 1-year death risk independent of the chart-HOMR score (Table 2). This illustrates that mortality risk prediction can be improved for particular patient groups by adding new covariates to the HOMR. Further work is required to determine how to incorporate these (and possibly other) covariates into the HOMR to create a unique chart-HOMR score. Finally, we found that a geriatrics assessment in the ED was associated with a significant (and notable) decrease in death risk. With these data, we are unable to indicate whether this association is causative. However, these findings indicate that the influence of emergency geriatric assessments on patient survival needs to be explored in more detail.
Several issues about our study should be considered when interpreting its results. First, this was a single-center study and the generalizability of our results to other centers is unknown. However, our study had the largest sample size of all primary data prognostic index validation studies1 ensuring that our results are, at the very least, internally reliable. In addition, our simple random sample ensured that we studied a broad assortment of patients to be certain that our results are representative of our institution. Second, we used a single abstractor for the study, which could limit the generalizability of our results. However, almost all the data points that were abstracted for our study were very objective.
In summary, our study shows that the HOMR score can be used to accurately predict 1-year death risk using data abstracted from the patient record. These findings will aid in individual patient prognostication for clinicians and researchers.
Disclosure
The authors report no financial conflicts of interest.
1. Yourman LC, Lee SJ, Schonberg MA, Widera EW, Smith AK. Prognostic indices for older adults: a systematic review. JAMA. 2012;307(2):182-192. PubMed
2. van Walraven C. The Hospital-patient One-year Mortality Risk score accurately predicts long term death risk in hospitalized patients. J Clin Epidemiol. 2014;67(9):1025-1034. PubMed
3. van Walraven C, McAlister FA, Bakal JA, Hawken S, Donzé J. External validation of the Hospital-patient One-year Mortality Risk (HOMR) model for predicting death within 1 year after hospital admission. CMAJ. 2015;187(10):725-733. PubMed
4. Walter LC, Brand RJ, Counsell SR, et al. Development and validation of a prognostic index for 1-year mortality in older adults after hospitalization. JAMA. 2001;285(23):2987-2994. PubMed
5. Clegg A, Young J, Iliffe S et al. Frailty in elderly people. The Lancet 2002;381:752-762. PubMed
6. Crowson CS, Atkinson EJ, Therneau TM. Assessing calibration of prognostic risk scores. Stat Methods Med Res. 2016;25(4):1692-1706. PubMed
7. Harrell FE Jr. Overview of Maximum Likelihood Estimation. Regression Modeling Strategies. New York, NY: Springer-Verlag; 2001: 179-212.
8. Justice AC, Covinsky KE, Berlin JA. Assessing the generalizability of prognostic information. Ann Intern Med. 1999;130(6):515-524. PubMed
9. Fischer SM, Gozansky WS, Sauaia A, Min SJ, Kutner JS, Kramer A. A practical tool to identify patients who may benefit from a palliative approach: the CARING criteria. J Pain Symptom Manage. 2006;31(4):285-292. PubMed
10. Inouye SK, Bogardus ST, Jr, Vitagliano G, et al. Burden of illness score for elderly persons: risk adjustment incorporating the cumulative impact of diseases, physiologic abnormalities, and functional impairments. Med Care. 2003;41(1):70-83. PubMed
11. Pilotto A, Ferrucci L, Franceschi M, et al. Development and validation of a multidimensional prognostic index for one-year mortality from comprehensive geriatric assessment in hospitalized older patients. Rejuvenation Res. 2008;11(1):151-161. PubMed
12. Teno JM, Harrell FE Jr, Knaus W, et al. Prediction of survival for older hospitalized patients: the HELP survival model. Hospitalized Elderly Longitudinal Project. J Am Geriatr Soc. 2000;48(5 suppl):S16-S24. PubMed
13. Dramé M, Novella JL, Lang PO, et al. Derivation and validation of a mortality-risk index from a cohort of frail elderly patients hospitalised in medical wards via emergencies: the SAFES study. Eur J Epidemiol. 2008;23(12):783-791. PubMed
14. Kieszak SM, Flanders WD, Kosinski AS, Shipp CC, Karp H. A comparison of the Charlson comorbidity index derived from medical record data and administrative billing data. J Clin Epidemiol. 1999;52(2):137-142. PubMed
15. Quan H, Parsons GA, Ghali WA. Validity of procedure codes in International Classification of Diseases, 9th revision, clinical modification administrative data. Med Care. 2004;42(8):801-809. PubMed
16. Zhang JX, Iwashyna TJ, Christakis NA. The performance of different lookback periods and sources of information for Charlson cComorbidity adjustment in Medicare claims. Med Care. 1999;37(11):1128-1139. PubMed
A patient’s prognosis can strongly influence their medical care. Decisions about diagnostic modalities, treatment options, and the use of preventive therapies can all be affected by the likelihood of a patient’s death in the near future. For example, patients with severely limited survival might forego prophylactic therapy, avoid interventions for asymptomatic issues, and cease screening interventions. Knowing survival probability would also be very helpful as a controlling variable in research analyses whenever death risk might be a possible confounder.
Sixteen indices that aim to predict patient death risk have been described by Yourman et al.1 They were all created from secondary analyses of clinical and administrative datasets, were applicable to patients in a variety of settings (including the community, nursing home, or hospital), and predicted survival probabilities in time horizons ranging from 6 months to 5 years. Prognostic factors that were most commonly included in these indices were comorbidity and functional status. In validation populations, the discrimination of these indices for 1-year survival in hospitalized patients was moderate (with C statistics that ranged from 0.64 to 0.79) with good calibration for broad prognostic ranges.
In 2014, we published the Hospitalized-patient One-year Mortality Risk (HOMR) score.2 This study used health administrative data for all adult Ontarians admitted in 2011 to hospital under nonpsychiatric services (n = 640,022) to estimate the probability of dying within 1 year of admission to hospital (which happened in 11.7% of people). The HOMR score included 12 patient and hospitalization factors (Table 1). It was highly discriminative (C statistic, 0.923; [0.922-0.924]) and well calibrated (the mean relative difference between observed and expected death risk was 2.0% [range, 0.0% to 7.0%]). It was externally validated in more than 3 million adults from Ontario, Alberta, and Boston in whom the C statistic ranged from 0.89 to 0.92 and calibration was excellent.3 We concluded from these studies that the HOMR score is excellent for prognosticating a diverse group of patients using health administrative data.
However, we do not know whether the HOMR score can be applied to patients using primary data (ie, those taken directly from the chart). This question is important for 2 reasons. First, if HOMR accurately predicts death risk using data abstracted from the medical record, it could be used in the clinical setting to assist in clinical decision-making. Second, HOMR uses multiple administrative datasets that are difficult to access and use by most clinical researchers; it is, therefore, important to determine if HOMR is accurate for clinical research based on primary medical record review. The primary objective of this study was to determine the accuracy of the HOMR score when calculated using data abstracted from clinical notes that were available when patients were admitted to hospital. Secondary objectives included determining whether functional measures abstracted were significantly associated with death risk beyond the HOMR score and whether HOMR scores calculated from chart review deviated from those calculated from administrative data.
METHODS
Study Cohort
The study, which was approved by our local research ethics board, took place at the Ottawa Hospital, a 1000-bed teaching hospital that is the primary referral center in our region. We used the hospital admission registry to identify all people 18 years or older who were admitted to a nonpsychiatric service at our hospital between January 1, 2011 and December 31, 2011 (this time frame corresponds with the year used to derive the HOMR score). We excluded overnight patients in the same-day surgery or the bone-marrow transplant units (since they would not have been included in the original study) and those without a valid health card number (which was required to link to provincial data to identify outcomes). From this list, we randomly selected 5000 patients.
Primary Data Collection
For each patient, we retrieved all data required to calculate the HOMR score from the medical record (Table 1). Patient registration information in our electronic medical record was used to identify patient age, sex, admitting service, number of emergency department (ED) visits in the previous year, number of admissions in the previous year (the nursing triage note was reviewed for each admission to determine if it was by ambulance), and whether or not the patient had been discharged from hospital in the previous 30 days. The admitting service consult note was used to determine the admitting diagnosis and whether or not the patient was admitted directly to the intensive care unit. If they were present, the emergency nursing triage note, the ED record of treatment, the admission consult note, the pre-operative consult note, and consult notes were all used to determine the patient’s comorbidities, living status, and home oxygen status. Admission urgency was determined using information from the patient registration information and the ED nursing triage note. All data were abstracted from information that had been registered prior to when the patient was physically transferred to their hospital bed. This ensured that we used only data available at the start of the admission.
Patient functional status has been shown to be strongly associated with survival4 but HOMR only indirectly captures functional information (through the patient’s living status). We, therefore, collected more detailed functional information from the medical record by determining if the patient was dependent for any activities of daily living (ADL) from the emergency nursing triage note, the ED record of treatment, the admission consult note, and the pre-operative consultation. We also collected information that might indicate frailty, which we defined per Clegg et al.5 as “a state of increased vulnerability to poor resolution of homeostasis following a stress.” This information included: delirium or more than 1 fall recorded on the emergency nursing triage note, the ED record of treatment, or the admission consultation note; or whether a geriatric nursing specialist assessment occurred in the ED in the previous 6 months. Finally, we recorded possible indicators of limited social support (no fixed address [from patient registration and nursing triage note], primary contact is not a family member [from the emergency notes, consult, and patient registration], and no religion noted in system [from patient registration]). Patients for whom religion status was missing were classified as having “no religion.”
Analysis
These data were encrypted and linked anonymously to population-based databases to determine whether patients died within 1 year of admission to hospital. We calculated the chart-HOMR score using information from the chart review and determined its association with the outcome using bivariate logistic regression. We compared observed and expected risk of death within 1 year of admission to hospital for each chart-HOMR score value, with expected risks determined from the external validation study.3 We regressed observed death risks on expected death risks for chart-HOMR scores (clustered into 22 groups to ensure adequate numbers in each group); and we gauged overall deviations from expected risk and the relationship between the observed and expected death risk (based on the chart-HOMR score) using the line’s intercept and slope, respectively.6 Next, we replicated methods from our studies2,3 to calculate the administrative-HOMR score in our study cohort using administrative databases. We compared these chart-HOMR and administrative-HOMR scores (and scores for each of its components). Finally, we determined which of the socio-functional factors were associated with 1-year death risk independent of the chart-HOMR score. We used the likelihood ratio test to determine whether these additional socio-functional factors significantly improved the model beyond the chart-HOMR score.7 This test subtracted the -2 logL value of the full model from that containing the chart-HOMR score alone, comparing its value to the χ2 distribution (with degrees of freedom equivalent to the number of additional parameters in the nested model) to determine statistical significance. All analyses were completed using SAS v9.4 (SAS Institute Inc., Cary, North Carolina).
RESULTS
There were 43,883 overnight hospitalizations at our hospital in 2011, and 38,886 hospitalizations were excluded: 1883 hospitalizations were in the same-day surgery or the bone-marrow transplant unit; 2485 did not have a valid health card number; 34,515 were not randomly selected; the records of 3 randomly selected patients had been blocked by our hospital’s privacy department; and 1 patient could not be linked with the population-based administrative datasets.
The 4996 study patients were middle-aged and predominantly female (Table 2). The extensive majority of patients was admitted from the community, was independent for ADL, had a family member as the principal contact, and had no admissions by ambulance in the previous year. Most people had no significant comorbidities or ED visits in the year prior to their admission. The mean chart-HOMR score was 22 (standard deviation [SD], 12), which is associated with a 1.2% expected risk of death within 1 year of hospital admission (Appendix 1).3
A total of 563 patients (11.3%) died within 1 year of admission to hospital (Table 2). In the study cohort, each chart-HOMR component was associated with death status. People who died were older, more likely to be male, had a greater number of important comorbidities, had more ED visits and admissions by ambulance in the previous year, and were more likely to have been discharged in the previous 30 days, and were admitted urgently, directly to the intensive care unit, or with complicated diagnoses. The mean chart-HOMR score differed extensively by survival status (37.4 [SD, 7.5] in those who died vs. 19.9 [SD, 12.2] in those who survived). Three of the socio-functional variables (delirium and falls noted on admission documents, and dependent for any ADL) also varied with death status.
The chart-HOMR score was strongly associated with the likelihood of death within 1 year of admission. When included in a logistic regression model having 1-year death as the outcome, a 1-point increase in the chart-HOMR score was associated with a 19% increase in the odds of death (P < 0.0001). This model (with only the chart-HOMR score) was highly discriminative (C statistic, 0.888) and well calibrated (Hosmer-Lemeshow test, 12.9 [8 df, P = 0.11]).
Observed and expected death risks by chart-HOMR score were similar (Figure 1). The observed total number of deaths (n = 563; 11.3%) exceeded the expected number of deaths (n = 437, 8.7%). When we regressed observed death risks on expected death risks for chart-HOMR scores (clustered into 22 groups), the Hosmer-Lemeshow test was significant, indicating that differences between observed and expected risks were beyond that expected by chance (Hosmer-Lemeshow test, 141.9, 21 df, P < 0.0001). The intercept of this model (0.035; 95% CI, 0.01-0.06) was statistically significant (P = 0.01), indicating that the observed number of cases significantly exceeded the expected; however, its calibration slope (1.02; 95% CI, 0.89-1.16) did not deviate significantly from unity, indicating that the relationship between the observed and expected death risk (based on the chart-HOMR score) remained intact (Figure 1).
The deviations between observed and expected death risks reflected deviations between the c chart-HOMR score and the administrative-HOMR score, with the former being significantly lower than the latter (Figure 2). Overall, the chart-HOMR score was 0.96 points lower (95% CI, 0.81-1.12) than the administrative-HOMR score. The HOMR score components that were notably underestimated using chart data included those for the age-Charlson Comorbidity Index interaction, living status, and admit points. Points for only 2 components (admitting service and admission urgency) were higher when calculated using chart data.
Four additional socio-functional variables collected from medical record review were significantly associated with 1-year death risk independent of the chart-HOMR score (Table 3). Admission documentation noting either delirium or falls were both associated with a significantly increased death risk (adjusted odds ratio [OR], 1.92 [95% CI, 1.24-2.96] and OR 1.96 [95% CI, 1.29-2.99], respectively). An independently increased death risk was also noted in patients who were dependent for any ADL (adjusted OR, 1.99 [95% CI, 1.24-3.19]). The presence of an ED geriatrics consultation within the previous 6 months was associated with a significantly decreased death risk of 60% (adjusted OR, 0.40 [95% CI, 0.20-0.81]). Adding these covariates to the logistic model with the chart-HOMR score significantly improved predictions (likelihood ratio statistic = 33.569, 4df, P < 0.00001).
DISCUSSION
In a large random sample of patients from our hospital, we found that the HOMR score using data abstracted from the medical record was significantly associated with 1-year death risk. The expected death risk based on the chart-HOMR score underestimated observed death risk but the relationship between the chart-HOMR score and death risk was similar to that in studies using administrative data. The HOMR score calculated using data from the chart was lower than that calculated using data from population-based administrative datasets; additional variables indicating patient frailty were significantly associated with 1-year death risk independent of the chart-HOMR score. Since the HOMR score was derived and initially validated using health administrative data, this study using data abstracted from the health record shows that the HOMR score has methodological generalizability.8
We think that our study has several notable findings. First, we found that data abstracted from the medical record can be used to calculate the HOMR score to accurately predict individual death risk. The chart-HOMR score discriminated very well between patients who did and did not die (C statistic, 0.88), which extensively exceeds the discrimination of published death risk indices (whose C statistics range between 0.69 and 0.82). It is also possible that chart abstraction for the HOMR score—without functional status—is simpler than other indices since its components are primarily very objective. (Other indices for hospital-based patients required factors that could be difficult to abstract reliably from the medical record including meeting more than 1 guideline for noncancer hospice care9; ambulation difficulties10; scales such as the Exton-Smith Scale or the Short Portable Mental Status Questionnaire11; weight loss12; functional status4; and pressure sore risk.13) Although expected risks for the chart-HOMR consistently underestimated observed risks (Figure 1), the mean deviation was small (with an absolute difference of 3.5% that can be used as a correction factor when determining expected risks with HOMR scores calculated from chart review), but it was an association between the chart-HOMR score and death risk that remained consistent through the cohort. Second, we found a small but significant decrease in the chart-HOMR score vs. the administrative-HOMR score (Figure 2). Some of these underestimates such as those for the number of ED visits or admissions by ambulance were expected since population-based health administrative databases would best capture such data. However, we were surprised that the comorbidity score was less when calculated using chart vs. database data (Figure 2). This finding is distinct from studies finding that particular comorbidities are documented in the chart are sometimes not coded.14,15 However, we identified comorbidities in the administrative databases using a 1-year ‘look-back’ period so that diagnostic codes from multiple hospitalizations (and from multiple hospitals) could be used to calculate the Charlson Comorbidity Index for a particular patient; this has been shown to increase the capture of comorbidities.16 Third, we found that variables from the chart review indicating frailty were predictive of 1-year death risk independent of the chart-HOMR score (Table 2). This illustrates that mortality risk prediction can be improved for particular patient groups by adding new covariates to the HOMR. Further work is required to determine how to incorporate these (and possibly other) covariates into the HOMR to create a unique chart-HOMR score. Finally, we found that a geriatrics assessment in the ED was associated with a significant (and notable) decrease in death risk. With these data, we are unable to indicate whether this association is causative. However, these findings indicate that the influence of emergency geriatric assessments on patient survival needs to be explored in more detail.
Several issues about our study should be considered when interpreting its results. First, this was a single-center study and the generalizability of our results to other centers is unknown. However, our study had the largest sample size of all primary data prognostic index validation studies1 ensuring that our results are, at the very least, internally reliable. In addition, our simple random sample ensured that we studied a broad assortment of patients to be certain that our results are representative of our institution. Second, we used a single abstractor for the study, which could limit the generalizability of our results. However, almost all the data points that were abstracted for our study were very objective.
In summary, our study shows that the HOMR score can be used to accurately predict 1-year death risk using data abstracted from the patient record. These findings will aid in individual patient prognostication for clinicians and researchers.
Disclosure
The authors report no financial conflicts of interest.
A patient’s prognosis can strongly influence their medical care. Decisions about diagnostic modalities, treatment options, and the use of preventive therapies can all be affected by the likelihood of a patient’s death in the near future. For example, patients with severely limited survival might forego prophylactic therapy, avoid interventions for asymptomatic issues, and cease screening interventions. Knowing survival probability would also be very helpful as a controlling variable in research analyses whenever death risk might be a possible confounder.
Sixteen indices that aim to predict patient death risk have been described by Yourman et al.1 They were all created from secondary analyses of clinical and administrative datasets, were applicable to patients in a variety of settings (including the community, nursing home, or hospital), and predicted survival probabilities in time horizons ranging from 6 months to 5 years. Prognostic factors that were most commonly included in these indices were comorbidity and functional status. In validation populations, the discrimination of these indices for 1-year survival in hospitalized patients was moderate (with C statistics that ranged from 0.64 to 0.79) with good calibration for broad prognostic ranges.
In 2014, we published the Hospitalized-patient One-year Mortality Risk (HOMR) score.2 This study used health administrative data for all adult Ontarians admitted in 2011 to hospital under nonpsychiatric services (n = 640,022) to estimate the probability of dying within 1 year of admission to hospital (which happened in 11.7% of people). The HOMR score included 12 patient and hospitalization factors (Table 1). It was highly discriminative (C statistic, 0.923; [0.922-0.924]) and well calibrated (the mean relative difference between observed and expected death risk was 2.0% [range, 0.0% to 7.0%]). It was externally validated in more than 3 million adults from Ontario, Alberta, and Boston in whom the C statistic ranged from 0.89 to 0.92 and calibration was excellent.3 We concluded from these studies that the HOMR score is excellent for prognosticating a diverse group of patients using health administrative data.
However, we do not know whether the HOMR score can be applied to patients using primary data (ie, those taken directly from the chart). This question is important for 2 reasons. First, if HOMR accurately predicts death risk using data abstracted from the medical record, it could be used in the clinical setting to assist in clinical decision-making. Second, HOMR uses multiple administrative datasets that are difficult to access and use by most clinical researchers; it is, therefore, important to determine if HOMR is accurate for clinical research based on primary medical record review. The primary objective of this study was to determine the accuracy of the HOMR score when calculated using data abstracted from clinical notes that were available when patients were admitted to hospital. Secondary objectives included determining whether functional measures abstracted were significantly associated with death risk beyond the HOMR score and whether HOMR scores calculated from chart review deviated from those calculated from administrative data.
METHODS
Study Cohort
The study, which was approved by our local research ethics board, took place at the Ottawa Hospital, a 1000-bed teaching hospital that is the primary referral center in our region. We used the hospital admission registry to identify all people 18 years or older who were admitted to a nonpsychiatric service at our hospital between January 1, 2011 and December 31, 2011 (this time frame corresponds with the year used to derive the HOMR score). We excluded overnight patients in the same-day surgery or the bone-marrow transplant units (since they would not have been included in the original study) and those without a valid health card number (which was required to link to provincial data to identify outcomes). From this list, we randomly selected 5000 patients.
Primary Data Collection
For each patient, we retrieved all data required to calculate the HOMR score from the medical record (Table 1). Patient registration information in our electronic medical record was used to identify patient age, sex, admitting service, number of emergency department (ED) visits in the previous year, number of admissions in the previous year (the nursing triage note was reviewed for each admission to determine if it was by ambulance), and whether or not the patient had been discharged from hospital in the previous 30 days. The admitting service consult note was used to determine the admitting diagnosis and whether or not the patient was admitted directly to the intensive care unit. If they were present, the emergency nursing triage note, the ED record of treatment, the admission consult note, the pre-operative consult note, and consult notes were all used to determine the patient’s comorbidities, living status, and home oxygen status. Admission urgency was determined using information from the patient registration information and the ED nursing triage note. All data were abstracted from information that had been registered prior to when the patient was physically transferred to their hospital bed. This ensured that we used only data available at the start of the admission.
Patient functional status has been shown to be strongly associated with survival4 but HOMR only indirectly captures functional information (through the patient’s living status). We, therefore, collected more detailed functional information from the medical record by determining if the patient was dependent for any activities of daily living (ADL) from the emergency nursing triage note, the ED record of treatment, the admission consult note, and the pre-operative consultation. We also collected information that might indicate frailty, which we defined per Clegg et al.5 as “a state of increased vulnerability to poor resolution of homeostasis following a stress.” This information included: delirium or more than 1 fall recorded on the emergency nursing triage note, the ED record of treatment, or the admission consultation note; or whether a geriatric nursing specialist assessment occurred in the ED in the previous 6 months. Finally, we recorded possible indicators of limited social support (no fixed address [from patient registration and nursing triage note], primary contact is not a family member [from the emergency notes, consult, and patient registration], and no religion noted in system [from patient registration]). Patients for whom religion status was missing were classified as having “no religion.”
Analysis
These data were encrypted and linked anonymously to population-based databases to determine whether patients died within 1 year of admission to hospital. We calculated the chart-HOMR score using information from the chart review and determined its association with the outcome using bivariate logistic regression. We compared observed and expected risk of death within 1 year of admission to hospital for each chart-HOMR score value, with expected risks determined from the external validation study.3 We regressed observed death risks on expected death risks for chart-HOMR scores (clustered into 22 groups to ensure adequate numbers in each group); and we gauged overall deviations from expected risk and the relationship between the observed and expected death risk (based on the chart-HOMR score) using the line’s intercept and slope, respectively.6 Next, we replicated methods from our studies2,3 to calculate the administrative-HOMR score in our study cohort using administrative databases. We compared these chart-HOMR and administrative-HOMR scores (and scores for each of its components). Finally, we determined which of the socio-functional factors were associated with 1-year death risk independent of the chart-HOMR score. We used the likelihood ratio test to determine whether these additional socio-functional factors significantly improved the model beyond the chart-HOMR score.7 This test subtracted the -2 logL value of the full model from that containing the chart-HOMR score alone, comparing its value to the χ2 distribution (with degrees of freedom equivalent to the number of additional parameters in the nested model) to determine statistical significance. All analyses were completed using SAS v9.4 (SAS Institute Inc., Cary, North Carolina).
RESULTS
There were 43,883 overnight hospitalizations at our hospital in 2011, and 38,886 hospitalizations were excluded: 1883 hospitalizations were in the same-day surgery or the bone-marrow transplant unit; 2485 did not have a valid health card number; 34,515 were not randomly selected; the records of 3 randomly selected patients had been blocked by our hospital’s privacy department; and 1 patient could not be linked with the population-based administrative datasets.
The 4996 study patients were middle-aged and predominantly female (Table 2). The extensive majority of patients was admitted from the community, was independent for ADL, had a family member as the principal contact, and had no admissions by ambulance in the previous year. Most people had no significant comorbidities or ED visits in the year prior to their admission. The mean chart-HOMR score was 22 (standard deviation [SD], 12), which is associated with a 1.2% expected risk of death within 1 year of hospital admission (Appendix 1).3
A total of 563 patients (11.3%) died within 1 year of admission to hospital (Table 2). In the study cohort, each chart-HOMR component was associated with death status. People who died were older, more likely to be male, had a greater number of important comorbidities, had more ED visits and admissions by ambulance in the previous year, and were more likely to have been discharged in the previous 30 days, and were admitted urgently, directly to the intensive care unit, or with complicated diagnoses. The mean chart-HOMR score differed extensively by survival status (37.4 [SD, 7.5] in those who died vs. 19.9 [SD, 12.2] in those who survived). Three of the socio-functional variables (delirium and falls noted on admission documents, and dependent for any ADL) also varied with death status.
The chart-HOMR score was strongly associated with the likelihood of death within 1 year of admission. When included in a logistic regression model having 1-year death as the outcome, a 1-point increase in the chart-HOMR score was associated with a 19% increase in the odds of death (P < 0.0001). This model (with only the chart-HOMR score) was highly discriminative (C statistic, 0.888) and well calibrated (Hosmer-Lemeshow test, 12.9 [8 df, P = 0.11]).
Observed and expected death risks by chart-HOMR score were similar (Figure 1). The observed total number of deaths (n = 563; 11.3%) exceeded the expected number of deaths (n = 437, 8.7%). When we regressed observed death risks on expected death risks for chart-HOMR scores (clustered into 22 groups), the Hosmer-Lemeshow test was significant, indicating that differences between observed and expected risks were beyond that expected by chance (Hosmer-Lemeshow test, 141.9, 21 df, P < 0.0001). The intercept of this model (0.035; 95% CI, 0.01-0.06) was statistically significant (P = 0.01), indicating that the observed number of cases significantly exceeded the expected; however, its calibration slope (1.02; 95% CI, 0.89-1.16) did not deviate significantly from unity, indicating that the relationship between the observed and expected death risk (based on the chart-HOMR score) remained intact (Figure 1).
The deviations between observed and expected death risks reflected deviations between the c chart-HOMR score and the administrative-HOMR score, with the former being significantly lower than the latter (Figure 2). Overall, the chart-HOMR score was 0.96 points lower (95% CI, 0.81-1.12) than the administrative-HOMR score. The HOMR score components that were notably underestimated using chart data included those for the age-Charlson Comorbidity Index interaction, living status, and admit points. Points for only 2 components (admitting service and admission urgency) were higher when calculated using chart data.
Four additional socio-functional variables collected from medical record review were significantly associated with 1-year death risk independent of the chart-HOMR score (Table 3). Admission documentation noting either delirium or falls were both associated with a significantly increased death risk (adjusted odds ratio [OR], 1.92 [95% CI, 1.24-2.96] and OR 1.96 [95% CI, 1.29-2.99], respectively). An independently increased death risk was also noted in patients who were dependent for any ADL (adjusted OR, 1.99 [95% CI, 1.24-3.19]). The presence of an ED geriatrics consultation within the previous 6 months was associated with a significantly decreased death risk of 60% (adjusted OR, 0.40 [95% CI, 0.20-0.81]). Adding these covariates to the logistic model with the chart-HOMR score significantly improved predictions (likelihood ratio statistic = 33.569, 4df, P < 0.00001).
DISCUSSION
In a large random sample of patients from our hospital, we found that the HOMR score using data abstracted from the medical record was significantly associated with 1-year death risk. The expected death risk based on the chart-HOMR score underestimated observed death risk but the relationship between the chart-HOMR score and death risk was similar to that in studies using administrative data. The HOMR score calculated using data from the chart was lower than that calculated using data from population-based administrative datasets; additional variables indicating patient frailty were significantly associated with 1-year death risk independent of the chart-HOMR score. Since the HOMR score was derived and initially validated using health administrative data, this study using data abstracted from the health record shows that the HOMR score has methodological generalizability.8
We think that our study has several notable findings. First, we found that data abstracted from the medical record can be used to calculate the HOMR score to accurately predict individual death risk. The chart-HOMR score discriminated very well between patients who did and did not die (C statistic, 0.88), which extensively exceeds the discrimination of published death risk indices (whose C statistics range between 0.69 and 0.82). It is also possible that chart abstraction for the HOMR score—without functional status—is simpler than other indices since its components are primarily very objective. (Other indices for hospital-based patients required factors that could be difficult to abstract reliably from the medical record including meeting more than 1 guideline for noncancer hospice care9; ambulation difficulties10; scales such as the Exton-Smith Scale or the Short Portable Mental Status Questionnaire11; weight loss12; functional status4; and pressure sore risk.13) Although expected risks for the chart-HOMR consistently underestimated observed risks (Figure 1), the mean deviation was small (with an absolute difference of 3.5% that can be used as a correction factor when determining expected risks with HOMR scores calculated from chart review), but it was an association between the chart-HOMR score and death risk that remained consistent through the cohort. Second, we found a small but significant decrease in the chart-HOMR score vs. the administrative-HOMR score (Figure 2). Some of these underestimates such as those for the number of ED visits or admissions by ambulance were expected since population-based health administrative databases would best capture such data. However, we were surprised that the comorbidity score was less when calculated using chart vs. database data (Figure 2). This finding is distinct from studies finding that particular comorbidities are documented in the chart are sometimes not coded.14,15 However, we identified comorbidities in the administrative databases using a 1-year ‘look-back’ period so that diagnostic codes from multiple hospitalizations (and from multiple hospitals) could be used to calculate the Charlson Comorbidity Index for a particular patient; this has been shown to increase the capture of comorbidities.16 Third, we found that variables from the chart review indicating frailty were predictive of 1-year death risk independent of the chart-HOMR score (Table 2). This illustrates that mortality risk prediction can be improved for particular patient groups by adding new covariates to the HOMR. Further work is required to determine how to incorporate these (and possibly other) covariates into the HOMR to create a unique chart-HOMR score. Finally, we found that a geriatrics assessment in the ED was associated with a significant (and notable) decrease in death risk. With these data, we are unable to indicate whether this association is causative. However, these findings indicate that the influence of emergency geriatric assessments on patient survival needs to be explored in more detail.
Several issues about our study should be considered when interpreting its results. First, this was a single-center study and the generalizability of our results to other centers is unknown. However, our study had the largest sample size of all primary data prognostic index validation studies1 ensuring that our results are, at the very least, internally reliable. In addition, our simple random sample ensured that we studied a broad assortment of patients to be certain that our results are representative of our institution. Second, we used a single abstractor for the study, which could limit the generalizability of our results. However, almost all the data points that were abstracted for our study were very objective.
In summary, our study shows that the HOMR score can be used to accurately predict 1-year death risk using data abstracted from the patient record. These findings will aid in individual patient prognostication for clinicians and researchers.
Disclosure
The authors report no financial conflicts of interest.
1. Yourman LC, Lee SJ, Schonberg MA, Widera EW, Smith AK. Prognostic indices for older adults: a systematic review. JAMA. 2012;307(2):182-192. PubMed
2. van Walraven C. The Hospital-patient One-year Mortality Risk score accurately predicts long term death risk in hospitalized patients. J Clin Epidemiol. 2014;67(9):1025-1034. PubMed
3. van Walraven C, McAlister FA, Bakal JA, Hawken S, Donzé J. External validation of the Hospital-patient One-year Mortality Risk (HOMR) model for predicting death within 1 year after hospital admission. CMAJ. 2015;187(10):725-733. PubMed
4. Walter LC, Brand RJ, Counsell SR, et al. Development and validation of a prognostic index for 1-year mortality in older adults after hospitalization. JAMA. 2001;285(23):2987-2994. PubMed
5. Clegg A, Young J, Iliffe S et al. Frailty in elderly people. The Lancet 2002;381:752-762. PubMed
6. Crowson CS, Atkinson EJ, Therneau TM. Assessing calibration of prognostic risk scores. Stat Methods Med Res. 2016;25(4):1692-1706. PubMed
7. Harrell FE Jr. Overview of Maximum Likelihood Estimation. Regression Modeling Strategies. New York, NY: Springer-Verlag; 2001: 179-212.
8. Justice AC, Covinsky KE, Berlin JA. Assessing the generalizability of prognostic information. Ann Intern Med. 1999;130(6):515-524. PubMed
9. Fischer SM, Gozansky WS, Sauaia A, Min SJ, Kutner JS, Kramer A. A practical tool to identify patients who may benefit from a palliative approach: the CARING criteria. J Pain Symptom Manage. 2006;31(4):285-292. PubMed
10. Inouye SK, Bogardus ST, Jr, Vitagliano G, et al. Burden of illness score for elderly persons: risk adjustment incorporating the cumulative impact of diseases, physiologic abnormalities, and functional impairments. Med Care. 2003;41(1):70-83. PubMed
11. Pilotto A, Ferrucci L, Franceschi M, et al. Development and validation of a multidimensional prognostic index for one-year mortality from comprehensive geriatric assessment in hospitalized older patients. Rejuvenation Res. 2008;11(1):151-161. PubMed
12. Teno JM, Harrell FE Jr, Knaus W, et al. Prediction of survival for older hospitalized patients: the HELP survival model. Hospitalized Elderly Longitudinal Project. J Am Geriatr Soc. 2000;48(5 suppl):S16-S24. PubMed
13. Dramé M, Novella JL, Lang PO, et al. Derivation and validation of a mortality-risk index from a cohort of frail elderly patients hospitalised in medical wards via emergencies: the SAFES study. Eur J Epidemiol. 2008;23(12):783-791. PubMed
14. Kieszak SM, Flanders WD, Kosinski AS, Shipp CC, Karp H. A comparison of the Charlson comorbidity index derived from medical record data and administrative billing data. J Clin Epidemiol. 1999;52(2):137-142. PubMed
15. Quan H, Parsons GA, Ghali WA. Validity of procedure codes in International Classification of Diseases, 9th revision, clinical modification administrative data. Med Care. 2004;42(8):801-809. PubMed
16. Zhang JX, Iwashyna TJ, Christakis NA. The performance of different lookback periods and sources of information for Charlson cComorbidity adjustment in Medicare claims. Med Care. 1999;37(11):1128-1139. PubMed
1. Yourman LC, Lee SJ, Schonberg MA, Widera EW, Smith AK. Prognostic indices for older adults: a systematic review. JAMA. 2012;307(2):182-192. PubMed
2. van Walraven C. The Hospital-patient One-year Mortality Risk score accurately predicts long term death risk in hospitalized patients. J Clin Epidemiol. 2014;67(9):1025-1034. PubMed
3. van Walraven C, McAlister FA, Bakal JA, Hawken S, Donzé J. External validation of the Hospital-patient One-year Mortality Risk (HOMR) model for predicting death within 1 year after hospital admission. CMAJ. 2015;187(10):725-733. PubMed
4. Walter LC, Brand RJ, Counsell SR, et al. Development and validation of a prognostic index for 1-year mortality in older adults after hospitalization. JAMA. 2001;285(23):2987-2994. PubMed
5. Clegg A, Young J, Iliffe S et al. Frailty in elderly people. The Lancet 2002;381:752-762. PubMed
6. Crowson CS, Atkinson EJ, Therneau TM. Assessing calibration of prognostic risk scores. Stat Methods Med Res. 2016;25(4):1692-1706. PubMed
7. Harrell FE Jr. Overview of Maximum Likelihood Estimation. Regression Modeling Strategies. New York, NY: Springer-Verlag; 2001: 179-212.
8. Justice AC, Covinsky KE, Berlin JA. Assessing the generalizability of prognostic information. Ann Intern Med. 1999;130(6):515-524. PubMed
9. Fischer SM, Gozansky WS, Sauaia A, Min SJ, Kutner JS, Kramer A. A practical tool to identify patients who may benefit from a palliative approach: the CARING criteria. J Pain Symptom Manage. 2006;31(4):285-292. PubMed
10. Inouye SK, Bogardus ST, Jr, Vitagliano G, et al. Burden of illness score for elderly persons: risk adjustment incorporating the cumulative impact of diseases, physiologic abnormalities, and functional impairments. Med Care. 2003;41(1):70-83. PubMed
11. Pilotto A, Ferrucci L, Franceschi M, et al. Development and validation of a multidimensional prognostic index for one-year mortality from comprehensive geriatric assessment in hospitalized older patients. Rejuvenation Res. 2008;11(1):151-161. PubMed
12. Teno JM, Harrell FE Jr, Knaus W, et al. Prediction of survival for older hospitalized patients: the HELP survival model. Hospitalized Elderly Longitudinal Project. J Am Geriatr Soc. 2000;48(5 suppl):S16-S24. PubMed
13. Dramé M, Novella JL, Lang PO, et al. Derivation and validation of a mortality-risk index from a cohort of frail elderly patients hospitalised in medical wards via emergencies: the SAFES study. Eur J Epidemiol. 2008;23(12):783-791. PubMed
14. Kieszak SM, Flanders WD, Kosinski AS, Shipp CC, Karp H. A comparison of the Charlson comorbidity index derived from medical record data and administrative billing data. J Clin Epidemiol. 1999;52(2):137-142. PubMed
15. Quan H, Parsons GA, Ghali WA. Validity of procedure codes in International Classification of Diseases, 9th revision, clinical modification administrative data. Med Care. 2004;42(8):801-809. PubMed
16. Zhang JX, Iwashyna TJ, Christakis NA. The performance of different lookback periods and sources of information for Charlson cComorbidity adjustment in Medicare claims. Med Care. 1999;37(11):1128-1139. PubMed
© 2017 Society of Hospital Medicine
Predicting 30-day pneumonia readmissions using electronic health record data
Pneumonia is a leading cause of hospitalizations in the U.S., accounting for more than 1.1 million discharges annually.1 Pneumonia is frequently complicated by hospital readmission, which is costly and potentially avoidable.2,3 Due to financial penalties imposed on hospitals for higher than expected 30-day readmission rates, there is increasing attention to implementing interventions to reduce readmissions in this population.4,5 However, because these programs are resource-intensive, interventions are thought to be most cost-effective if they are targeted to high-risk individuals who are most likely to benefit.6-8
Current pneumonia-specific readmission risk-prediction models that could enable identification of high-risk patients suffer from poor predictive ability, greatly limiting their use, and most were validated among older adults or by using data from single academic medical centers, limiting their generalizability.9-14 A potential reason for poor predictive accuracy is the omission of known robust clinical predictors of pneumonia-related outcomes, including pneumonia severity of illness and stability on discharge.15-17 Approaches using electronic health record (EHR) data, which include this clinically granular data, could enable hospitals to more accurately and pragmatically identify high-risk patients during the index hospitalization and enable interventions to be initiated prior to discharge.
An alternative strategy to identifying high-risk patients for readmission is to use a multi-condition risk-prediction model. Developing and implementing models for every condition may be time-consuming and costly. We have derived and validated 2 multi-condition risk-prediction models using EHR data—1 using data from the first day of hospital admission (‘first-day’ model), and the second incorporating data from the entire hospitalization (‘full-stay’ model) to reflect in-hospital complications and clinical stability at discharge.18,19 However, it is unknown if a multi-condition model for pneumonia would perform as well as a disease-specific model.
This study aimed to develop 2 EHR-based pneumonia-specific readmission risk-prediction models using data routinely collected in clinical practice—a ‘first-day’ and a ‘full-stay’ model—and compare the performance of each model to: 1) one another; 2) the corresponding multi-condition EHR model; and 3) to other potentially useful models in predicting pneumonia readmissions (the Centers for Medicare and Medicaid Services [CMS] pneumonia model, and 2 commonly used pneumonia severity of illness scores validated for predicting mortality). We hypothesized that the pneumonia-specific EHR models would outperform other models; and the full-stay pneumonia-specific model would outperform the first-day pneumonia-specific model.
METHODS
Study Design, Population, and Data Sources
We conducted an observational study using EHR data collected from 6 hospitals (including safety net, community, teaching, and nonteaching hospitals) in north Texas between November 2009 and October 2010, All hospitals used the Epic EHR (Epic Systems Corporation, Verona, WI). Details of this cohort have been published.18,19
We included consecutive hospitalizations among adults 18 years and older discharged from any medicine service with principal discharge diagnoses of pneumonia (ICD-9-CM codes 480-483, 485, 486-487), sepsis (ICD-9-CM codes 038, 995.91, 995.92, 785.52), or respiratory failure (ICD-9-CM codes 518.81, 518.82, 518.84, 799.1) when the latter 2 were also accompanied by a secondary diagnosis of pneumonia.20 For individuals with multiple hospitalizations during the study period, we included only the first hospitalization. We excluded individuals who died during the index hospitalization or within 30 days of discharge, were transferred to another acute care facility, or left against medical advice.
Outcomes
The primary outcome was all-cause 30-day readmission, defined as a nonelective hospitalization within 30 days of discharge to any of 75 acute care hospitals within a 100-mile radius of Dallas, ascertained from an all-payer regional hospitalization database.
Predictor Variables for the Pneumonia-Specific Readmission Models
The selection of candidate predictors was informed by our validated multi-condition risk-prediction models using EHR data available within 24 hours of admission (‘first-day’ multi-condition EHR model) or during the entire hospitalization (‘full-stay’ multi-condition EHR model).18,19 For the pneumonia-specific models, we included all variables in our published multi-condition models as candidate predictors, including sociodemographics, prior utilization, Charlson Comorbidity Index, select laboratory and vital sign abnormalities, length of stay, hospital complications (eg, venous thromboembolism), vital sign instabilities, and disposition status (see Supplemental Table 1 for complete list of variables). We also assessed additional variables specific to pneumonia for inclusion that were: (1) available in the EHR of all participating hospitals; (2) routinely collected or available at the time of admission or discharge; and (3) plausible predictors of adverse outcomes based on literature and clinical expertise. These included select comorbidities (eg, psychiatric conditions, chronic lung disease, history of pneumonia),10,11,21,22 the pneumonia severity index (PSI),16,23,24 intensive care unit stay, and receipt of invasive or noninvasive ventilation. We used a modified PSI score because certain data elements were missing. The modified PSI (henceforth referred to as PSI) did not include nursing home residence and included diagnostic codes as proxies for the presence of pleural effusion (ICD-9-CM codes 510, 511.1, and 511.9) and altered mental status (ICD-9-CM codes 780.0X, 780.97, 293.0, 293.1, and 348.3X).
Statistical Analysis
Model Derivation. Candidate predictor variables were classified as available in the EHR within 24 hours of admission and/or at the time of discharge. For example, socioeconomic factors could be ascertained within the first day of hospitalization, whereas length of stay would not be available until the day of discharge. Predictors with missing values were assumed to be normal (less than 1% missing for each variable). Univariate relationships between readmission and each candidate predictor were assessed in the overall cohort using a pre-specified significance threshold of P ≤ 0.10. Significant variables were entered in the respective first-day and full-stay pneumonia-specific multivariable logistic regression models using stepwise-backward selection with a pre-specified significance threshold of P ≤ 0.05. In sensitivity analyses, we alternately derived our models using stepwise-forward selection, as well as stepwise-backward selection minimizing the Bayesian information criterion and Akaike information criterion separately. These alternate modeling strategies yielded identical predictors to our final models.
Model Validation. Model validation was performed using 5-fold cross-validation, with the overall cohort randomly divided into 5 equal-size subsets.25 For each cycle, 4 subsets were used for training to estimate model coefficients, and the fifth subset was used for validation. This cycle was repeated 5 times with each randomly-divided subset used once as the validation set. We repeated this entire process 50 times and averaged the C statistic estimates to derive an optimism-corrected C statistic. Model calibration was assessed qualitatively by comparing predicted to observed probabilities of readmission by quintiles of predicted risk, and with the Hosmer-Lemeshow goodness-of-fit test.
Comparison to Other Models. The main comparisons of the first-day and full-stay pneumonia-specific EHR model performance were to each other and the corresponding multi-condition EHR model.18,19 The multi-condition EHR models were separately derived and validated within the larger parent cohort from which this study cohort was derived, and outperformed the CMS all-cause model, the HOSPITAL model, and the LACE index.19 To further triangulate our findings, given the lack of other rigorously validated pneumonia-specific risk-prediction models for readmission,14 we compared the pneumonia-specific EHR models to the CMS pneumonia model derived from administrative claims data,10 and 2 commonly used risk-prediction scores for short-term mortality among patients with community-acquired pneumonia, the PSI and CURB-65 scores.16 Although derived and validated using patient-level data, the CMS model was developed to benchmark hospitals according to hospital-level readmission rates.10 The CURB-65 score in this study was also modified to include the same altered mental status diagnostic codes according to the modified PSI as a proxy for “confusion.” Both the PSI and CURB-65 scores were calculated using the most abnormal values within the first 24 hours of admission. The ‘updated’ PSI and the ‘updated’ CURB-65 were calculated using the most abnormal values within 24 hours prior to discharge, or the last known observation prior to discharge if no results were recorded within this time period. A complete list of variables for each of the comparison models are shown in Supplemental Table 1.
We assessed model performance by calculating the C statistic, integrated discrimination index, and net reclassification index (NRI) compared to our pneumonia-specific models. The integrated discrimination index is the difference in the mean predicted probability of readmission between patients who were and were not actually readmitted between 2 models, where more positive values suggest improvement in model performance compared to a reference model.26 The NRI is defined as the sum of the net proportions of correctly reclassified persons with and without the event of interest.27 Here, we calculated a category-based NRI to evaluate the performance of pneumonia-specific models in correctly classifying individuals with and without readmissions into the 2 highest readmission risk quintiles vs the lowest 3 risk quintiles compared to other models.27 This pre-specified cutoff is relevant for hospitals interested in identifying the highest risk individuals for targeted intervention.7 Finally, we assessed calibration of comparator models in our cohort by comparing predicted probability to observed probability of readmission by quintiles of risk for each model. We conducted all analyses using Stata 12.1 (StataCorp, College Station, Texas). This study was approved by the University of Texas Southwestern Medical Center Institutional Review Board.
RESULTS
Of 1463 index hospitalizations (Supplemental Figure 1), the 30-day all-cause readmission rate was 13.6%. Individuals with a 30-day readmission had markedly different sociodemographic and clinical characteristics compared to those not readmitted (Table 1; see Supplemental Table 2 for additional clinical characteristics).
Derivation, Validation, and Performance of the Pneumonia-Specific Readmission Risk-Prediction Models
The final first-day pneumonia-specific EHR model included 7 variables, including sociodemographic characteristics; prior hospitalizations; thrombocytosis, and PSI (Table 2). The first-day pneumonia-specific model had adequate discrimination (C statistic, 0.695; optimism-corrected C statistic 0.675, 95% confidence interval [CI], 0.667-0.685; Table 3). It also effectively stratified individuals across a broad range of risk (average predicted decile of risk ranged from 4% to 33%; Table 3) and was well calibrated (Supplemental Table 3).
The final full-stay pneumonia-specific EHR readmission model included 8 predictors, including 3 variables from the first-day model (median income, thrombocytosis, and prior hospitalizations; Table 2). The full-stay pneumonia-specific EHR model also included vital sign instabilities on discharge, updated PSI, and disposition status (ie, being discharged with home health or to a post-acute care facility was associated with greater odds of readmission, and hospice with lower odds). The full-stay pneumonia-specific EHR model had good discrimination (C statistic, 0.731; optimism-corrected C statistic, 0.714; 95% CI, 0.706-0.720), and stratified individuals across a broad range of risk (average predicted decile of risk ranged from 3% to 37%; Table 3), and was also well calibrated (Supplemental Table 3).
First-Day Pneumonia-Specific EHR Model vs First-Day Multi-Condition EHR Model
The first-day pneumonia-specific EHR model outperformed the first-day multi-condition EHR model with better discrimination (P = 0.029) and more correctly classified individuals in the top 2 highest risk quintiles vs the bottom 3 risk quintiles (Table 3, Supplemental Table 4, and Supplemental Figure 2A). With respect to calibration, the first-day multi-condition EHR model overestimated risk among the highest quintile risk group compared to the first-day pneumonia-specific EHR model (Figure 1A, 1B).
Full-Stay Pneumonia-Specific EHR Model vs Other Models
The full-stay pneumonia-specific EHR model comparatively outperformed the corresponding full-stay multi-condition EHR model, as well as the first-day pneumonia-specific EHR model, the CMS pneumonia model, the updated PSI, and the updated CURB-65 (Table 3, Supplemental Table 5, Supplemental Table 6, and Supplemental Figures 2B and 2C). Compared to the full-stay multi-condition and first-day pneumonia-specific EHR models, the full-stay pneumonia-specific EHR model had better discrimination, better reclassification (NRI, 0.09 and 0.08, respectively), and was able to stratify individuals across a broader range of readmission risk (Table 3). It also had better calibration in the highest quintile risk group compared to the full-stay multi-condition EHR model (Figure 1C and 1D).
Updated vs First-Day Modified PSI and CURB-65 Scores
The updated PSI was more strongly predictive of readmission than the PSI calculated on the day of admission (Wald test, 9.83; P = 0.002). Each 10-point increase in the updated PSI was associated with a 22% increased odds of readmission vs an 11% increase for the PSI calculated upon admission (Table 2). The improved predictive ability of the updated PSI and CURB-65 scores was also reflected in the superior discrimination and calibration vs the respective first-day pneumonia severity of illness scores (Table 3).
DISCUSSION
Using routinely available EHR data from 6 diverse hospitals, we developed 2 pneumonia-specific readmission risk-prediction models that aimed to allow hospitals to identify patients hospitalized with pneumonia at high risk for readmission. Overall, we found that a pneumonia-specific model using EHR data from the entire hospitalization outperformed all other models—including the first-day pneumonia-specific model using data present only on admission, our own multi-condition EHR models, and the CMS pneumonia model based on administrative claims data—in all aspects of model performance (discrimination, calibration, and reclassification). We found that socioeconomic status, prior hospitalizations, thrombocytosis, and measures of clinical severity and stability were important predictors of 30-day all-cause readmissions among patients hospitalized with pneumonia. Additionally, an updated discharge PSI score was a stronger independent predictor of readmissions compared to the PSI score calculated upon admission; and inclusion of the updated PSI in our full-stay pneumonia model led to improved prediction of 30-day readmissions.
The marked improvement in performance of the full-stay pneumonia-specific EHR model compared to the first-day pneumonia-specific model suggests that clinical stability and trajectory during hospitalization (as modeled through disposition status, updated PSI, and vital sign instabilities at discharge) are important predictors of 30-day readmission among patients hospitalized for pneumonia, which was not the case for our EHR-based multi-condition models.19 With the inclusion of these measures, the full-stay pneumonia-specific model correctly reclassified an additional 8% of patients according to their true risk compared to the first-day pneumonia-specific model. One implication of these findings is that hospitals interested in targeting their highest risk individuals with pneumonia for transitional care interventions could do so using the first-day pneumonia-specific EHR model and could refine their targeted strategy at the time of discharge by using the full-stay pneumonia model. This staged risk-prediction strategy would enable hospitals to initiate transitional care interventions for high-risk individuals in the inpatient setting (ie, patient education).7 Then, hospitals could enroll both persistent and newly identified high-risk individuals for outpatient interventions (ie, follow-up telephone call) in the immediate post-discharge period, an interval characterized by heightened vulnerability for adverse events,28 based on patients’ illness severity and stability at discharge. This approach can be implemented by hospitals by building these risk-prediction models directly into the EHR, or by extracting EHR data in near real time as our group has done successfully for heart failure.7
Another key implication of our study is that, for pneumonia, a disease-specific modeling approach has better predictive ability than using a multi-condition model. Compared to multi-condition models, the first-day and full-stay pneumonia-specific EHR models correctly reclassified an additional 6% and 9% of patients, respectively. Thus, hospitals interested in identifying the highest risk patients with pneumonia for targeted interventions should do so using the disease-specific models, if the costs and resources of doing so are within reach of the healthcare system.
An additional novel finding of our study is the added value of an updated PSI for predicting adverse events. Studies of pneumonia severity of illness scores have calculated the PSI and CURB-65 scores using data present only on admission.16,24 While our study also confirms that the PSI calculated upon admission is a significant predictor of readmission,23,29 this study extends this work by showing that an updated PSI score calculated at the time of discharge is an even stronger predictor for readmission, and its inclusion in the model significantly improves risk stratification and prognostication.
Our study was noteworthy for several strengths. First, we used data from a common EHR system, thus potentially allowing for the implementation of the pneumonia-specific models in real time across a number of hospitals. The use of routinely collected data for risk-prediction modeling makes this approach scalable and sustainable, because it obviates the need for burdensome data collection and entry. Second, to our knowledge, this is the first study to measure the additive influence of illness severity and stability at discharge on the readmission risk among patients hospitalized with pneumonia. Third, our study population was derived from 6 hospitals diverse in payer status, age, race/ethnicity, and socioeconomic status. Fourth, our models are less likely to be overfit to the idiosyncrasies of our data given that several predictors included in our final pneumonia-specific models have been associated with readmission in this population, including marital status,13,30 income,11,31 prior hospitalizations,11,13 thrombocytosis,32-34 and vital sign instabilities on discharge.17 Lastly, the discrimination of the CMS pneumonia model in our cohort (C statistic, 0.64) closely matched the discrimination observed in 4 independent cohorts (C statistic, 0.63), suggesting adequate generalizability of our study setting and population.10,12
Our results should be interpreted in the context of several limitations. First, generalizability to other regions beyond north Texas is unknown. Second, although we included a diverse cohort of safety net, community, teaching, and nonteaching hospitals, the pneumonia-specific models were not externally validated in a separate cohort, which may lead to more optimistic estimates of model performance. Third, PSI and CURB-65 scores were modified to use diagnostic codes for altered mental status and pleural effusion, and omitted nursing home residence. Thus, the independent associations for the PSI and CURB-65 scores and their predictive ability are likely attenuated. Fourth, we were unable to include data on medications (antibiotics and steroid use) and outpatient visits, which may influence readmission risk.2,9,13,35-40 Fifth, we included only the first pneumonia hospitalization per patient in this study. Had we included multiple hospitalizations per patient, we anticipate better model performance for the 2 pneumonia-specific EHR models since prior hospitalization was a robust predictor of readmission.
In conclusion, the full-stay pneumonia-specific EHR readmission risk-prediction model outperformed the first-day pneumonia-specific model, multi-condition EHR models, and the CMS pneumonia model. This suggests that: measures of clinical severity and stability at the time of discharge are important predictors for identifying patients at highest risk for readmission; and that EHR data routinely collected for clinical practice can be used to accurately predict risk of readmission among patients hospitalized for pneumonia.
Acknowledgments
The authors would like to acknowledge Ruben Amarasingham, MD, MBA, president and chief executive officer of Parkland Center for Clinical Innovation, and Ferdinand Velasco, MD, chief health information officer at Texas Health Resources for their assistance in assembling the 6-hospital cohort used in this study.
Disclosures
This work was supported by the Agency for Healthcare Research and Quality-funded UT Southwestern Center for Patient-Centered Outcomes Research (R24 HS022418-01); the Commonwealth Foundation (#20100323); the UT Southwestern KL2 Scholars Program supported by the National Institutes of Health (KL2 TR001103 to ANM and OKN); and the National Center for Advancing Translational Sciences at the National Institute of Health (U54 RFA-TR-12-006 to E.A.H.). The study sponsors had no role in design and conduct of the study; collection, management, analysis, and interpretation of the data; and preparation, review, or approval of the manuscript. The authors have no financial conflicts of interest to disclose
1. Centers for Disease Control and Prevention. Pneumonia. http://www.cdc.gov/nchs/fastats/pneumonia.htm. Accessed January 26, 2016.
33. Prina E, Ferrer M, Ranzani OT, et al. Thrombocytosis is a marker of poor outcome in community-acquired pneumonia. Chest. 2013;143(3):767-775. PubMed
34. Violi F, Cangemi R, Calvieri C. Pneumonia, thrombosis and vascular disease. J Thromb Haemost. 2014;12(9):1391-1400. PubMed
35. Weinberger M, Oddone EZ, Henderson WG. Does increased access to primary care reduce hospital readmissions? Veterans Affairs Cooperative Study Group on Primary Care and Hospital Readmission. N Engl J Med. 1996;334(22):1441-1447. PubMed
36. Field TS, Ogarek J, Garber L, Reed G, Gurwitz JH. Association of early post-discharge follow-up by a primary care physician and 30-day rehospitalization among older adults. J Gen Intern Med. 2015;30(5):565-571. PubMed
37. Spatz ES, Sheth SD, Gosch KL, et al. Usual source of care and outcomes following acute myocardial infarction. J Gen Intern Med. 2014;29(6):862-869. PubMed
38. Brooke BS, Stone DH, Cronenwett JL, et al. Early primary care provider follow-up and readmission after high-risk surgery. JAMA Surg. 2014;149(8):821-828. PubMed
39. Adamuz J, Viasus D, Campreciós-Rodriguez P, et al. A prospective cohort study of healthcare visits and rehospitalizations after discharge of patients with community-acquired pneumonia. Respirology. 2011;16(7):1119-1126. PubMed
40. Shorr AF, Zilberberg MD, Reichley R, et al. Readmission following hospitalization for pneumonia: the impact of pneumonia type and its implication for hospitals. Clin Infect Dis. 2013;57(3):362-367. PubMed
Pneumonia is a leading cause of hospitalizations in the U.S., accounting for more than 1.1 million discharges annually.1 Pneumonia is frequently complicated by hospital readmission, which is costly and potentially avoidable.2,3 Due to financial penalties imposed on hospitals for higher than expected 30-day readmission rates, there is increasing attention to implementing interventions to reduce readmissions in this population.4,5 However, because these programs are resource-intensive, interventions are thought to be most cost-effective if they are targeted to high-risk individuals who are most likely to benefit.6-8
Current pneumonia-specific readmission risk-prediction models that could enable identification of high-risk patients suffer from poor predictive ability, greatly limiting their use, and most were validated among older adults or by using data from single academic medical centers, limiting their generalizability.9-14 A potential reason for poor predictive accuracy is the omission of known robust clinical predictors of pneumonia-related outcomes, including pneumonia severity of illness and stability on discharge.15-17 Approaches using electronic health record (EHR) data, which include this clinically granular data, could enable hospitals to more accurately and pragmatically identify high-risk patients during the index hospitalization and enable interventions to be initiated prior to discharge.
An alternative strategy to identifying high-risk patients for readmission is to use a multi-condition risk-prediction model. Developing and implementing models for every condition may be time-consuming and costly. We have derived and validated 2 multi-condition risk-prediction models using EHR data—1 using data from the first day of hospital admission (‘first-day’ model), and the second incorporating data from the entire hospitalization (‘full-stay’ model) to reflect in-hospital complications and clinical stability at discharge.18,19 However, it is unknown if a multi-condition model for pneumonia would perform as well as a disease-specific model.
This study aimed to develop 2 EHR-based pneumonia-specific readmission risk-prediction models using data routinely collected in clinical practice—a ‘first-day’ and a ‘full-stay’ model—and compare the performance of each model to: 1) one another; 2) the corresponding multi-condition EHR model; and 3) to other potentially useful models in predicting pneumonia readmissions (the Centers for Medicare and Medicaid Services [CMS] pneumonia model, and 2 commonly used pneumonia severity of illness scores validated for predicting mortality). We hypothesized that the pneumonia-specific EHR models would outperform other models; and the full-stay pneumonia-specific model would outperform the first-day pneumonia-specific model.
METHODS
Study Design, Population, and Data Sources
We conducted an observational study using EHR data collected from 6 hospitals (including safety net, community, teaching, and nonteaching hospitals) in north Texas between November 2009 and October 2010, All hospitals used the Epic EHR (Epic Systems Corporation, Verona, WI). Details of this cohort have been published.18,19
We included consecutive hospitalizations among adults 18 years and older discharged from any medicine service with principal discharge diagnoses of pneumonia (ICD-9-CM codes 480-483, 485, 486-487), sepsis (ICD-9-CM codes 038, 995.91, 995.92, 785.52), or respiratory failure (ICD-9-CM codes 518.81, 518.82, 518.84, 799.1) when the latter 2 were also accompanied by a secondary diagnosis of pneumonia.20 For individuals with multiple hospitalizations during the study period, we included only the first hospitalization. We excluded individuals who died during the index hospitalization or within 30 days of discharge, were transferred to another acute care facility, or left against medical advice.
Outcomes
The primary outcome was all-cause 30-day readmission, defined as a nonelective hospitalization within 30 days of discharge to any of 75 acute care hospitals within a 100-mile radius of Dallas, ascertained from an all-payer regional hospitalization database.
Predictor Variables for the Pneumonia-Specific Readmission Models
The selection of candidate predictors was informed by our validated multi-condition risk-prediction models using EHR data available within 24 hours of admission (‘first-day’ multi-condition EHR model) or during the entire hospitalization (‘full-stay’ multi-condition EHR model).18,19 For the pneumonia-specific models, we included all variables in our published multi-condition models as candidate predictors, including sociodemographics, prior utilization, Charlson Comorbidity Index, select laboratory and vital sign abnormalities, length of stay, hospital complications (eg, venous thromboembolism), vital sign instabilities, and disposition status (see Supplemental Table 1 for complete list of variables). We also assessed additional variables specific to pneumonia for inclusion that were: (1) available in the EHR of all participating hospitals; (2) routinely collected or available at the time of admission or discharge; and (3) plausible predictors of adverse outcomes based on literature and clinical expertise. These included select comorbidities (eg, psychiatric conditions, chronic lung disease, history of pneumonia),10,11,21,22 the pneumonia severity index (PSI),16,23,24 intensive care unit stay, and receipt of invasive or noninvasive ventilation. We used a modified PSI score because certain data elements were missing. The modified PSI (henceforth referred to as PSI) did not include nursing home residence and included diagnostic codes as proxies for the presence of pleural effusion (ICD-9-CM codes 510, 511.1, and 511.9) and altered mental status (ICD-9-CM codes 780.0X, 780.97, 293.0, 293.1, and 348.3X).
Statistical Analysis
Model Derivation. Candidate predictor variables were classified as available in the EHR within 24 hours of admission and/or at the time of discharge. For example, socioeconomic factors could be ascertained within the first day of hospitalization, whereas length of stay would not be available until the day of discharge. Predictors with missing values were assumed to be normal (less than 1% missing for each variable). Univariate relationships between readmission and each candidate predictor were assessed in the overall cohort using a pre-specified significance threshold of P ≤ 0.10. Significant variables were entered in the respective first-day and full-stay pneumonia-specific multivariable logistic regression models using stepwise-backward selection with a pre-specified significance threshold of P ≤ 0.05. In sensitivity analyses, we alternately derived our models using stepwise-forward selection, as well as stepwise-backward selection minimizing the Bayesian information criterion and Akaike information criterion separately. These alternate modeling strategies yielded identical predictors to our final models.
Model Validation. Model validation was performed using 5-fold cross-validation, with the overall cohort randomly divided into 5 equal-size subsets.25 For each cycle, 4 subsets were used for training to estimate model coefficients, and the fifth subset was used for validation. This cycle was repeated 5 times with each randomly-divided subset used once as the validation set. We repeated this entire process 50 times and averaged the C statistic estimates to derive an optimism-corrected C statistic. Model calibration was assessed qualitatively by comparing predicted to observed probabilities of readmission by quintiles of predicted risk, and with the Hosmer-Lemeshow goodness-of-fit test.
Comparison to Other Models. The main comparisons of the first-day and full-stay pneumonia-specific EHR model performance were to each other and the corresponding multi-condition EHR model.18,19 The multi-condition EHR models were separately derived and validated within the larger parent cohort from which this study cohort was derived, and outperformed the CMS all-cause model, the HOSPITAL model, and the LACE index.19 To further triangulate our findings, given the lack of other rigorously validated pneumonia-specific risk-prediction models for readmission,14 we compared the pneumonia-specific EHR models to the CMS pneumonia model derived from administrative claims data,10 and 2 commonly used risk-prediction scores for short-term mortality among patients with community-acquired pneumonia, the PSI and CURB-65 scores.16 Although derived and validated using patient-level data, the CMS model was developed to benchmark hospitals according to hospital-level readmission rates.10 The CURB-65 score in this study was also modified to include the same altered mental status diagnostic codes according to the modified PSI as a proxy for “confusion.” Both the PSI and CURB-65 scores were calculated using the most abnormal values within the first 24 hours of admission. The ‘updated’ PSI and the ‘updated’ CURB-65 were calculated using the most abnormal values within 24 hours prior to discharge, or the last known observation prior to discharge if no results were recorded within this time period. A complete list of variables for each of the comparison models are shown in Supplemental Table 1.
We assessed model performance by calculating the C statistic, integrated discrimination index, and net reclassification index (NRI) compared to our pneumonia-specific models. The integrated discrimination index is the difference in the mean predicted probability of readmission between patients who were and were not actually readmitted between 2 models, where more positive values suggest improvement in model performance compared to a reference model.26 The NRI is defined as the sum of the net proportions of correctly reclassified persons with and without the event of interest.27 Here, we calculated a category-based NRI to evaluate the performance of pneumonia-specific models in correctly classifying individuals with and without readmissions into the 2 highest readmission risk quintiles vs the lowest 3 risk quintiles compared to other models.27 This pre-specified cutoff is relevant for hospitals interested in identifying the highest risk individuals for targeted intervention.7 Finally, we assessed calibration of comparator models in our cohort by comparing predicted probability to observed probability of readmission by quintiles of risk for each model. We conducted all analyses using Stata 12.1 (StataCorp, College Station, Texas). This study was approved by the University of Texas Southwestern Medical Center Institutional Review Board.
RESULTS
Of 1463 index hospitalizations (Supplemental Figure 1), the 30-day all-cause readmission rate was 13.6%. Individuals with a 30-day readmission had markedly different sociodemographic and clinical characteristics compared to those not readmitted (Table 1; see Supplemental Table 2 for additional clinical characteristics).
Derivation, Validation, and Performance of the Pneumonia-Specific Readmission Risk-Prediction Models
The final first-day pneumonia-specific EHR model included 7 variables, including sociodemographic characteristics; prior hospitalizations; thrombocytosis, and PSI (Table 2). The first-day pneumonia-specific model had adequate discrimination (C statistic, 0.695; optimism-corrected C statistic 0.675, 95% confidence interval [CI], 0.667-0.685; Table 3). It also effectively stratified individuals across a broad range of risk (average predicted decile of risk ranged from 4% to 33%; Table 3) and was well calibrated (Supplemental Table 3).
The final full-stay pneumonia-specific EHR readmission model included 8 predictors, including 3 variables from the first-day model (median income, thrombocytosis, and prior hospitalizations; Table 2). The full-stay pneumonia-specific EHR model also included vital sign instabilities on discharge, updated PSI, and disposition status (ie, being discharged with home health or to a post-acute care facility was associated with greater odds of readmission, and hospice with lower odds). The full-stay pneumonia-specific EHR model had good discrimination (C statistic, 0.731; optimism-corrected C statistic, 0.714; 95% CI, 0.706-0.720), and stratified individuals across a broad range of risk (average predicted decile of risk ranged from 3% to 37%; Table 3), and was also well calibrated (Supplemental Table 3).
First-Day Pneumonia-Specific EHR Model vs First-Day Multi-Condition EHR Model
The first-day pneumonia-specific EHR model outperformed the first-day multi-condition EHR model with better discrimination (P = 0.029) and more correctly classified individuals in the top 2 highest risk quintiles vs the bottom 3 risk quintiles (Table 3, Supplemental Table 4, and Supplemental Figure 2A). With respect to calibration, the first-day multi-condition EHR model overestimated risk among the highest quintile risk group compared to the first-day pneumonia-specific EHR model (Figure 1A, 1B).
Full-Stay Pneumonia-Specific EHR Model vs Other Models
The full-stay pneumonia-specific EHR model comparatively outperformed the corresponding full-stay multi-condition EHR model, as well as the first-day pneumonia-specific EHR model, the CMS pneumonia model, the updated PSI, and the updated CURB-65 (Table 3, Supplemental Table 5, Supplemental Table 6, and Supplemental Figures 2B and 2C). Compared to the full-stay multi-condition and first-day pneumonia-specific EHR models, the full-stay pneumonia-specific EHR model had better discrimination, better reclassification (NRI, 0.09 and 0.08, respectively), and was able to stratify individuals across a broader range of readmission risk (Table 3). It also had better calibration in the highest quintile risk group compared to the full-stay multi-condition EHR model (Figure 1C and 1D).
Updated vs First-Day Modified PSI and CURB-65 Scores
The updated PSI was more strongly predictive of readmission than the PSI calculated on the day of admission (Wald test, 9.83; P = 0.002). Each 10-point increase in the updated PSI was associated with a 22% increased odds of readmission vs an 11% increase for the PSI calculated upon admission (Table 2). The improved predictive ability of the updated PSI and CURB-65 scores was also reflected in the superior discrimination and calibration vs the respective first-day pneumonia severity of illness scores (Table 3).
DISCUSSION
Using routinely available EHR data from 6 diverse hospitals, we developed 2 pneumonia-specific readmission risk-prediction models that aimed to allow hospitals to identify patients hospitalized with pneumonia at high risk for readmission. Overall, we found that a pneumonia-specific model using EHR data from the entire hospitalization outperformed all other models—including the first-day pneumonia-specific model using data present only on admission, our own multi-condition EHR models, and the CMS pneumonia model based on administrative claims data—in all aspects of model performance (discrimination, calibration, and reclassification). We found that socioeconomic status, prior hospitalizations, thrombocytosis, and measures of clinical severity and stability were important predictors of 30-day all-cause readmissions among patients hospitalized with pneumonia. Additionally, an updated discharge PSI score was a stronger independent predictor of readmissions compared to the PSI score calculated upon admission; and inclusion of the updated PSI in our full-stay pneumonia model led to improved prediction of 30-day readmissions.
The marked improvement in performance of the full-stay pneumonia-specific EHR model compared to the first-day pneumonia-specific model suggests that clinical stability and trajectory during hospitalization (as modeled through disposition status, updated PSI, and vital sign instabilities at discharge) are important predictors of 30-day readmission among patients hospitalized for pneumonia, which was not the case for our EHR-based multi-condition models.19 With the inclusion of these measures, the full-stay pneumonia-specific model correctly reclassified an additional 8% of patients according to their true risk compared to the first-day pneumonia-specific model. One implication of these findings is that hospitals interested in targeting their highest risk individuals with pneumonia for transitional care interventions could do so using the first-day pneumonia-specific EHR model and could refine their targeted strategy at the time of discharge by using the full-stay pneumonia model. This staged risk-prediction strategy would enable hospitals to initiate transitional care interventions for high-risk individuals in the inpatient setting (ie, patient education).7 Then, hospitals could enroll both persistent and newly identified high-risk individuals for outpatient interventions (ie, follow-up telephone call) in the immediate post-discharge period, an interval characterized by heightened vulnerability for adverse events,28 based on patients’ illness severity and stability at discharge. This approach can be implemented by hospitals by building these risk-prediction models directly into the EHR, or by extracting EHR data in near real time as our group has done successfully for heart failure.7
Another key implication of our study is that, for pneumonia, a disease-specific modeling approach has better predictive ability than using a multi-condition model. Compared to multi-condition models, the first-day and full-stay pneumonia-specific EHR models correctly reclassified an additional 6% and 9% of patients, respectively. Thus, hospitals interested in identifying the highest risk patients with pneumonia for targeted interventions should do so using the disease-specific models, if the costs and resources of doing so are within reach of the healthcare system.
An additional novel finding of our study is the added value of an updated PSI for predicting adverse events. Studies of pneumonia severity of illness scores have calculated the PSI and CURB-65 scores using data present only on admission.16,24 While our study also confirms that the PSI calculated upon admission is a significant predictor of readmission,23,29 this study extends this work by showing that an updated PSI score calculated at the time of discharge is an even stronger predictor for readmission, and its inclusion in the model significantly improves risk stratification and prognostication.
Our study was noteworthy for several strengths. First, we used data from a common EHR system, thus potentially allowing for the implementation of the pneumonia-specific models in real time across a number of hospitals. The use of routinely collected data for risk-prediction modeling makes this approach scalable and sustainable, because it obviates the need for burdensome data collection and entry. Second, to our knowledge, this is the first study to measure the additive influence of illness severity and stability at discharge on the readmission risk among patients hospitalized with pneumonia. Third, our study population was derived from 6 hospitals diverse in payer status, age, race/ethnicity, and socioeconomic status. Fourth, our models are less likely to be overfit to the idiosyncrasies of our data given that several predictors included in our final pneumonia-specific models have been associated with readmission in this population, including marital status,13,30 income,11,31 prior hospitalizations,11,13 thrombocytosis,32-34 and vital sign instabilities on discharge.17 Lastly, the discrimination of the CMS pneumonia model in our cohort (C statistic, 0.64) closely matched the discrimination observed in 4 independent cohorts (C statistic, 0.63), suggesting adequate generalizability of our study setting and population.10,12
Our results should be interpreted in the context of several limitations. First, generalizability to other regions beyond north Texas is unknown. Second, although we included a diverse cohort of safety net, community, teaching, and nonteaching hospitals, the pneumonia-specific models were not externally validated in a separate cohort, which may lead to more optimistic estimates of model performance. Third, PSI and CURB-65 scores were modified to use diagnostic codes for altered mental status and pleural effusion, and omitted nursing home residence. Thus, the independent associations for the PSI and CURB-65 scores and their predictive ability are likely attenuated. Fourth, we were unable to include data on medications (antibiotics and steroid use) and outpatient visits, which may influence readmission risk.2,9,13,35-40 Fifth, we included only the first pneumonia hospitalization per patient in this study. Had we included multiple hospitalizations per patient, we anticipate better model performance for the 2 pneumonia-specific EHR models since prior hospitalization was a robust predictor of readmission.
In conclusion, the full-stay pneumonia-specific EHR readmission risk-prediction model outperformed the first-day pneumonia-specific model, multi-condition EHR models, and the CMS pneumonia model. This suggests that: measures of clinical severity and stability at the time of discharge are important predictors for identifying patients at highest risk for readmission; and that EHR data routinely collected for clinical practice can be used to accurately predict risk of readmission among patients hospitalized for pneumonia.
Acknowledgments
The authors would like to acknowledge Ruben Amarasingham, MD, MBA, president and chief executive officer of Parkland Center for Clinical Innovation, and Ferdinand Velasco, MD, chief health information officer at Texas Health Resources for their assistance in assembling the 6-hospital cohort used in this study.
Disclosures
This work was supported by the Agency for Healthcare Research and Quality-funded UT Southwestern Center for Patient-Centered Outcomes Research (R24 HS022418-01); the Commonwealth Foundation (#20100323); the UT Southwestern KL2 Scholars Program supported by the National Institutes of Health (KL2 TR001103 to ANM and OKN); and the National Center for Advancing Translational Sciences at the National Institute of Health (U54 RFA-TR-12-006 to E.A.H.). The study sponsors had no role in design and conduct of the study; collection, management, analysis, and interpretation of the data; and preparation, review, or approval of the manuscript. The authors have no financial conflicts of interest to disclose
Pneumonia is a leading cause of hospitalizations in the U.S., accounting for more than 1.1 million discharges annually.1 Pneumonia is frequently complicated by hospital readmission, which is costly and potentially avoidable.2,3 Due to financial penalties imposed on hospitals for higher than expected 30-day readmission rates, there is increasing attention to implementing interventions to reduce readmissions in this population.4,5 However, because these programs are resource-intensive, interventions are thought to be most cost-effective if they are targeted to high-risk individuals who are most likely to benefit.6-8
Current pneumonia-specific readmission risk-prediction models that could enable identification of high-risk patients suffer from poor predictive ability, greatly limiting their use, and most were validated among older adults or by using data from single academic medical centers, limiting their generalizability.9-14 A potential reason for poor predictive accuracy is the omission of known robust clinical predictors of pneumonia-related outcomes, including pneumonia severity of illness and stability on discharge.15-17 Approaches using electronic health record (EHR) data, which include this clinically granular data, could enable hospitals to more accurately and pragmatically identify high-risk patients during the index hospitalization and enable interventions to be initiated prior to discharge.
An alternative strategy to identifying high-risk patients for readmission is to use a multi-condition risk-prediction model. Developing and implementing models for every condition may be time-consuming and costly. We have derived and validated 2 multi-condition risk-prediction models using EHR data—1 using data from the first day of hospital admission (‘first-day’ model), and the second incorporating data from the entire hospitalization (‘full-stay’ model) to reflect in-hospital complications and clinical stability at discharge.18,19 However, it is unknown if a multi-condition model for pneumonia would perform as well as a disease-specific model.
This study aimed to develop 2 EHR-based pneumonia-specific readmission risk-prediction models using data routinely collected in clinical practice—a ‘first-day’ and a ‘full-stay’ model—and compare the performance of each model to: 1) one another; 2) the corresponding multi-condition EHR model; and 3) to other potentially useful models in predicting pneumonia readmissions (the Centers for Medicare and Medicaid Services [CMS] pneumonia model, and 2 commonly used pneumonia severity of illness scores validated for predicting mortality). We hypothesized that the pneumonia-specific EHR models would outperform other models; and the full-stay pneumonia-specific model would outperform the first-day pneumonia-specific model.
METHODS
Study Design, Population, and Data Sources
We conducted an observational study using EHR data collected from 6 hospitals (including safety net, community, teaching, and nonteaching hospitals) in north Texas between November 2009 and October 2010, All hospitals used the Epic EHR (Epic Systems Corporation, Verona, WI). Details of this cohort have been published.18,19
We included consecutive hospitalizations among adults 18 years and older discharged from any medicine service with principal discharge diagnoses of pneumonia (ICD-9-CM codes 480-483, 485, 486-487), sepsis (ICD-9-CM codes 038, 995.91, 995.92, 785.52), or respiratory failure (ICD-9-CM codes 518.81, 518.82, 518.84, 799.1) when the latter 2 were also accompanied by a secondary diagnosis of pneumonia.20 For individuals with multiple hospitalizations during the study period, we included only the first hospitalization. We excluded individuals who died during the index hospitalization or within 30 days of discharge, were transferred to another acute care facility, or left against medical advice.
Outcomes
The primary outcome was all-cause 30-day readmission, defined as a nonelective hospitalization within 30 days of discharge to any of 75 acute care hospitals within a 100-mile radius of Dallas, ascertained from an all-payer regional hospitalization database.
Predictor Variables for the Pneumonia-Specific Readmission Models
The selection of candidate predictors was informed by our validated multi-condition risk-prediction models using EHR data available within 24 hours of admission (‘first-day’ multi-condition EHR model) or during the entire hospitalization (‘full-stay’ multi-condition EHR model).18,19 For the pneumonia-specific models, we included all variables in our published multi-condition models as candidate predictors, including sociodemographics, prior utilization, Charlson Comorbidity Index, select laboratory and vital sign abnormalities, length of stay, hospital complications (eg, venous thromboembolism), vital sign instabilities, and disposition status (see Supplemental Table 1 for complete list of variables). We also assessed additional variables specific to pneumonia for inclusion that were: (1) available in the EHR of all participating hospitals; (2) routinely collected or available at the time of admission or discharge; and (3) plausible predictors of adverse outcomes based on literature and clinical expertise. These included select comorbidities (eg, psychiatric conditions, chronic lung disease, history of pneumonia),10,11,21,22 the pneumonia severity index (PSI),16,23,24 intensive care unit stay, and receipt of invasive or noninvasive ventilation. We used a modified PSI score because certain data elements were missing. The modified PSI (henceforth referred to as PSI) did not include nursing home residence and included diagnostic codes as proxies for the presence of pleural effusion (ICD-9-CM codes 510, 511.1, and 511.9) and altered mental status (ICD-9-CM codes 780.0X, 780.97, 293.0, 293.1, and 348.3X).
Statistical Analysis
Model Derivation. Candidate predictor variables were classified as available in the EHR within 24 hours of admission and/or at the time of discharge. For example, socioeconomic factors could be ascertained within the first day of hospitalization, whereas length of stay would not be available until the day of discharge. Predictors with missing values were assumed to be normal (less than 1% missing for each variable). Univariate relationships between readmission and each candidate predictor were assessed in the overall cohort using a pre-specified significance threshold of P ≤ 0.10. Significant variables were entered in the respective first-day and full-stay pneumonia-specific multivariable logistic regression models using stepwise-backward selection with a pre-specified significance threshold of P ≤ 0.05. In sensitivity analyses, we alternately derived our models using stepwise-forward selection, as well as stepwise-backward selection minimizing the Bayesian information criterion and Akaike information criterion separately. These alternate modeling strategies yielded identical predictors to our final models.
Model Validation. Model validation was performed using 5-fold cross-validation, with the overall cohort randomly divided into 5 equal-size subsets.25 For each cycle, 4 subsets were used for training to estimate model coefficients, and the fifth subset was used for validation. This cycle was repeated 5 times with each randomly-divided subset used once as the validation set. We repeated this entire process 50 times and averaged the C statistic estimates to derive an optimism-corrected C statistic. Model calibration was assessed qualitatively by comparing predicted to observed probabilities of readmission by quintiles of predicted risk, and with the Hosmer-Lemeshow goodness-of-fit test.
Comparison to Other Models. The main comparisons of the first-day and full-stay pneumonia-specific EHR model performance were to each other and the corresponding multi-condition EHR model.18,19 The multi-condition EHR models were separately derived and validated within the larger parent cohort from which this study cohort was derived, and outperformed the CMS all-cause model, the HOSPITAL model, and the LACE index.19 To further triangulate our findings, given the lack of other rigorously validated pneumonia-specific risk-prediction models for readmission,14 we compared the pneumonia-specific EHR models to the CMS pneumonia model derived from administrative claims data,10 and 2 commonly used risk-prediction scores for short-term mortality among patients with community-acquired pneumonia, the PSI and CURB-65 scores.16 Although derived and validated using patient-level data, the CMS model was developed to benchmark hospitals according to hospital-level readmission rates.10 The CURB-65 score in this study was also modified to include the same altered mental status diagnostic codes according to the modified PSI as a proxy for “confusion.” Both the PSI and CURB-65 scores were calculated using the most abnormal values within the first 24 hours of admission. The ‘updated’ PSI and the ‘updated’ CURB-65 were calculated using the most abnormal values within 24 hours prior to discharge, or the last known observation prior to discharge if no results were recorded within this time period. A complete list of variables for each of the comparison models are shown in Supplemental Table 1.
We assessed model performance by calculating the C statistic, integrated discrimination index, and net reclassification index (NRI) compared to our pneumonia-specific models. The integrated discrimination index is the difference in the mean predicted probability of readmission between patients who were and were not actually readmitted between 2 models, where more positive values suggest improvement in model performance compared to a reference model.26 The NRI is defined as the sum of the net proportions of correctly reclassified persons with and without the event of interest.27 Here, we calculated a category-based NRI to evaluate the performance of pneumonia-specific models in correctly classifying individuals with and without readmissions into the 2 highest readmission risk quintiles vs the lowest 3 risk quintiles compared to other models.27 This pre-specified cutoff is relevant for hospitals interested in identifying the highest risk individuals for targeted intervention.7 Finally, we assessed calibration of comparator models in our cohort by comparing predicted probability to observed probability of readmission by quintiles of risk for each model. We conducted all analyses using Stata 12.1 (StataCorp, College Station, Texas). This study was approved by the University of Texas Southwestern Medical Center Institutional Review Board.
RESULTS
Of 1463 index hospitalizations (Supplemental Figure 1), the 30-day all-cause readmission rate was 13.6%. Individuals with a 30-day readmission had markedly different sociodemographic and clinical characteristics compared to those not readmitted (Table 1; see Supplemental Table 2 for additional clinical characteristics).
Derivation, Validation, and Performance of the Pneumonia-Specific Readmission Risk-Prediction Models
The final first-day pneumonia-specific EHR model included 7 variables, including sociodemographic characteristics; prior hospitalizations; thrombocytosis, and PSI (Table 2). The first-day pneumonia-specific model had adequate discrimination (C statistic, 0.695; optimism-corrected C statistic 0.675, 95% confidence interval [CI], 0.667-0.685; Table 3). It also effectively stratified individuals across a broad range of risk (average predicted decile of risk ranged from 4% to 33%; Table 3) and was well calibrated (Supplemental Table 3).
The final full-stay pneumonia-specific EHR readmission model included 8 predictors, including 3 variables from the first-day model (median income, thrombocytosis, and prior hospitalizations; Table 2). The full-stay pneumonia-specific EHR model also included vital sign instabilities on discharge, updated PSI, and disposition status (ie, being discharged with home health or to a post-acute care facility was associated with greater odds of readmission, and hospice with lower odds). The full-stay pneumonia-specific EHR model had good discrimination (C statistic, 0.731; optimism-corrected C statistic, 0.714; 95% CI, 0.706-0.720), and stratified individuals across a broad range of risk (average predicted decile of risk ranged from 3% to 37%; Table 3), and was also well calibrated (Supplemental Table 3).
First-Day Pneumonia-Specific EHR Model vs First-Day Multi-Condition EHR Model
The first-day pneumonia-specific EHR model outperformed the first-day multi-condition EHR model with better discrimination (P = 0.029) and more correctly classified individuals in the top 2 highest risk quintiles vs the bottom 3 risk quintiles (Table 3, Supplemental Table 4, and Supplemental Figure 2A). With respect to calibration, the first-day multi-condition EHR model overestimated risk among the highest quintile risk group compared to the first-day pneumonia-specific EHR model (Figure 1A, 1B).
Full-Stay Pneumonia-Specific EHR Model vs Other Models
The full-stay pneumonia-specific EHR model comparatively outperformed the corresponding full-stay multi-condition EHR model, as well as the first-day pneumonia-specific EHR model, the CMS pneumonia model, the updated PSI, and the updated CURB-65 (Table 3, Supplemental Table 5, Supplemental Table 6, and Supplemental Figures 2B and 2C). Compared to the full-stay multi-condition and first-day pneumonia-specific EHR models, the full-stay pneumonia-specific EHR model had better discrimination, better reclassification (NRI, 0.09 and 0.08, respectively), and was able to stratify individuals across a broader range of readmission risk (Table 3). It also had better calibration in the highest quintile risk group compared to the full-stay multi-condition EHR model (Figure 1C and 1D).
Updated vs First-Day Modified PSI and CURB-65 Scores
The updated PSI was more strongly predictive of readmission than the PSI calculated on the day of admission (Wald test, 9.83; P = 0.002). Each 10-point increase in the updated PSI was associated with a 22% increased odds of readmission vs an 11% increase for the PSI calculated upon admission (Table 2). The improved predictive ability of the updated PSI and CURB-65 scores was also reflected in the superior discrimination and calibration vs the respective first-day pneumonia severity of illness scores (Table 3).
DISCUSSION
Using routinely available EHR data from 6 diverse hospitals, we developed 2 pneumonia-specific readmission risk-prediction models that aimed to allow hospitals to identify patients hospitalized with pneumonia at high risk for readmission. Overall, we found that a pneumonia-specific model using EHR data from the entire hospitalization outperformed all other models—including the first-day pneumonia-specific model using data present only on admission, our own multi-condition EHR models, and the CMS pneumonia model based on administrative claims data—in all aspects of model performance (discrimination, calibration, and reclassification). We found that socioeconomic status, prior hospitalizations, thrombocytosis, and measures of clinical severity and stability were important predictors of 30-day all-cause readmissions among patients hospitalized with pneumonia. Additionally, an updated discharge PSI score was a stronger independent predictor of readmissions compared to the PSI score calculated upon admission; and inclusion of the updated PSI in our full-stay pneumonia model led to improved prediction of 30-day readmissions.
The marked improvement in performance of the full-stay pneumonia-specific EHR model compared to the first-day pneumonia-specific model suggests that clinical stability and trajectory during hospitalization (as modeled through disposition status, updated PSI, and vital sign instabilities at discharge) are important predictors of 30-day readmission among patients hospitalized for pneumonia, which was not the case for our EHR-based multi-condition models.19 With the inclusion of these measures, the full-stay pneumonia-specific model correctly reclassified an additional 8% of patients according to their true risk compared to the first-day pneumonia-specific model. One implication of these findings is that hospitals interested in targeting their highest risk individuals with pneumonia for transitional care interventions could do so using the first-day pneumonia-specific EHR model and could refine their targeted strategy at the time of discharge by using the full-stay pneumonia model. This staged risk-prediction strategy would enable hospitals to initiate transitional care interventions for high-risk individuals in the inpatient setting (ie, patient education).7 Then, hospitals could enroll both persistent and newly identified high-risk individuals for outpatient interventions (ie, follow-up telephone call) in the immediate post-discharge period, an interval characterized by heightened vulnerability for adverse events,28 based on patients’ illness severity and stability at discharge. This approach can be implemented by hospitals by building these risk-prediction models directly into the EHR, or by extracting EHR data in near real time as our group has done successfully for heart failure.7
Another key implication of our study is that, for pneumonia, a disease-specific modeling approach has better predictive ability than using a multi-condition model. Compared to multi-condition models, the first-day and full-stay pneumonia-specific EHR models correctly reclassified an additional 6% and 9% of patients, respectively. Thus, hospitals interested in identifying the highest risk patients with pneumonia for targeted interventions should do so using the disease-specific models, if the costs and resources of doing so are within reach of the healthcare system.
An additional novel finding of our study is the added value of an updated PSI for predicting adverse events. Studies of pneumonia severity of illness scores have calculated the PSI and CURB-65 scores using data present only on admission.16,24 While our study also confirms that the PSI calculated upon admission is a significant predictor of readmission,23,29 this study extends this work by showing that an updated PSI score calculated at the time of discharge is an even stronger predictor for readmission, and its inclusion in the model significantly improves risk stratification and prognostication.
Our study was noteworthy for several strengths. First, we used data from a common EHR system, thus potentially allowing for the implementation of the pneumonia-specific models in real time across a number of hospitals. The use of routinely collected data for risk-prediction modeling makes this approach scalable and sustainable, because it obviates the need for burdensome data collection and entry. Second, to our knowledge, this is the first study to measure the additive influence of illness severity and stability at discharge on the readmission risk among patients hospitalized with pneumonia. Third, our study population was derived from 6 hospitals diverse in payer status, age, race/ethnicity, and socioeconomic status. Fourth, our models are less likely to be overfit to the idiosyncrasies of our data given that several predictors included in our final pneumonia-specific models have been associated with readmission in this population, including marital status,13,30 income,11,31 prior hospitalizations,11,13 thrombocytosis,32-34 and vital sign instabilities on discharge.17 Lastly, the discrimination of the CMS pneumonia model in our cohort (C statistic, 0.64) closely matched the discrimination observed in 4 independent cohorts (C statistic, 0.63), suggesting adequate generalizability of our study setting and population.10,12
Our results should be interpreted in the context of several limitations. First, generalizability to other regions beyond north Texas is unknown. Second, although we included a diverse cohort of safety net, community, teaching, and nonteaching hospitals, the pneumonia-specific models were not externally validated in a separate cohort, which may lead to more optimistic estimates of model performance. Third, PSI and CURB-65 scores were modified to use diagnostic codes for altered mental status and pleural effusion, and omitted nursing home residence. Thus, the independent associations for the PSI and CURB-65 scores and their predictive ability are likely attenuated. Fourth, we were unable to include data on medications (antibiotics and steroid use) and outpatient visits, which may influence readmission risk.2,9,13,35-40 Fifth, we included only the first pneumonia hospitalization per patient in this study. Had we included multiple hospitalizations per patient, we anticipate better model performance for the 2 pneumonia-specific EHR models since prior hospitalization was a robust predictor of readmission.
In conclusion, the full-stay pneumonia-specific EHR readmission risk-prediction model outperformed the first-day pneumonia-specific model, multi-condition EHR models, and the CMS pneumonia model. This suggests that: measures of clinical severity and stability at the time of discharge are important predictors for identifying patients at highest risk for readmission; and that EHR data routinely collected for clinical practice can be used to accurately predict risk of readmission among patients hospitalized for pneumonia.
Acknowledgments
The authors would like to acknowledge Ruben Amarasingham, MD, MBA, president and chief executive officer of Parkland Center for Clinical Innovation, and Ferdinand Velasco, MD, chief health information officer at Texas Health Resources for their assistance in assembling the 6-hospital cohort used in this study.
Disclosures
This work was supported by the Agency for Healthcare Research and Quality-funded UT Southwestern Center for Patient-Centered Outcomes Research (R24 HS022418-01); the Commonwealth Foundation (#20100323); the UT Southwestern KL2 Scholars Program supported by the National Institutes of Health (KL2 TR001103 to ANM and OKN); and the National Center for Advancing Translational Sciences at the National Institute of Health (U54 RFA-TR-12-006 to E.A.H.). The study sponsors had no role in design and conduct of the study; collection, management, analysis, and interpretation of the data; and preparation, review, or approval of the manuscript. The authors have no financial conflicts of interest to disclose
1. Centers for Disease Control and Prevention. Pneumonia. http://www.cdc.gov/nchs/fastats/pneumonia.htm. Accessed January 26, 2016.
33. Prina E, Ferrer M, Ranzani OT, et al. Thrombocytosis is a marker of poor outcome in community-acquired pneumonia. Chest. 2013;143(3):767-775. PubMed
34. Violi F, Cangemi R, Calvieri C. Pneumonia, thrombosis and vascular disease. J Thromb Haemost. 2014;12(9):1391-1400. PubMed
35. Weinberger M, Oddone EZ, Henderson WG. Does increased access to primary care reduce hospital readmissions? Veterans Affairs Cooperative Study Group on Primary Care and Hospital Readmission. N Engl J Med. 1996;334(22):1441-1447. PubMed
36. Field TS, Ogarek J, Garber L, Reed G, Gurwitz JH. Association of early post-discharge follow-up by a primary care physician and 30-day rehospitalization among older adults. J Gen Intern Med. 2015;30(5):565-571. PubMed
37. Spatz ES, Sheth SD, Gosch KL, et al. Usual source of care and outcomes following acute myocardial infarction. J Gen Intern Med. 2014;29(6):862-869. PubMed
38. Brooke BS, Stone DH, Cronenwett JL, et al. Early primary care provider follow-up and readmission after high-risk surgery. JAMA Surg. 2014;149(8):821-828. PubMed
39. Adamuz J, Viasus D, Campreciós-Rodriguez P, et al. A prospective cohort study of healthcare visits and rehospitalizations after discharge of patients with community-acquired pneumonia. Respirology. 2011;16(7):1119-1126. PubMed
40. Shorr AF, Zilberberg MD, Reichley R, et al. Readmission following hospitalization for pneumonia: the impact of pneumonia type and its implication for hospitals. Clin Infect Dis. 2013;57(3):362-367. PubMed
1. Centers for Disease Control and Prevention. Pneumonia. http://www.cdc.gov/nchs/fastats/pneumonia.htm. Accessed January 26, 2016.
33. Prina E, Ferrer M, Ranzani OT, et al. Thrombocytosis is a marker of poor outcome in community-acquired pneumonia. Chest. 2013;143(3):767-775. PubMed
34. Violi F, Cangemi R, Calvieri C. Pneumonia, thrombosis and vascular disease. J Thromb Haemost. 2014;12(9):1391-1400. PubMed
35. Weinberger M, Oddone EZ, Henderson WG. Does increased access to primary care reduce hospital readmissions? Veterans Affairs Cooperative Study Group on Primary Care and Hospital Readmission. N Engl J Med. 1996;334(22):1441-1447. PubMed
36. Field TS, Ogarek J, Garber L, Reed G, Gurwitz JH. Association of early post-discharge follow-up by a primary care physician and 30-day rehospitalization among older adults. J Gen Intern Med. 2015;30(5):565-571. PubMed
37. Spatz ES, Sheth SD, Gosch KL, et al. Usual source of care and outcomes following acute myocardial infarction. J Gen Intern Med. 2014;29(6):862-869. PubMed
38. Brooke BS, Stone DH, Cronenwett JL, et al. Early primary care provider follow-up and readmission after high-risk surgery. JAMA Surg. 2014;149(8):821-828. PubMed
39. Adamuz J, Viasus D, Campreciós-Rodriguez P, et al. A prospective cohort study of healthcare visits and rehospitalizations after discharge of patients with community-acquired pneumonia. Respirology. 2011;16(7):1119-1126. PubMed
40. Shorr AF, Zilberberg MD, Reichley R, et al. Readmission following hospitalization for pneumonia: the impact of pneumonia type and its implication for hospitals. Clin Infect Dis. 2013;57(3):362-367. PubMed
© 2017 Society of Hospital Medicine
Evaluating automated rules for rapid response system alarm triggers in medical and surgical patients
Patients typically show signs and symptoms of deterioration hours to days prior to cardiorespiratory arrest.1,2 The rate of inhospital cardiorespiratory arrest (CRA) requiring cardiopulmonary resuscitation is estimated to be 0.174 per bed per year in the United States.3 After CRA, survival to discharge is estimated to be as low as 18%.3,4 Efforts to predict and prevent arrest could prove beneficial.1,2
Rapid response systems (RRS) have been proposed as a means of identifying clinical deterioration and facilitating a timely response. These systems were designed to bring clinicians with critical care expertise to the bedside to prevent unnecessary deaths. They typically include an afferent limb (detects deteriorating patients), an efferent limb (responds to calls and acts to avoid further deterioration), and administrative and data analysis limbs.5,6 Automatic provision of recommendations and computer-based systems are desirable components of the afferent limb of the detection system.6 Both are independent predictors of improved clinical practices for clinical decision support systems.7 However, the existing early warning scores (EWS) may not be ready for automation due to low positive predictive values (PPV) and sensitivities.8
It is possible that the low discriminatory accuracy of the published EWS may be secondary to the use of aggregate patient populations for derivation of scores. We hypothesized that these EWS perform differently in medical and in surgical subpopulations. Also, the EWS need to be tested in a time-dependent manner to serve as a realistic clinical support tool for hospitalized patients.
STUDY AIM
The aim of this study was to evaluate the differential performance of widely used EWS in medical vs surgical patients.
METHODS
Site
The study was conducted in an academic center with 2 hospitals in Southeastern Minnesota totaling approximately 1500 general care nonintensive care unit (ICU) beds. The Mayo Clinic Institutional Review Board approved the research proposal.
Subjects
Our retrospective cohort was comprised of all adult inpatients discharged from 2 academic hospitals between January 1, 2011 and December 31, 2011 who spent any time in a general care (non-ICU) unit. We excluded patients younger than 18 years, psychiatric or rehabilitation inpatients, those without research authorization, and patients admitted for research purposes.
Study patients were divided into medical and surgical cohorts. Hospitalizations were considered surgical if patients had surgery at any time during their hospital stay according to billing data. A trigger was an instance in which a patient met the conditions of a specific rule (score/vital sign exceeded the published/defined threshold).
A resuscitation call was defined as a call for cardiopulmonary resuscitation when a patient has a CRA.
An event was an occurrence of 1 of the following in a general care setting: unplanned transfer to the ICU, resuscitation call, or RRS activation.
The RRS activation criteria consisted of an “acute and persistent change” in any 1 or more of the following: oxygen saturations less than 90%, heart rate less than 40 or greater than 130 beats/minute, systolic blood pressure less than 90 mm Hg, or respiratory rate less than 10 or greater than 28 breaths/minute. The RRS activation requires health provider action; they are not electronically generated. Nurses and physicians may also activate the RRS if they are concerned about a patient, even if calling criteria are not met. This is in contrast to the EWS analyzed, which are aggregate composites of multiple parameters. However, whether or not a derangement in vital signs is considered an “acute and persistent change” still involves clinical judgment. Any movement from a general care bed to an ICU bed, or from a general care bed to a procedure area, and from there to an ICU, was considered unplanned. Transfers to the ICU directly from the emergency department or operating room (OR) were not considered as an unplanned transfer and were not included in the analyses.
Coverage time was the period observed for events after a rule was triggered. In this analysis, a coverage time of 24 hours was considered, with a 1-hour look-back. A trigger was counted as a true positive if an event occurred during the following 24 hours. The 1-hour look-back was included to take into account the nursing clinical process of prioritizing a call to the RRS followed by documentation of the altered vital signs that prompted the call.
An episode was the continuous time on the general care floor within a hospitalization, excluding times when a patient was in the OR or ICU. For example, if a patient was admitted to a general bed on a surgery floor, subsequently went to the OR, and then returned to the surgery floor, the 2 episodes were considered separate: the time on the floor before surgery, and the time on the floor after surgery.
Assessment of implementation of RRS in our hospitals showed a significant drop in the failure-to-rescue rate (issues considered related to delay or failure to identify or intervene appropriately when a patient was deteriorating, as identified through mortality review) and a decrease in non-ICU mortality.9,10 This suggests that our current process captures many of the relevant episodes of acute deterioration when a rapid response team is needed and supports using RRS activation as outcomes.
Data Sources
We developed a time-stamped longitudinal database of patient data from the electronic health record, including vital signs, laboratory test results, demographics (age, sex), administrative data (including length of stay), comorbidities, resuscitation code status, location in hospital, and at the minute level throughout each patient’s hospital stay. Physiologically impossible values (eg, blood pressures of 1200 mm Hg) were considered entered in error and eliminated from the database. Time spent in the OR or ICU was excluded because RRS activation would not be applied in these already highly monitored areas. SAS Statistical software (SAS Institute Inc. Cary, North Carolina) was used for database creation.
We applied the current RRS calling criteria in our institution and calculated the Kirkland score,11 along with some of the most widely used early warning scores:12 Modified Early Warning System (MEWS),13 Standardized Early Warning Scoring System (SEWS),14 Global Modified Early Warning Score (GMEWS),15 Worthing physiologic scoring system,16 National Early Warning Score (NEWS),17 and VitaPAC Early Warning Score (ViEWS).18 Published thresholds for these scores were used to create rule triggers in the data. Once a trigger was created to calculate the number of false positives and true positives, all subsequent triggers were ignored until the end of the episode or until 24 hours elapsed. We calculated triggers in a rolling fashion throughout the episodes of care. The EWS score was updated every time a new parameter was entered into the analytical electronic health record, and the most recent value for each was used to calculate the score. SAS statistical software was used for calculation of scores and identification of outcomes.
For our analysis, events were treated as dependent variables, and triggers were independent variables. We calculated the score for each EWS to the minute level throughout our retrospective database. If the score for a specific EWS was higher than the published/recommended threshold for that EWS, an alert was considered to have been issued, and the patient was followed for 24 hours. If the patient had an event in the subsequent 24 hours, or 1 hour before (1-hour look-back), the alert was considered a true positive; if not, a false positive. Events that were not preceded by an alert were false negatives, and 24-hour intervals without either an alert or an event were considered true negatives. This simulation exercise was performed for each EWS in both subcohorts (medical and surgical). Clusters of RRS calls followed by transfers to the ICU within 3 hours were considered as a single adverse event (RRS calls, as it was the first event to occur) to avoid double counting. We have described how well this simulation methodology,8 correlates with results from prospective studies.19
Statistical Analysis
To calculate whether results were statistically significant for subgroups, a jackknife method of calculating variance20 was used. The jackknife method calculates variance by repeating the calculations of the statistic leaving out 1 sample at a time. In our case, we repeated the calculation of sensitivity and PPV leaving out 1 patient at a time. Once the simulation method had been run and the false/true positives/negatives had been assigned, calculation of each metric (PPV and sensitivity) was repeated for n subsamples, each leaving out 1 patient. The variance was calculated and 2 Student t tests were performed for each EWS: 1 for PPV and another for sensitivity. SAS statistical software v 9.3 was used for the simulation analysis; R statistical software v 3.0.2 (The R Foundation, Vienna, Austria) was used for the calculation of the statistical significance of results. A univariable analysis was also performed to assess the sensitivity and PPVs for the published thresholds of the most common variables in each EWS: respiratory rate, systolic blood pressure, heart rate, temperature, and mental status as measured by the modified Richmond Agitation Sedation Score.21
RESULTS
The initial cohort included 60,020 hospitalizations, of which the following were excluded: 2751 because of a lack of appropriate research authorization; 6433 because the patients were younger than 18 years; 2129 as psychiatric admissions; 284 as rehabilitation admissions; 872 as research purposes-only admissions; and 1185 because the patient was never in a general care bed (eg, they were either admitted directly to the ICU, or they were admitted for an outpatient surgical procedure and spent time in the postanesthesia care unit).
Table 1 summarizes patient and trigger characteristics, overall and by subgroup. The final cohort included 75,240 total episodes in 46,366 hospitalizations, from 34,898 unique patients, of which 48.7% were male. There were 23,831 medical and 22,535 surgical hospitalizations. Median length of episode was 2 days both for medical and surgical patients. Median length of stay was 3 days, both for medical and for surgical patients.
There were 3332 events in total, of which 1709 were RRS calls, 185 were resuscitation calls, and 1438 were unscheduled transfers to the ICU. The rate of events was 4.67 events per 100 episodes in the aggregate adult population. There were 3.93 events per 100 episodes for surgical hospitalizations, and 5.86 events per 100 episodes for medical hospitalizations (P < .001). The number of CRAs in our cohort was 0.27 per 100 episodes, 0.128 per hospital bed per year, or 4.37 per 1000 hospital admissions, similar to other reported numbers in the literature.3, 22,23
The total number of EWS triggers varied greatly between EWS rules, with the volume ranging during the study year from 1363 triggers with the GMEWS rule to 77,711 triggers with the ViEWS score.
All scores had PPVs less than 25%. As seen in Table 2 and shown graphically in the Figure, all scores performed better on medical patients (blue) than on surgical patients (yellow). The P value was < .0001 for both PPV and sensitivity. The Worthing score had the highest sensitivity (0.78 for medical and 0.68 for surgical) but a very low PPV (0.04 for medical and 0.03 for surgical), while GMEWS was the opposite: low sensitivity (0.10 and 0.07) but the highest PPV (0.22 and 0.18).
The results of the univariable analysis can be seen in Table 3. Most of the criteria performed better (higher sensitivity and PPV) as predictors in the medical hospitalizations than in the surgical hospitalizations.
DISCUSSION
We hypothesized that EWS may perform differently when applied to medical rather than surgical patients. Studies had not analyzed this in a time-dependent manner,24-26 which limited the applicability of the results.8
All analyzed scores performed better in medical patients than in surgical patients (Figure). This could reflect a behavioral difference by the teams on surgical and medical floors in the decision to activate the RRS, or a bias of the clinicians who designed the scores (mostly nonsurgeons). The difference could also mean that physiological deteriorations are intrinsically different in patients who have undergone anesthesia and surgery. For example, in surgical patients, a bleeding episode is more likely to be the cause of their physiological deterioration, or the lingering effects of anesthesia could mask underlying deterioration. Such patients would benefit from scores where variables such as heart rate, blood pressure, or hemoglobin had more influence.
When comparing the different scores, it was much easier for a patient to meet the alerting score with the Worthing score than with GMEWS. In the Worthing score, a respiratory rate greater than 22 breaths per minute, or a systolic blood pressure less than 100 mm Hg, already meet alerting criteria. Similar vital signs result in 0 and 1 points (respectively) in GMEWS, far from its alerting score of 5. This reflects the intrinsic tradeoff of EWS: as the threshold for considering a patient “at risk” drops, not only does the number of true positives (and the sensitivity) increase, but also the number of false positives, thus lowering the PPV.
However, none of the scores analyzed were considered to perform well based on their PPV and sensitivity, particularly in the surgical subpopulation. Focusing on another metric, the area under the receiver operator curve can give misleadingly optimistic results.24,27 However, the extremely low prevalence of acute physiological deterioration can produce low PPVs even when specificity seems acceptable, which is why it is important to evaluate PPV directly.28
To use EWS effectively to activate RRS, they need to be combined with clinical judgment to avoid high levels of false alerts, particularly in surgical patients. It has been reported that RRS is activated only 30% of the time a patient meets RRS calling criteria.29 While there may be cultural characteristics inhibiting the decision to call,30 our study hints at another explanation: if RRS was activated every time a patient met calling criteria based on the scores analyzed, the number of RRS calls would be very high and difficult to manage. So health providers may be doing the right thing when “filtering” RRS calls and not applying the criteria strictly, but in conjunction with clinical judgment.
A limitation of any study like this is how to define “acute physiological deterioration.” We defined an event as recognized episodes of acute physiological deterioration that are signaled by escalations of care (eg, RRS, resuscitation calls, or transfers to an ICU) or unexpected death. By definition, our calculated PPV is affected by clinicians’ recognition of clinical deteriorations. This definition, common in the literature, has the limitation of potentially underestimating EWS’ performance by missing some events that are resolved by the primary care team without an escalation of care. However, we believe our interpretation is not unreasonable since the purpose of EWS is to trigger escalations of care in a timely fashion. Prospective studies could define an event in a way that is less affected by the clinicians’ judgment.
Regarding patient demographics, age was similar between the 2 groups (average, 58.2 years for medical vs 58.9 years for surgical), and there was only a small difference in gender ratios (45.1% male in the medical vs 51.4% in the surgical group). These differences are unlikely to have affected the results significantly, but unknown differences in demographics or other patient characteristics between groups may account for differences in score performance between surgical and medical patients.
Several of the EWS analyzed had overlapping trigger criteria with our own RRS activation criteria (although as single-parameter triggers and not as aggregate). To test how these potential biases could affect our results, we performed a post hoc sensitivity analysis eliminating calls to the RRS as an outcome (so using the alternative outcome of unexpected transfers to the ICU and resuscitation calls). The results are similar to those of our main analysis, with all analyzed scores having lower sensitivity and PPV in surgical hospitalizations when compared to medical hospitalizations.
Our study suggests that, to optimize detection of physiological deterioration events, EWS should try to take into account different patient types, with the most basic distinction being surgical vs medical. This tailoring will make EWS more complex, and less suited for paper-based calculation, but new electronic health records are increasingly able to incorporate decision support, and some EWS have been developed for electronic calculation only. Of particular interest in this regard is the score developed by Escobar et al,31 which groups patients into categories according to the reason for admission, and calculates a different subscore based on that category. While the score by Escobar et al. does not split patients based on medical or surgical status, a more general interpretation of our results suggests that a score may be more accurate if it classifies patients into subgroups with different subscores. This seems to be confirmed by the fact that the score by Escobar et al performs better than MEWS.28 Unfortunately, the paper describing it does not provide enough detail to use it in our database.
A recent systematic review showed increasing evidence that RRS may be effective in reducing CRAs occurring in a non-ICU setting and, more important, overall inhospital mortality.32 While differing implementation strategies (eg, different length of the educational effort, changes in the frequency of vital signs monitoring) can impact the success of such an initiative, it has been speculated that the afferent limb (which often includes an EWS) might be the most critical part of the system.33 Our results show that the most widely used EWS perform significantly worse on surgical patients, and suggest that a way to improve the accuracy of EWS would be to tailor the risk calculation to different patient subgroups (eg, medical and surgical patients). Plausible next steps would be to demonstrate that tailoring risk calculation to medical and surgical patients separately can improve risk predictions and accuracy of EWS.
Disclosure
The authors report no financial conflicts of interest.
1. Buist MD, Jarmolowski E, Burton PR, Bernard SA, Waxman BP, Anderson J. Recognising clinical instability in hospital patients before cardiac arrest or unplanned admission to intensive care. A pilot study in a tertiary-care hospital. Med J Aust. 1999; 171(1):22-25. PubMed
2. Schein RM, Hazday N, Pena M, Ruben BH, Sprung CL. Clinical antecedents to in-hospital cardiopulmonary arrest. Chest. 1990;98(6):1388-1392. PubMed
3. Peberdy MA, Kaye W, Ornato JP, Larkin GL, Nadkarni V, Mancini ME, et al. Cardiopulmonary resuscitation of adults in the hospital: a report of 14720 cardiac arrests from the National Registry of Cardiopulmonary Resuscitation. Resuscitation. 2003; 58(3):297-308. PubMed
4. Nadkarni VM, Larkin GL, Peberdy MA, Carey SM, Kaye W, Mancini ME, et al. First documented rhythm and clinical outcome from in-hospital cardiac arrest among children and adults. JAMA. 2006;295(1):50-57. PubMed
5. Devita MA, Bellomo R, Hillman K, Kellum J, Rotondi A, Teres D, et al. Findings of the first consensus conference on medical emergency teams. Crit Care Med. 2006;34(9):2463-2478. PubMed
6. DeVita MA, Smith GB, Adam SK, Adams-Pizarro I, Buist M, Bellomo R, et al. “Identifying the hospitalised patient in crisis”--a consensus conference on the afferent limb of rapid response systems. Resuscitation. 2010;81(4):375-382. PubMed
7. Kawamoto K, Houlihan CA, Balas EA, Lobach DF. Improving clinical practice using clinical decision support systems: a systematic review of trials to identify features critical to success. BMJ. 2005;330(7494):765. PubMed
8. Romero-Brufau S, Huddleston JM, Naessens JM, Johnson MG, Hickman J, Morlan BW, et al. Widely used track and trigger scores: are they ready for automation in practice? Resuscitation. 2014;85(4):549-552. PubMed
9. Huddleston JM, Diedrich DA, Kinsey GC, Enzler MJ, Manning DM. Learning from every death. J Patient Saf. 2014;10(1):6-12. PubMed
10. Moriarty JP, Schiebel NE, Johnson MG, Jensen JB, Caples SM, Morlan BW, et al. Evaluating implementation of a rapid response team: considering alternative outcome measures. Int J Qual Health Care. 2014;26(1):49-57. PubMed
11. Kirkland LL, Malinchoc M, O’Byrne M, Benson JT, Kashiwagi DT, Burton MC, et al. A clinical deterioration prediction tool for internal medicine patients. Am J Med Qual. 2013;28(2):135-142. PubMed
12. Griffiths JR, Kidney EM. Current use of early warning scores in UK emergency departments. Emerg Med J. 2012;29(1):65-66. PubMed
13. Subbe CP, Kruger M, Rutherford P, Gemmel L. Validation of a modified Early Warning Score in medical admissions. QJM. 2001;94(10):521-526. PubMed
14. Paterson R, MacLeod DC, Thetford D, Beattie A, Graham C, Lam S, et al.. Prediction of in-hospital mortality and length of stay using an early warning scoring system: clinical audit. Clin Med (Lond). 2006;6(3):281-284. PubMed
15. Harrison GA, Jacques T, McLaws ML, Kilborn G. Combinations of early signs of critical illness predict in-hospital death–the SOCCER study (signs of critical conditions and emergency responses). Resuscitation. 2006;71(3):327-334. PubMed
16. Duckitt RW, Buxton-Thomas R, Walker J, Cheek E, Bewick V, Venn R, et al. Worthing physiological scoring system: derivation and validation of a physiological early-warning system for medical admissions. An observational, population-based single-centre study. Br J Anaesth. 2007; 98(6):769-774. PubMed
17. Smith GB, Prytherch DR, Meredith P, Schmidt PE, Featherstone PI. The ability of the National Early Warning Score (NEWS) to discriminate patients at risk of early cardiac arrest, unanticipated intensive care unit admission, and death. Resuscitation. 2013;84(4):465-470. PubMed
18. Prytherch DR, Smith GB, Schmidt PE, Featherstone PI. ViEWS--Towards a national early warning score for detecting adult inpatient deterioration. Resuscitation. 2010;81(8):932-937. PubMed
19. Romero-Brufau S, Huddleston JM. Reply to letter: widely used track and trigger scores: are they ready for automation in practice? Resuscitation. 2014;85(10):e159. PubMed
20. Efron B, Stein C. The jackknife estimate of variance. Annals of Statistics. 1981;586-596.
21. Sessler CN, Gosnell MS, Grap MJ, Brophy GM, O’Neal PV, Keane KA, et al. The Richmond Agitation-Sedation Scale: validity and reliability in adult intensive care unit patients. Am J Respir Crit Care Med. 2002;166(10):1338-1344. PubMed
22. DeVita MA, Braithwaite RS, Mahidhara R, Stuart S, Foraida M, Simmons RL. Medical Emergency Response Improvement Team (MERIT). Use of medical emergency team responses to reduce hospital cardiopulmonary arrests. Qual Saf Health Care. 2004;13(4):251-254. PubMed
23. Goncales PD, Polessi JA, Bass LM, Santos Gde P, Yokota PK, Laselva CR, et al. Reduced frequency of cardiopulmonary arrests by rapid response teams. Einstein (Sao Paulo). 2012;10(4):442-448. PubMed
24. Cuthbertson BH, Boroujerdi M, McKie L, Aucott L, Prescott G. Can physiological variables and early warning scoring systems allow early recognition of the deteriorating surgical patient? Crit Care Med. 2007;35(2):402-409. PubMed
25. Gardner-Thorpe J, Love N, Wrightson J, Walsh S, Keeling N. The value of Modified Early Warning Score (MEWS) in surgical in-patients: a prospective observational study. Ann R Coll Surg Engl. 2006;88(6):571-575. PubMed
26. Stenhouse C, Coates S, Tivey M, Allsop P, Parker T. Prospective evaluation of a modified Early Warning Score to aid earlier detection of patients developing critical illness on a general surgical ward. British Journal of Anaesthesia. 2000;84(5):663-663.
27. Smith GB, Prytherch DR, Schmidt PE, Featherstone PI. Review and performance evaluation of aggregate weighted ‘track and trigger’ systems. Resuscitation. 2008;77(2):170-179. PubMed
28. Romero-Brufau S, Huddleston JM, Escobar GJ, Liebow M. Why the C-statistic is not informative to evaluate early warning scores and what metrics to use. Crit Care. 2015; 19:285. PubMed
29. Hillman K, Chen J, Cretikos M, Bellomo R, Brown D, Doig G, et al. Introduction of the medical emergency team (MET) system: a cluster-randomised controlled trial. Lancet. 2005;365(9477):2091-2097. PubMed
30. Shearer B, Marshall S, Buist MD, Finnigan M, Kitto S, Hore T, et al. What stops hospital clinical staff from following protocols? An analysis of the incidence and factors behind the failure of bedside clinical staff to activate the rapid response system in a multi-campus Australian metropolitan healthcare service. BMJ Qual Saf. 2012;21(7):569-575. PubMed
31. Escobar GJ, LaGuardia JC, Turk BJ, Ragins A, Kipnis P, Draper D. Early detection of impending physiologic deterioration among patients who are not in intensive care: development of predictive models using data from an automated electronic medical record. J Hosp Med. 2012;7(5):388-395. PubMed
32. Winters BD, Weaver SJ, Pfoh ER, Yang T, Pham JC, Dy SM. Rapid-response systems as a patient safety strategy: a systematic review. Ann Intern Med. 2013;158(5 pt 2):417-425. PubMed
33. Jones DA, DeVita MA, Bellomo R. Rapid-response teams. N Engl J Med. 2011;365(2):139-146. PubMed
Patients typically show signs and symptoms of deterioration hours to days prior to cardiorespiratory arrest.1,2 The rate of inhospital cardiorespiratory arrest (CRA) requiring cardiopulmonary resuscitation is estimated to be 0.174 per bed per year in the United States.3 After CRA, survival to discharge is estimated to be as low as 18%.3,4 Efforts to predict and prevent arrest could prove beneficial.1,2
Rapid response systems (RRS) have been proposed as a means of identifying clinical deterioration and facilitating a timely response. These systems were designed to bring clinicians with critical care expertise to the bedside to prevent unnecessary deaths. They typically include an afferent limb (detects deteriorating patients), an efferent limb (responds to calls and acts to avoid further deterioration), and administrative and data analysis limbs.5,6 Automatic provision of recommendations and computer-based systems are desirable components of the afferent limb of the detection system.6 Both are independent predictors of improved clinical practices for clinical decision support systems.7 However, the existing early warning scores (EWS) may not be ready for automation due to low positive predictive values (PPV) and sensitivities.8
It is possible that the low discriminatory accuracy of the published EWS may be secondary to the use of aggregate patient populations for derivation of scores. We hypothesized that these EWS perform differently in medical and in surgical subpopulations. Also, the EWS need to be tested in a time-dependent manner to serve as a realistic clinical support tool for hospitalized patients.
STUDY AIM
The aim of this study was to evaluate the differential performance of widely used EWS in medical vs surgical patients.
METHODS
Site
The study was conducted in an academic center with 2 hospitals in Southeastern Minnesota totaling approximately 1500 general care nonintensive care unit (ICU) beds. The Mayo Clinic Institutional Review Board approved the research proposal.
Subjects
Our retrospective cohort was comprised of all adult inpatients discharged from 2 academic hospitals between January 1, 2011 and December 31, 2011 who spent any time in a general care (non-ICU) unit. We excluded patients younger than 18 years, psychiatric or rehabilitation inpatients, those without research authorization, and patients admitted for research purposes.
Study patients were divided into medical and surgical cohorts. Hospitalizations were considered surgical if patients had surgery at any time during their hospital stay according to billing data. A trigger was an instance in which a patient met the conditions of a specific rule (score/vital sign exceeded the published/defined threshold).
A resuscitation call was defined as a call for cardiopulmonary resuscitation when a patient has a CRA.
An event was an occurrence of 1 of the following in a general care setting: unplanned transfer to the ICU, resuscitation call, or RRS activation.
The RRS activation criteria consisted of an “acute and persistent change” in any 1 or more of the following: oxygen saturations less than 90%, heart rate less than 40 or greater than 130 beats/minute, systolic blood pressure less than 90 mm Hg, or respiratory rate less than 10 or greater than 28 breaths/minute. The RRS activation requires health provider action; they are not electronically generated. Nurses and physicians may also activate the RRS if they are concerned about a patient, even if calling criteria are not met. This is in contrast to the EWS analyzed, which are aggregate composites of multiple parameters. However, whether or not a derangement in vital signs is considered an “acute and persistent change” still involves clinical judgment. Any movement from a general care bed to an ICU bed, or from a general care bed to a procedure area, and from there to an ICU, was considered unplanned. Transfers to the ICU directly from the emergency department or operating room (OR) were not considered as an unplanned transfer and were not included in the analyses.
Coverage time was the period observed for events after a rule was triggered. In this analysis, a coverage time of 24 hours was considered, with a 1-hour look-back. A trigger was counted as a true positive if an event occurred during the following 24 hours. The 1-hour look-back was included to take into account the nursing clinical process of prioritizing a call to the RRS followed by documentation of the altered vital signs that prompted the call.
An episode was the continuous time on the general care floor within a hospitalization, excluding times when a patient was in the OR or ICU. For example, if a patient was admitted to a general bed on a surgery floor, subsequently went to the OR, and then returned to the surgery floor, the 2 episodes were considered separate: the time on the floor before surgery, and the time on the floor after surgery.
Assessment of implementation of RRS in our hospitals showed a significant drop in the failure-to-rescue rate (issues considered related to delay or failure to identify or intervene appropriately when a patient was deteriorating, as identified through mortality review) and a decrease in non-ICU mortality.9,10 This suggests that our current process captures many of the relevant episodes of acute deterioration when a rapid response team is needed and supports using RRS activation as outcomes.
Data Sources
We developed a time-stamped longitudinal database of patient data from the electronic health record, including vital signs, laboratory test results, demographics (age, sex), administrative data (including length of stay), comorbidities, resuscitation code status, location in hospital, and at the minute level throughout each patient’s hospital stay. Physiologically impossible values (eg, blood pressures of 1200 mm Hg) were considered entered in error and eliminated from the database. Time spent in the OR or ICU was excluded because RRS activation would not be applied in these already highly monitored areas. SAS Statistical software (SAS Institute Inc. Cary, North Carolina) was used for database creation.
We applied the current RRS calling criteria in our institution and calculated the Kirkland score,11 along with some of the most widely used early warning scores:12 Modified Early Warning System (MEWS),13 Standardized Early Warning Scoring System (SEWS),14 Global Modified Early Warning Score (GMEWS),15 Worthing physiologic scoring system,16 National Early Warning Score (NEWS),17 and VitaPAC Early Warning Score (ViEWS).18 Published thresholds for these scores were used to create rule triggers in the data. Once a trigger was created to calculate the number of false positives and true positives, all subsequent triggers were ignored until the end of the episode or until 24 hours elapsed. We calculated triggers in a rolling fashion throughout the episodes of care. The EWS score was updated every time a new parameter was entered into the analytical electronic health record, and the most recent value for each was used to calculate the score. SAS statistical software was used for calculation of scores and identification of outcomes.
For our analysis, events were treated as dependent variables, and triggers were independent variables. We calculated the score for each EWS to the minute level throughout our retrospective database. If the score for a specific EWS was higher than the published/recommended threshold for that EWS, an alert was considered to have been issued, and the patient was followed for 24 hours. If the patient had an event in the subsequent 24 hours, or 1 hour before (1-hour look-back), the alert was considered a true positive; if not, a false positive. Events that were not preceded by an alert were false negatives, and 24-hour intervals without either an alert or an event were considered true negatives. This simulation exercise was performed for each EWS in both subcohorts (medical and surgical). Clusters of RRS calls followed by transfers to the ICU within 3 hours were considered as a single adverse event (RRS calls, as it was the first event to occur) to avoid double counting. We have described how well this simulation methodology,8 correlates with results from prospective studies.19
Statistical Analysis
To calculate whether results were statistically significant for subgroups, a jackknife method of calculating variance20 was used. The jackknife method calculates variance by repeating the calculations of the statistic leaving out 1 sample at a time. In our case, we repeated the calculation of sensitivity and PPV leaving out 1 patient at a time. Once the simulation method had been run and the false/true positives/negatives had been assigned, calculation of each metric (PPV and sensitivity) was repeated for n subsamples, each leaving out 1 patient. The variance was calculated and 2 Student t tests were performed for each EWS: 1 for PPV and another for sensitivity. SAS statistical software v 9.3 was used for the simulation analysis; R statistical software v 3.0.2 (The R Foundation, Vienna, Austria) was used for the calculation of the statistical significance of results. A univariable analysis was also performed to assess the sensitivity and PPVs for the published thresholds of the most common variables in each EWS: respiratory rate, systolic blood pressure, heart rate, temperature, and mental status as measured by the modified Richmond Agitation Sedation Score.21
RESULTS
The initial cohort included 60,020 hospitalizations, of which the following were excluded: 2751 because of a lack of appropriate research authorization; 6433 because the patients were younger than 18 years; 2129 as psychiatric admissions; 284 as rehabilitation admissions; 872 as research purposes-only admissions; and 1185 because the patient was never in a general care bed (eg, they were either admitted directly to the ICU, or they were admitted for an outpatient surgical procedure and spent time in the postanesthesia care unit).
Table 1 summarizes patient and trigger characteristics, overall and by subgroup. The final cohort included 75,240 total episodes in 46,366 hospitalizations, from 34,898 unique patients, of which 48.7% were male. There were 23,831 medical and 22,535 surgical hospitalizations. Median length of episode was 2 days both for medical and surgical patients. Median length of stay was 3 days, both for medical and for surgical patients.
There were 3332 events in total, of which 1709 were RRS calls, 185 were resuscitation calls, and 1438 were unscheduled transfers to the ICU. The rate of events was 4.67 events per 100 episodes in the aggregate adult population. There were 3.93 events per 100 episodes for surgical hospitalizations, and 5.86 events per 100 episodes for medical hospitalizations (P < .001). The number of CRAs in our cohort was 0.27 per 100 episodes, 0.128 per hospital bed per year, or 4.37 per 1000 hospital admissions, similar to other reported numbers in the literature.3, 22,23
The total number of EWS triggers varied greatly between EWS rules, with the volume ranging during the study year from 1363 triggers with the GMEWS rule to 77,711 triggers with the ViEWS score.
All scores had PPVs less than 25%. As seen in Table 2 and shown graphically in the Figure, all scores performed better on medical patients (blue) than on surgical patients (yellow). The P value was < .0001 for both PPV and sensitivity. The Worthing score had the highest sensitivity (0.78 for medical and 0.68 for surgical) but a very low PPV (0.04 for medical and 0.03 for surgical), while GMEWS was the opposite: low sensitivity (0.10 and 0.07) but the highest PPV (0.22 and 0.18).
The results of the univariable analysis can be seen in Table 3. Most of the criteria performed better (higher sensitivity and PPV) as predictors in the medical hospitalizations than in the surgical hospitalizations.
DISCUSSION
We hypothesized that EWS may perform differently when applied to medical rather than surgical patients. Studies had not analyzed this in a time-dependent manner,24-26 which limited the applicability of the results.8
All analyzed scores performed better in medical patients than in surgical patients (Figure). This could reflect a behavioral difference by the teams on surgical and medical floors in the decision to activate the RRS, or a bias of the clinicians who designed the scores (mostly nonsurgeons). The difference could also mean that physiological deteriorations are intrinsically different in patients who have undergone anesthesia and surgery. For example, in surgical patients, a bleeding episode is more likely to be the cause of their physiological deterioration, or the lingering effects of anesthesia could mask underlying deterioration. Such patients would benefit from scores where variables such as heart rate, blood pressure, or hemoglobin had more influence.
When comparing the different scores, it was much easier for a patient to meet the alerting score with the Worthing score than with GMEWS. In the Worthing score, a respiratory rate greater than 22 breaths per minute, or a systolic blood pressure less than 100 mm Hg, already meet alerting criteria. Similar vital signs result in 0 and 1 points (respectively) in GMEWS, far from its alerting score of 5. This reflects the intrinsic tradeoff of EWS: as the threshold for considering a patient “at risk” drops, not only does the number of true positives (and the sensitivity) increase, but also the number of false positives, thus lowering the PPV.
However, none of the scores analyzed were considered to perform well based on their PPV and sensitivity, particularly in the surgical subpopulation. Focusing on another metric, the area under the receiver operator curve can give misleadingly optimistic results.24,27 However, the extremely low prevalence of acute physiological deterioration can produce low PPVs even when specificity seems acceptable, which is why it is important to evaluate PPV directly.28
To use EWS effectively to activate RRS, they need to be combined with clinical judgment to avoid high levels of false alerts, particularly in surgical patients. It has been reported that RRS is activated only 30% of the time a patient meets RRS calling criteria.29 While there may be cultural characteristics inhibiting the decision to call,30 our study hints at another explanation: if RRS was activated every time a patient met calling criteria based on the scores analyzed, the number of RRS calls would be very high and difficult to manage. So health providers may be doing the right thing when “filtering” RRS calls and not applying the criteria strictly, but in conjunction with clinical judgment.
A limitation of any study like this is how to define “acute physiological deterioration.” We defined an event as recognized episodes of acute physiological deterioration that are signaled by escalations of care (eg, RRS, resuscitation calls, or transfers to an ICU) or unexpected death. By definition, our calculated PPV is affected by clinicians’ recognition of clinical deteriorations. This definition, common in the literature, has the limitation of potentially underestimating EWS’ performance by missing some events that are resolved by the primary care team without an escalation of care. However, we believe our interpretation is not unreasonable since the purpose of EWS is to trigger escalations of care in a timely fashion. Prospective studies could define an event in a way that is less affected by the clinicians’ judgment.
Regarding patient demographics, age was similar between the 2 groups (average, 58.2 years for medical vs 58.9 years for surgical), and there was only a small difference in gender ratios (45.1% male in the medical vs 51.4% in the surgical group). These differences are unlikely to have affected the results significantly, but unknown differences in demographics or other patient characteristics between groups may account for differences in score performance between surgical and medical patients.
Several of the EWS analyzed had overlapping trigger criteria with our own RRS activation criteria (although as single-parameter triggers and not as aggregate). To test how these potential biases could affect our results, we performed a post hoc sensitivity analysis eliminating calls to the RRS as an outcome (so using the alternative outcome of unexpected transfers to the ICU and resuscitation calls). The results are similar to those of our main analysis, with all analyzed scores having lower sensitivity and PPV in surgical hospitalizations when compared to medical hospitalizations.
Our study suggests that, to optimize detection of physiological deterioration events, EWS should try to take into account different patient types, with the most basic distinction being surgical vs medical. This tailoring will make EWS more complex, and less suited for paper-based calculation, but new electronic health records are increasingly able to incorporate decision support, and some EWS have been developed for electronic calculation only. Of particular interest in this regard is the score developed by Escobar et al,31 which groups patients into categories according to the reason for admission, and calculates a different subscore based on that category. While the score by Escobar et al. does not split patients based on medical or surgical status, a more general interpretation of our results suggests that a score may be more accurate if it classifies patients into subgroups with different subscores. This seems to be confirmed by the fact that the score by Escobar et al performs better than MEWS.28 Unfortunately, the paper describing it does not provide enough detail to use it in our database.
A recent systematic review showed increasing evidence that RRS may be effective in reducing CRAs occurring in a non-ICU setting and, more important, overall inhospital mortality.32 While differing implementation strategies (eg, different length of the educational effort, changes in the frequency of vital signs monitoring) can impact the success of such an initiative, it has been speculated that the afferent limb (which often includes an EWS) might be the most critical part of the system.33 Our results show that the most widely used EWS perform significantly worse on surgical patients, and suggest that a way to improve the accuracy of EWS would be to tailor the risk calculation to different patient subgroups (eg, medical and surgical patients). Plausible next steps would be to demonstrate that tailoring risk calculation to medical and surgical patients separately can improve risk predictions and accuracy of EWS.
Disclosure
The authors report no financial conflicts of interest.
Patients typically show signs and symptoms of deterioration hours to days prior to cardiorespiratory arrest.1,2 The rate of inhospital cardiorespiratory arrest (CRA) requiring cardiopulmonary resuscitation is estimated to be 0.174 per bed per year in the United States.3 After CRA, survival to discharge is estimated to be as low as 18%.3,4 Efforts to predict and prevent arrest could prove beneficial.1,2
Rapid response systems (RRS) have been proposed as a means of identifying clinical deterioration and facilitating a timely response. These systems were designed to bring clinicians with critical care expertise to the bedside to prevent unnecessary deaths. They typically include an afferent limb (detects deteriorating patients), an efferent limb (responds to calls and acts to avoid further deterioration), and administrative and data analysis limbs.5,6 Automatic provision of recommendations and computer-based systems are desirable components of the afferent limb of the detection system.6 Both are independent predictors of improved clinical practices for clinical decision support systems.7 However, the existing early warning scores (EWS) may not be ready for automation due to low positive predictive values (PPV) and sensitivities.8
It is possible that the low discriminatory accuracy of the published EWS may be secondary to the use of aggregate patient populations for derivation of scores. We hypothesized that these EWS perform differently in medical and in surgical subpopulations. Also, the EWS need to be tested in a time-dependent manner to serve as a realistic clinical support tool for hospitalized patients.
STUDY AIM
The aim of this study was to evaluate the differential performance of widely used EWS in medical vs surgical patients.
METHODS
Site
The study was conducted in an academic center with 2 hospitals in Southeastern Minnesota totaling approximately 1500 general care nonintensive care unit (ICU) beds. The Mayo Clinic Institutional Review Board approved the research proposal.
Subjects
Our retrospective cohort was comprised of all adult inpatients discharged from 2 academic hospitals between January 1, 2011 and December 31, 2011 who spent any time in a general care (non-ICU) unit. We excluded patients younger than 18 years, psychiatric or rehabilitation inpatients, those without research authorization, and patients admitted for research purposes.
Study patients were divided into medical and surgical cohorts. Hospitalizations were considered surgical if patients had surgery at any time during their hospital stay according to billing data. A trigger was an instance in which a patient met the conditions of a specific rule (score/vital sign exceeded the published/defined threshold).
A resuscitation call was defined as a call for cardiopulmonary resuscitation when a patient has a CRA.
An event was an occurrence of 1 of the following in a general care setting: unplanned transfer to the ICU, resuscitation call, or RRS activation.
The RRS activation criteria consisted of an “acute and persistent change” in any 1 or more of the following: oxygen saturations less than 90%, heart rate less than 40 or greater than 130 beats/minute, systolic blood pressure less than 90 mm Hg, or respiratory rate less than 10 or greater than 28 breaths/minute. The RRS activation requires health provider action; they are not electronically generated. Nurses and physicians may also activate the RRS if they are concerned about a patient, even if calling criteria are not met. This is in contrast to the EWS analyzed, which are aggregate composites of multiple parameters. However, whether or not a derangement in vital signs is considered an “acute and persistent change” still involves clinical judgment. Any movement from a general care bed to an ICU bed, or from a general care bed to a procedure area, and from there to an ICU, was considered unplanned. Transfers to the ICU directly from the emergency department or operating room (OR) were not considered as an unplanned transfer and were not included in the analyses.
Coverage time was the period observed for events after a rule was triggered. In this analysis, a coverage time of 24 hours was considered, with a 1-hour look-back. A trigger was counted as a true positive if an event occurred during the following 24 hours. The 1-hour look-back was included to take into account the nursing clinical process of prioritizing a call to the RRS followed by documentation of the altered vital signs that prompted the call.
An episode was the continuous time on the general care floor within a hospitalization, excluding times when a patient was in the OR or ICU. For example, if a patient was admitted to a general bed on a surgery floor, subsequently went to the OR, and then returned to the surgery floor, the 2 episodes were considered separate: the time on the floor before surgery, and the time on the floor after surgery.
Assessment of implementation of RRS in our hospitals showed a significant drop in the failure-to-rescue rate (issues considered related to delay or failure to identify or intervene appropriately when a patient was deteriorating, as identified through mortality review) and a decrease in non-ICU mortality.9,10 This suggests that our current process captures many of the relevant episodes of acute deterioration when a rapid response team is needed and supports using RRS activation as outcomes.
Data Sources
We developed a time-stamped longitudinal database of patient data from the electronic health record, including vital signs, laboratory test results, demographics (age, sex), administrative data (including length of stay), comorbidities, resuscitation code status, location in hospital, and at the minute level throughout each patient’s hospital stay. Physiologically impossible values (eg, blood pressures of 1200 mm Hg) were considered entered in error and eliminated from the database. Time spent in the OR or ICU was excluded because RRS activation would not be applied in these already highly monitored areas. SAS Statistical software (SAS Institute Inc. Cary, North Carolina) was used for database creation.
We applied the current RRS calling criteria in our institution and calculated the Kirkland score,11 along with some of the most widely used early warning scores:12 Modified Early Warning System (MEWS),13 Standardized Early Warning Scoring System (SEWS),14 Global Modified Early Warning Score (GMEWS),15 Worthing physiologic scoring system,16 National Early Warning Score (NEWS),17 and VitaPAC Early Warning Score (ViEWS).18 Published thresholds for these scores were used to create rule triggers in the data. Once a trigger was created to calculate the number of false positives and true positives, all subsequent triggers were ignored until the end of the episode or until 24 hours elapsed. We calculated triggers in a rolling fashion throughout the episodes of care. The EWS score was updated every time a new parameter was entered into the analytical electronic health record, and the most recent value for each was used to calculate the score. SAS statistical software was used for calculation of scores and identification of outcomes.
For our analysis, events were treated as dependent variables, and triggers were independent variables. We calculated the score for each EWS to the minute level throughout our retrospective database. If the score for a specific EWS was higher than the published/recommended threshold for that EWS, an alert was considered to have been issued, and the patient was followed for 24 hours. If the patient had an event in the subsequent 24 hours, or 1 hour before (1-hour look-back), the alert was considered a true positive; if not, a false positive. Events that were not preceded by an alert were false negatives, and 24-hour intervals without either an alert or an event were considered true negatives. This simulation exercise was performed for each EWS in both subcohorts (medical and surgical). Clusters of RRS calls followed by transfers to the ICU within 3 hours were considered as a single adverse event (RRS calls, as it was the first event to occur) to avoid double counting. We have described how well this simulation methodology,8 correlates with results from prospective studies.19
Statistical Analysis
To calculate whether results were statistically significant for subgroups, a jackknife method of calculating variance20 was used. The jackknife method calculates variance by repeating the calculations of the statistic leaving out 1 sample at a time. In our case, we repeated the calculation of sensitivity and PPV leaving out 1 patient at a time. Once the simulation method had been run and the false/true positives/negatives had been assigned, calculation of each metric (PPV and sensitivity) was repeated for n subsamples, each leaving out 1 patient. The variance was calculated and 2 Student t tests were performed for each EWS: 1 for PPV and another for sensitivity. SAS statistical software v 9.3 was used for the simulation analysis; R statistical software v 3.0.2 (The R Foundation, Vienna, Austria) was used for the calculation of the statistical significance of results. A univariable analysis was also performed to assess the sensitivity and PPVs for the published thresholds of the most common variables in each EWS: respiratory rate, systolic blood pressure, heart rate, temperature, and mental status as measured by the modified Richmond Agitation Sedation Score.21
RESULTS
The initial cohort included 60,020 hospitalizations, of which the following were excluded: 2751 because of a lack of appropriate research authorization; 6433 because the patients were younger than 18 years; 2129 as psychiatric admissions; 284 as rehabilitation admissions; 872 as research purposes-only admissions; and 1185 because the patient was never in a general care bed (eg, they were either admitted directly to the ICU, or they were admitted for an outpatient surgical procedure and spent time in the postanesthesia care unit).
Table 1 summarizes patient and trigger characteristics, overall and by subgroup. The final cohort included 75,240 total episodes in 46,366 hospitalizations, from 34,898 unique patients, of which 48.7% were male. There were 23,831 medical and 22,535 surgical hospitalizations. Median length of episode was 2 days both for medical and surgical patients. Median length of stay was 3 days, both for medical and for surgical patients.
There were 3332 events in total, of which 1709 were RRS calls, 185 were resuscitation calls, and 1438 were unscheduled transfers to the ICU. The rate of events was 4.67 events per 100 episodes in the aggregate adult population. There were 3.93 events per 100 episodes for surgical hospitalizations, and 5.86 events per 100 episodes for medical hospitalizations (P < .001). The number of CRAs in our cohort was 0.27 per 100 episodes, 0.128 per hospital bed per year, or 4.37 per 1000 hospital admissions, similar to other reported numbers in the literature.3, 22,23
The total number of EWS triggers varied greatly between EWS rules, with the volume ranging during the study year from 1363 triggers with the GMEWS rule to 77,711 triggers with the ViEWS score.
All scores had PPVs less than 25%. As seen in Table 2 and shown graphically in the Figure, all scores performed better on medical patients (blue) than on surgical patients (yellow). The P value was < .0001 for both PPV and sensitivity. The Worthing score had the highest sensitivity (0.78 for medical and 0.68 for surgical) but a very low PPV (0.04 for medical and 0.03 for surgical), while GMEWS was the opposite: low sensitivity (0.10 and 0.07) but the highest PPV (0.22 and 0.18).
The results of the univariable analysis can be seen in Table 3. Most of the criteria performed better (higher sensitivity and PPV) as predictors in the medical hospitalizations than in the surgical hospitalizations.
DISCUSSION
We hypothesized that EWS may perform differently when applied to medical rather than surgical patients. Studies had not analyzed this in a time-dependent manner,24-26 which limited the applicability of the results.8
All analyzed scores performed better in medical patients than in surgical patients (Figure). This could reflect a behavioral difference by the teams on surgical and medical floors in the decision to activate the RRS, or a bias of the clinicians who designed the scores (mostly nonsurgeons). The difference could also mean that physiological deteriorations are intrinsically different in patients who have undergone anesthesia and surgery. For example, in surgical patients, a bleeding episode is more likely to be the cause of their physiological deterioration, or the lingering effects of anesthesia could mask underlying deterioration. Such patients would benefit from scores where variables such as heart rate, blood pressure, or hemoglobin had more influence.
When comparing the different scores, it was much easier for a patient to meet the alerting score with the Worthing score than with GMEWS. In the Worthing score, a respiratory rate greater than 22 breaths per minute, or a systolic blood pressure less than 100 mm Hg, already meet alerting criteria. Similar vital signs result in 0 and 1 points (respectively) in GMEWS, far from its alerting score of 5. This reflects the intrinsic tradeoff of EWS: as the threshold for considering a patient “at risk” drops, not only does the number of true positives (and the sensitivity) increase, but also the number of false positives, thus lowering the PPV.
However, none of the scores analyzed were considered to perform well based on their PPV and sensitivity, particularly in the surgical subpopulation. Focusing on another metric, the area under the receiver operator curve can give misleadingly optimistic results.24,27 However, the extremely low prevalence of acute physiological deterioration can produce low PPVs even when specificity seems acceptable, which is why it is important to evaluate PPV directly.28
To use EWS effectively to activate RRS, they need to be combined with clinical judgment to avoid high levels of false alerts, particularly in surgical patients. It has been reported that RRS is activated only 30% of the time a patient meets RRS calling criteria.29 While there may be cultural characteristics inhibiting the decision to call,30 our study hints at another explanation: if RRS was activated every time a patient met calling criteria based on the scores analyzed, the number of RRS calls would be very high and difficult to manage. So health providers may be doing the right thing when “filtering” RRS calls and not applying the criteria strictly, but in conjunction with clinical judgment.
A limitation of any study like this is how to define “acute physiological deterioration.” We defined an event as recognized episodes of acute physiological deterioration that are signaled by escalations of care (eg, RRS, resuscitation calls, or transfers to an ICU) or unexpected death. By definition, our calculated PPV is affected by clinicians’ recognition of clinical deteriorations. This definition, common in the literature, has the limitation of potentially underestimating EWS’ performance by missing some events that are resolved by the primary care team without an escalation of care. However, we believe our interpretation is not unreasonable since the purpose of EWS is to trigger escalations of care in a timely fashion. Prospective studies could define an event in a way that is less affected by the clinicians’ judgment.
Regarding patient demographics, age was similar between the 2 groups (average, 58.2 years for medical vs 58.9 years for surgical), and there was only a small difference in gender ratios (45.1% male in the medical vs 51.4% in the surgical group). These differences are unlikely to have affected the results significantly, but unknown differences in demographics or other patient characteristics between groups may account for differences in score performance between surgical and medical patients.
Several of the EWS analyzed had overlapping trigger criteria with our own RRS activation criteria (although as single-parameter triggers and not as aggregate). To test how these potential biases could affect our results, we performed a post hoc sensitivity analysis eliminating calls to the RRS as an outcome (so using the alternative outcome of unexpected transfers to the ICU and resuscitation calls). The results are similar to those of our main analysis, with all analyzed scores having lower sensitivity and PPV in surgical hospitalizations when compared to medical hospitalizations.
Our study suggests that, to optimize detection of physiological deterioration events, EWS should try to take into account different patient types, with the most basic distinction being surgical vs medical. This tailoring will make EWS more complex, and less suited for paper-based calculation, but new electronic health records are increasingly able to incorporate decision support, and some EWS have been developed for electronic calculation only. Of particular interest in this regard is the score developed by Escobar et al,31 which groups patients into categories according to the reason for admission, and calculates a different subscore based on that category. While the score by Escobar et al. does not split patients based on medical or surgical status, a more general interpretation of our results suggests that a score may be more accurate if it classifies patients into subgroups with different subscores. This seems to be confirmed by the fact that the score by Escobar et al performs better than MEWS.28 Unfortunately, the paper describing it does not provide enough detail to use it in our database.
A recent systematic review showed increasing evidence that RRS may be effective in reducing CRAs occurring in a non-ICU setting and, more important, overall inhospital mortality.32 While differing implementation strategies (eg, different length of the educational effort, changes in the frequency of vital signs monitoring) can impact the success of such an initiative, it has been speculated that the afferent limb (which often includes an EWS) might be the most critical part of the system.33 Our results show that the most widely used EWS perform significantly worse on surgical patients, and suggest that a way to improve the accuracy of EWS would be to tailor the risk calculation to different patient subgroups (eg, medical and surgical patients). Plausible next steps would be to demonstrate that tailoring risk calculation to medical and surgical patients separately can improve risk predictions and accuracy of EWS.
Disclosure
The authors report no financial conflicts of interest.
1. Buist MD, Jarmolowski E, Burton PR, Bernard SA, Waxman BP, Anderson J. Recognising clinical instability in hospital patients before cardiac arrest or unplanned admission to intensive care. A pilot study in a tertiary-care hospital. Med J Aust. 1999; 171(1):22-25. PubMed
2. Schein RM, Hazday N, Pena M, Ruben BH, Sprung CL. Clinical antecedents to in-hospital cardiopulmonary arrest. Chest. 1990;98(6):1388-1392. PubMed
3. Peberdy MA, Kaye W, Ornato JP, Larkin GL, Nadkarni V, Mancini ME, et al. Cardiopulmonary resuscitation of adults in the hospital: a report of 14720 cardiac arrests from the National Registry of Cardiopulmonary Resuscitation. Resuscitation. 2003; 58(3):297-308. PubMed
4. Nadkarni VM, Larkin GL, Peberdy MA, Carey SM, Kaye W, Mancini ME, et al. First documented rhythm and clinical outcome from in-hospital cardiac arrest among children and adults. JAMA. 2006;295(1):50-57. PubMed
5. Devita MA, Bellomo R, Hillman K, Kellum J, Rotondi A, Teres D, et al. Findings of the first consensus conference on medical emergency teams. Crit Care Med. 2006;34(9):2463-2478. PubMed
6. DeVita MA, Smith GB, Adam SK, Adams-Pizarro I, Buist M, Bellomo R, et al. “Identifying the hospitalised patient in crisis”--a consensus conference on the afferent limb of rapid response systems. Resuscitation. 2010;81(4):375-382. PubMed
7. Kawamoto K, Houlihan CA, Balas EA, Lobach DF. Improving clinical practice using clinical decision support systems: a systematic review of trials to identify features critical to success. BMJ. 2005;330(7494):765. PubMed
8. Romero-Brufau S, Huddleston JM, Naessens JM, Johnson MG, Hickman J, Morlan BW, et al. Widely used track and trigger scores: are they ready for automation in practice? Resuscitation. 2014;85(4):549-552. PubMed
9. Huddleston JM, Diedrich DA, Kinsey GC, Enzler MJ, Manning DM. Learning from every death. J Patient Saf. 2014;10(1):6-12. PubMed
10. Moriarty JP, Schiebel NE, Johnson MG, Jensen JB, Caples SM, Morlan BW, et al. Evaluating implementation of a rapid response team: considering alternative outcome measures. Int J Qual Health Care. 2014;26(1):49-57. PubMed
11. Kirkland LL, Malinchoc M, O’Byrne M, Benson JT, Kashiwagi DT, Burton MC, et al. A clinical deterioration prediction tool for internal medicine patients. Am J Med Qual. 2013;28(2):135-142. PubMed
12. Griffiths JR, Kidney EM. Current use of early warning scores in UK emergency departments. Emerg Med J. 2012;29(1):65-66. PubMed
13. Subbe CP, Kruger M, Rutherford P, Gemmel L. Validation of a modified Early Warning Score in medical admissions. QJM. 2001;94(10):521-526. PubMed
14. Paterson R, MacLeod DC, Thetford D, Beattie A, Graham C, Lam S, et al.. Prediction of in-hospital mortality and length of stay using an early warning scoring system: clinical audit. Clin Med (Lond). 2006;6(3):281-284. PubMed
15. Harrison GA, Jacques T, McLaws ML, Kilborn G. Combinations of early signs of critical illness predict in-hospital death–the SOCCER study (signs of critical conditions and emergency responses). Resuscitation. 2006;71(3):327-334. PubMed
16. Duckitt RW, Buxton-Thomas R, Walker J, Cheek E, Bewick V, Venn R, et al. Worthing physiological scoring system: derivation and validation of a physiological early-warning system for medical admissions. An observational, population-based single-centre study. Br J Anaesth. 2007; 98(6):769-774. PubMed
17. Smith GB, Prytherch DR, Meredith P, Schmidt PE, Featherstone PI. The ability of the National Early Warning Score (NEWS) to discriminate patients at risk of early cardiac arrest, unanticipated intensive care unit admission, and death. Resuscitation. 2013;84(4):465-470. PubMed
18. Prytherch DR, Smith GB, Schmidt PE, Featherstone PI. ViEWS--Towards a national early warning score for detecting adult inpatient deterioration. Resuscitation. 2010;81(8):932-937. PubMed
19. Romero-Brufau S, Huddleston JM. Reply to letter: widely used track and trigger scores: are they ready for automation in practice? Resuscitation. 2014;85(10):e159. PubMed
20. Efron B, Stein C. The jackknife estimate of variance. Annals of Statistics. 1981;586-596.
21. Sessler CN, Gosnell MS, Grap MJ, Brophy GM, O’Neal PV, Keane KA, et al. The Richmond Agitation-Sedation Scale: validity and reliability in adult intensive care unit patients. Am J Respir Crit Care Med. 2002;166(10):1338-1344. PubMed
22. DeVita MA, Braithwaite RS, Mahidhara R, Stuart S, Foraida M, Simmons RL. Medical Emergency Response Improvement Team (MERIT). Use of medical emergency team responses to reduce hospital cardiopulmonary arrests. Qual Saf Health Care. 2004;13(4):251-254. PubMed
23. Goncales PD, Polessi JA, Bass LM, Santos Gde P, Yokota PK, Laselva CR, et al. Reduced frequency of cardiopulmonary arrests by rapid response teams. Einstein (Sao Paulo). 2012;10(4):442-448. PubMed
24. Cuthbertson BH, Boroujerdi M, McKie L, Aucott L, Prescott G. Can physiological variables and early warning scoring systems allow early recognition of the deteriorating surgical patient? Crit Care Med. 2007;35(2):402-409. PubMed
25. Gardner-Thorpe J, Love N, Wrightson J, Walsh S, Keeling N. The value of Modified Early Warning Score (MEWS) in surgical in-patients: a prospective observational study. Ann R Coll Surg Engl. 2006;88(6):571-575. PubMed
26. Stenhouse C, Coates S, Tivey M, Allsop P, Parker T. Prospective evaluation of a modified Early Warning Score to aid earlier detection of patients developing critical illness on a general surgical ward. British Journal of Anaesthesia. 2000;84(5):663-663.
27. Smith GB, Prytherch DR, Schmidt PE, Featherstone PI. Review and performance evaluation of aggregate weighted ‘track and trigger’ systems. Resuscitation. 2008;77(2):170-179. PubMed
28. Romero-Brufau S, Huddleston JM, Escobar GJ, Liebow M. Why the C-statistic is not informative to evaluate early warning scores and what metrics to use. Crit Care. 2015; 19:285. PubMed
29. Hillman K, Chen J, Cretikos M, Bellomo R, Brown D, Doig G, et al. Introduction of the medical emergency team (MET) system: a cluster-randomised controlled trial. Lancet. 2005;365(9477):2091-2097. PubMed
30. Shearer B, Marshall S, Buist MD, Finnigan M, Kitto S, Hore T, et al. What stops hospital clinical staff from following protocols? An analysis of the incidence and factors behind the failure of bedside clinical staff to activate the rapid response system in a multi-campus Australian metropolitan healthcare service. BMJ Qual Saf. 2012;21(7):569-575. PubMed
31. Escobar GJ, LaGuardia JC, Turk BJ, Ragins A, Kipnis P, Draper D. Early detection of impending physiologic deterioration among patients who are not in intensive care: development of predictive models using data from an automated electronic medical record. J Hosp Med. 2012;7(5):388-395. PubMed
32. Winters BD, Weaver SJ, Pfoh ER, Yang T, Pham JC, Dy SM. Rapid-response systems as a patient safety strategy: a systematic review. Ann Intern Med. 2013;158(5 pt 2):417-425. PubMed
33. Jones DA, DeVita MA, Bellomo R. Rapid-response teams. N Engl J Med. 2011;365(2):139-146. PubMed
1. Buist MD, Jarmolowski E, Burton PR, Bernard SA, Waxman BP, Anderson J. Recognising clinical instability in hospital patients before cardiac arrest or unplanned admission to intensive care. A pilot study in a tertiary-care hospital. Med J Aust. 1999; 171(1):22-25. PubMed
2. Schein RM, Hazday N, Pena M, Ruben BH, Sprung CL. Clinical antecedents to in-hospital cardiopulmonary arrest. Chest. 1990;98(6):1388-1392. PubMed
3. Peberdy MA, Kaye W, Ornato JP, Larkin GL, Nadkarni V, Mancini ME, et al. Cardiopulmonary resuscitation of adults in the hospital: a report of 14720 cardiac arrests from the National Registry of Cardiopulmonary Resuscitation. Resuscitation. 2003; 58(3):297-308. PubMed
4. Nadkarni VM, Larkin GL, Peberdy MA, Carey SM, Kaye W, Mancini ME, et al. First documented rhythm and clinical outcome from in-hospital cardiac arrest among children and adults. JAMA. 2006;295(1):50-57. PubMed
5. Devita MA, Bellomo R, Hillman K, Kellum J, Rotondi A, Teres D, et al. Findings of the first consensus conference on medical emergency teams. Crit Care Med. 2006;34(9):2463-2478. PubMed
6. DeVita MA, Smith GB, Adam SK, Adams-Pizarro I, Buist M, Bellomo R, et al. “Identifying the hospitalised patient in crisis”--a consensus conference on the afferent limb of rapid response systems. Resuscitation. 2010;81(4):375-382. PubMed
7. Kawamoto K, Houlihan CA, Balas EA, Lobach DF. Improving clinical practice using clinical decision support systems: a systematic review of trials to identify features critical to success. BMJ. 2005;330(7494):765. PubMed
8. Romero-Brufau S, Huddleston JM, Naessens JM, Johnson MG, Hickman J, Morlan BW, et al. Widely used track and trigger scores: are they ready for automation in practice? Resuscitation. 2014;85(4):549-552. PubMed
9. Huddleston JM, Diedrich DA, Kinsey GC, Enzler MJ, Manning DM. Learning from every death. J Patient Saf. 2014;10(1):6-12. PubMed
10. Moriarty JP, Schiebel NE, Johnson MG, Jensen JB, Caples SM, Morlan BW, et al. Evaluating implementation of a rapid response team: considering alternative outcome measures. Int J Qual Health Care. 2014;26(1):49-57. PubMed
11. Kirkland LL, Malinchoc M, O’Byrne M, Benson JT, Kashiwagi DT, Burton MC, et al. A clinical deterioration prediction tool for internal medicine patients. Am J Med Qual. 2013;28(2):135-142. PubMed
12. Griffiths JR, Kidney EM. Current use of early warning scores in UK emergency departments. Emerg Med J. 2012;29(1):65-66. PubMed
13. Subbe CP, Kruger M, Rutherford P, Gemmel L. Validation of a modified Early Warning Score in medical admissions. QJM. 2001;94(10):521-526. PubMed
14. Paterson R, MacLeod DC, Thetford D, Beattie A, Graham C, Lam S, et al.. Prediction of in-hospital mortality and length of stay using an early warning scoring system: clinical audit. Clin Med (Lond). 2006;6(3):281-284. PubMed
15. Harrison GA, Jacques T, McLaws ML, Kilborn G. Combinations of early signs of critical illness predict in-hospital death–the SOCCER study (signs of critical conditions and emergency responses). Resuscitation. 2006;71(3):327-334. PubMed
16. Duckitt RW, Buxton-Thomas R, Walker J, Cheek E, Bewick V, Venn R, et al. Worthing physiological scoring system: derivation and validation of a physiological early-warning system for medical admissions. An observational, population-based single-centre study. Br J Anaesth. 2007; 98(6):769-774. PubMed
17. Smith GB, Prytherch DR, Meredith P, Schmidt PE, Featherstone PI. The ability of the National Early Warning Score (NEWS) to discriminate patients at risk of early cardiac arrest, unanticipated intensive care unit admission, and death. Resuscitation. 2013;84(4):465-470. PubMed
18. Prytherch DR, Smith GB, Schmidt PE, Featherstone PI. ViEWS--Towards a national early warning score for detecting adult inpatient deterioration. Resuscitation. 2010;81(8):932-937. PubMed
19. Romero-Brufau S, Huddleston JM. Reply to letter: widely used track and trigger scores: are they ready for automation in practice? Resuscitation. 2014;85(10):e159. PubMed
20. Efron B, Stein C. The jackknife estimate of variance. Annals of Statistics. 1981;586-596.
21. Sessler CN, Gosnell MS, Grap MJ, Brophy GM, O’Neal PV, Keane KA, et al. The Richmond Agitation-Sedation Scale: validity and reliability in adult intensive care unit patients. Am J Respir Crit Care Med. 2002;166(10):1338-1344. PubMed
22. DeVita MA, Braithwaite RS, Mahidhara R, Stuart S, Foraida M, Simmons RL. Medical Emergency Response Improvement Team (MERIT). Use of medical emergency team responses to reduce hospital cardiopulmonary arrests. Qual Saf Health Care. 2004;13(4):251-254. PubMed
23. Goncales PD, Polessi JA, Bass LM, Santos Gde P, Yokota PK, Laselva CR, et al. Reduced frequency of cardiopulmonary arrests by rapid response teams. Einstein (Sao Paulo). 2012;10(4):442-448. PubMed
24. Cuthbertson BH, Boroujerdi M, McKie L, Aucott L, Prescott G. Can physiological variables and early warning scoring systems allow early recognition of the deteriorating surgical patient? Crit Care Med. 2007;35(2):402-409. PubMed
25. Gardner-Thorpe J, Love N, Wrightson J, Walsh S, Keeling N. The value of Modified Early Warning Score (MEWS) in surgical in-patients: a prospective observational study. Ann R Coll Surg Engl. 2006;88(6):571-575. PubMed
26. Stenhouse C, Coates S, Tivey M, Allsop P, Parker T. Prospective evaluation of a modified Early Warning Score to aid earlier detection of patients developing critical illness on a general surgical ward. British Journal of Anaesthesia. 2000;84(5):663-663.
27. Smith GB, Prytherch DR, Schmidt PE, Featherstone PI. Review and performance evaluation of aggregate weighted ‘track and trigger’ systems. Resuscitation. 2008;77(2):170-179. PubMed
28. Romero-Brufau S, Huddleston JM, Escobar GJ, Liebow M. Why the C-statistic is not informative to evaluate early warning scores and what metrics to use. Crit Care. 2015; 19:285. PubMed
29. Hillman K, Chen J, Cretikos M, Bellomo R, Brown D, Doig G, et al. Introduction of the medical emergency team (MET) system: a cluster-randomised controlled trial. Lancet. 2005;365(9477):2091-2097. PubMed
30. Shearer B, Marshall S, Buist MD, Finnigan M, Kitto S, Hore T, et al. What stops hospital clinical staff from following protocols? An analysis of the incidence and factors behind the failure of bedside clinical staff to activate the rapid response system in a multi-campus Australian metropolitan healthcare service. BMJ Qual Saf. 2012;21(7):569-575. PubMed
31. Escobar GJ, LaGuardia JC, Turk BJ, Ragins A, Kipnis P, Draper D. Early detection of impending physiologic deterioration among patients who are not in intensive care: development of predictive models using data from an automated electronic medical record. J Hosp Med. 2012;7(5):388-395. PubMed
32. Winters BD, Weaver SJ, Pfoh ER, Yang T, Pham JC, Dy SM. Rapid-response systems as a patient safety strategy: a systematic review. Ann Intern Med. 2013;158(5 pt 2):417-425. PubMed
33. Jones DA, DeVita MA, Bellomo R. Rapid-response teams. N Engl J Med. 2011;365(2):139-146. PubMed
© 2017 Society of Hospital Medicine
Onetime, Nondrug Treatment May Be Better for Some MS Patients
High-dose immunosuppressive therapy and autologous hematopoietic cell transplant (HDIT/HCT) has had “highly promising” results for patients with relapsing-remitting multiple sclerosis (MS), according to researchers from the HALT-MS trial. In the 5-year study, 69% of 24 participants survived without progression of disability, relapse, or new brain lesions, despite not taking MS medications.
Findings published at the 3-year mark were encouraging. The event-free survival rate was 78%. The extended findings suggest that “onetime treatment with HDIT/HCT may be substantially more effective than long-term treatment with the best available medications” for these patients, said NIAID Director Anthony Fauci, MD.
The treatment “resets” the immune system, the researchers say. First, doctors collect the patient’s blood-forming stem cells, then give the patient chemotherapy to deplete the immune system. Finally, the doctors return the patient’s stem cells to rebuild the immune system.
Adverse events were consistent with those routinely observed after HDIT/HCT. Adverse effects recorded at 4 and 5 years were not related to the transplant and were not considered severe. Three patients died, but their deaths were not related to the study treatments.
Five years later, most trial participants remained in remission and stabilized, and some showed improvements, such as recovering mobility.
High-dose immunosuppressive therapy and autologous hematopoietic cell transplant (HDIT/HCT) has had “highly promising” results for patients with relapsing-remitting multiple sclerosis (MS), according to researchers from the HALT-MS trial. In the 5-year study, 69% of 24 participants survived without progression of disability, relapse, or new brain lesions, despite not taking MS medications.
Findings published at the 3-year mark were encouraging. The event-free survival rate was 78%. The extended findings suggest that “onetime treatment with HDIT/HCT may be substantially more effective than long-term treatment with the best available medications” for these patients, said NIAID Director Anthony Fauci, MD.
The treatment “resets” the immune system, the researchers say. First, doctors collect the patient’s blood-forming stem cells, then give the patient chemotherapy to deplete the immune system. Finally, the doctors return the patient’s stem cells to rebuild the immune system.
Adverse events were consistent with those routinely observed after HDIT/HCT. Adverse effects recorded at 4 and 5 years were not related to the transplant and were not considered severe. Three patients died, but their deaths were not related to the study treatments.
Five years later, most trial participants remained in remission and stabilized, and some showed improvements, such as recovering mobility.
High-dose immunosuppressive therapy and autologous hematopoietic cell transplant (HDIT/HCT) has had “highly promising” results for patients with relapsing-remitting multiple sclerosis (MS), according to researchers from the HALT-MS trial. In the 5-year study, 69% of 24 participants survived without progression of disability, relapse, or new brain lesions, despite not taking MS medications.
Findings published at the 3-year mark were encouraging. The event-free survival rate was 78%. The extended findings suggest that “onetime treatment with HDIT/HCT may be substantially more effective than long-term treatment with the best available medications” for these patients, said NIAID Director Anthony Fauci, MD.
The treatment “resets” the immune system, the researchers say. First, doctors collect the patient’s blood-forming stem cells, then give the patient chemotherapy to deplete the immune system. Finally, the doctors return the patient’s stem cells to rebuild the immune system.
Adverse events were consistent with those routinely observed after HDIT/HCT. Adverse effects recorded at 4 and 5 years were not related to the transplant and were not considered severe. Three patients died, but their deaths were not related to the study treatments.
Five years later, most trial participants remained in remission and stabilized, and some showed improvements, such as recovering mobility.
The State of Hepatitis C Care in the VA
As the largest single provider of hepatitis C virus infection (HCV) care in the U.S., the VA provided care to > 174,000 veterans with chronic HCV in 2013. Identifying veterans most likely to be infected with HCV, particularly those in the highrisk birth cohort born between 1945 and 1965, is a priority given recent CDC and U.S. Preventive Services Task Force (USPSTF) recommendations.1,2 The availability of new, all-oral HCV antiviral regimens with shorter treatment durations and improved tolerability are expected to greatly increase the number of veterans with HCV who could be treated successfully. In order to effectively reach those who are undiagnosed and to ensure that those diagnosed with HCV are evaluated and offered treatment, expanded reliance on primary care providers (PCPs) is essential. This article provides a population view of the current state of VA care for this large HCV-infected population and the important role PCPs share in disease identification and management.
Data Source
Data regarding the state of HCV care in the VA comes from the VA National Clinical Case Registry (CCR) for HCV.3 The VA HCV CCR is an extract of the VA electronic medical record that contains laboratory results, pharmacy information, provider information, and ICD-9 diagnosis codes from inpatient hospitalizations, outpatient visits, and problem lists of veterans with HCV seen at all VAMCs.
Screening and Prevalence of HCV
It is estimated that 2.3 to 2.7 million Americans are living with HCV, with 45% to 85% of those unaware of their infection.4,5 Nearly 75% of those infected are expected to have been born between 1945 and 1965; thus, the CDC and USPSTF now recommend onetime HCV screening for this birth cohort.1,2 Among nearly 5.6 million veterans with a VA outpatient visit in 2013, 56% have been screened for HCV. The HCV screening rate was 42% for those born prior to 1945, 65% for those born during 1945-1965, and 59% for those born after 1965. HCV infection prevalence overall in the VA was 5.8% but differed markedly among the birth cohorts: 1.6% for those born prior to 1945, 9.5% for those born during 1945-1965, and 1.2% for those born after 1965. The prevalence rate of veterans born in the 1945-1965 birth cohort (9.5%) is almost 3 times higher than that of the general U.S. population in this birth cohort (2.4%). The high prevalence serves as a reminder of the greater HCV disease burden in veterans and largely represents Vietman era veterans. Although HCV screening rates in VA have increased over 25% since 2002, the high prevalence among veterans in this birth cohort underscores the importance of continued screening efforts.
Veterans with Chronic HCV Infection
The VA Office of Public Health/Population Health generates national HCV reports annually from the HCV CCR describing the population of veterans with chronic HCV infection receiving VA care. These reports are intended to inform about patient care activities, clinician and patient education, prevention activities, and research directed at continuous improvement of veteran care. The first step in providing responsive care is understanding the affected population, and summarized herein is a description of the veterans with chronic HCV who received VA care in 2013.
In 2013, 174,302 veterans had laboratory evidence of HCV viremia and could be characterized as having chronic HCV. HCV treatment regimens and response depend on the viral genotype. Among veterans with genotype testing, 107,144 (80%) have genotype 1; 15,486 (12%) genotype 2; 9,745 (7%) genotype 3; 1333 (1%) genotype 4; and 63 (< 1%) genotype 5 or 6.
In terms of demographics, most veterans with chronic HCV in VA care in 2013 were men (97%); however, > 5,000 women veterans with chronic HCV received care from the VA. Over half (54%) of veterans with chronic HCV are white, and about one-third (34%) are black. The proportion of blacks within the HCV-infected veteran population is substantially greater than the proportion of blacks in the overall veteran population in VA care (15%) and highlights the disproportionately large burden of HCV that black veterans bear. A smaller proportion of the VA HCV population is Hispanic (6%), and the remaining veterans are other races, multiple races, or unknown.
The HCV-infected veteran population is aging. The mean age of veterans with chronic HCV in 2013 was 59.7 years and for the first time, more veterans with HCV were aged 60 years (Figure 1).
Among the comorbidities that may have historically prevented veterans from receiving HCV antiviral therapy, 2 of the most pervasive are mental health conditions and alcohol use. The rates of mental illness among veterans overall is high, but mental illness is particularly high in veterans with HCV. Depression has affected 60%; of this population anxiety, 37%; posttraumatic stress disorder, 28%; and schizophrenia, 10%. Alcohol use disorders are also common among veterans with HCV in care. Active mental health conditions and substance use may affect medication adherence or follow-up visit adherence thereby limiting treatment candidacy. Integrating care of these individuals with mental health providers and substance-use treatment specialists is an important aspect of HCV care and is a priority in VA.
Three-quarters (76%) of the HCV-infected veteran population has been screened for HIV and HIV-HCV co-infection is present in 5733 (3%) of veterans with HCV. HIV-HCV co-infection is associated with an increased progression of liver disease and may have implications for the selection of HCV antiviral agents because of drug interactions. Hepatitis B virus (HBV)-HCV co-infection rates are higher at 7%. HBV vaccination or documentation of HBV immunity in those without HBV infection is 78%.
With regard to specific liver complications, 5% to 20% of those infected with chronic HCV will develop cirrhosis over a period of 20 to 30 years, and 1% to 5% will die of hepatocellular carcinoma (HCC) or cirrhosis.6 Given the natural history of chronic HCV and the aging HCV veteran cohort, increasing numbers of conditions related to progression of liver disease are expected over time. This is most evident in the number of veterans with a diagnosis of cirrhosis, which has increased from approximately 10,000 veterans (8%) in care in 2001 to nearly 30,000 veterans (17%) in care in 2013 (Figure 2).
Antiviral Therapy for Chronic HCV
Prior to mid-2011, the standard of care for HCV treatment was the combination of pegylated interferon and ribavirin. From 2011 through 2013, direct-acting antiviral (DAA) regimens containing boceprevir and telaprevir in combination with pegylated interferon and ribavirin became standard of care for genotype 1 while
the standard of care remained pegylated interferon and ribavirin for genotypes other than genotype 1. Recent advances in HCV antiviral therapy offer higher cure rates and fewer adverse events (AEs) compared with peginterferon-containing treatment. The expected ease and tolerability of these all-oral combination regimens is anticipated to greatly increase the number of veterans with HCV who could be treated successfully.
Because of the poor tolerability, prolonged treatment durations, serious AEs, and relative or absolute contraindications to peginterferon-based therapy, many veterans were not previously candidates for treatment. Of the 174,302 veterans with chronic HCV in care in 2013, 39,388 (23%) had received at least 1 course of HCV antiviral treatment. This largely reflects the time when peginterferon-based therapy was the standard of care. Since the approval of boceprevir and telaprevir 5,732 veterans (5.8%) in care in 2013 had ever received boceprevir or telaprevir-based regimens.
While recognizing that all veterans should be considered for HCV treatment, the urgency for treatment may be greater in those with advanced liver disease, because these patients are at the highest risk of developing decompensated cirrhosis or dying of liver-related disease. In 2013, there were 28,945 veterans in care that had advanced liver disease who might be considered potential HCV treatment candidates with an urgency to treat.
Duration of treatment and anticipated rates of treatment success with the all-oral regimens depend in part on a patient’s prior treatment status in addition to whether the patient has a diagnosis of advanced liver disease/cirrhosis. Regardless of HCV genotype, among all veterans approximately 85% are treatment-naïve and 15% are treatment-experienced. Advanced liver disease is present in 24% of treatment-naïve and 31% of treatment-experienced veterans with HCV genotype 1; 23% and 24% of veterans with HCV genotype 2, respectively; and 34% and 43% of veterans with HCV genotype 3, respectively.
Further understanding the population of veterans with HCV, including prior treatment status and stage of liver disease, is useful in identifying the target population for treatment. The VA uses these data to project treatment costs and assess capacity across the system in preparation for expected uptake of new regimens.
Sustained Virologic Response After HCV Antiviral Treatment
The goal of HCV antiviral therapy is to eradicate HCV and reduce the progression of liver disease and death from HCV infection. Successful antiviral treatment of HCV is determined by achieving a sustained virologic response (SVR) defined as an undetectable HCV viral load 12 weeks after the end of treatment. Of the 39,388 veterans in VA care in 2013 who have ever received antiviral therapy, SVR could be assessed in 32,815 veterans, and the overall SVR rate in this population was 42%. This SVR rate is similar to that observed in phase III trials of pegylated interferon-based regimens, where 42% to 46% of those infected with HCV genotype 1 achieved SVR.7,8 Although most veterans with genotype 1 infection received boceprevir or telaprevir-based regimens in 2013 and achieved higher SVRs of 50% to 52%, the overwhelming majority of veterans in care in 2013 received prior treatment with only peginterferon and ribavirin.9 Although SVR rates are expected to increase with newer all-oral HCV regimens, differences between clinical efficacy and real-world effectiveness will continue to be apparent,
and patient and provider expectations should be tempered accordingly.
The Role of Primary Care in HCV
Primary care providers have held the responsibility for multiple roles in HCV care since the discovery of the virus—particularly for HCV risk factor assessment, screening, and diagnosis. HCV antiviral treatment, however, was largely placed in the hands of specialists, given the complexities of patient selection, frequent reliance on a liver biopsy for determining need for treatment, and the toxicities of peginterferon and ribavirin therapy.
There are discussions both inside and outside the VA about potentially expanding the role of PCPs in HCV care. First, primary care is the major setting where the CDC and USPSTF recommendations for birth cohort screening are being implemented, and thus PCPs will be identifying veterans previously undiagnosed with HCV.1,2 Second, the ease and tolerability of the new all-oral combination regimens is causing a shift in the paradigm for HCV treatment, from a highly individualized approach, toward a more uniform approach.
Expanding the role of primary care would have multiple benefits to patients and the health care system as a whole. Only approximately 9% of HCV-infected veterans in VA care have been successfully treated at this time, largely due to low eligibility rates and the poor response rates, but other barriers have also contributed to the low success rate, one of which has been limited access to specialists. Furthermore, veterans who are referred to specialists are often noncompliant with the referral.10 If seeing an HCV specialist is required for treatment, the time to treat the HCV population will be much greater, more costly, and less efficient. Therefore, if the prospect of delivering HCV treatment to the majority of HCV patients is to be accomplished, it is necessary to consider providing treatment in the primary care setting as well as the specialist setting.
Treatment provided by nonspecialists has been evaluated in patients receiving peginterferon and ribavirin regimens and has shown that with adequate education and support, SVR rates were equivalent in the specialist and nonspecialist setting.11 To develop programs to provide primary care with such support, the VA has implemented the Specialty Care Access Network-Extension of Community Healthcare Outcomes program initiative, with casebased learning along with real-time consultation.
Currently, the majority of HCV-infected patients have never seen an HCV specialist, thus PCPs are already providing the majority of HCV care beyond HCV antiviral
treatment.12 Primary care providers are, therefore, key to addressing multiple important aspects of HCV care, including (1) counseling patients on transmission, prevention, lifestyle, and the role of substance use; (2) providing hepatitis A and B vaccination as well as appropriate general vaccinations for any patient with chronic liver disease; (3) modifying comorbidities that could accelerate fibrosis progression, such as diabetes mellitus, obesity and hyperlipidemia; (4) reducing risk from ongoing alcohol, drug, and tobacco use; (5) monitoring patients for fibrosis progression and identifying the presence of cirrhosis; and (6) providing general care for patients with cirrhosis, including HCC screening. These are critical aspects of HCV care, and many PCPs may still need additional education for these roles. The VA provides education and support for PCPs in their current role and is enhancing efforts to expand delivery of HCV treatment to the primary care setting as well.
Conclusions
In 2013, the typical veteran with chronic HCV was white, aged 60 years, and male, with a history of comorbidities, including hypertension, depression, and current or prior alcohol abuse. The proportion of veterans with advanced liver disease including cirrhosis (17%) and HCC (3%), has grown significantly over the past 10 years. By the end of 2013, almost 40,000 veterans had received antiviral therapy for HCV, more than 5,700 of whom had received DAAs. Overall SVR rates have been about 42% among those who were treated. Of veterans who are still potential treatment candidates, 85% are treatment-naive and about one-quarter have advanced liver disease.
Although HCV screening rates in veterans are higher than reported in other health care settings, particularly among those in the critical 1945-1965 birth cohort (65% screening rate), substantial numbers of veterans still require testing. The burden of disease, the lack of specialists, the ease and tolerability of new HCV antiviral medications, and the interplay of HCV with other traditional primary care efforts underly an increased role for PCPs in the care of veterans with HCV. Together, this information helps to construct a view of historical, current, and future HCV care in veterans.
Author disclosures
The authors report no actual or potential conflicts of interest with regard to this article.
Disclaimer
The opinions expressed herein are those of the authors and do not necessarily reflect those of Federal Practitioner, Frontline Medical Communications Inc., the U.S. Government, or any of its agencies. This article may discuss unlabeled or investigational use of certain drugs. Please review complete prescribing information for specific drugs or drug combinations—including indications, contraindications, warnings, and adverse effects—before administering pharmacologic therapy to patient.
1. Smith BD, Morgan RL, Beckett GA, et al; Centers for Disease Control and Prevention. Recommendations for the identification of chronic hepatitis C virus infection among persons born during 1945-1965 [published correction appears in MMWR Recomm Rep. 2012;61(43):886]. MMWR Recomm Rep. 2012;61(RR-4):1-32.
2. Moyer VA; U.S. Preventive Services Task Force. Screening for hepatitis C virus infection in adults: U.S. Preventive Services Task Force recommendation statement. Ann Intern Med. 2013;159(5):349-357.
3. Backus LI, Gavrilov S, Loomis TP, et al. Clinical Case Registries: Simultaneous local and national disease registries for population quality management. J Am Med Inform Assoc. 2009;16(6):775-783.
4. Kabiri M, Jazwinski AB, Roberts MS, Schaefer AJ, Chhatwal J. The changing burden of hepatitis C virus infection in the United States: Model-based predictions. Ann Intern Med. 2014;161(3):170-180
5. Denniston MM, Jiles RB, Drobeniuc J, et al. Chronic hepatitis C virus infection in the United States, National Health and Nutrition Examination Survey 2003 to 2010. Ann Intern Med. 2014;160(5):293-300
6. Davila JA, Morgan RO, Shaib Y, McGlynn KA, El-Serag HB. Hepatitis C infection and the increasing incidence of hepatocellular carcinoma: A population-based study. Gastroenterology. 2004;127(5):1372-1380.
7. Manns MP, McHutchison JG, Gordon SC, et al. Peginterferon alfa-2b plus ribavirin compared with interferon alfa-2b plus ribavirin for initial treatment of chronic hepatitis C: A randomised trial. Lancet. 2001;358(9286):958-965.
8. Fried MW, Shiffman ML, Reddy KR, et al. Peginterferon alfa-2a plus ribavirin for chronic hepatitis C virus infection. N Engl J Med. 2002;347(13):975-982.
9. Backus LI, Belperio PS, Shahoumian TA, Cheung R, Mole LA. Comparative effectiveness of the hepatitis C virus protease inhibitors boceprevir and telaprevir in a large U.S. cohort. Aliment Pharmacol Ther. 2014;39(1):93-103.
10. Brady CW, Coffman CJ, Provenzale D. Compliance with referral for hepatitis C evaluation among veterans. J Clin Gastroenterol. 2007;41(10):927-931.
11. Arora S, Thornton K, Murata G, et al. Outcomes of treatment for hepatitis C infection by primary care providers. N Engl J Med. 2011;364(23):2199-2207.
12. Holmberg SD, Spradling PR, Moorman AC, Denniston MM. Hepatitis C in the United States. N Engl J Med. 2013;368(20):1859-1861.
As the largest single provider of hepatitis C virus infection (HCV) care in the U.S., the VA provided care to > 174,000 veterans with chronic HCV in 2013. Identifying veterans most likely to be infected with HCV, particularly those in the highrisk birth cohort born between 1945 and 1965, is a priority given recent CDC and U.S. Preventive Services Task Force (USPSTF) recommendations.1,2 The availability of new, all-oral HCV antiviral regimens with shorter treatment durations and improved tolerability are expected to greatly increase the number of veterans with HCV who could be treated successfully. In order to effectively reach those who are undiagnosed and to ensure that those diagnosed with HCV are evaluated and offered treatment, expanded reliance on primary care providers (PCPs) is essential. This article provides a population view of the current state of VA care for this large HCV-infected population and the important role PCPs share in disease identification and management.
Data Source
Data regarding the state of HCV care in the VA comes from the VA National Clinical Case Registry (CCR) for HCV.3 The VA HCV CCR is an extract of the VA electronic medical record that contains laboratory results, pharmacy information, provider information, and ICD-9 diagnosis codes from inpatient hospitalizations, outpatient visits, and problem lists of veterans with HCV seen at all VAMCs.
Screening and Prevalence of HCV
It is estimated that 2.3 to 2.7 million Americans are living with HCV, with 45% to 85% of those unaware of their infection.4,5 Nearly 75% of those infected are expected to have been born between 1945 and 1965; thus, the CDC and USPSTF now recommend onetime HCV screening for this birth cohort.1,2 Among nearly 5.6 million veterans with a VA outpatient visit in 2013, 56% have been screened for HCV. The HCV screening rate was 42% for those born prior to 1945, 65% for those born during 1945-1965, and 59% for those born after 1965. HCV infection prevalence overall in the VA was 5.8% but differed markedly among the birth cohorts: 1.6% for those born prior to 1945, 9.5% for those born during 1945-1965, and 1.2% for those born after 1965. The prevalence rate of veterans born in the 1945-1965 birth cohort (9.5%) is almost 3 times higher than that of the general U.S. population in this birth cohort (2.4%). The high prevalence serves as a reminder of the greater HCV disease burden in veterans and largely represents Vietman era veterans. Although HCV screening rates in VA have increased over 25% since 2002, the high prevalence among veterans in this birth cohort underscores the importance of continued screening efforts.
Veterans with Chronic HCV Infection
The VA Office of Public Health/Population Health generates national HCV reports annually from the HCV CCR describing the population of veterans with chronic HCV infection receiving VA care. These reports are intended to inform about patient care activities, clinician and patient education, prevention activities, and research directed at continuous improvement of veteran care. The first step in providing responsive care is understanding the affected population, and summarized herein is a description of the veterans with chronic HCV who received VA care in 2013.
In 2013, 174,302 veterans had laboratory evidence of HCV viremia and could be characterized as having chronic HCV. HCV treatment regimens and response depend on the viral genotype. Among veterans with genotype testing, 107,144 (80%) have genotype 1; 15,486 (12%) genotype 2; 9,745 (7%) genotype 3; 1333 (1%) genotype 4; and 63 (< 1%) genotype 5 or 6.
In terms of demographics, most veterans with chronic HCV in VA care in 2013 were men (97%); however, > 5,000 women veterans with chronic HCV received care from the VA. Over half (54%) of veterans with chronic HCV are white, and about one-third (34%) are black. The proportion of blacks within the HCV-infected veteran population is substantially greater than the proportion of blacks in the overall veteran population in VA care (15%) and highlights the disproportionately large burden of HCV that black veterans bear. A smaller proportion of the VA HCV population is Hispanic (6%), and the remaining veterans are other races, multiple races, or unknown.
The HCV-infected veteran population is aging. The mean age of veterans with chronic HCV in 2013 was 59.7 years and for the first time, more veterans with HCV were aged 60 years (Figure 1).
Among the comorbidities that may have historically prevented veterans from receiving HCV antiviral therapy, 2 of the most pervasive are mental health conditions and alcohol use. The rates of mental illness among veterans overall is high, but mental illness is particularly high in veterans with HCV. Depression has affected 60%; of this population anxiety, 37%; posttraumatic stress disorder, 28%; and schizophrenia, 10%. Alcohol use disorders are also common among veterans with HCV in care. Active mental health conditions and substance use may affect medication adherence or follow-up visit adherence thereby limiting treatment candidacy. Integrating care of these individuals with mental health providers and substance-use treatment specialists is an important aspect of HCV care and is a priority in VA.
Three-quarters (76%) of the HCV-infected veteran population has been screened for HIV and HIV-HCV co-infection is present in 5733 (3%) of veterans with HCV. HIV-HCV co-infection is associated with an increased progression of liver disease and may have implications for the selection of HCV antiviral agents because of drug interactions. Hepatitis B virus (HBV)-HCV co-infection rates are higher at 7%. HBV vaccination or documentation of HBV immunity in those without HBV infection is 78%.
With regard to specific liver complications, 5% to 20% of those infected with chronic HCV will develop cirrhosis over a period of 20 to 30 years, and 1% to 5% will die of hepatocellular carcinoma (HCC) or cirrhosis.6 Given the natural history of chronic HCV and the aging HCV veteran cohort, increasing numbers of conditions related to progression of liver disease are expected over time. This is most evident in the number of veterans with a diagnosis of cirrhosis, which has increased from approximately 10,000 veterans (8%) in care in 2001 to nearly 30,000 veterans (17%) in care in 2013 (Figure 2).
Antiviral Therapy for Chronic HCV
Prior to mid-2011, the standard of care for HCV treatment was the combination of pegylated interferon and ribavirin. From 2011 through 2013, direct-acting antiviral (DAA) regimens containing boceprevir and telaprevir in combination with pegylated interferon and ribavirin became standard of care for genotype 1 while
the standard of care remained pegylated interferon and ribavirin for genotypes other than genotype 1. Recent advances in HCV antiviral therapy offer higher cure rates and fewer adverse events (AEs) compared with peginterferon-containing treatment. The expected ease and tolerability of these all-oral combination regimens is anticipated to greatly increase the number of veterans with HCV who could be treated successfully.
Because of the poor tolerability, prolonged treatment durations, serious AEs, and relative or absolute contraindications to peginterferon-based therapy, many veterans were not previously candidates for treatment. Of the 174,302 veterans with chronic HCV in care in 2013, 39,388 (23%) had received at least 1 course of HCV antiviral treatment. This largely reflects the time when peginterferon-based therapy was the standard of care. Since the approval of boceprevir and telaprevir 5,732 veterans (5.8%) in care in 2013 had ever received boceprevir or telaprevir-based regimens.
While recognizing that all veterans should be considered for HCV treatment, the urgency for treatment may be greater in those with advanced liver disease, because these patients are at the highest risk of developing decompensated cirrhosis or dying of liver-related disease. In 2013, there were 28,945 veterans in care that had advanced liver disease who might be considered potential HCV treatment candidates with an urgency to treat.
Duration of treatment and anticipated rates of treatment success with the all-oral regimens depend in part on a patient’s prior treatment status in addition to whether the patient has a diagnosis of advanced liver disease/cirrhosis. Regardless of HCV genotype, among all veterans approximately 85% are treatment-naïve and 15% are treatment-experienced. Advanced liver disease is present in 24% of treatment-naïve and 31% of treatment-experienced veterans with HCV genotype 1; 23% and 24% of veterans with HCV genotype 2, respectively; and 34% and 43% of veterans with HCV genotype 3, respectively.
Further understanding the population of veterans with HCV, including prior treatment status and stage of liver disease, is useful in identifying the target population for treatment. The VA uses these data to project treatment costs and assess capacity across the system in preparation for expected uptake of new regimens.
Sustained Virologic Response After HCV Antiviral Treatment
The goal of HCV antiviral therapy is to eradicate HCV and reduce the progression of liver disease and death from HCV infection. Successful antiviral treatment of HCV is determined by achieving a sustained virologic response (SVR) defined as an undetectable HCV viral load 12 weeks after the end of treatment. Of the 39,388 veterans in VA care in 2013 who have ever received antiviral therapy, SVR could be assessed in 32,815 veterans, and the overall SVR rate in this population was 42%. This SVR rate is similar to that observed in phase III trials of pegylated interferon-based regimens, where 42% to 46% of those infected with HCV genotype 1 achieved SVR.7,8 Although most veterans with genotype 1 infection received boceprevir or telaprevir-based regimens in 2013 and achieved higher SVRs of 50% to 52%, the overwhelming majority of veterans in care in 2013 received prior treatment with only peginterferon and ribavirin.9 Although SVR rates are expected to increase with newer all-oral HCV regimens, differences between clinical efficacy and real-world effectiveness will continue to be apparent,
and patient and provider expectations should be tempered accordingly.
The Role of Primary Care in HCV
Primary care providers have held the responsibility for multiple roles in HCV care since the discovery of the virus—particularly for HCV risk factor assessment, screening, and diagnosis. HCV antiviral treatment, however, was largely placed in the hands of specialists, given the complexities of patient selection, frequent reliance on a liver biopsy for determining need for treatment, and the toxicities of peginterferon and ribavirin therapy.
There are discussions both inside and outside the VA about potentially expanding the role of PCPs in HCV care. First, primary care is the major setting where the CDC and USPSTF recommendations for birth cohort screening are being implemented, and thus PCPs will be identifying veterans previously undiagnosed with HCV.1,2 Second, the ease and tolerability of the new all-oral combination regimens is causing a shift in the paradigm for HCV treatment, from a highly individualized approach, toward a more uniform approach.
Expanding the role of primary care would have multiple benefits to patients and the health care system as a whole. Only approximately 9% of HCV-infected veterans in VA care have been successfully treated at this time, largely due to low eligibility rates and the poor response rates, but other barriers have also contributed to the low success rate, one of which has been limited access to specialists. Furthermore, veterans who are referred to specialists are often noncompliant with the referral.10 If seeing an HCV specialist is required for treatment, the time to treat the HCV population will be much greater, more costly, and less efficient. Therefore, if the prospect of delivering HCV treatment to the majority of HCV patients is to be accomplished, it is necessary to consider providing treatment in the primary care setting as well as the specialist setting.
Treatment provided by nonspecialists has been evaluated in patients receiving peginterferon and ribavirin regimens and has shown that with adequate education and support, SVR rates were equivalent in the specialist and nonspecialist setting.11 To develop programs to provide primary care with such support, the VA has implemented the Specialty Care Access Network-Extension of Community Healthcare Outcomes program initiative, with casebased learning along with real-time consultation.
Currently, the majority of HCV-infected patients have never seen an HCV specialist, thus PCPs are already providing the majority of HCV care beyond HCV antiviral
treatment.12 Primary care providers are, therefore, key to addressing multiple important aspects of HCV care, including (1) counseling patients on transmission, prevention, lifestyle, and the role of substance use; (2) providing hepatitis A and B vaccination as well as appropriate general vaccinations for any patient with chronic liver disease; (3) modifying comorbidities that could accelerate fibrosis progression, such as diabetes mellitus, obesity and hyperlipidemia; (4) reducing risk from ongoing alcohol, drug, and tobacco use; (5) monitoring patients for fibrosis progression and identifying the presence of cirrhosis; and (6) providing general care for patients with cirrhosis, including HCC screening. These are critical aspects of HCV care, and many PCPs may still need additional education for these roles. The VA provides education and support for PCPs in their current role and is enhancing efforts to expand delivery of HCV treatment to the primary care setting as well.
Conclusions
In 2013, the typical veteran with chronic HCV was white, aged 60 years, and male, with a history of comorbidities, including hypertension, depression, and current or prior alcohol abuse. The proportion of veterans with advanced liver disease including cirrhosis (17%) and HCC (3%), has grown significantly over the past 10 years. By the end of 2013, almost 40,000 veterans had received antiviral therapy for HCV, more than 5,700 of whom had received DAAs. Overall SVR rates have been about 42% among those who were treated. Of veterans who are still potential treatment candidates, 85% are treatment-naive and about one-quarter have advanced liver disease.
Although HCV screening rates in veterans are higher than reported in other health care settings, particularly among those in the critical 1945-1965 birth cohort (65% screening rate), substantial numbers of veterans still require testing. The burden of disease, the lack of specialists, the ease and tolerability of new HCV antiviral medications, and the interplay of HCV with other traditional primary care efforts underly an increased role for PCPs in the care of veterans with HCV. Together, this information helps to construct a view of historical, current, and future HCV care in veterans.
Author disclosures
The authors report no actual or potential conflicts of interest with regard to this article.
Disclaimer
The opinions expressed herein are those of the authors and do not necessarily reflect those of Federal Practitioner, Frontline Medical Communications Inc., the U.S. Government, or any of its agencies. This article may discuss unlabeled or investigational use of certain drugs. Please review complete prescribing information for specific drugs or drug combinations—including indications, contraindications, warnings, and adverse effects—before administering pharmacologic therapy to patient.
As the largest single provider of hepatitis C virus infection (HCV) care in the U.S., the VA provided care to > 174,000 veterans with chronic HCV in 2013. Identifying veterans most likely to be infected with HCV, particularly those in the highrisk birth cohort born between 1945 and 1965, is a priority given recent CDC and U.S. Preventive Services Task Force (USPSTF) recommendations.1,2 The availability of new, all-oral HCV antiviral regimens with shorter treatment durations and improved tolerability are expected to greatly increase the number of veterans with HCV who could be treated successfully. In order to effectively reach those who are undiagnosed and to ensure that those diagnosed with HCV are evaluated and offered treatment, expanded reliance on primary care providers (PCPs) is essential. This article provides a population view of the current state of VA care for this large HCV-infected population and the important role PCPs share in disease identification and management.
Data Source
Data regarding the state of HCV care in the VA comes from the VA National Clinical Case Registry (CCR) for HCV.3 The VA HCV CCR is an extract of the VA electronic medical record that contains laboratory results, pharmacy information, provider information, and ICD-9 diagnosis codes from inpatient hospitalizations, outpatient visits, and problem lists of veterans with HCV seen at all VAMCs.
Screening and Prevalence of HCV
It is estimated that 2.3 to 2.7 million Americans are living with HCV, with 45% to 85% of those unaware of their infection.4,5 Nearly 75% of those infected are expected to have been born between 1945 and 1965; thus, the CDC and USPSTF now recommend onetime HCV screening for this birth cohort.1,2 Among nearly 5.6 million veterans with a VA outpatient visit in 2013, 56% have been screened for HCV. The HCV screening rate was 42% for those born prior to 1945, 65% for those born during 1945-1965, and 59% for those born after 1965. HCV infection prevalence overall in the VA was 5.8% but differed markedly among the birth cohorts: 1.6% for those born prior to 1945, 9.5% for those born during 1945-1965, and 1.2% for those born after 1965. The prevalence rate of veterans born in the 1945-1965 birth cohort (9.5%) is almost 3 times higher than that of the general U.S. population in this birth cohort (2.4%). The high prevalence serves as a reminder of the greater HCV disease burden in veterans and largely represents Vietman era veterans. Although HCV screening rates in VA have increased over 25% since 2002, the high prevalence among veterans in this birth cohort underscores the importance of continued screening efforts.
Veterans with Chronic HCV Infection
The VA Office of Public Health/Population Health generates national HCV reports annually from the HCV CCR describing the population of veterans with chronic HCV infection receiving VA care. These reports are intended to inform about patient care activities, clinician and patient education, prevention activities, and research directed at continuous improvement of veteran care. The first step in providing responsive care is understanding the affected population, and summarized herein is a description of the veterans with chronic HCV who received VA care in 2013.
In 2013, 174,302 veterans had laboratory evidence of HCV viremia and could be characterized as having chronic HCV. HCV treatment regimens and response depend on the viral genotype. Among veterans with genotype testing, 107,144 (80%) have genotype 1; 15,486 (12%) genotype 2; 9,745 (7%) genotype 3; 1333 (1%) genotype 4; and 63 (< 1%) genotype 5 or 6.
In terms of demographics, most veterans with chronic HCV in VA care in 2013 were men (97%); however, > 5,000 women veterans with chronic HCV received care from the VA. Over half (54%) of veterans with chronic HCV are white, and about one-third (34%) are black. The proportion of blacks within the HCV-infected veteran population is substantially greater than the proportion of blacks in the overall veteran population in VA care (15%) and highlights the disproportionately large burden of HCV that black veterans bear. A smaller proportion of the VA HCV population is Hispanic (6%), and the remaining veterans are other races, multiple races, or unknown.
The HCV-infected veteran population is aging. The mean age of veterans with chronic HCV in 2013 was 59.7 years and for the first time, more veterans with HCV were aged 60 years (Figure 1).
Among the comorbidities that may have historically prevented veterans from receiving HCV antiviral therapy, 2 of the most pervasive are mental health conditions and alcohol use. The rates of mental illness among veterans overall is high, but mental illness is particularly high in veterans with HCV. Depression has affected 60%; of this population anxiety, 37%; posttraumatic stress disorder, 28%; and schizophrenia, 10%. Alcohol use disorders are also common among veterans with HCV in care. Active mental health conditions and substance use may affect medication adherence or follow-up visit adherence thereby limiting treatment candidacy. Integrating care of these individuals with mental health providers and substance-use treatment specialists is an important aspect of HCV care and is a priority in VA.
Three-quarters (76%) of the HCV-infected veteran population has been screened for HIV and HIV-HCV co-infection is present in 5733 (3%) of veterans with HCV. HIV-HCV co-infection is associated with an increased progression of liver disease and may have implications for the selection of HCV antiviral agents because of drug interactions. Hepatitis B virus (HBV)-HCV co-infection rates are higher at 7%. HBV vaccination or documentation of HBV immunity in those without HBV infection is 78%.
With regard to specific liver complications, 5% to 20% of those infected with chronic HCV will develop cirrhosis over a period of 20 to 30 years, and 1% to 5% will die of hepatocellular carcinoma (HCC) or cirrhosis.6 Given the natural history of chronic HCV and the aging HCV veteran cohort, increasing numbers of conditions related to progression of liver disease are expected over time. This is most evident in the number of veterans with a diagnosis of cirrhosis, which has increased from approximately 10,000 veterans (8%) in care in 2001 to nearly 30,000 veterans (17%) in care in 2013 (Figure 2).
Antiviral Therapy for Chronic HCV
Prior to mid-2011, the standard of care for HCV treatment was the combination of pegylated interferon and ribavirin. From 2011 through 2013, direct-acting antiviral (DAA) regimens containing boceprevir and telaprevir in combination with pegylated interferon and ribavirin became standard of care for genotype 1 while
the standard of care remained pegylated interferon and ribavirin for genotypes other than genotype 1. Recent advances in HCV antiviral therapy offer higher cure rates and fewer adverse events (AEs) compared with peginterferon-containing treatment. The expected ease and tolerability of these all-oral combination regimens is anticipated to greatly increase the number of veterans with HCV who could be treated successfully.
Because of the poor tolerability, prolonged treatment durations, serious AEs, and relative or absolute contraindications to peginterferon-based therapy, many veterans were not previously candidates for treatment. Of the 174,302 veterans with chronic HCV in care in 2013, 39,388 (23%) had received at least 1 course of HCV antiviral treatment. This largely reflects the time when peginterferon-based therapy was the standard of care. Since the approval of boceprevir and telaprevir 5,732 veterans (5.8%) in care in 2013 had ever received boceprevir or telaprevir-based regimens.
While recognizing that all veterans should be considered for HCV treatment, the urgency for treatment may be greater in those with advanced liver disease, because these patients are at the highest risk of developing decompensated cirrhosis or dying of liver-related disease. In 2013, there were 28,945 veterans in care that had advanced liver disease who might be considered potential HCV treatment candidates with an urgency to treat.
Duration of treatment and anticipated rates of treatment success with the all-oral regimens depend in part on a patient’s prior treatment status in addition to whether the patient has a diagnosis of advanced liver disease/cirrhosis. Regardless of HCV genotype, among all veterans approximately 85% are treatment-naïve and 15% are treatment-experienced. Advanced liver disease is present in 24% of treatment-naïve and 31% of treatment-experienced veterans with HCV genotype 1; 23% and 24% of veterans with HCV genotype 2, respectively; and 34% and 43% of veterans with HCV genotype 3, respectively.
Further understanding the population of veterans with HCV, including prior treatment status and stage of liver disease, is useful in identifying the target population for treatment. The VA uses these data to project treatment costs and assess capacity across the system in preparation for expected uptake of new regimens.
Sustained Virologic Response After HCV Antiviral Treatment
The goal of HCV antiviral therapy is to eradicate HCV and reduce the progression of liver disease and death from HCV infection. Successful antiviral treatment of HCV is determined by achieving a sustained virologic response (SVR) defined as an undetectable HCV viral load 12 weeks after the end of treatment. Of the 39,388 veterans in VA care in 2013 who have ever received antiviral therapy, SVR could be assessed in 32,815 veterans, and the overall SVR rate in this population was 42%. This SVR rate is similar to that observed in phase III trials of pegylated interferon-based regimens, where 42% to 46% of those infected with HCV genotype 1 achieved SVR.7,8 Although most veterans with genotype 1 infection received boceprevir or telaprevir-based regimens in 2013 and achieved higher SVRs of 50% to 52%, the overwhelming majority of veterans in care in 2013 received prior treatment with only peginterferon and ribavirin.9 Although SVR rates are expected to increase with newer all-oral HCV regimens, differences between clinical efficacy and real-world effectiveness will continue to be apparent,
and patient and provider expectations should be tempered accordingly.
The Role of Primary Care in HCV
Primary care providers have held the responsibility for multiple roles in HCV care since the discovery of the virus—particularly for HCV risk factor assessment, screening, and diagnosis. HCV antiviral treatment, however, was largely placed in the hands of specialists, given the complexities of patient selection, frequent reliance on a liver biopsy for determining need for treatment, and the toxicities of peginterferon and ribavirin therapy.
There are discussions both inside and outside the VA about potentially expanding the role of PCPs in HCV care. First, primary care is the major setting where the CDC and USPSTF recommendations for birth cohort screening are being implemented, and thus PCPs will be identifying veterans previously undiagnosed with HCV.1,2 Second, the ease and tolerability of the new all-oral combination regimens is causing a shift in the paradigm for HCV treatment, from a highly individualized approach, toward a more uniform approach.
Expanding the role of primary care would have multiple benefits to patients and the health care system as a whole. Only approximately 9% of HCV-infected veterans in VA care have been successfully treated at this time, largely due to low eligibility rates and the poor response rates, but other barriers have also contributed to the low success rate, one of which has been limited access to specialists. Furthermore, veterans who are referred to specialists are often noncompliant with the referral.10 If seeing an HCV specialist is required for treatment, the time to treat the HCV population will be much greater, more costly, and less efficient. Therefore, if the prospect of delivering HCV treatment to the majority of HCV patients is to be accomplished, it is necessary to consider providing treatment in the primary care setting as well as the specialist setting.
Treatment provided by nonspecialists has been evaluated in patients receiving peginterferon and ribavirin regimens and has shown that with adequate education and support, SVR rates were equivalent in the specialist and nonspecialist setting.11 To develop programs to provide primary care with such support, the VA has implemented the Specialty Care Access Network-Extension of Community Healthcare Outcomes program initiative, with casebased learning along with real-time consultation.
Currently, the majority of HCV-infected patients have never seen an HCV specialist, thus PCPs are already providing the majority of HCV care beyond HCV antiviral
treatment.12 Primary care providers are, therefore, key to addressing multiple important aspects of HCV care, including (1) counseling patients on transmission, prevention, lifestyle, and the role of substance use; (2) providing hepatitis A and B vaccination as well as appropriate general vaccinations for any patient with chronic liver disease; (3) modifying comorbidities that could accelerate fibrosis progression, such as diabetes mellitus, obesity and hyperlipidemia; (4) reducing risk from ongoing alcohol, drug, and tobacco use; (5) monitoring patients for fibrosis progression and identifying the presence of cirrhosis; and (6) providing general care for patients with cirrhosis, including HCC screening. These are critical aspects of HCV care, and many PCPs may still need additional education for these roles. The VA provides education and support for PCPs in their current role and is enhancing efforts to expand delivery of HCV treatment to the primary care setting as well.
Conclusions
In 2013, the typical veteran with chronic HCV was white, aged 60 years, and male, with a history of comorbidities, including hypertension, depression, and current or prior alcohol abuse. The proportion of veterans with advanced liver disease including cirrhosis (17%) and HCC (3%), has grown significantly over the past 10 years. By the end of 2013, almost 40,000 veterans had received antiviral therapy for HCV, more than 5,700 of whom had received DAAs. Overall SVR rates have been about 42% among those who were treated. Of veterans who are still potential treatment candidates, 85% are treatment-naive and about one-quarter have advanced liver disease.
Although HCV screening rates in veterans are higher than reported in other health care settings, particularly among those in the critical 1945-1965 birth cohort (65% screening rate), substantial numbers of veterans still require testing. The burden of disease, the lack of specialists, the ease and tolerability of new HCV antiviral medications, and the interplay of HCV with other traditional primary care efforts underly an increased role for PCPs in the care of veterans with HCV. Together, this information helps to construct a view of historical, current, and future HCV care in veterans.
Author disclosures
The authors report no actual or potential conflicts of interest with regard to this article.
Disclaimer
The opinions expressed herein are those of the authors and do not necessarily reflect those of Federal Practitioner, Frontline Medical Communications Inc., the U.S. Government, or any of its agencies. This article may discuss unlabeled or investigational use of certain drugs. Please review complete prescribing information for specific drugs or drug combinations—including indications, contraindications, warnings, and adverse effects—before administering pharmacologic therapy to patient.
1. Smith BD, Morgan RL, Beckett GA, et al; Centers for Disease Control and Prevention. Recommendations for the identification of chronic hepatitis C virus infection among persons born during 1945-1965 [published correction appears in MMWR Recomm Rep. 2012;61(43):886]. MMWR Recomm Rep. 2012;61(RR-4):1-32.
2. Moyer VA; U.S. Preventive Services Task Force. Screening for hepatitis C virus infection in adults: U.S. Preventive Services Task Force recommendation statement. Ann Intern Med. 2013;159(5):349-357.
3. Backus LI, Gavrilov S, Loomis TP, et al. Clinical Case Registries: Simultaneous local and national disease registries for population quality management. J Am Med Inform Assoc. 2009;16(6):775-783.
4. Kabiri M, Jazwinski AB, Roberts MS, Schaefer AJ, Chhatwal J. The changing burden of hepatitis C virus infection in the United States: Model-based predictions. Ann Intern Med. 2014;161(3):170-180
5. Denniston MM, Jiles RB, Drobeniuc J, et al. Chronic hepatitis C virus infection in the United States, National Health and Nutrition Examination Survey 2003 to 2010. Ann Intern Med. 2014;160(5):293-300
6. Davila JA, Morgan RO, Shaib Y, McGlynn KA, El-Serag HB. Hepatitis C infection and the increasing incidence of hepatocellular carcinoma: A population-based study. Gastroenterology. 2004;127(5):1372-1380.
7. Manns MP, McHutchison JG, Gordon SC, et al. Peginterferon alfa-2b plus ribavirin compared with interferon alfa-2b plus ribavirin for initial treatment of chronic hepatitis C: A randomised trial. Lancet. 2001;358(9286):958-965.
8. Fried MW, Shiffman ML, Reddy KR, et al. Peginterferon alfa-2a plus ribavirin for chronic hepatitis C virus infection. N Engl J Med. 2002;347(13):975-982.
9. Backus LI, Belperio PS, Shahoumian TA, Cheung R, Mole LA. Comparative effectiveness of the hepatitis C virus protease inhibitors boceprevir and telaprevir in a large U.S. cohort. Aliment Pharmacol Ther. 2014;39(1):93-103.
10. Brady CW, Coffman CJ, Provenzale D. Compliance with referral for hepatitis C evaluation among veterans. J Clin Gastroenterol. 2007;41(10):927-931.
11. Arora S, Thornton K, Murata G, et al. Outcomes of treatment for hepatitis C infection by primary care providers. N Engl J Med. 2011;364(23):2199-2207.
12. Holmberg SD, Spradling PR, Moorman AC, Denniston MM. Hepatitis C in the United States. N Engl J Med. 2013;368(20):1859-1861.
1. Smith BD, Morgan RL, Beckett GA, et al; Centers for Disease Control and Prevention. Recommendations for the identification of chronic hepatitis C virus infection among persons born during 1945-1965 [published correction appears in MMWR Recomm Rep. 2012;61(43):886]. MMWR Recomm Rep. 2012;61(RR-4):1-32.
2. Moyer VA; U.S. Preventive Services Task Force. Screening for hepatitis C virus infection in adults: U.S. Preventive Services Task Force recommendation statement. Ann Intern Med. 2013;159(5):349-357.
3. Backus LI, Gavrilov S, Loomis TP, et al. Clinical Case Registries: Simultaneous local and national disease registries for population quality management. J Am Med Inform Assoc. 2009;16(6):775-783.
4. Kabiri M, Jazwinski AB, Roberts MS, Schaefer AJ, Chhatwal J. The changing burden of hepatitis C virus infection in the United States: Model-based predictions. Ann Intern Med. 2014;161(3):170-180
5. Denniston MM, Jiles RB, Drobeniuc J, et al. Chronic hepatitis C virus infection in the United States, National Health and Nutrition Examination Survey 2003 to 2010. Ann Intern Med. 2014;160(5):293-300
6. Davila JA, Morgan RO, Shaib Y, McGlynn KA, El-Serag HB. Hepatitis C infection and the increasing incidence of hepatocellular carcinoma: A population-based study. Gastroenterology. 2004;127(5):1372-1380.
7. Manns MP, McHutchison JG, Gordon SC, et al. Peginterferon alfa-2b plus ribavirin compared with interferon alfa-2b plus ribavirin for initial treatment of chronic hepatitis C: A randomised trial. Lancet. 2001;358(9286):958-965.
8. Fried MW, Shiffman ML, Reddy KR, et al. Peginterferon alfa-2a plus ribavirin for chronic hepatitis C virus infection. N Engl J Med. 2002;347(13):975-982.
9. Backus LI, Belperio PS, Shahoumian TA, Cheung R, Mole LA. Comparative effectiveness of the hepatitis C virus protease inhibitors boceprevir and telaprevir in a large U.S. cohort. Aliment Pharmacol Ther. 2014;39(1):93-103.
10. Brady CW, Coffman CJ, Provenzale D. Compliance with referral for hepatitis C evaluation among veterans. J Clin Gastroenterol. 2007;41(10):927-931.
11. Arora S, Thornton K, Murata G, et al. Outcomes of treatment for hepatitis C infection by primary care providers. N Engl J Med. 2011;364(23):2199-2207.
12. Holmberg SD, Spradling PR, Moorman AC, Denniston MM. Hepatitis C in the United States. N Engl J Med. 2013;368(20):1859-1861.