Keywords adverse drug events - natural language processing - clinical decision support systems
Background and Significance
Background and Significance
Adverse drug reactions (ADRs) comprise a serious health care problem, causing substantial morbidity and mortality.[1 ]
[2 ]
[3 ] They generate preventable emergency department visits and hospital admissions, prolong hospital stays, and increase health care costs.[1 ]
[2 ]
[3 ]
[4 ]
[5 ]
[6 ]
[7 ] Providers often fail to recognize ADRs quickly enough or at all.[1 ]
[2 ] As new drugs become available and patients take more medications, providers will face increasing challenges in recognizing their patients' ADRs. This article describes the development and evaluation of the Adverse Drug Effect Recognizer (ADER). Authors hypothesized that ADER could help physicians recognize and address patients' ongoing preadmission ADRs.
Forty years ago, the World Health Organization[8 ] (WHO) defined ADR as “a response to a drug that is noxious and unintended and occurs at doses normally used in man for the prophylaxis, diagnosis or therapy of disease, or for modification of physiological function.” Edwards and Aronson[9 ] later defined an ADR as “an appreciably harmful or unpleasant reaction, … related to the use of a medicinal product, which predicts hazard from future administration and warrants prevention or specific treatment, or alteration of the dosage regimen, or withdrawal of the product.” They explained that adverse drug effect and ADR refer to the same concept, but while ADR is from the point of view of the patient, adverse drug effect is from the point of view of the drug.[9 ] The ADER system identifies electronic health record (EHR)-based patient findings that match known adverse effects of patients' preadmission medications. Only the patient's physician can determine whether matched findings actually represent true ADRs, since a patient's underlying disorder might alternatively have caused the finding. For clarity, the authors use the term adverse effect for known medication side effects. Authors use ADR for clinician-confirmed true ADRs. Correspondingly, authors refer to findings mentioned in the EHR that match known adverse effects as potential ADRs .
A 2008 systematic review of 25 studies found that, on average, 5% of hospital admissions were associated with ADRs.[5 ] A 2014 retrospective analysis of 7 million U.S. hospital discharge abstracts similarly found a combined prevalence of 5.6% for preadmission and inpatient ADRs.[10 ] The Centers for Disease Control and Prevention (CDC) noted that 82% of American adults take at least one medication, and 29% take five or more.[2 ] A 2014 study found those taking three or more medications were at higher risk for ADRs, with an increased risk for those taking more medications.[11 ] Additionally, the CDC estimated that ADRs result in more than one million emergency department visits and 280,00 hospitalizations in the United States annually.[1 ]
[2 ] Thus, ADRs generate substantial health care costs.[12 ]
[13 ]
[14 ]
[15 ]
[16 ] In 2006, the Institute of Medicine estimated that the U.S. spent $3.5 billion on extra medical costs due to ADRs, and that 40% of the costs of outpatient ADRs were preventable.[2 ]
Over the past three decades, investigators have developed automated approaches to ADR detection and prevention. Early systems identified ADRs that clinicians already had recognized (and presumably addressed). Those systems surveyed inpatient records for sudden discontinuations of medications, administrations of known antidotes, and billing codes indicating ADRs.[17 ]
[18 ] In the 1990s, University of Utah researchers implemented such a system; it detected significantly more verified ADRs than traditional reporting methods.[19 ]
[20 ]
[21 ] Bates et al later showed that automated ADR detection using diagnosis codes, allergy rules, and text analysis were effective, although with lower predictive values than intensive retrospective manual chart reviews.[22 ]
[23 ]
[24 ] Nevertheless, they also found that combining discharge summaries, laboratory results, and medication records could identify valuable ADR-related information.[25 ]
[26 ]
[27 ] Their text-searching tools revealed more potential ADRs than simply reviewing laboratory results or diagnostic codes.[23 ]
Since 2000, improved natural language processing (NLP) techniques have enabled researchers to better identify ADRs from clinical documents. For example, researchers from Columbia University applied NLP to discharge summaries and the biomedical literature to retrospectively detect potentially unrecognized ADRs and other disease–drug associations.[28 ]
[29 ]
[30 ]
[31 ]
[32 ] Researchers at Stanford University also utilized NLP to extract ADR information for pharmacovigilance studies.[33 ]
[34 ]
[35 ]
[36 ]
[37 ] The latter studies demonstrated the need to apply additional drug knowledge to distinguish ADRs from confounders such as indications for drug therapy.
Past work utilizing NLP in combination with data on known adverse effects forms the conceptual basis for the current ADER study. Previous studies[38 ]
[39 ] have validated NLP and text-mining methods for retrospectively identifying adverse effects from EHR notes. On admission, ADER automatically extracts inpatients' medications, symptoms, and previous diagnoses using NLP of provider-generated admission history and physical examination (H&P) notes. The system then cross-references these data against a set of known adverse effects and generates patient-specific alerts in the EHR. This article describes an intervention study to determine whether ADER could improve providers' recognition of ADRs. The current study is one of the first to evaluate a system that uses NLP to warn clinicians of potential ongoing ADRs at the time of hospital admission.
Objective
The goal of this project was development of the ADER system and preliminary evaluation of whether it could improve inpatient providers' recognition of ongoing ADRs due to outpatient medications. Authors hypothesized that physicians receiving ADER alerts would hold or discontinue suspected ADR-causing medications at a greater rate than a retrospective control group of similar physicians who did not receive alerts. The study assessed ADER's effect on ADR recognition by comparing admission medications to inpatient and discharge medication orders for patients in both the intervention and control groups.
Methods
The ADER system utilizes several previously validated and publicly available NLP tools. SecTag[40 ] identifies section headers in H&P notes; the KnowledgeMap Concept Identifier (KMCI)[41 ]
[42 ] extracts clinical concepts from text in a manner similar to MetaMap,[43 ] and MedEx[44 ] extracts medication names and dosage information from clinical texts. MedEx was originally trained on data similar to that used for the ADER evaluation.[44 ] A previous study confirmed that H&P notes at the authors' institution accurately capture patients' preadmission medications.[45 ] In the current study, KMCI and ADER used the 2013ab version of the Unified Medical Language System (UMLS); ADER and MedEx utilized the 3 February 2014 release of RxNorm.[46 ]
[47 ]
ADER Design
Upon each patient's hospital admission, clinicians perform a thorough evaluation recorded in the H&P notes. During the current study, immediately upon generation of a new H&P in the hospital EHR, ADER copied the note to its database for processing. First, ADER ran SecTag to parse sections of the note, such as Chief Complaint , History of Present Illness , Past Medical History , Medications , and Family Medical History . Next, ADER used KMCI to identify the patient's symptoms, findings, and previous diagnoses (diseases) mentioned in each section and mapped them to the Systematized Nomenclature of Medicine – Clinical Terms (SNOMED-CT) clinical vocabulary via UMLS.[48 ] Using an implementation of the NegEx algorithm,[49 ] KMCI identified whether any identified concepts were characterized as absent (e.g., “no fever,” “patient denied cough,” etc.). The ADER system ignores such negated items and also terms mentioned in sections unlikely to represent the patient's own findings (e.g., Family History as identified by SecTag). Additionally, ADER used regular expressions to extract numerical values from the H&P Vital Signs section, mapping abnormal findings to SNOMED-CT (e.g., hypotension for systolic blood pressure < 90). Similarly, ADER mapped abnormal admission of EHR laboratory results to SNOMED-CT concepts (e.g., hypokalemia for serum potassium < 3.2 mEq/L). Next, ADER utilized MedEx to identify the patient's current medications from the H&P's Medications section. It automatically mapped drugs to their generic ingredients and represents them using RxNorm.[47 ]
Finally, ADER cross-referenced the patient's recorded medications and clinical findings against a database of known adverse effects. Thus, ADER identified potential ongoing ADRs due to the patient's preadmission (outpatient) medications. The ADER adverse effect database derives from the authors' previous work on the Drug Evidence Base (DEB2), a knowledge base containing drug indications and adverse effects extracted from five public domain sources using automated methods.[50 ] See details in the “ADER Evaluation” section below.
A patient's underlying medical conditions can also potentially explain findings that ADER labels as potential ADRs. For example, a patient taking Lisinopril might have a cough due to chronic obstructive pulmonary disease, and not due to an ADR. Prior to the current study, the authors calculated the frequency of mentions of each potential ADR finding in a corpus of 350,000 historical H&P notes. Authors then discerned how often such mentions occurred in patients taking potentially ADR-causing medications, as well as how often such mentions were associated with patients' various discharge diagnoses when no ADR-causing medication was recorded. Based on those calculated relative risk scores, ADER determined whether a patient's current diagnoses plausibly might have caused a finding that ADER identified as a potential ADR. The system warned clinicians of such confounding diseases in its ADR alerts. [Fig. 1 ] provides an overview of ADER workflow.
Fig. 1 Complete Adverse Drug Effect Recognizer (ADER) workflow diagram. HPI, history of present illness; PE, physical exam; ROS, review of systems.
Within a few minutes of a clinician creating an H&P note, ADER identified any potential ADRs within the note and generated patient-specific alerts. Each alert displayed the potential ADRs, suspected causative medications, specific evidence extracted from the chart, and any possible confounding diseases. [Fig. 2 ] shows sample ADER alerts. Physicians could review the alerts in multiple places in Vanderbilt's EHR, “StarPanel”:[51 ] the EHR Team Summary (used by all care providers); in StarPanel EHR progress note template forms (used to create clinicians' daily progress notes); and, in frequently used, patient-specific, paper-based reports (used by physicians for daily rounds). As part of the current study, alerts included a brief embedded questionnaire to collect feedback from intervention providers.
Fig. 2 Several Adverse Drug Effect Recognizer (ADER) alerts from the study, including potential false positive alerts (renal insufficiency, cough, and edema are potential ADRs, but could also be caused by the patient's kidney failure, chronic obstructive pulmonary disease [COPD], and heart failure).
To prevent alert fatigue,[52 ] and also at the request of physicians contributing to study design, ADER used noninterruptive alert delivery mechanisms (i.e., its passive displays did not break EHR workflows). The system automatically removed alerts from the EHR when: (1) a provider responded to the alert questionnaire; (2) the patient was discharged; or (3) alert-related questionnaire went unanswered for 5 days.
ADER Evaluation
The ADER intervention study tested the hypothesis that ADER could help providers to better recognize and respond to inpatients' preadmission ongoing ADRs. The study focused on one common class of medications, antihypertensives, which have a myriad of known side effects. The complete set of antihypertensive adverse effects detectable by ADER appears in the Supplementary Material ([Supplementary Tables S1 ]
[S2 ]
[S3 ], available in the online version). A team of clinical pharmacists reviewed the antihypertensive adverse effects from the previously mentioned DEB2 database, making any necessary clinically relevant revisions prior to initiation of the study. After the target adverse effects had been determined, the authors utilized the aforementioned corpus of 350,000 previous-to-current-study H&Ps to refine KMCI's recognition of those specific effects. This process captured common misspellings, abbreviations, and lexical variants, which were used to improve KMCI recognition of them. To determine the accuracy of concepts recognized by ADER's NLP during the study, the authors analyzed a sample of 100 H&P notes where alerts had been issued. This manual audit attempted to verify that all ADER-identified clinical findings and medications mentioned in alerts were actually present in patient's records.
The Vanderbilt University Institutional Review Board (IRB) approved the study (IRB #141341). Indirectly, the study involved participating house officers' attending physicians and adult inpatients; they did not require informed consent because the intervention simply informed house officers of potential ADRs. The system did not directly change patients' medications. All treatment and medication decisions remained entirely with the care team.
The ADER intervention study began on August 1 and ended on October 31, 2015. The intervention group included consented interns and residents on the General Internal Medicine wards of the Vanderbilt University Hospital. The study length was based on pragmatic convenience—authors were uncertain prior to the study whether the number of alerts that ADER generated would cause alert fatigue. Therefore, consented participating residents and interns, as well as the Director of the Internal Medicine Residency program, agreed to a 3-month trial period as the maximum tolerable duration (in the event that warnings became onerous). The retrospective control group consisted of Internal Medicine interns and residents assigned to the same wards as intervention physicians during a 3-month period just prior to the activation of the system (April 1 through June 30, 2015). To enable comparison of intervention physicians' behaviors to that of controls, the authors ran ADER on the retrospective control group's previous H&P notes, identifying instances when alerts would have been issued had the system been active at the time.
For both intervention and control groups, the study analyzed all clinicians' prescribing behaviors using NLP-based analysis of admission notes, inpatient medication orders, and discharge summaries. This process determined whether preadmission medications from the H&P were continued, held or discontinued during the hospital stay and at discharge. The study compared medication hold/discontinue rates of all intervention clinicians who received alerts and all control physicians who would have received alerts had the system been active. The study also determined separately whether the subgroup of intervention providers who completed the alert-embedded survey (and therefore definitely had seen the alerts) were more likely to change suspected ADR-causing medications compared with all nonalerted control group providers. When evaluating discharge medications, the study only analyzed notes for those patients both admitted and discharged by study subjects (i.e., Internal Medicine house staff). Patients transferred to the care of physicians on another service (e.g., Surgery) were not included in analysis of discharge medications.
When comparing preadmission medications to inpatient medication records , the authors evaluated the difference between intervention and control groups for suspected ADR-causing medications (1) ordered ever during the inpatient stay, (2) ordered in the first 24 hours of the stay, (3) ordered, then stopped or held, as well as (4) suspected ADR-causing medications stopped or held after intervention providers responded to the alert survey versus the overall hold rate in control group. When comparing preadmission medications to discharge summaries , the authors evaluated the difference between intervention and control groups for suspected ADR-causing medications (5) held at discharge and (6) held at discharge for intervention providers who responded to the alert survey. The primary outcome was (1), the difference between groups for suspected ADR-causing medications ordered ever during the inpatient stay.
Whenever intervention subjects' survey responses indicated that they would make or had made changes to suspected ADR-causing medications, the study identified (by analyzing actual orders issued) whether the provider actually followed-through with the medication order changes. An original long version of the survey questionnaire had a low response rate (12%); subjects reported that it was cumbersome and had too many questions. Therefore, a shorter survey ([Fig. 2 ]) replaced the longer version approximately 3 weeks into the study. Only survey responses obtained during the 9 weeks using the shorter-form survey appear in the results. The authors reviewed open-ended survey comments submitted by intervention subjects and classified them based upon similar themes. Additionally, the authors met with study subjects, including the Chief Resident and Director of the Internal Medicine Residency Program, to discuss their experiences and suggestions at the completion of the intervention.
Statistical Analysis
The study characterized categorical variables using frequencies and proportions and used medians and interquartile range (IQR) to describe continuous variables. To determine whether ADER alerts affected providers' behaviors, authors analyzed medication order changes during the inpatient stay and at discharge. The analysis compared the rates of medication changes between the control and intervention groups using chi-square tests with a significance level of 0.05. When performing secondary analyses on data, the authors used Bonferroni correction to adjust the significance level for all comparisons to account for multiple testing. All statistical analyses used R version 2.15.2.
Results
[Table 1 ] illustrates ADER study results for both the intervention and retrospective control groups. The 3-month preintervention control group included 2,666 H&P notes written by 138 interns and residents. During the 3-month intervention period, the ADER scanned 2,312 H&P notes written by 137 interns and residents. No significant differences occurred between the intervention and control groups for any metrics shown in [Table 1 ] (p > 0.05). The ADER system detected potential ADRs in 30 to 32% of both groups' notes, with 78 to 79% triggered by H&P text findings and 21 to 22% by laboratory results. The median time from when an intervention group physician saved an admission note to the EHR to when ADER posted an alert was 8 minutes (IQR = 6–9 minutes). The median time between ADER issuing an alert and an intervention subject responding to the embedded survey questions was 24 hours (IQR = 17–37 hours).
Table 1
ADER study results for intervention and control groups
H&Ps
Unique patients
Notes with potential ADRs
Average number of antihypertensive medications per note
N
N
%
N
%
Notes without potential ADRs
Notes with potential ADRs
Intervention
2,312
2,049
89
701
30
0.73
2.4
Control
2,666
2,352
88
867
32
0.76
2.4
Potential ADRs per alert
Adverse effect source
Number of potential ADRs found per alert
H&P text
Labs
Min.
25%
Median
75%
Max
Intervention
2.3
79
21
1
1
2
3
15
Control
2.4
78
22
1
1
2
3
23
Admissions with discharge note
Unique patients with discharge note
Average length of stay
Number of providers
N
%
N
%
Total
With discharge
Intervention
1,835
79
1,646
90
115 h
137
137
Control
2,160
81
1,916
89
112 h
138
128
Abbreviations: ADER, Adverse Drug Effect Recognizer; ADR, adverse drug reaction; H&P, history and physical examination.
ADER NLP Accuracy
Manual expert review of 100 intervention group H&P notes for which ADER had issued potential ADR alerts revealed that 98% of the 157 medications that ADER identified as current therapies were correct. In three instances, ADER incorrectly identified drugs that the H&Ps described as previously discontinued. Of 136 total adverse effects (findings or diseases) that ADER identified from H&P notes, 116 (86%) were confirmed as “present” in the chart. Of the 20 incorrectly identified adverse effects, 15 were mentioned as being absent but were not recognized as “negated” by KMCI due to use of nonstandard wording. The remaining misidentified concepts were mentioned only as possible points of differential diagnosis or past inactive diagnoses. Of the 30 clinical findings the ADER recognized from laboratory results, 29 (97%) were accurate. The only incorrectly classified numerical laboratory result was “hyperkalemia” that had been noted in comments to be due to a hemolyzed sample.
Comparing Admission Medications to Inpatient Medication Orders
The intervention group ordered significantly fewer suspected ADR-causing preadmission medications at any point during the inpatient stay compared with the control group (47% vs. 58%, p < 0.001). Similarly, during the first 24 hours after admission, intervention physicians ordered only 28% of 1,140 suspected ADR-causing preadmission medications flagged by ADER; the control group had ordered 39% of 1,468 suspected ADR-causing preadmission medications during the same interval (28% vs. 39%, p < 0.001). See [Table 2 ].
Table 2
Preadmission medications compared with ordered inpatient medications, testing the null hypothesis that the proportion for the intervention group was equal to the proportion for the control group
Suspected ADR-causing medications
Ordered ever during inpatient stay
Ordered in first 24 hours
Ordered, then stopped or held
N
Proportion
N
Proportion
N
Proportion
Intervention
1,140
532
0.467
315
0.276
91
0.171
Control
1,468
846
0.576
570
0.39
125
0.148
Difference
(95% CI)
p -value
–0.109
(–0.149, –0.070)
p < 0.001
–0.114
(–0.149, –0.075)
p < 0.001
0.023
(–0.018, 0.065)
p = 0.28
Abbreviations: ADR, adverse drug reaction; CI, confidence interval.
Among suspected ADR-causing medications ordered during the inpatient stay, 84% (711/846) were ordered within 24 hours of admission in the control group; in the intervention group, 77% (410/532) were ordered within 24 hours of admission (p = 0.002). Suspected ADR-causing medications were stopped before patient discharge at rates of 14 and 16% for the control and intervention groups, respectively (p = 0.40). Among medications started after 24 hours into the inpatient stay, subsequent hold rates were 19 and 20% in the control and intervention groups, respectively (p = 1.00).
Among suspected ADR-causing preadmission medications ordered at any point in the inpatient stay, the rate at which they were subsequently held or discontinued did not vary significantly between groups (see [Table 2 ], p = 0.28). However, secondary analysis, as shown in [Table 3 ], revealed that the subset of intervention providers who responded to alert-embedded surveys—and therefore who definitely viewed the alerts—subsequently held or discontinued 25% (41/167) of the then-active suspected ADR-causing preadmission medications. This rate is significantly different than the aforementioned control group rate (25% vs. 15%, p = 0.003). In the intervention group, the median time between responding to the survey and subsequent medication changes was 25 hours (IQR = 13–66).
Table 3
Preadmission medications ordering during inpatient stay later stopped or held, testing the null hypothesis that the proportion for the intervention group after survey response was equal to the overall proportion for the control group
Suspected ADR-causing meds
N
N
Proportion
Intervention
Active order at time of ADER alert survey response
167
Stopped or held after survey response
41
0.246
Control
Ordered ever during the inpatient stay
846
Ordered ever, then stopped or held
125
0.148
Difference
(95% CI)
p -value
0.098
(0.025, 0.171)
p = 0.003
Abbreviations: ADER, Adverse Drug Effect Recognizer; ADR, adverse drug reaction; CI, confidence interval.
Comparing Admission Medications to Discharge Medications
Approximately 20% of intervention and control group discharge summaries were not generated by study subjects and thus were excluded from review. [Table 4 ] shows hold/discontinuation rates of suspected ADR-causing preadmission medications at discharge for the remaining 80% of discharge summaries.
Table 4
Preadmission medications compared with discharge medications testing the null hypothesis that the proportion for the intervention group was equal to the proportion for the control group
Suspected ADR-causing preadmission medications: full intervention group versus full control group
Suspected ADR-causing preadmission medications; intervention group survey respondents versus full control group
Total
Medication missing from discharge note
Held or discontinued at discharge (explicitly)
Total
Held or discontinued at discharge (explicitly)
N
N
Proportion
N
Proportion
N
N
Proportion
Intervention
927
104
0.112
184
0.198
258
97
0.376
Control
1,225
75
0.061
287
0.234
1,225
287
0.234
Difference
(95% CI)
p -value
0.051 (0.025, 0.076)
p < 0.001
–0.036 (–0.072, 0.0001)
p = 0.053
0.142
(0.076, 0.208)
p < 0.001
Abbreviations: ADR, adverse drug reaction; CI, confidence interval.
When discharge notes did not mention previously alerted preadmission medications, the study could not definitively determine whether the medications had been held or stopped at discharge (104/927 = 11% from intervention, 75/1,225 = 6.1% from control, p < 0.001). Actual specific mentions in discharge notes of holds/discontinuations for alerted medications did not differ significantly: 20% (184/927) in the intervention group and 23% (287/1225) in the control group (p = 0.053). However, intervention providers who responded to the alert-embedded survey held or discontinued 38% (97/258) of suspected ADR-causing medications at discharge, significantly different (p < 0.001) than the overall 23% rate for the control group.
Survey Questionnaire Results and Provider Comments
Intervention subjects responded to 53% (298/560) of the short-version alert-embedded questionnaires (see [Table 5 ]). For survey respondents who stated that they would change the patient's therapy due to identified ADRs, subsequent medication holds or dosage decreases occurred more than 90% of the time. For the 38 intervention subjects who responded that they were uncertain whether they would make a change, analysis identified 24 instances (63%) where one or more suspected ADR-causing medications was subsequently changed. Nearly three-quarters of survey responses indicated alerts either somewhat or fully merited consideration for their patients. More than half said alerts were helpful or possibly helpful in managing the patient.
Table 5
Responses to ADER alert survey questionnaires (n = 298)
Question
Response
N
(%)
N
(%)
N
(%)
Did any of these alerts merit at least passing consideration for this patient?
Yes
144
(48)
Somewhat
74
(24)
No
80
(26)
Were any of these alerts helpful in managing this patient?
Yes
87
(29)
Possibly
80
(26)
No
131
(43)
Will you (or did you already) change patient's therapy due to these possible ADRs?
Yes
91
(30)
Uncertain
38
(12)
No
169
(56)
Abbreviations: ADER, Adverse Drug Effect Recognizer; ADR, adverse drug reaction.
Approximately 30% of the 52 optional free-text survey comments indicated that ADER was correct in identifying an ADR. Conversely, 30% of the survey comments stated that claimed adverse effects were more likely due the patient's underlying medical conditions (usually those recognized by ADER as possible confounders). For example, when ADER identified renal insufficiency or acute kidney injury as an ADR, providers often indicated that the patient's existing end-stage renal disease was the more likely cause of impaired renal function.
In ∼20% of survey comments, providers indicated that the medications were held, but not due to the ADER-recognized ADRs. In at least two cases, however, the H&P Plan section indicated that a provider had in fact held the medications for the reasons identified by ADER. In 10% of the comments, providers indicated that the patients were not suffering from the suspected ADR. In one of those cases, ADER had recognized “gout” as a possible ADR. The patient had a previous diagnosis of gout, but no acute gout exacerbation was present on admission. In another case, ADER “incorrectly” identified that the patient had a cough because of the phrase, “Hurts with movement, cough, sneezing, position changes.” Several survey respondents suggested that they had already recognized the ADR suggested by ADER.
Study subjects also provided feedback to the authors directly. Participating providers generally viewed ADER favorably, but suggested that ADER might be more useful for those medications that internists prescribe less frequently than antihypertensives, such as psychiatric medications. They also expressed that potential confounders included in ADER alerts were helpful in assessing the veracity of suspected ADRs and in considering whether medications were possibly exacerbating the recognized findings.
Medication Holds Analyzed by ADR
Among the 940 intervention group alerts, 41 were distinct adverse effects (patient findings/disorders). The most frequent alert-related conditions (occurring more than 50 times) were: renal insufficiency, light headedness, edema, dizziness, hypotension, and syncope. Suspected causative medications were held at discharge at rates of 35% for hypotension, 28% for renal insufficiency, 42% for hyperkalemia, 36% for hypokalemia, and 50% for hyponatremia. The Supplementary Material ([Supplementary Tables S1 ]
[S2 ]
[S3 ], available in the online version) provides hold/discontinue rates for all medications with detected ADRs. The online Supplementary Material also reports comparisons of hold rates between suspected ADR-causing medications and those not suspected of causing ADRs.
Discussion
The current study illustrates the potential for automated alerting systems like ADER to help providers identify unrecognized ADRs. Physicians receiving ADER alerts held or discontinued more suspected ADR-causing medications at multiple points in the patient care workflow compared with physicians who did not receive alerts. In survey responses, physicians stated that ADER alerts more than half the time helped with patient management. Analysis of ADER's accuracy revealed that the majority EHR-extracted findings and medications were correct, but also revealed opportunities for improvements in NLP, particularly in regards to negation detection.
Intervention providers receiving ADER alerts ordered significantly fewer suspected ADR-causing preadmission medications in the first 24 hours after admission and at any time during the inpatient stay. As a whole, intervention providers receiving alerts were no more likely to discontinue suspected ADR-causing preadmission medications at discharge than the control group (although they had ordered fewer such medications overall at admission). Nevertheless, providers who responded to the alert-embedded survey were more likely to hold or discontinue ADR-causing medications during the patient's hospitalization and at discharge. Two explanations are possible—either providers who believed that a potential ADR alert was valid were more likely to respond to the survey, or providers who believed ADER alerts were true more often addressed potential ADR-causing medications at discharge. Further study could determine whether ADER alerts would be more effective if they interrupted workflows, since the current study used noninterruptive alerts. Why intervention physicians in some instances ordered previously alerted ADR-causing medications later during an admission is also a topic for further research. A possible explanation is that when ADR-related findings resolved, re-starting the medication was appropriate (e.g., hypotension in patients with baseline hypertension).
From a technical standpoint, the ADER approach is generalizable to many inpatient facilities. The system only requires that a hospital have an EHR system that can electronically export (under appropriate security) admission H&P notes and laboratory values as they are generated to a generic desktop-class computer running the ADER software. Another key requirement is the hospital EHR's ability to incorporate ADER-generated alerts into the EHR workflow in locations likely to be viewed by clinicians.
Comparison with Prior Research
Use of NLP to identify potential ongoing ADRs from outpatient medications at the time of admission is novel. A substantial body of work exists regarding other methods of ADR detection and associated determinations of physicians' responses. In 2017, Dexheimer et al reported that providers in a pediatric hospital, after receiving inpatient dosing, allergy, and drug–drug interaction alerts cancelled or modified medication orders an average of 12.6, 7.0, and 16.6%, respectively.[53 ] A 2003 study of five adult primary care practices found that physicians accepted only 8.8% of drug allergy and 10.6% of high-severity drug–drug interaction alerts.[54 ] Further analysis revealed physicians were more likely to override alerts for renewals than new prescriptions and that house officers were less likely to prescribe alerted medications than other physicians. A different study suggested that house staff were more likely to override ADR alerts than staff physicians.[55 ] In a classic 1976 study, McDonald evaluated provider responses to electronic reminders, including suggestions to observe a physical finding or ask about a specific symptom, order a diagnostic study or change or start a therapeutic regimen.[56 ] The study found physicians responded to 51% of events when given electronic reminders compared with 22% of events when they did not receive reminders. Of note, few such reminders involved medication changes due to potential ADRs. The study indicated that physicians comply with a higher proportion of electronic recommendations when protocols are more precise.
The current study's physician alert response rates exceed the aforementioned previous studies regarding other types of ADR alerts. A possible explanation is that the current study named in its alerts the suspected ADR symptoms that were present in the patient at admission. The ADER methodology is significantly different than that of the above-mentioned efforts, as well as the adverse effect detection studies discussed in the background. The previous efforts utilized sudden discontinuations of medications, administrations of known antidotes, and ADR-related billing codes, all of which indicate that providers had already recognized the ADRs. They also warned about dosing errors, allergies, and possible drug–drug interactions, which often do not take into account the actual symptoms being experienced by the patient. The ADER system is novel in that it utilizes NLP of provider-generated admission notes to warn clinicians about potential ongoing ADRs, as evidenced by documented symptoms and findings, due to outpatient medications. It has the potential to detect unrecognized ADRs and provide a safeguard to ensure all providers are aware of the ADRs their patients may be experiencing.
Limitations and Future Work
Several limitations pertain to the study. The intervention only included Internal Medicine interns and residents as alert recipients. That survey responses followed alerts by ∼24 hours suggests that house officers may have waited to discuss alerts with the ward team and attending physicians before acting. Additionally, house staff often rotate “off service” so that the physicians writing discharge orders may not have been the ones who received the initial ADER alerts at the time of admission, even though the discharging physician qualified as study participants by membership in the Internal Medicine house staff. A wider range of subjects (e.g., academic attending physicians, community-based physicians, nurse practitioners, physician assistants) should be evaluated in teaching and nonteaching settings. The study limited the ADER intervention to antihypertensive medications so that potential alert fatigue would not confound initial evaluation of ADER's utility—as might occur if all categories of medications generated alerts. The study targeted the commonly prescribed antihypertensive class because they cause a wide variety of ADRs. Admittedly, providers who regularly prescribe these medications are more likely to be aware of their potential ADRs, and study participants suggested that ADER might be more useful for medications they prescribe less frequently. In the future, ADER should be expanded, in consultation with clinical experts, to cover additional drug classes and their related ADRs, with particular focus on uncommon or severe ADRs.
The study analyzed 3 months of intervention data compared with a similar-duration retrospective control group. While analysis showed the control group to be comparable to the intervention group in nearly all respects, the groups contained different participants and the admissions occurred at somewhat different times in the academic year (controls were soon-to-be promoted house officers; intervention subjects had arrived in their current positions 1 month prior to study initiation). A randomized controlled trial involving multiple institutions and all drug classes could provide a more definitive result than the current study. Before undertaking such a study, the potentially correctable flaws in ADER's NLP methods (identified in this study) should be addressed.
To evaluate the potential of this methodology, this study focused only on the accuracy of the ADER alerts that were presented. Future studies with additional drug classes should consider both precision and recall to determine which potential ADRs ADER misses (i.e., fails to generate alerts when appropriate to warn clinicians). To reduce alert fatigue, more advanced confounder detection should be studied with the goal of filtering out false-positive alerts when potential ADRs have a significantly more likely clinical cause. Future studies will also include more in-depth qualitative analysis of providers' impressions and resultant behavior.
Conclusion
Intervention subjects reduced ordering of suspected ADR-causing medications at multiple points during the admission compared with the control group. Systems like ADER have the potential to improve both recognition of adverse effects and discontinuation of medications causing preadmission ADRs. Future ADER-like systems must supply relevant information to care providers at the right place and time to improve patient safety, potentially decrease cost, and improve the quality of care.
Clinical Relevance Statement
Clinical Relevance Statement
This study shows that using NLP extraction of patient preadmission medications and findings to generate real-time ADR alerts can affect provider ordering behavior. This approach could be generalized to other facilities that use electronic health records. Clinical decision support systems like ADER could help to improve the safety of care delivery in the future.
Multiple Choice Questions
Multiple Choice Questions
Approximately what percentage of ADER-identified adverse drug reactions were found in the text of the admission note, as compared with solely derived from laboratory test results?
How often did intervention clinicians receiving an alert order the potentially offending medication during the hospital stay, and how often did control physicians do so?