Keywords
decision support - clinical informatics - health information
Background and Significance
Background and Significance
Patient-centered clinical decision support (CDS) has the potential to contribute substantially
to the quadruple aim of advancing the health of populations, enhancing patient experience,
reducing costs, and caring for caregivers. Nonetheless, CDS systems have variable
effectiveness at improving adherence to evidence-based practices that lead to better
outcomes.[1]
[2]
[3]
[4]
[5]
[6]
[7] The “five rights” CDS framework,[8] GUIDES checklist,[9] and similar tools[10] can help identify success factors of CDS systems that are associated with effective
interventions. To incorporate these success factors reliably into the creation and
maintenance of CDS systems, health care systems must coordinate CDS design, measurement,
and governance. Many organizations lack the health information technology (IT) resources
to consistently support high-quality CDS. This gap has led to calls for a CDS maturity
model to guide organizational investment in functions and processes that would help
health care systems use CDS more effectively.[11]
Maturity models can help drive development by providing a roadmap to organizations
for more effective use of health IT.[12] For example, Knosp et al demonstrate the use of a maturity index and a deployment
index to assess academic health centers' capacity to provide research IT services
and determine organizational investment priorities.[13] Multiple health care IT maturity models exist that concentrate on functional capacities
related to a specific technology or a set of tasks[14]
[15]
[16]
[17] or an organization's acceptance of and regular use of a set of methodologies.[18]
[19] To our knowledge, no standardized maturity model yet exists for health care organizations'
CDS capabilities.[11]
CDS is defined as a process for enhancing health-related decisions and actions with
pertinent, organized clinical knowledge and patient information to improve health
and health care delivery.[8] Nearly all CDS processes aim to change behavior in complex environments where there
are multiple interacting components and stakeholders at different organizational levels.[20] Developing organizational competence in any one specific technology or methodology
is likely inadequate to understand processes and intervene effectively in such complex
settings. Thus, a maturity model that focuses uniquely on characteristics of organizational
use of methodologies or specific functional capacities may not capture the interactions
between these elements to promote CDS effectiveness.
In this article, we took the perspective of an organization trying to determine how
to maximize the impact of their operational investment in CDS. We aimed to describe
a roadmap toward mature organizational functions and processes to help health care
systems use CDS more effectively to drive better outcomes.
Objectives
After reading this article, the reader should be able to:
-
Identify the functions of a CDS operational team, its key stakeholders, and its interactions
with other organizational entities.
-
Understand the uses of a maturity model to promote organizational development and
discuss applicability of a proposed CDS maturity model to health care organizations.
-
Describe key components of a proposed CDS maturity model, the organizational benefits
of maturing along each pillar, and advantages of developing capacities in tandem across
the components.
Model Development
We developed a maturity model through notes taken from conversations with CDS stakeholders
at U.S. health care organizations in the context of discussing interest in CDS analytics
and governance. Of note, these conversations were not intended as rigorous qualitative
research, but rather to synthesize lessons learned about effective CDS operations
and gaps that could be addressed. Therefore, conversations were not recorded, although
notes were taken and reviewed with participants for accuracy and appropriateness (member-checking).[21]
Stakeholder Outreach
We identified CDS team members through internet searches for health care organization
chief medical information officers and reviewed online profiles of staff with job
titles or descriptions of “head of informatics,” “knowledge manager,” “knowledge engineer,”
or any mention of “clinical decision support.” We did not have explicit criteria such
as years of CDS experience to determine who to contact. We contacted these individuals
via e-mail to discuss CDS analytics and governance to inform potential software tools
and scheduled webinars (or in-person meetings when practical) with interested parties.
Additional CDS team members were identified via snowball sampling.[22] Discussions were generally held with multiple people from a single institution at
the same time in group settings.
We arranged in-person or webinar meetings with CDS team members from a total of 80
organizations ([Table 1]). Each meeting was grossly structured in three phases: (1) current assessment of
the organizations' CDS governance structure, capabilities, and practices (as defined
by the stakeholders); (2) discussion of the organizations' goals and needs to further
develop CDS capacity; (3) demonstration and feedback on a software application developed
by several authors (E.W.O., N.M., D.F.F., M.D.Z., and M.C.T.) for CDS analytics and
governance in use at their home institution at the time (Children's Hospital of Philadelphia,
CHOP). We did not use a set list of questions but, similar to grounded theory,[21] adjusted questions over time, based on learnings from prior conversations, with
the goal of synthesizing organizational needs to create and maintain high-quality
CDS.
Table 1
Roles of CDS team members participating in discussion
|
Role[a]
|
N (%)
|
|
Chief medical information officer (CMIO)
|
49 (39)
|
|
Physician informaticist (not CMIO)
|
47 (37)
|
|
Nonphysician informaticist
|
11 (9)
|
|
Knowledge manager/engineer
|
7 (5)
|
|
Other
|
13 (10)
|
Abbreviation: CDS, clinical decision support.
a Each individual categorized into a single role based on online profiles.
Model Synthesis
After reviewing notes from the 80 meetings, five authors (four clinical informaticists
[E.W.O., N.M., D.F.F., and M.C.T.] and one business officer [M.D.Z.]) iteratively
developed the proposed CDS maturity model. First, M.D.Z. and M.C.T. synthesized conversations
into a candidate model. This candidate model was then reviewed with E.W.O., N.M.,
and D.F.F. while reviewing notes from approximately five meetings at a time. After
each meeting, M.C.T. and M.D.Z. adjusted the candidate model until all meeting notes
had been reviewed by these five authors with no suggested changes. After achieving
agreement, the proposed model was reviewed with 19 additional organizations and presented
with case examples at a regional informatics meeting (Penn Healthcare IT Roundtable)
and two national informatics meetings (AMIA Annual Symposium 2018[23] and AMIA Clinical Informatics Conference 2019)[24] with feedback from each presentation reviewed by the same five authors. We ceased
iterations when feedback from three consecutive organizations did not result in any
changes to the model, which was achieved after review with the first six organizations
(and maintained for the subsequent 13 organizations).
Proposed Maturity Model
The proposed CDS maturity model ([Fig. 1]) describes a roadmap of interdependent organizational functions and processes to
help health care systems use CDS to drive better outcomes. Three “pillars” focus on
distinct domains of an organization's approach to CDS: “Content Creation,” “Analytics
and Reporting,” and “Governance and Management.” The “roof” represents the advanced
CDS functions that become attainable after progressing along each of the three pillars.
The initial state for all three pillars is “ad hoc”—for example, in content creation,
ad hoc would indicate that a clinician asks for a particular CDS artifact (e.g., alert,
order set, documentation template), and it is created as per their wishes without
further effort to understand the problem or care process. Advancing along each pillar
gives the CDS team a greater understanding of the problem to facilitate improved CDS
processes.
Fig. 1 Clinical decision support operations maturity model.
Of note, the organizational effort required to advance one step is likely different
for each point in the model. An organization can advance in one pillar separately
from another—for example, an organization could develop advanced CDS utilization analytical
reports that associate CDS artifact use with outcomes, but may not have the governance
in place to engage providers to use those data in designing and optimizing CDS processes.
Similarly, an organization may develop a new capacity such as user-centered design
but not yet apply it systematically or comprehensively across all CDS artifacts or
processes. Advancing in tandem along the columns can confer additional advantages;
for example, a robust analytics platform can yield greater benefits when coupled with
interdisciplinary engagement, monitoring and maintenance processes, and user-centered
design.
Additionally, organizations frequently advance within an individual pillar in a different
sequence than described in the model. While this practice may be appropriate for specific
CDS initiatives, we hypothesize that going in order increases the value of organizational
investment in higher levels of CDS maturity.
Content Creation
The “Content Creation” pillar describes advancement from a focus on CDS tools to a
more comprehensive understanding of the work system[25]
[26] to inform CDS design ([Table 2]). In the first three levels, specific CDS artifacts are built with increasing level
of sophistication—initially to satisfy the ideas of individual clinicians (e.g., through
personal order sets), then with expert review, and then to promote evidence-based
practices endorsed by applicable guidelines. In level 4, CDS teams attempt to gain
a greater understanding of work as done (not only work as imagined) by systematically
incorporating front-line user feedback regarding use of the CDS artifact in context.[27]
[28] In level 5, decision support processes are designed based on more formal analysis
of the users and their tasks.[29] Additionally, rather than obtaining feedback by showing users a CDS artifact and
simply asking their opinion, front-line users are asked to simulate their work. Data
on their performance informs adjustments to the design.
Table 2
Content creation pillar of the proposed CDS maturity model
|
Level
|
Capability
|
Description
|
|
5
|
User-centered design
|
Design based on user- and task-analysis with scenario-based testing of prototypes
performed systematically for high-risk CDS prior to implementation
|
|
4
|
Front-line user feedback
|
Feedback from front-line users on each CDS artifact is systematically solicited, reviewed,
and incorporated
|
|
3
|
Guideline/literature based
|
Guidelines and academic literature are consistently reviewed and incorporated
|
|
2
|
Expert-based
|
Building based on local experts' consensus
|
|
1
|
Ad hoc build
|
Human and technical resources to build and customize CDS when needed
|
Abbreviation: CDS, clinical decision support.
Analytics and Reporting
The “Analytics and Reporting” pillar describes advancement from resource-intensive
data requests to more rapid and more useful evaluation of CDS process and outcome
measures, and finally to automating maintenance functions. In levels 2 to 4, CDS teams
are able to answer more sophisticated questions through self-service without having
to enter a data request and wait on organizational resources. Level 2 represents the
ability for improvement advocates in the organization to rapidly determine what CDS is being used without a dedicated data request. Level 3 represents the ability
to rapidly determine how CDS is being used as well as the stakeholders to contact to determine why it is being used a particular way. Level 4 moves from examination of the CDS artifacts
alone to the intended goals of the decision support (e.g., process or outcome metrics).
Level 5 invokes automated processes to identify anomalies in the use of CDS or process
or outcome metrics ([Table 3]).[30]
[31]
[32]
[33]
Table 3
Analytics and Reporting pillar of the proposed CDS maturity model
|
Level
|
Capability
|
Description
|
|
5
|
Anomaly detection
|
Automatically monitor data to identify broken, malfunctioning, or suboptimal CDS
|
|
4
|
Outcome association
|
Systematically associate clinical measures with CDS performance
|
|
3
|
Data exploration and visualization
|
User is easily able to manipulate CDS data to move from understanding what is happening to why it is happening
|
|
2
|
Self-service access
|
Basic, descriptive reports are available on-demand for consumers
|
|
1
|
Ad hoc requests
|
Consumers of CDS data receive reports from another person or team after request
|
Abbreviation: CDS, clinical decision support.
Governance and Management
The “Governance and Management” pillar describes increasing engagement of organizational
stakeholders with the CDS team to accomplish organizational goals. Increasing from
level 1 to level 4 represents more regular review of CDS processes. Additionally,
wider representation and engagement with the CDS team makes it easier to identify
“owners” of a CDS process who have the appropriate clinical expertise and connections
to ensure adequate review. This interdisciplinary approach also helps avoid unintended
consequences across a variety of clinical workflows. At level 5, strategic alignment,
the organization understands that CDS processes are critical to accomplishing goals
related to the quadruple aim, leading to increased investment in people, processes,
and technology to ensure CDS team involvement in all key organizational initiatives
([Table 4]).
Table 4
Governance and Management pillar of the proposed CDS maturity model
|
Level
|
Capability
|
Description
|
|
5
|
Strategic alignment
|
The CDS team is leveraged throughout the organization as a vital partner in helping
reach organizational strategic goals
|
|
4
|
Monitoring and maintenance
|
Standardized review and change recommendation processes for all existing CDS
|
|
3
|
Interdisciplinary
engagement
|
Cross-department representation in governance processes
|
|
2
|
Governance committee
|
Defined governance group and stakeholders that have established intake processes;
CDS team as “gatekeepers”
|
|
1
|
Ad hoc governance
|
Governance occurs only as needed (e.g., for regulatory or safety events)
|
Abbreviation: CDS, clinical decision support.
Advanced Capabilities
Achieving advanced levels across all three pillars creates new opportunities to use
CDS to advance care. For example, in addition to automated anomaly detection, advanced
analytics can also uncover opportunities for improvement with CDS by identifying variation
in care practices or through interorganizational benchmarking. However, to take advantage
of these insights, the organization must also have advanced governance processes to
evaluate these opportunities and adequate resources to create high-quality CDS content.
Achieving all of these milestones creates a “learning health system for CDS,” in which
the organization is capable of continuously generating, interpreting, and acting on
CDS data to promote evidence-based practices in pursuit of the quadruple aim. Of note,
this “learning health system for CDS” is not equivalent to the global aim of The Learning
Healthcare System from the Institute of Medicine which also integrates knowledge discovery,
innovation, patient engagement, and other domains which are traditionally not the
subject of CDS alone ([Table 5]).[34]
Table 5
Advanced capabilities of the proposed CDS maturity model
|
Capability
|
Description
|
|
Learning health system
|
Continuous integration of user input, clinical data, and organizational knowledge
into CDS
|
|
Interorganizational benchmarking and sharing
|
Ability to compare CDS metrics to different organizations—both for specific processes
(e.g., sepsis order set usage) and for system-wide metrics (e.g., alert burden)
|
|
Proactive opportunity discovery
|
Automated identification of potential applications for CDS based on variation in care
as well as cross-department and/or cross-institution benchmarking
|
Abbreviation: CDS, clinical decision support.
Case Studies: Maturity Model in Practice
Case Studies: Maturity Model in Practice
Vanderbilt University Medical Center
Vanderbilt University Medical Center (VUMC) is one of the largest academic medical
centers in the Southeast, managing more than 2 million patient visits each year. In
November 2017, VUMC fully implemented a new electronic health record (EHR; Epic Systems)
throughout all inpatient hospitals and outpatient clinics. CDS in the form of evidence-based
order sets and alerts facilitates guideline-concordant care across the care continuum.
Robust content creation and governance processes are in place to support lifecycle
management of all deployed CDS artifacts. For CDS content creation, VUMC leverages
user-centered design principles (level 5 of the “Content Creation” pillar) through
an iterative approach that involves end-user feedback, driven by Osheroff's five “rights”
for CDS.[8] Specifically, design goals for the CDS interventions include simple presentation
of relevant data with recommended actions, optimal workflow integration via appropriate
triggers minimizing unnecessary alerting, and delivering evidence-based guidance according
to VUMC's content governance framework. Utilizing user-centered design principles,
VUMC recently deployed CDS to implement pneumonia management guidelines in the pediatric
emergency department (PED). The CDS team conducted observations of clinician users
(e.g., PED faculty, fellows, residents, and nurse practitioners) investigating clinical
workflows, decision-making processes, current use of technology and artifacts, and
communication patterns. Structured cognitive interviews identified factors affecting
decisions, care goals, and contingencies and exceptions as well as strengths and limitations
of proposed care algorithms. These data informed key user-interface design requirements
for several CDS candidates. Subsequently, the CDS team conducted a usability evaluation
of the candidate user-interface designs by presenting PED providers with prototypes
(i.e., formative usability testing) to refine design elements. The CDS design that
scored the highest among all participating providers was selected for further refinements
and eventual implementation. The final version consisted of three streamlined data
input forms to confirm (or obtain) CDS algorithm inputs (patient clinical variables)
that triggered the recommended actions—laboratory tests and antibiotic orders.
Despite following best practices for CDS design, analysis of usage data suggests that
the adoption rates for some CDS artifacts remain low. Ongoing assessment of users'
compliance with CDS recommendations and rapid identification of the root cause for
low rates of adoption is vital for sustainable CDS. In addition to employing user-centered
design principles for initial design, VUMC utilizes targeted postimplementation user
feedback (level 4 of “Content Creation” pillar) informed by periodic review of CDS
usage data on providers. In addition, postdeployment interviews were conducted with
PED providers to determine reasons for certain user behaviors and not using the CDS
during certain situations as well as modifications and solutions to resolve the related
issues. Extending the user-centered CDS design principles to include ongoing evaluation
of CDS has been an integral part of CDS lifecycle management and ongoing CDS optimization.
Children's Healthcare of Atlanta
Children's Healthcare of Atlanta (CHOA) is the largest pediatric health system in
Georgia, managing over 1 million patient visits per year as of 2018. The organization's
EHR (Epic Systems) is integrated throughout the enterprise and achieved the Healthcare
Information and Management Systems Society (HIMSS) EMRAM Acute Care Stage 7 in 2016.
CHOA has developed 90 clinical practice guidelines to standardize care and improve
quality, with all of these supported by CDS artifacts such as order sets, information
resources, and/or alerts. To facilitate plan, do, study, act cycles, CHOA has developed
a self-service analytics application using QlikView that allows users to quickly view
a set of 21 quality outcomes (e.g., length of stay, readmission rate, % intensive
care unit [ICU] transfers) for populations meeting guideline criteria, and to filter
by whether the intended order set for that population was activated. Thus, CHOA has
the ability to associate CDS with outcomes (level 4 of the “Analytics and Reporting”
pillar). For example, in mid-2017, the CDS team discovered that patients diagnosed
with croup were noted to have a brief elevation in ICU transfer rates while use of
the order set was decreasing to approximately 50% of encounters meeting criteria for
the croup guideline. These data informed educational interventions to increase awareness
of the order set and its association with better outcomes, which led to increased
utilization of up to 75 to 80%. Over the same time period, ICU transfers and readmission
within 7 days both decreased.
While the capacity to associate CDS use with outcomes has yielded some successes,
the analytics tools at CHOA do not allow for more detailed exploration and visualization
of CDS utilization data. For example, among patient encounters meeting criteria for
the inpatient uncomplicated pneumonia guideline in 2018, only 23% had the pneumonia
order set used, limiting progress toward promoting narrow-spectrum antibiotics. Determining
how users selected orders within the order set or any rationale behind choices not
to follow a guideline would require a separate data request. Similarly, there is no
systematic mechanism for obtaining and incorporating front-line user feedback, slowing
the investigation into why adherence might be low to design more effective processes. Thus, future efforts will
aim to solidify CDS data exploration and visualization capabilities (level 3 of the
“Analytics and Reporting” pillar) as well as front-line user feedback (level 4 of
the “Content Creation” pillar) to more rapidly iterate on CDS design to achieve its
intended outcomes.
Children's Hospital of Philadelphia
CHOP is the community hospital and primary care hub for children in West and South
Philadelphia and a major tertiary referral center for the Greater Delaware Valley
area. Annually, CHOP has nearly 30,000 inpatient admissions, more than 81,000 emergency
room visits, and approximately 1.24 million outpatient visits. CHOP has been able
to develop its CDS governance practices over multiple years after implementing its
current EHR (Epic Systems) in the ambulatory care network in 2000 and in the inpatient
and emergent care settings in 2011. A centralized governance structure (level 2 of
the “Governance and Management” pillar) was established around the time of system
roll-out, with biweekly meeting of a CDS governance committee with multidisciplinary
representation (level 3) from pharmacy, nursing, physicians, and information services.
Initially, this centralized governance primarily served to mitigate unintended consequences,
especially with substantial new build at the time of system implementation. Engaging
a broad group of interested stakeholders, however, also provided the opportunity to
develop a deep pool of informatics expertise. For example, there are currently over
20 physicians at CHOP who are board certified in clinical informatics.
Once this pool of experts was available, the clinical informatics leadership was able
to partner and align CDS efforts with other strategic initiatives in the organization
(level 5 of “Governance and Management”). The development of a clinical pathways program
was paired with the creation of order sets and other decision support tools to achieve
standardized care. This led to further investment in clinical informatics resources
by quality-improvement leadership, and the dissemination of quality-improvement best
practices to clinical informatics. In recent work, the CHOP Sepsis Quality Improvement
program aimed to reduce time to antimicrobial administration for patients with suspected
sepsis in the pediatric intensive care unit (PICU). PICU sepsis leadership partnered
with quality improvement and clinical informatics teams to identify this cohort of
patients in real time and, when “stat” ordering was not used, prompt the clinician
to change the ordering priority to “stat.” Clinical informatics helped refine design
of an EHR alert. Analysis after the intervention using statistical process control
charts demonstrated special cause variation and 20% center-line shift improvements
in “stat” ordering and time to antimicrobial administration.
CHOP only recently established system-wide review processes for CDS tools (level 4
of the “Governance and Management” pillar). While this illustrates the ability for
any given organization to move out of sequence in the maturity model, the review process
also demonstrated opportunities for improvement that could have allowed for more optimal
engagement with quality improvement and patient safety (level 5) had it been in place
earlier.
Discussion
We developed a maturity model to guide organizations' development of data-driven CDS
processes to improve population health, enhance patient experience, reduce costs,
and care for caregivers. Informed by discussions with 80 organizations, the model
provides a structure to assess current capacity in the domains of CDS content creation,
analytics, and governance. Of note, these domains are focused on organizational capacities
that support the creation and maintenance of high-quality CDS, rather than success
factors for a specific CDS implementation. Within each of these domains, we describe
a natural progression to help organizations identify a “zone of proximal development”[35] where investment of resources could yield immediate benefits in addition to building
CDS infrastructure. Advancing in tandem across domains confers additional benefits,
and a “roof” represents advanced capabilities that become possible through achievement
of high levels of maturity across the three pillars. At the pinnacle of the model,
organizations create a “learning health system for CDS” in which CDS designs are continuously
adjusted based on a thorough understanding of the work system and the impact of the
CDS on outcomes of interest.
The model as defined suggests that capabilities must be achieved serially and that
each capability is either implemented or not. However, organizations may have different
levels of deployment[13] for each capability described in the model—for example, the ability to visualize
and explore EHR use patterns for alerts but not order sets. Additionally, as described
in the case studies, many organizations develop “higher” capabilities prior to “lower”
ones or work to develop multiple areas simultaneously. Nonetheless, the imposed order
is intended to reflect that going in order maximizes the benefit of higher levels.
For example, if an organization has the ability to do anomaly detection but cannot
associate CDS artifacts with outcomes, they would not gain the ability to monitor
for unexpected changes in the outcome (for example, due to changes in diagnosis codes).[36] Similarly, if an organization has optimized the process of incorporating front-line
user feedback but does not base CDS processes on guidelines and literature, the resulting
system may reflect what many clinicians want but may diverge from evidence or established
best practice, leading to poorer patient outcomes. Intentional progression through
these stages can limit unnecessary disruption to existing processes while still transforming
the organization positively.[37]
To our knowledge, no previous studies have systematically assessed organizations'
outlook on CDS maturity or synthesized insights into a CDS maturity model to support
CDS operations. Many studies have evaluated the impact of specific CDS content[1] or characterized organizational variability in use and effectiveness of specific
CDS tool types such as order sets[38] and computerized reminders or alerts.[39] Numerous models and textbooks describe IT governance and management strategies,[40] but these models have not been adapted for health care knowledge management. Health
Catalyst has developed a maturity model for health care analytics, but the model is
intended for general use across the health care enterprise and not specific to CDS
systems.[41]
While this maturity model was informed by discussions with CDS team members from 80
organizations, we did not employ rigorous methods for qualitative research such as
transcribing interviews or independent identification of themes.[21] Similarly, while we obtained feedback on the proposed model from 19 organizations
as well as participants in a regional and national health informatics meeting, we
did not use validated methods such as consensus mapping[42] or Delphi methods.[43] The proposed model may therefore be skewed toward a specific vision of operationally
focused, data-driven CDS that may differ from many health care organizations' goals.
Conclusion
Effective CDS has the potential to transform health care and improve patient, population,
and caregiver outcomes at reduced costs. However, achieving these aims requires operational
coordination of CDS content creation, analytics, and governance. The proposed CDS
maturity model can help organizations assess their current capacity and guide investment
in CDS capabilities to accomplish organizational goals. Future efforts will include
creation and validation of assessment tools that help individual health care organizations
identify CDS improvement opportunities and drive policy that facilitates effective
CDS implementation.
Clinical Relevance Statement
Clinical Relevance Statement
To leverage the potential of CDS to improve outcomes, health care organizations must
develop operational capacities to create high-quality decision support processes,
streamline data analytics, and develop robust governance strategies. However, informatics
teams with limited resources must be intentional when investing time and resources
into building CDS capacity. We propose a maturity model for CDS operations based on
discussion with 80 organizations that can help CDS teams assess their current capacity,
create a roadmap for development, and identify attainable next steps that yield immediate
benefits while building CDS infrastructure.
Multiple Choice Questions
Multiple Choice Questions
-
Anomaly detection is an important component of a robust clinical decision support
(CDS) analytics group because:
-
It pushes otherwise hidden insights to the CDS team.
-
It identifies patients that have the best outcomes.
-
It allows a CDS team to visualize their data.
-
It aggregates the data into a single location.
Correct Answer: The correct answer is option a. Modern CDS systems provide a large amount of performance
log data. CDS analytics typically follow a “pull” approach where an analyst manually
retrieves pertinent insight that they are seeking. Anomaly detection enables insights
around malfunctions or other significant changes to be readily available to the CDS
team without needing to explicitly seek it out.
-
A hospital system creates order sets by convening discussions between groups of attending
clinicians in which they explain their preferences for a given order set, the order
set is built as described by the clinical decision support (CDS) team, and the order
set is reviewed once more by a representative clinician before going to production.
What stage of CDS maturity in the Content Creation domain does this represent?
Correct Answer: The correct answer is option b. In this example, the hospital system's CDS content
creation is based on discussion between clinical experts. This is an improvement over
ad hoc build (in which anyone in the organization can ask for an order set and it
is built to their specification) because it requires groups of similar clinicians
to come to consensus on their workflow. The system as described does not enforce use
of guidelines or evidence from the literature to inform order-set design. Similarly,
this system does not incorporate feedback from actual users of the order set after
it has been deployed to adjust the design. Finally, no formal user or task analysis
was performed, nor was there any scenario-based testing of users' interaction with
the order set to inform design.
-
A health care organization assigns clinical owners to each order set and alert in
their system and provides those owners with data on use patterns. A critical care
physician notices that out of 150 items in their critical care order set, only 50
have been used at least once in the last year. The physician requests removal of the
extraneous items to produce a leaner, more relevant order set. What stage of CDS maturity
in the Governance and Management domain does this represent?
Correct Answer: The correct answer is option d. In this example, the health care organization has
provided clinical owners of decision support artifacts with data allowing them to
monitor how those artifacts are used and make adjustments. In addition to interdisciplinary
engagement from naming clinical owners, empowering owners with data simplifies the
monitoring and maintenance process, allowing clinical owners to adjust the design
of the artifacts based on use patterns. This example does not yet demonstrate strategic
alignment as the CDS team was not tied directly to organizational strategic goals
in this example.
-
Which of the following examples demonstrates the use of outcome association to inform
clinical decision support design?
-
An organization reports central-line associated bloodstream infection (CLABSI) rates
to its executive board. When CLABSI rates increase, the board sends e-mails to all
clinical leaders asking them to inform nurses and providers to follow hospital policy
when caring for patients with central lines.
-
A critical care physician sees that, of the 150 items in their admission order set,
only 50 have been used in the last year. The physician requests that extraneous items
be removed, producing a leaner, easier-to-use order set that front-line users prefer.
-
An organization's anomaly detection algorithm notifies the clinical decision support
(CDS) team that a venous thromboembolism prophylaxis alert, which used to fire on
average 30 times per day, is now firing 220 times per day. The CDS team investigates
and sees that a change in codes is leading the alert to fire inappropriately. The
team adjusts the codes and alert frequency returns to the previous level.
-
An organization notices that among patients admitted for community-acquired pneumonia,
those in whom a pneumonia order set was used were no more likely to receive narrow-spectrum
antibiotics compared with patients where the specified order set was not used. After
changing defaults, admissions in which the order set was used achieve higher rates
of narrow-spectrum antibiotic use.
Correct Answer: The correct answer is option d. In example (d), the clinical decision support (CDS)
team is monitoring the outcome of interest for which the CDS was developed—namely
to increase use of narrow-spectrum antibiotics for community-acquired pneumonia. By
associating CDS use with this outcome, the organization is able to determine that
the CDS is not leading to the desired outcome, take action, and produce an improvement.
Example (a) involves reporting of an outcome, but there is no association with CDS.
Example (b) demonstrates use of self-service access, data exploration and visualization,
and monitoring and maintenance—however, there is no clinical outcome that is used
to inform the design of the order set. Example (c) demonstrates use of anomaly detection,
but again no clinical outcome is used to inform the design.