Abstract
Existing reviews have identified factors influencing Clinical Decision Support (CDS) adoption by clinicians in practice but overlook the dynamic and evolving nature of technology and users’ needs over time. This review aimed to identify factors that influence early, mid-term, and sustained acceptance and use of CDS in hospital settings. Five databases were searched from 2007 to January 2024 and 67 papers were included. Factors were extracted and synthesised according to the time that data were collected following CDS implementation. Factors relating to the CDS intervention (e.g. utility) and inner setting (e.g. fit with workflows) were reported across all time periods. Perceived outcomes were more often identified in the first year of use, and individual factors after the first 6 months of use. Strategies to work around CDS limitations were reported 5 years after implementation. Our review provides guidance for developing, implementing, and supporting ongoing use of CDS systems.
Similar content being viewed by others
Introduction
Clinical Decision Support (CDS) systems offer many opportunities to improve patient care in hospitals1. However, the impact of CDS on workflows and clinical outcomes is generally reported to be low in practice2,3,4. Clinicians’ uptake of CDS, an essential step to realising these outcomes, was recently reported to be just 34.2% in a meta-analysis conducted across 60 CDS study arms5.
Existing reviews have identified the factors that influence CDS success in depth, providing insight into why some CDS systems are more likely to be used by clinicians than others6,7,8,9,10. Factors commonly reported include the systems’ usefulness and ease of use, its fit with existing workflows, and the provision of resources to support users7,10. However, existing evidence syntheses have conceptualised these factors in a static, cross-sectional nature that assumes they remain equally relevant from clinicians’ initial uptake of CDS through to routine, sustained use. This assumption has been challenged in several studies. For example, one study found clinicians’ perceptions corresponded to their level of exposure to CDS11 and another found different issues were relevant to clinicians at different points in time following implementation of a system containing decision support features12. Taken together with theories of technology adoption, such as the diffusion of innovations theory13,14 and normalisation process theory15, that describe the temporal nature of embedding complex interventions into routine practice, evidence indicates that CDS use is likely to be a dynamic process where user needs unfold and change over time.
Understanding factors that influence acceptance and use of CDS across the system lifecycle would allow for the deployment of targeted, adaptive, and relevant strategies to anticipate user needs, encouraging both initial uptake and sustained use over time5. However, this has not yet been systematically examined. The current study aimed to address this gap by systematically reviewing the literature to identify factors that influence early, mid-term, and sustained acceptance and use of CDS in hospital settings.
Results
A flowchart of the search strategy, selection process and exclusions is presented in Fig. 1. Out of 67 studies included in the review, 23 studies (34%) contained entirely relevant results and all results were extracted from these studies for analysis. Forty four studies (66%) contained partially relevant results, i.e. some results met inclusion criteria and were extracted for analysis, while other results were excluded from analysis. Common reasons for excluding results are detailed in Supplementary Table 2.
Study characteristics and methods
Study characteristics and methods are presented in Supplementary Table 4. Over half of the studies were conducted in the United States (n = 38), with remaining studies conducted across 17 other countries (Supplementary Table 4). Methods used to evaluate clinicians’ acceptance and use of CDS included surveys and questionnaires (52%), interviews (33%), analysis of system data (19%), focus groups (12%) and observations (7%). Twelve studies used two or more methods at the same point in time, with interviews and observations being the most common pairing (n = 5).
Interventions
CDS interventions were heterogeneous, with interruptive alerts targeting medication management being the most common form of CDS evaluated (n = 14) (see Table 1). Forty-five studies evaluated knowledge-based CDS systems (i.e. guideline, rule or algorithm based), 13 studied non-knowledge (i.e. AI or ML based systems), 2 evaluated CDS with knowledge and non-knowledge-based components and 7 studies did not report CDS to this level of detail (Table 1).
Quality assessment
Quality assessment scores for each study are provided in Supplementary Table 5. Using the Mixed Methods Appraisal Tool (MMAT), 32 studies met all 5 quality criteria (48%), 9 met 4 criteria (13%), 10 met 3 criteria (15%), 10 met 2 criteria (15%), 5 met only 1 criterion (7%) and 1 met no criteria (1%). All studies were retained for analysis.
Time following CDS implementation
CDS were evaluated from 0 months (i.e., immediately following CDS implementation), up to 18 years post-implementation (see Supplementary Table 4). In 5 papers, results were reported within multiple timeframes following CDS implementation due to multiple methods being used or methods being repeated at different time-points. These were included in the review as separate “study time-points”. There were three instances where multiple papers reported findings from the same CDS implementation within the same timeframe following implementation. These six papers were consolidated into three study time-points. This resulted in 70 separate study time-points included in the review. The timeframe containing the highest number of study time-points was the first 6 months following CDS implementation (n = 33, 47%).
Factors influencing acceptance and use over time
A total of 132 unique factors were identified and mapped across six Consolidated Framework for Implementation Research (CFIR) domains (Table 2). Overall, factors within the intervention and inner setting domains were the most frequently reported across studies. Thirty-three unique intervention factors (total n = 247; 44% of factors) and 45 unique inner setting factors (total n = 164; 29% of factors) were identified. Fewer factors were identified relating to outcomes (n = 83, 15%), individuals (n = 69, 12%), process (n = 21, 4%), and outer setting (n = 6, 1%) domains.
As shown in Fig. 2, factors relating to the intervention were prominent in studies conducted at all timeframes following implementation, while those within the inner setting were more often reported after 1 year. Though less frequently reported, factors relating to outcomes (a new domain, not previously in CFIR) were predominantly reported in studies conducted in the first year following CDS implementation, whereas factors relating to individuals were more often reported after 6 months post-implementation.
m months, y years, n number of study time-points. The proportion of factors identified in each CFIR domain are presented relative to the total number of factors identified within each timeframe. Barriers, facilitators, and moderators were counted once per study time-point (see factor count calculation in Supplementary Table 3) and summed across all study time-points within each timeframe.
Figure 3 shows the key constructs reported in each timeframe. Relative advantage (n = 60), design quality and packaging (n = 48), task and work context (a new construct, not previously in CFIR, n = 43) and compatibility (n = 41) contained the most barriers, facilitators and moderators reported in study timepoints across timeframes. In Table 3, examples of specific barriers and facilitators identified within these key constructs are presented over time.
Key constructs presented in this figure were identified as barriers, facilitators or moderators in over 25% of study time-points within a given timeframe. The proportion % and number () of study time-points where a construct was identified as a barrier (B), facilitator (F) or moderators (M) to CDS acceptance and use, relative to the total number of study timepoints identified within a given timeframe, are presented. For example, complexity appeared as a barrier in 4 study time-points conducted between 0–6 months, representing 14% of the total 28 study time-points included in this timeframe. Colour saturation was based on the proportion that constructs were reported within each timeframe (i.e. lighter = lower proportion, darker = higher proportion), with red gradients representing barriers, green representing facilitators and blue representing moderating factors. Constructs were counted once per study time-point (see construct count calculation in Supplementary Table 3) and summed across all study time-points within each timeframe.
0–6 months following CDS implementation
Twenty-eight studies were conducted in the first 6 months following CDS implementation16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43. Four of these studies16,19,30,31 evaluated acceptance and/or use at multiple months post-implementation, yielding 33 study time-points within this timeframe (see Table 1). Factors identified (n = 245) most often related to the intervention, followed by the inner setting, outcomes, individuals, process and outer setting, as shown in Fig. 2. Supplementary Fig. 1A, B shows the proportion of barriers and facilitators identified in each domain over monthly intervals. Both barriers and facilitators in the intervention domain remained most frequently reported at most monthly intervals, though those relating to outcomes trended upward across this timeframe.
The proportion of barriers and facilitators identified in each domain across all timeframes are presented in Fig. 4a, b. In the first 6 months post-implementation, barriers (n = 80) within the intervention and inner setting domains were more frequent relative to other domains (Supplementary Fig. 1A). In contrast, facilitators (n = 146) within the intervention and outcomes domains were more often identified (Supplementary Fig. 1B). Examples of common barriers and facilitators identified during the first 6 months following CDS implementation are presented in Supplementary Table 6.
m months, y years, n number of study time-points. a The proportion of barriers identified in each CFIR domain are presented relative to the total number of barriers identified within each timeframe. Barriers were counted once per study time-point (see factor count calculation in Supplementary Table 3) and summed across all study timepoints within each timeframe. b The proportion of facilitators identified in each CFIR domain are presented relative to the total number of facilitators identified within each timeframe. Facilitators were counted once per study time-point (see factor count calculation in Supplementary Table 3) and summed across all study timepoints within each timeframe.
7–12 months following CDS implementation
Thirteen studies, evaluating 12 unique systems, were conducted between 7–12 months following implementation, resulting in 12 study time-points (see Table 1)44,45,46,47,48,49,50,51,52,53,54,55,56. Factors (n = 137) relating to the intervention remained most prevalent, despite decreasing from the previous timeframe. Those relating to individuals increased, whereas those in the inner setting domain decreased and those in outcomes, process and outer setting domains remained relatively stable (Fig. 2).
Barriers (n = 55) were most frequently identified within intervention and inner setting domains, though those relating to individuals and outcomes were also prevalent in studies conducted during this time (Fig. 4a). Despite decreasing in incidence from 0–6 months post-implementation, facilitators (n = 78) in the intervention domain remained the most reported (Fig. 4b). Facilitators relating to outcomes and individuals were also frequently reported.
1–2 years following CDS implementation
Ten studies were conducted between 1–2 years following CDS implementation57,58,59,60,61,62,63,64,65,66. Factors (identified n = 90) remained most frequent in the intervention domain, closely followed by the inner setting, increasing from the previous timeframe. Those reported in individuals and outcomes domains decreased, whereas those in process and outer setting domains remained low (Fig. 2).
Barriers (n = 46) within the intervention and inner setting domains remained the most frequently identified during this timeframe, followed by those in the individuals domain (Fig. 4a). Facilitators (n = 33) identified in the intervention and inner setting domains were most prevalent, both increasing in incidence from the previous timeframe, while those relating to outcomes and individuals decreased (Fig. 4b).
2–5 years following CDS implementation
Six studies were conducted between 2–5 years following CDS implementation, with 5 study time-points identified (see Table 1)66,67,68,69,70,71. One study66 reported factors between 1–2 years, and between 2–5 years. Factors (n = 41) within the inner setting domain slightly increased, followed by those within the intervention domain which slightly decreased (Fig. 2). Within the individuals domain, the incidence of factors increased, whereas those within the outcomes domain decreased. There were no factors identified within process or outer setting domains.
Barriers (n = 18) were most frequently identified within the intervention domain and inner setting despite decreasing from the previous timeframe, while the incidence of barriers in the individuals domain tripled (Fig. 4a). Facilitators (n = 14) within the intervention and individuals domains were most frequently identified, both increasing in incidence during this timeframe (Fig. 4b). Those relating to the inner setting and outcomes domains however, decreased.
Over 5 years following CDS implementation
Eleven studies were conducted 5 or more years following CDS implementation, with 10 study time-points identified (see Table 1)72,73,74,75,76,77,78,79,80,81,82. Factors (n = 51) relating to the intervention were most frequently reported, rising from the previous timeframe (Fig. 2). This was followed by factors relating to the inner setting, and individuals, which decreased from the previous timeframe, and those relating to outcomes, which increased. Consistent with the previous timeframe, no factors relating to the process or outer setting were identified.
Barriers (n = 20) within the intervention domain were most frequently identified during this timeframe, followed by those related to the inner setting. Those relating to individuals and outcomes domains were low (Fig. 4a). Facilitators (n = 19) were most frequently identified in the intervention domain, decreasing from the previous timeframe. This was followed by the inner setting and individuals, remaining relatively stable compared to the previous timeframe, and outcomes, which increased (Fig. 4b).
Moderating factors influencing CDS acceptance and use
Moderating factors (n = 55) were more often identified in studies conducted over a year following CDS implementation (between 50–60% of studies conducted during this time) (Fig. 3). Moderators were primarily identified in the ‘task and work context’, including the clinical user role e.g. nurses typically held more favourable views than doctors18,36 and junior doctors typically held more favourable views than senior doctors68. The department, unit or ward, the patient/population and type of shift where CDS was used e.g. whether clinicians were on call, also influenced acceptance and use (Table 2).
Discussion
We conducted a systematic review of studies that reported factors influencing clinicians’ acceptance and use of CDS systems following their implementation in hospital settings. Our findings align with previous reviews of studies evaluating clinicians’ experiences of CDS, which highlight technological and organisational factors, such as usefulness, usability and fit with workflows, as key issues6,7,8,9,10. To our knowledge, this is the first review to collate and synthesise factors according to the point in time they were reported post-implementation. In doing so, we expand on previous work by identifying important themes relating to clinicians’ acceptance and use of CDS over time. We discuss these themes in the context of existing research and outline their implications for design, implementation, and evaluation of CDS systems, and reporting of future research.
Firstly, our synthesis suggests that certain barriers can arise early post-implementation and may continue to be experienced if not actively addressed at an early stage83. We found barriers relating to the system and inner setting, including poor design quality and packaging, relative advantage, and compatibility between the CDS and existing workflows, were frequently reported across all timeframes (Figs. 3 and 4a). Specific barriers within these constructs, such as poor integration and interoperability between CDS and the EMR, fit with workflows, and system performance, were reported in studies conducted early after implementation and up to 2 years post (Table 3). Identifying and addressing these barriers soon after they emerge may improve the likelihood of uptake and sustained use.
Some barriers may be experienced more prominently in the immediate period following CDS implementation and resolve as clinicians become increasingly familiar with the system, as the system is adapted to meet local needs, and as clinicians develop strategies to overcome system limitations over time84. We found issues in the intervention, inner setting, and outcomes domains, such as limited transparency of CDS, a lack of resources to support CDS use, and reduced efficiency, to be primary concerns in the first 6 months post-implementation (Table 3). However, in the following timeframe (7–12 months), there was an increase in facilitators, and decrease in barriers, relating to clinicians’ skills to use CDS appropriately. Additionally, workarounds to overcome system limitations were rarely identified until later phases of use. These findings align with existing research evaluating computerised provider order entry (CPOE) systems, where clinicians’ inexperience and unfamiliarity contributed to an increase in barriers and errors reported in the immediate post-implementation period12,85. Our review, like other studies12,86, suggests users develop ways to work around system limitations over time. Though workarounds can be positive, allowing for users to overcome design and workflow inefficiencies, they may also increase the risk of errors occurring86. Thus, enhancing CDS design early may minimise workarounds87, and providing clinicians with enhanced support, such as ongoing training and information sessions, may be helpful to overcome challenges associated with a lack of familiarity during the early phases of use56.
Our review also suggests that clinicians’ ability to recognise certain barriers and outcomes increases over time as they become more experienced with CDS. We found concerns about the accuracy of data inputs driving CDS recommendations to be rarely identified until 7–12 months post-implementation, coinciding with an increase in users’ understanding of the system. For example, in Lichtner et al.70, clinicians became more ‘watchful’ of automated behaviour with increased use. Similarly, the prevalence of outcomes identified in our review increased over the first 6 months following implementation (Supplementary Fig. 1A, B) and up to 1-year post-implementation (Figs. 2 and 3). We found some negative outcomes, including delays in care and new system-related errors to be reported only in studies conducted between 7–12 months after implementation. These findings echo existing research evaluating CPOE systems over time12,84.
Additionally, changes to the work system and associated context may impact issues experienced over time. In Salwei et al.53 and Campion et al.73, changes made to related clinical information systems (CIS) resulted in the disruption of CDS workflows at 1 year, and over 5 years, following CDS implementation. Building on recommendations from previous reviews, these findings exemplify the importance of engaging clinicians not only during CDS development, but on an ongoing basis to identify and address both expected and unanticipated issues that may arise over time6. Furthermore, engaging users prior to changing existing, or deploying new, CIS systems may help to uncover potential workflow impacts to systems already in use. Despite this, no factors related to the implementation process were identified beyond 2 years post-implementation, indicating that strategies such as user feedback and system monitoring to address persistent or late-emerging barriers are rarely utilised long-term.
Staffing changes, such as the rotation of clinicians, and new users of CDS are inevitable and likely to impact how CDS is accepted and used over time13. However, only one study reported the impact of the organisations’ transitory workforce53 and no studies investigated how new users adopted existing CDS. The lack of factors relating to available resources identified in later timeframes suggests limited training and education opportunities to support later adopters. Similarly, there were very few studies that reported factors relating to the outer setting environment, with none discussing regulatory or clinical guideline changes that could affect CDS acceptance and use over time. Such topics warrant future research.
Results showed that outcomes may become less visible to clinicians over a prolonged period of time. Interestingly, both positive and negative outcomes were rarely identified in papers beyond 1-year post-implementation (Fig. 2). Such findings could reflect the process of ‘normalisation’, in which a system becomes increasingly integrated into routine practice and consequently ‘disappears from view’15. Changes to the CDS system however, may spark new benefits realisation. A slight peak in the prevalence of positive outcomes reported in studies conducted beyond 5 years after implementation coincided with intervention factors commonly reported during this time, such as ongoing design needs and modifications (Fig. 4b and Table 3). Our findings suggest that those looking to evaluate perceived benefits of CDS systems, should do so within the first year following implementation, before CDS becomes normalised. Future studies should also explicitly explore how perceived outcomes of CDS change over time and whether and how perceived CDS outcomes are sustained, given the lack of outcomes identified in later phases of use may reflect a lack of long-term benefits evaluation in existing studies.
Lastly, clinicians appeared to be able to better understand advantages and limitations of CDS with increased use. Factors relating to clinicians’ self-efficacy to use CDS increased over time, with studies conducted between 2–5 years post-implementation often reporting that clinicians combined their clinical judgement, intuition and experience with CDS recommendations, and rejected recommendations where CDS did not align (Fig. 3 and Table 3)67,68,70. This finding is particularly interesting, given increasing concerns of automation bias leading to over-reliance on CDS88. While a few studies reported that clinicians were concerned about the ‘potential’ for over-reliance on CDS21,51,56, findings from the review support the theory that clinicians can more accurately consider limitations with increased use of and familiarity with the system over time88.
Almost half the studies included in this review were conducted during the first 6 months following implementation (28/67) and fewer factors were identified in studies conducted in later timeframes, particularly >2 years post-implementation (92/556 factors). This indicates a need for further research to comprehensively explore the factors driving sustainable use of CDS systems. Furthermore, there were limited studies that evaluated acceptance and use of CDS at different points in time within a single study. Studies that reported findings at multiple points in time, did so as a consequence of employing multiple methods of evaluation, rather than purposefully exploring changes over time. Though a few studies identified in our search explored clinicians’ acceptance of CDS over time, these studies did not provide the point-in-time that CDS systems were evaluated11 or reported findings on a broader CIS implementation (i.e. did not report findings related specifically to CDS systems)12,84,89, and thus were excluded from this review.
Though the CFIR provided a useful lens to systematically consider factors related to CDS acceptance and use, we identified additional factors that did not fit within the existing framework. These included the quality of data inputs and outputs of the CDS system in the intervention domain, the task and work context in the inner setting domain, and perceived outcomes, which should be considered when evaluating acceptance and use of CDS systems. Importantly, while our review suggests that factors influencing clinicians’ acceptance and use CDS systems can change over time, 196 papers that would have otherwise been included did not report the time following implementation that evaluation was completed and were excluded from the review. We therefore urge future studies examining acceptance and use of CDS and other digital health interventions to report the time of data collection in relation to implementation. We also recommend that research reporting guidelines be updated to make reporting of time between implementation and evaluation of interventions a requirement.
A key limitation of this review was the between-studies design and heterogeneity of included studies. Thus, factors identified within timeframes may have been influenced by differences in CDS systems, users, settings, methods used, and specific focus of studies conducted at each point in time. Further research employing longitudinal, within studies designs are required to confirm and expand upon the findings laid out in this review. Additionally, as the majority of included studies were conducted in the US and other high-income countries, findings may have limited generalisability to other settings, particularly developing countries.
While we only included studies that specified the time of data collection following implementation, ‘implementation’ may have been interpreted and reported inconsistently between studies. For example, CDS systems may be implemented in a limited capacity before full implementation whereas others may be implemented using a ‘big bang’ approach90. Likewise, CDS systems are often updated and adapted following their initial ‘go-live’ date89, but this detail is rarely provided in publications. This indicates a need for future research that evaluates what and how changes, such as adaptations to CDS and new users, can impact clinicians’ acceptance and use of CDS over time.
Our review provides practical guidance to assist stakeholders in anticipating and identifying issues likely to impact CDS acceptance and use over time. We emphasise the importance of engaging clinicians early after implementation, and on an ongoing basis, to ensure issues that develop over time are promptly and successfully addressed. We must move away from episodic evaluations of clinicians’ acceptance and use of CDS systems and towards a framework that considers the complexity of factors, including how they emerge, interact, and change over time. Doing so will allow for more efficient and nuanced approaches that target the issues clinicians experience at different points in time, increasing the likelihood of sustained system success. Reporting the time of data collection post-implementation and employing longitudinal designs in future research is necessary to achieve this goal.
Methods
This systematic review is reported following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines91. The protocol for this review is registered in PROSPERO (CRD42022325469). Two related reviews, each with minor variations in study inclusion criteria, have been previously published as conference papers. These papers aimed to review the methods used to evaluate clinicians’ acceptance and use of CDS over time92, and the use of approaches to involve clinicians in CDS design on post-implementation acceptance and use93, which differed from the aims of the current review.
Search strategy
Ovid MEDLINE, Embase, Web of Science, CINAHL and PsycINFO were systematically searched on 17 March 2022, with an additional search conducted on 19 January 2024, to identify studies reporting clinicians’ acceptance and use of CDS following implementation in hospital settings. Our search was restricted to studies published within 15 years of the initial search date (i.e. from January 2007) to ensure studies reflected current CDS systems and organisational environments. A professional librarian was consulted in the development of the search strategy (see Supplementary Note 1). A manual search of reference lists of relevant reviews was also conducted.
Inclusion and exclusion criteria
CDS systems were defined as electronic systems that aim to enhance clinical decisions with targeted clinical knowledge and patient information to support individual patient care94. All types of CDS systems (e.g., alerts, dashboards) were considered in scope, however eligible CDS must have been integrated with a CIS (e.g., Electronic Medical or Health Records, and Computerised Provider Order Entry systems). Our population of interest included any hospital-based clinicians (e.g. doctors, nurses) who were end users of a CDS system, targeting any health condition or patient group. Eligible studies reported factors influencing clinicians’ perceptions of (acceptance), and/or actual interactions (use) with a CDS system to support patient care in inpatient or outpatient hospital settings. Studies that evaluated CDS as part of a broader system were only included if CDS-specific results were reported. We included peer-reviewed original research and case studies that employed qualitative, quantitative or mixed-methods designs, and were available in English. To capture the point-in-time that factors emerged, eligible studies needed to report the specific timing of data collection in relation to CDS implementation.
Study selection
After removing duplicates using EndNote 20 software95, titles and abstracts were imported into Covidence (www.covidence.org) and independently screened for inclusion by two authors (NN and AB, RF or MB). Full texts of potentially relevant articles were screened against inclusion criteria. A sample of full texts were independently screened by review pairs (NN and AB, RF or MB), until Cohen’s kappa of >0.81, representing ‘almost perfect’ interrater reliability, was achieved96. Disagreements were resolved through discussion between the review pair and if required, discussion and consensus among four authors. Remaining texts were screened by one reviewer each (NN, AB, RF or MB).
Data extraction
Data were extracted independently by two authors (NN and AB, RF, AT or MB) using a structured data collection form in Microsoft Excel. The form was developed and iteratively refined following extraction of data from a sample of studies. Data extracted included study details and identifiers (e.g. authors, year), participant role, setting (e.g. department, unit), CDS description (e.g. type, AI vs. non-AI based), acceptance and/or use measurement, factors associated with clinicians’ acceptance and/or use of CDS, and the time of data collection following implementation. Missing or unclear information were recorded as not reported (NR). Disagreements were resolved as described above.
Quality assessment
The methodological quality of included studies was independently appraised by two authors (NN and AB, RF, AT or MB) using the well-established MMAT97. Quality was assessed only for study methods and results that met inclusion criteria (see Supplementary Note 3). Disagreements were resolved as described above. As we aimed to comprehensively identify the factors observed to influence CDS acceptance and use over time, quality assessment scores were not used as a basis for exclusion of studies but to guide the interpretation of findings98,99.
Data analysis and synthesis of factors over time
We used the CFIR to synthesise findings100. The CFIR is an implementation science framework that is widely used in healthcare settings to evaluate individual, technological and contextual factors influencing the implementation of complex interventions. The CFIR was selected for this review because of its comprehensive structure, ensuring factors were captured via a whole of system approach. Factors were synthesised using a convergent integrative approach101, where a word or short phrase that captured the meaning and direction (i.e. positive, negative, no direction) of factors reported in studies was noted under one of five major domains in the CFIR framework: intervention, outer setting, inner setting, individuals, and process.
Following extraction of all data, five reviewers (NN, AB, RF, MB, AT) participated in 7 workshops totalling 11 h to synthesise factors under CFIR constructs. This included discussion of each original factor documented in the data extraction form and in some cases, merging and/or renaming factors, before allocating to constructs. This created a framework of CFIR domains, constructs and factors that influenced CDS acceptance and use. In cases where factors did not align with any existing CFIR constructs or domains, new constructs or domains were created. All factors were classified as barriers, facilitators or moderators. Moderating factors were those that influenced the level of CDS acceptance and/or use.
The point-in-time that data were collected in each study was grouped into a time category (‘timeframe’) in line with inclusion criteria detailing time specificity that is described in Supplementary Note 2 and Supplementary Table 1. Graphs were created in Microsoft Excel and Tableau102 to visualise the occurrence and proportion of overall factors, constructs, barriers, facilitators reported in different domains at each timeframe.
Data availability
The datasets analysed during the current study are available from the corresponding author upon request.
References
Sutton, R. T. et al. An overview of clinical decision support systems: benefits, risks, and strategies for success. NPJ Digital Med. 3, 17 (2020).
Roshanov, P. S. et al. Features of effective computerised clinical decision support systems: meta-regression of 162 randomised trials. BMJ 346, f657 (2013).
Kwan, J. L. et al. Computerised clinical decision support systems and absolute improvements in care: meta-analysis of controlled clinical trials. BMJ 370, m3216 (2020).
Bright, T. J. et al. Effect of clinical decision-support systems: a systematic review. Ann. Intern. Med. 157, 29–43 (2012).
Kouri, A., Yamada, J., Lam Shin Cheung, J., Van de Velde, S. & Gupta, S. Do providers use computerized clinical decision support systems? A systematic review and meta-regression of clinical decision support uptake. Implement. Sci. 17, 21 (2022).
Khairat, S., Marc, D., Crosby, W. & Al Sanousi, A. Reasons for physicians not adopting clinical decision support systems: critical analysis. JMIR Med. Inf. 6, e24 (2018).
Kilsdonk, E., Peute, L. W. & Jaspers, M. W. Factors influencing implementation success of guideline-based clinical decision support systems: a systematic review and gaps analysis. Int. J. Med. Inf. 98, 56–64 (2017).
Devaraj, S., Sharma, S. K., Fausto, D. J., Viernes, S. & Kharrazi, H. Barriers and facilitators to clinical decision support systems adoption: a systematic review. J. Bus. Adm. Res. 3, 36 (2014).
Camacho, J., Zanoletti-Mannello, M., Landis-Lewis, Z., Kane-Gill, S. L. & Boyce, R. D. A conceptual framework to study the implementation of clinical decision support systems (BEAR): literature review and concept mapping. J. Med. Internet Res. 22, e18388 (2020).
Abell, B. et al. Identifying barriers and facilitators to successful implementation of computerized clinical decision support systems in hospitals: a NASSS framework-informed scoping review. Implement. Sci. 18, 32 (2023).
Liberati, E. G. et al. What hinders the uptake of computerized decision support systems in hospitals? A qualitative study and framework for implementation. Implement. Sci. 12, 113 (2017).
Baysari, M. T. et al. Longitudinal study of user experiences of a CPOE system in a pediatric hospital. Int. J. Med. Inform. 109, 5–14 (2018).
Rogers, E. M. Diffusion of Innovations 5 edn (Free Press, 2003).
Greenhalgh, T., Robert, G., Macfarlane, F., Bate, P. & Kyriakidou, O. Diffusion of innovations in service organizations: systematic review and recommendations. Milbank Q. 82, 581–629 (2004).
Murray, E. et al. Normalisation process theory: a framework for developing, evaluating and implementing complex interventions. BMC Med. 8, 63 (2010).
Castellanos, I., Kraus, S., Toddenroth, D., Prokosch, H.-U. & Burkle, T. Using Arden syntax medical logic modules to reduce overutilization of laboratory tests for detection of bacterial infections-Success or failure?. Artif. Intell. Med. 92, 43–50 (2018).
Grauer, A. et al. Indication alerts to improve problem list documentation. J. Am. Med. Inform. Assoc. 29, 909–917 (2022).
Guidi, J. L. et al. Clinician perception of the effectiveness of an automated early warning and response system for sepsis in an academic medical center. Ann. Am. Thorac. Soc. 12, 1514–1519 (2015).
Sauro, K. M. et al. Improving the adoption of optimal venous thromboembolism prophylaxis in critically ill patients: a process evaluation of a complex quality improvement initiative. J. Crit. Care 50, 111–117 (2019).
Tsai, W.-C. et al. Design and implementation of a comprehensive AI dashboard for real-time prediction of adverse prognosis of ED patients. Healthcare10, 1498–N.PAG (2022).
De Bie, A. J. R. et al. Intelligent checklists improve checklist compliance in the intensive care unit: a prospective before-and-after mixed-method study. Br. J. Anaesth. 126, 404–414 (2021).
Harrison, A. M. et al. Comparison of methods of alert acknowledgement by critical care clinicians in the ICU setting. PeerJ 5, e3083 (2017).
Petersen, J. D. et al. Clinical decision support for hyperbilirubinemia risk assessment in the electronic health record. Acad. Pediatr. 20, 857–862 (2020).
Thayer, J. G. et al. Human-centered development of an electronic health record-embedded, interactive information visualization in the emergency department using fast healthcare interoperability resources. J. Am. Med. Inform. Assoc. 28, 1401–1410 (2021).
Berge, G. T. et al. Machine learning-driven clinical decision support system for concept-based searching: a field trial in a Norwegian hospital. BMC Med. Inform. Decis. Mak. 23, 5 (2023).
Casey, S. D. et al. Physicians’ perceptions of clinical decision support to treat patients with heart failure in the ED. JAMA Netw. Open 6, E2344393 (2023).
Chadwick, D. R. et al. A feasibility study for a clinical decision support system prompting HIV testing. HIV Med. 18, 435–439 (2017).
Huang, H.-Y., Lee, T.-T., Hsu, T.-C., Mills, M. E. & Tzeng, I. S. Evaluation of the pressure injury prevention information system. Compu. Inform. Nurs.38, 625–632 (2020).
Jenssen, B. P. et al. Clinical decision support tool for parental tobacco treatment in hospitalized children. Appl. Clin. Inform. 7, 399–411 (2016).
Keim-Malpass, J. et al. Advancing continuous predictive analytics monitoring: moving from implementation to clinical action in a learning health system. Crit. Care Nurs. Clin. North Am. 30, 273–287 (2018).
Mahabee-Gittens, E. M. et al. An electronic health record-based strategy to address child tobacco smoke exposure. Am. J. Prev. Med. 54, 64–71 (2018).
Rosenthal, B. et al. Integration of physical abuse clinical decision support at 2 general emergency departments. J. Am. Med. Inf. Assoc. 26, 1020–1029 (2019).
Yoon, D. H. et al. Effect of an artificial-intelligent chest radiographs reporting system in an emergency department. Signa Vitae 19, 144–151 (2023).
Feldstein, D. A. et al. Disseminating child abuse clinical decision support among commercial electronic health records: effects on clinical practice. JAMIA Open 6, ooad022 (2023).
Suresh, S. et al. Improved detection of child maltreatment with routine screening in a tertiary care paediatric hospital. J. Pediatr. 243, 181–187 (2022).
Ginestra, J. C. et al. Clinician perception of a machine learning-based early warning system designed to predict severe sepsis and septic shock. Crit. Care Med. 47, 1477–1484 (2019).
Holroyd-Leduc, J. M. et al. A pragmatic study exploring the prevention of delirium among hospitalized older hip fracture patients: applying evidence to routine clinical practice using clinical decision support. Implement. Sci.5, 81 (2010).
Rabinovich, D., Mosquera, C., Torrens, P., Aineseder, M. & Benitez, S. User satisfaction with an AI system for chest X-ray analysis implemented in a hospital’s emergency setting…32nd Medical Informatics Europe Conference (MIE2022), 27-30 May, 2022, Nice, Fran. Stud. Health Technol. Inform. 294, 8–12 (2022).
Bellodi, E., Vagnoni, E., Bonvento, B. & Lamma, E. Economic and organizational impact of a clinical decision support system on laboratory test ordering. BMC Med. Inform. Decis. Mak. 17, 179 (2017).
English, D., Ankem, K. & English, K. Acceptance of clinical decision support surveillance technology in the clinical pharmacy. Inform. Health Soc. Care 42, 135–152 (2017).
Hoekstra, M. et al. Implementation and evaluation of a nurse-centered computerized potassium regulation protocol in the intensive care unit—a before and after analysis. BMC Med. Inform. Decis. Mak. 10, 5 (2010).
Jones, B. E. et al. CDS in a learning health care system: identifying physicians’ reasons for rejection of best-practice recommendations in pneumonia through computerized clinical decision support. Appl. Clin. Inform. 10, 1–9 (2019).
Uppot, R. N., Yu, A. Y. C., Samadi, K., Pino, R. M. & Lee, J. Let the EHR talk loudly: an EHR-connected verbal surgical safety checklist for medical procedures in the intensive care unit. J. Patient Saf. 18, e136–e139 (2022).
Agostini, J. V., Concato, J. & Inouye, S. K. Improving sedative-hypnotic prescribing in older hospitalized patients: provider-perceived benefits and barriers of a computer-based reminder. J. Gen. Intern-. Med. 23, 32–36 (2008).
Bell, H., Garfield, S., Khosla, S., Patel, C. & Franklin, B. D. Mixed methods study of medication-related decision support alerts experienced during electronic prescribing for inpatients at an English hospital. Eur. J. Hosp. Pharm. Sci. Pract. 26, 318–322 (2019).
Cho, I., Park, I., Kim, E., Lee, E. & Bates, D. W. Using EHR data to predict hospital-acquired pressure ulcers: a prospective study of a Bayesian Network model. Int. J. Med. Inform. 82, 1059–1067 (2013).
Groshaus, H., Boscan, A., Khandwala, F. & Holroyd-Leduc, J. Use of clinical decision support to improve the quality of care provided to older hospitalized patients. Appl. Clin. Inform. 3, 94–102 (2012).
Jauk, S. et al. Technology acceptance of a machine learning algorithm predicting delirium in a clinical setting: a mixed-methods study. J. Med. Syst. 45, 48 (2021).
Lytle, K. S., Short, N. M., Richesson, R. L. & Horvath, M. M. Clinical decision support for nurses: a fall risk and prevention example. Comput. Inform. Nurs.33, 530–E531 (2015).
Neame, M. et al. The impact of paediatric dose range checking software. Eur. J. Hosp. Pharm. 28, E18–E22 (2021).
Nydert, P., Veg, A., Bastholm-Rahmner, P. & Lindemalm, S. Pediatricians’ understanding and experiences of an electronic clinical-decision-support-system. Online J. Public Health Inform. 9, e200 (2017).
Pirnejad, H., Niazkhanlab, Z., Aarts, J. & Bal, R. What makes an information system more preferable for clinicians? A qualitative comparison of two systems. Stud. Health Technol. Inform. 169, 392–396 (2011).
Salwei, M. E. et al. Workflow integration analysis of a human factors-based clinical decision support in the emergency department. Appl. Ergon.97, 103498 (2021).
Salwei, M. E. et al. Post-implementation usability evaluation of a human factors-based clinical decision support for pulmonary embolism (PE) diagnosis (Dx): PE Dx Study Part 1. Hum. Factors Healthc. 4, 100056 (2023).
Stutman, H. R., Fineman, R., Meyer, K. & Jones, D. Optimizing the acceptance of medication-based alerts by physicians during CPOE implementation in a community hospital environment. AMIA Annu Symp Proc. 701–705 (2007).
Henry, K. E. et al. Human-machine teaming is key to AI adoption: clinicians’ experiences with a deployed machine learning system. NPJ Digital Med. 5, 97 (2022).
Bersani, K. et al. Use, perceived usability, and barriers to implementation of a patient safety dashboard integrated within a vendor EHR. Appl. Clin. Inf. 11, 34–45 (2020).
Eden, R. et al. Digitising an Australian university hospital: qualitative analysis of staff-reported impacts. Aust. Health Rev.44, 677–689 (2020).
Frymoyer, A. et al. Model-informed precision dosing of vancomycin in hospitalized children: implementation and adoption at an academic children’s hospital. Front. Pharm. 11, 551 (2020).
Goldstein, J. E., Guo, X., Swenor, B. K., Boland, M. V. & Smith, K. Using electronic clinical decision support to examine vision rehabilitation referrals and practice guidelines in ophthalmology. Transl. Vis. Sci. Technol. 11, 8 (2022).
Hum, R. S. et al. Developing clinical decision support within a commercial electronic health record system to improve antimicrobial prescribing in the neonatal ICU. Appl. Clin. Inform. 5, 368–387 (2014).
Salwei, M. E. et al. Usability of a human factors-based clinical decision support in the emergency department: lessons learned for design and implementation. Hum. Factors. 66, 647–657 (2022).
Scheepers-Hoeks, A.-M. J., Grouls, R. J., Neef, C., Ackerman, E. W. & Korsten, E. H. Physicians’ responses to clinical decision support on an intensive care unit—comparison of four different alerting methods. Artif. Intell. Med. 59, 33–38 (2013).
Short, B. et al. Implementation of lung protective ventilation order to improve adherence to low tidal volume ventilation: a RE-AIM evaluation. J. Crit. Care 63, 167–174 (2021).
Zhai, Y., Yu, Z., Zhang, Q. & Zhang, Y. Barriers and facilitators to implementing a nursing clinical decision support system in a tertiary hospital setting: a qualitative study using the FITT framework. Int. J. Med. Inf. 166, 104841 (2022).
Chow, A. L. et al. Implementation hurdles of an interactive, integrated, point-of-care computerised decision support system for hospital antibiotic prescription. Int. J. Antimicrob. Agents 47, 132–139 (2016).
Campion, T. R. Jr. et al. Characteristics and effects of nurse dosing over-rides on computer-based intensive insulin therapy protocol performance. J. Am. Med. Inform. Assoc.18, 251–258 (2011).
Chow, A., Lye, D. C. B. & Arab, O. A. Psychosocial determinants of physicians’ acceptance of recommendations by antibiotic computerised decision support systems: a mixed methods study. Int. J. Antimicrob. Agents 45, 295–304 (2015).
Galanter, W. L., Hier, D. B., Jao, C. & Sarne, D. Computerized physician order entry of medications and clinical decision support can improve problem list documentation compliance. Int. J. Med. Inf. 79, 332–338 (2010).
Lichtner, V., Franklin, B. D., Dalla-Pozza, L. & Westbrook, J. I. Electronic ordering and the management of treatment interdependencies: a qualitative study of paediatric chemotherapy. BMC Med. Inf. Decis. Mak. 20, 193 (2020).
Lin, Y.-C., Chang, C.-S., Yeh, C.-J. & Wu, Y.-C. The appropriateness and physician compliance of platelet usage by a computerized transfusion decision support system in a medical center. Transfusion 50, 2565–2570 (2010).
Beeler, P. E., John Orav, E., Seger, D. L., Dykes, P. C. & Bates, D. W. Provider variation in responses to warnings: do the same providers run stop signs repeatedly?. J. Am. Med. Inform. Assoc. 23, 93–98 (2016).
Campion, T. R. Jr., Waitman, L. R., Lorenzi, N. M., May, A. K. & Gadd, C. S. Barriers and facilitators to the use of computer-based intensive insulin therapy. Int. J. Med. Inform. 80, 863–871 (2011).
Choi, K. S., Lee, E. & Rhie, S. J. Impact of pharmacists’ interventions on physicians’ decision of a knowledge-based renal dosage adjustment system. Int. J. Clin. Pharm. 41, 424–433 (2019).
Choudhury, A. Factors influencing clinicians’ willingness to use an AI-based clinical decision support system. Front. Digital Health 4, 920662 (2022).
Choudhury, A. & Asan, O. Impact of cognitive workload and situation awareness on clinicians’ willingness to use an artificial intelligence system in clinical practice. IISE Trans. Healthc. Syst. Eng. 13, 89–100 (2023).
Luna, D. R., Rizzato Lede, D. A., Otero, C. M., Risk, M. R. & Gonzalez Bernaldo de Quiros, F. User-centered design improves the usability of drug-drug interaction alerts: experimental comparison of interfaces. J. Biomed. Inform. 66, 204–213 (2017).
Ng, H. J. H. et al. Optimizing Best Practice Advisory alerts in electronic medical records with a multi-pronged strategy at a tertiary care hospital in Singapore. JAMIA Open 6, ooad056 (2023).
Pontefract, S. K. et al. The impact of computerised physician order entry and clinical decision support on pharmacist-physician communication in the hospital setting: a qualitative study. PLoS ONE 13, e0207450 (2018).
Van De Sijpe, G. et al. Overall performance of a drug-drug interaction clinical decision support system: quantitative evaluation and end-user survey. BMC Med. Inform. Decis. Mak. 22, 48 (2022).
Wong, A. et al. Comparison of overridden medication-related clinical decision support in the intensive care unit between a commercial system and a legacy system. Appl. Clin. Inform. 8, 866–879 (2017).
Wright, A. et al. Reduced effectiveness of interruptive drug-drug interaction alerts after conversion to a commercial electronic health record. J. Gen. Intern. Med. 33, 1868–1876 (2018).
Markus, M. L. Technochange management: using IT to drive organizational change. J. Inf. Technol.19, 4–20 (2004).
Cresswell, K. M. et al. Sustained user engagement in health information technology: the long road from implementation to system optimization of computerized physician order entry and clinical decision support systems for prescribing in hospitals in England. Health Serv. Res. 52, 1928–1957 (2017).
Westbrook, J. et al. Immediate and long-term effects of an electronic medication management system on paediatric prescribing errors: a stepped-wedge cluster randomised controlled trial with one-year follow-up. NPJ Digital Med. 5, 179 (2022).
Kinlay, M. et al. An analysis of incident reports related to electronic medication management: how they change over time. J. Patient Saf. 20, 202–208 (2024).
Koppel, R., Wetterneck, T., Telles, J. L. & Karsh, B. T. Workarounds to barcode medication administration systems: their occurrences, causes, and threats to patient safety. J. Am. Med. Inf. Assoc. 15, 408–423 (2008).
Khera, R., Simon, M. A. & Ross, J. S. Automation bias and assistive AI: risk of harm from AI-driven clinical decision support. JAMA 330, 2255–2257 (2023).
Wiegel, V. et al. A systematic analysis of the optimization of computerized physician order entry and clinical decision support systems: a qualitative study in English hospitals. Health Inform. J. 26, 1118–1132 (2020).
Owens, K. EMR implementation: big bang or a phased approach?. J. Med. Pract. Manag.23, 279–281 (2008).
Page, M. J. et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 372, n71 (2021).
Newton, N., Bamgboje-Ayodele, A., Forsyth, R., Tariq, A. & Baysari, M. T. How are clinicians’ acceptance and use of clinical decision support systems evaluated over time? A systematic review. Stud. Health Technol. Inform. 310, 259–263 (2024).
Newton, N., Bamgboje-Ayodele, A., Forsyth, R., Tariq, A. & Baysari, M. T. Does involving clinicians in decision support development facilitate system use over time? A systematic review. Stud. Health Technol. Inform. 304, 11–15 (2023).
Osheroff, J. A. et al. Improving Outcomes with Clinical Decision Support: An Implementer’s Guide 2nd edn (HIMSS Publishing, 2012).
EndNote v. EndNote 20 (Clarivate, 2020).
McHugh, M. L. Interrater reliability: the kappa statistic. Biochem. Med.22, 276–282 (2012).
Hong, F. et al. The Mixed Methods Appraisal Tool (MMAT) version 2018 for information professionals and researchers. Educ. Inf. 34, 285–291 (2018).
Hassan Assem, M. et al. Barriers and facilitators to improving patient safety learning systems: a systematic review of qualitative studies and meta-synthesis. BMJ Open Qual. 12, e002134 (2023).
Dixon-Woods, M., Fitzpatrick, R. & Roberts, K. Including qualitative research in systematic reviews: opportunities and problems. J. Eval. Clin. Pr. 7, 125–133 (2001).
Damschroder, L. J. et al. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement. Sci. 4, 50 (2009).
Stern, C. et al. Methodological guidance for the conduct of mixed methods systematic reviews. JBI Evid. Implement. 19, 120–129 (2021).
Tableau Software. Tableau version 2021.3 [software]. Salesforce, Seattle, WA, USA (2021).
Damschroder, L. J., Reardon, C. M., Widerquist, M. A. O. & Lowery, J. The updated consolidated framework for implementation research based on user feedback. Implement. Sci. 17, 75 (2022).
Acknowledgements
This research was supported by Digital Health CRC Limited (“DHCRC”). DHCRC is funded under the Commonwealth’s Cooperative Research Centres (CRC) Program. The funder played no role in study design, data collection, analysis and interpretation of data, or the writing of this manuscript.
Author information
Authors and Affiliations
Contributions
All authors conceived and planned the review, including the search strategy. N.N. searched databases and conducted manual searches. N.N., M.B., A.B. and R.F. screened identified articles for eligibility. All authors (N.N., M.B., A.B., R.F. and A.T.) extracted data and participated in workshops to synthesise factors under the CFIR framework. N.N. (with input from all authors) analysed data and created graphs. The manuscript was drafted by N.N., with all authors contributing to subsequent edits and revisions. All authors approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Newton, N., Bamgboje-Ayodele, A., Forsyth, R. et al. A systematic review of clinicians’ acceptance and use of clinical decision support systems over time. npj Digit. Med. 8, 309 (2025). https://doi.org/10.1038/s41746-025-01662-7
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41746-025-01662-7






