Abstract
Electronic health record (EHR) systems are critical to modern healthcare delivery, yet the dynamic workflows that govern electronic order processing remain underexplored. Inefficiencies in these digital pathways can cause delays in care, repetitive workloads, and even patient harm. This study presents a discrete-event simulation framework used to reconstruct and evaluate EHR-based order workflows in a large integrated healthcare system. Using real-world data extracted from the Veterans Health Administration’s Corporate Data Warehouse, the authors mapped order events to standardized state transitions and modeled their progression across different facilities of varying complexity levels. After being calibrated with empirical distributions of transition times and validated against observed time-in-system metrics, the simulation demonstrates close alignment with historical performance. Scenario analyses reveal that resource capacity constraints significantly amplify the impact of electronic order surges, which are reflected in the disproportionate growth in backlogs and processing delays. Adjustments in transition probabilities further increased recirculation and extended workflow paths. Network-based analysis identified Reserved, InProgress, and Completed as structurally critical states that function as hubs within the process network but the transitions in-between also act as major bottlenecks. These results showcased the effectiveness of simulation-based approaches in monitoring EHR order processing performance and evaluating consequences of workflow changes on healthcare network resources planning. The proposed simulation framework provides a scalable data-driven tool to support operational decision-making and improve the efficiency of electronic order management in complex healthcare environments.
Similar content being viewed by others
Introduction
The electronic health record (EHR) system is a cornerstone of Health Information Technology (HIT). EHRs enable digital management of patient health information and support the coordination of care across complex organizations. In the United States, the Veterans Health Administration (VHA) operates the country’s largest integrated healthcare system, serving over 9 million enrolled veterans through 1380 facilities, including 170 medical centers and 1193 outpatient sites across the country 1. Since the 1990s, the Veterans Health Information Systems and Technology Architecture (VistA)2 has provided the backbone of the VHA’s HIT (e.g., the Corporate Data Warehouse [CDW]3) to facilitate data exchange and care coordination. Recognizing persistent challenges related to system fragmentation and interoperability, the US Department of Veterans Affairs (VA) launched a multi-billion-dollar modernization initiative in 2017 to transition from VistA to Federal EHR 4,5. This effort aims to streamline order management workflows among healthcare providers by unifying EHR systems between the VHA and the US Department of Defense to provide a single, longitudinal health record that contains a given veteran’s full medical history, which will help reduce prescription errors, prevent redundant testing, and ultimately improve the quality of veteran healthcare. Although the anticipated benefits of a consolidated EHR system are substantial, the transition has introduced complexities and challenges in areas such as data migration, clinician training 6, workflow adaptation, and system usability 7,8. These challenges underscore the need for tools to monitor operational performance and evaluate the impact of change to maximize the benefits of this modernization effort.
The dynamic and complex nature of healthcare systems, especially within large integrated networks such as those of the VHA, presents significant challenges in predicting the downstream effects of operational changes. Traditional evaluation methods (e.g., retrospective data analysis, observational studies) often fall short in capturing the intricate interdependencies among clinical process, resource allocation, and patient care outcomes. For instance, process mining research in health care domain has been reviewed 9 from three novel dimensions of process mining project stages, involvement of domain expertise and key performance indicators. It is found that process redesign is rarely part of a process mining project, domain experts are mostly asked for validating insights, and less than half of the published papers consider one or more specific performance indicators to direct the analysis. To address these challenges, simulation-based approaches such as discrete-event simulation (DES), agent-based modeling (ABM), and system dynamics (SD) have been widely employed in healthcare research because they can provide valuable insights into operational efficiency optimization (e.g., patient wait time reduction) and care delivery improvement10. Among these simulation approaches, the number of publications that describe the use of DES methods has increased significantly since 200411. For example, Kalton et al. developed a multi-agent simulation to evaluate care coordination for patients with serious mental illnesses by simulating interactions between patients, healthcare providers, social services, and the criminal justice system. Using real-world data and expert input, the model demonstrates how enhanced coordination reduces crises and improves patient outcomes12. Vahdat et al. examined the effects of increased provider-patient encounter times caused by EHR implementation in an outpatient dermatology clinic. The results show that even small increases in in-room documentation time (1–10 minutes) led to significant delays, with wait times rising by 22%–373% and length of stay increasing by 12%–193%. To offset these delays, clinics would need to reduce patient volume by up to 50% to maintain performance13. Simwita et al. applied DES to evaluate resource flexibility in an orthopedic care process at Bugando Medical Centre in Tanzania14 and demonstrated that reallocating mid-level healthcare workers for routine tasks to optimize the use of specialized surgeons can reduce patient waiting times by up to 72.7% and increase throughput by 94%. Despite DES being increasingly applied to operational healthcare issues, fewer than 10% of simulation studies progress beyond the modeling stage to real-world implementation15. In addition to standalone simulation methods, hybrid simulation approaches have emerged to capture multilevel healthcare processes. Djanatliev et al. introduced a multi-paradigm simulation approach that forms an SD-ABM-DES framework to support prospective health technology assessment16. After reviewing the hybrid simulation applications, Santos et al. reported the most popular combination is DES-SD, with AnyLogic as the predominant simulation tool17.
Although simulation-based research helps enable testing interventions and examine rare events that may be inaccessible in reality, it faces challenges in validity and applicability. However, integrating simulation with other methods can enhance its applicability and make it a promising approach for healthcare improvement research18,19. After reviewing 231 papers on DES modeling in healthcare15, around 66% of theoretical studies incorporate DES alongside other analytical techniques (e.g., optimization and Lean/Six Sigma). For example, DES is integrated with lean management and a business game to improve patient flow in an outpatient hematology-oncology clinic during a Kaizen event20. Kovalchuk et al. combined process mining, data mining, and DES21 to simulate patient flow in acute coronary syndrome care and show how empirical data can enhance simulation fidelity and predictive capacity. As another example, PartiSim is a multi-methodology framework that combines DES with a soft systems methodology to enhance stakeholder involvement throughout the study life cycle22. The framework features a comprehensive six-stage process that guides stakeholders from problem identification to solution implementation and has been successfully applied in two real-world studies. Kuruppu Appuhamilage et al. developed a dual-layer digital twin framework that combines DES and Azure cloud infrastructure23 to capture critical care workflows in real time, where DES serves as both a data collection mechanism (via barcode scanning) and a workflow structuring tool that maps these events into a conceptual digital twin for continuous tracking and visualization of care processes. Data-driven process simulation has been explored for capacity management applications in healthcare24, with a case study in a Belgian hospital’s radiology department demonstrating its utility in informing decisions on equipment requirements, waiting area capacity, and staffing needs.
Although patient flow simulations capture the movement of individuals through physical care environments, current studies often overlook the digital workflows that underpin clinical decision-making, such as the processing of medication orders, lab tests, and diagnostic imaging requests within the EHR system. This gap is particularly significant given the critical role that electronic order management plays in reducing delays and minimizing errors. Moreover, HIT-related challenges, such as poor system interoperability, human data entry errors, and unintended EHR system functionalities, further complicate electronic order processing. Issues such as system downtimes25,26, alert fatigue25,27,28, and inefficient clinical decision support mechanisms have been identified as persistent threats to patient safety and workflow efficiency in EHR environments29,30. This study takes the first step in modeling the flow of electronic orders within EHR systems.
To address the gap, this study proposes a DES framework tailored to reconstruct and evaluate electronic order workflows. First, the general simulation-driven order flow decision framework is introduced. The simulation models are then developed for the EHR order flows of five healthcare facilities based on exemplary data mapping (e.g., OASIS). The simulation is parameterized with empirical transition time distributions and validated against observed time-in-system metrics. Second, using one VHA healthcare facility as an example, the study describes how the impact of electronic order surges, resource capacity constraints, and transition probabilities can be evaluated and how order trajectory changes are visualized based on simulated results under different configurations. Finally, network analysis further reveals the relative centrality and importance of specific states and transition edges within the workflow. The approach provides a scalable data-driven foundation for evaluating EHR workflow performance under different settings, identifying trajectory variations, and supporting real-time anomaly detection and alerts.
Results
Five VHA healthcare facilities were selected to evaluate the applicability of the proposed simulation framework (further details are shown in Fig.11). To standardize and interpret the heterogeneous order action logs generated by VHA’s EHR system, the data mapping protocol OASIS (Organization for the Advancement of Structured Information Standards) was adopted. OASIS is a nonprofit consortium that drives the development, convergence and adoption of open standards for the global information society31. It provides a generalized state-transition model widely used in business process management. OASIS-derived state mappings (further details are shown in Fig. 13) were applied on the raw EHR Consult order actions to improve the interpretability of the order logs, which standardizes the order processing states as Created, Ready, Reserved, InProgress, and Completed and the abnormal states as Failed, Exited, and Error. The simulation mirrors the state transitions and samples empirical transition durations from historical data. The following results include a model validation step followed by what-if scenario analyses on one facility from three aspects—electronic order surge, transition dynamics, and network analysis—to better understand the system dynamics and performance bottlenecks.
Model validation
The transition probability for each order transition path in the simulation was estimated by using the relative frequencies of observed transitions in historical order flow data, and the transition duration was sampled directly from empirical distributions to capture variability without imposing parametric assumptions. Alternatively, the statistical distributions can be fit for transition durations.
Model performance was assessed by comparing simulated and observed time-in-system distributions, a key operational metric in electronic order processing. Time-in-system refers to the total elapsed time between when an order is first created and when it reaches its terminal state (e.g., Completed or abnormal states). This metric captures all intermediate waiting, processing, and decision steps associated with the handling of the order. Quantile-quantile (Q-Q) plots were used to visually assess concordance between the simulated and empirical distributions (Fig. 1). Across all five facilities, the Q-Q plots demonstrated strong agreement in the lower and central quantiles, and these results suggest that the models can accurately replicate typical processing patterns. Slight deviations in the upper quantiles were also noted, indicating that the simulations occasionally produced longer-than-observed process times. However, this is consistent with the use of some heavily tailed empirical samples. Given the potential operational impact of extreme values in healthcare (e.g., prolonged delays in diagnostic or therapeutic orders), we did not exclude outliers or truncate distributions during simulation in order to fully retain the risk landscape. This could enable the simulation to surface rare but critical scenarios which may warrant targeted analysis in other applications like exception handling.
a Facility A. b Facility B. c Facility C. d Facility D. e Facility E. The Q-Q plots illustrate the correspondence between simulated and empirical datasets of time-in-system. Each point represents a corresponding quantile in the two datasets, and the red-dashed line indicates perfect concordance where the two datasets are identical.
To better statistically quantify the model accuracy, two sample t-tests were performed to compare the mean time-in-system between simulated and observed data. These tests yielded t-statistics ranging from -1.62 to 1.02 with p-values between 0.11 and 0.31 across all facilities. Thus, it is safe to conclude that no statistically significant differences were detected at the 95% confidence level.
Supplemental to the t-tests for comparing mean performance, two-sample Kolmogorov-Smirnov (K-S) tests were also conducted to assess differences between the simulated and real time-in-system distribution. Empirical cumulative distribution functions for each facility are shown in Fig. 2, with the K-S statistics ranging from 0.0154 to 0.0423. These K-S statistics represent the maximum vertical distance between the two cumulative distributions and are uniformly small, which indicates close agreement in distributional shape. Although the corresponding p-values are extremely small, this is a known sensitivity issue of K-S test under large sample sizes32,33 and does not necessarily imply meaningful practical differences.
a Facility A. b Facility B. c Facility C. d Facility D. e Facility E. The ECDF plots compare the distributional alignment between simulated and real time-in-system data. Each curve represents the cumulative probability of order duration for one dataset. The degree of vertical separation between the curves reflects the maximum difference in cumulative distributions, which is quantified by the K-S test statistic.
Additionally, We examined the Interquartile Range (IQR) for each facility to assess how well the simulation reproduces the spread of order durations. Table 1 summarizes the Q1, Q3, and IQR values for both real and simulated time-in-system data. Because the IQR captures the dispersion of the central 50% of the distribution, it provides a robust measure of system variability. Across all facilities, the IQR ratio fell within a narrow ± 10% range for the system level analysis, although no universal threshold exists for simulation validation.
In summary, these results indicate that the models could reproduce both the structure and variability of real-world electronic order workflows and the simulation outputs are valid for downstream scenario analysis.
Performance comparison
To further characterize order process behavior, conditional transition probabilities (Fig. 3) and order loop-back ratios at different workflow states (Fig. 4) were analyzed by using simulated data. Note that the empirical transition probabilities in real data can be directly computed from the frequency directly follows graph; for example, the probability values for Facility D can be derived from the flows shown in Fig. 15 for comparison. Although the overall workflow structure was generally consistent across facilities, as shown in Fig. 3, the transitions from Created to Ready occurred with near certainty, which reflects a potential standardized order initial processing procedure. However, operational variations emerged in transition dynamics and resolution patterns for orders in abnormal states. For example, the orders move from Reserved to InProgress with probabilities varying from 0.55 to 0.75 across facilities. The handling of orders in the Failed state notably differed: in Facility A, Failed orders led to Exited more often, whereas in other facilities, it was typically resolved to Completed, which suggests large variations in order-retry mechanisms or correction policies.
Based on the user-defined logical sequence of Created → Ready → Reserved → InProgress → Completed → Failed → Exited → Error, the loop-back ratio in Fig. 4 is constructed based on Fig. 3 by row-wise aggregating the transition probabilities to previous states. This ratio provides additional insight into order rework frequency. Additionally, for facilities B–E, the Failed state consistently exhibited the highest loop-back ratios (often > 0.7), indicating more frequent reentry into prior workflow stages (mostly to InProgress or Completed). This is consistent with order abnormal recovery or rework cycles. However, the loop-back ratio of Failed orders in Facility A is lower (< 0.2), and these orders have a high probability of flowing to Exited (Fig. 3). States such as InProgress also showed moderate loop-back patterns with ratios ranging from 0.3 to 0.45, which could indicate order reassessments or additional tasks. In contrast, terminal states such as Completed and Exited were rarely revisited, which reflects a proper end-state designation within the order life cycle.
These differences in process handling across facilities likely reflect the underlying heterogeneity in operational procedures, staff levels, and workload intensity. Moreover, identifying those states with high variability or recurrent loops may inform targeted interventions to reduce potential repetitive work and streamline transitions.
In the following three subsections, a series of simulation experiments was conducted to answer what-if questions such as how will the system respond if the electronic order increases or the processing resources have been limited? Facility D was selected as a representative example for the following scenario analyses. The simulation period was set to 2 years with a 1 h time resolution.
Electronic order surge
Multiple order surge scenarios were simulated by increasing the order arrival rate and adjusting resource capacity constraints. The following scenarios were considered:
-
max_capacity_baseline: This baseline scenario imposes no resource capacity limitations on any transition path. It allows all paths to concurrently process an unlimited number of orders. max_capacity_arrival_+20% represents the same configuration with a 20% increase in order arrival rate.
-
2000_capacity_baseline: In this scenario, a resource capacity limitation of 2000 orders is imposed on the Reserved → InProgress and InProgress → Completed transitions, meaning that no more than 2000 orders can be processed simultaneously on these two paths. Any additional orders are held in the InProgress queue for Reserved → InProgress and the Completed queue for InProgress → Completed. 2000_capacity_arrival_+20% represents this same configuration with a 20% increase in order arrival rate.
-
1000_capacity_baseline: This scenario is similar to the previous scenario but uses a stricter resource capacity limitation of 1000 orders on the Reserved → InProgress and InProgress → Completed paths. 1000_capacity_arrival_+20% represents the same configuration with a 20% increase in order arrival rate.
Figure 5 shows the instantaneous order number in the system with a 1-month rolling average trend for each scenario’s simulation. In queuing theory, the average number-in-system, L, is also calculated from the simulated results. Theoretically, the average number-in-system in a stable system follows Little’s Law: L = λ ⋅ W, where λ is the average arrival rate, and W is the average time-in-system34. The results demonstrate that under baseline arrival rates, the order numbers in the system stabilize at a manageable level for max_capacity_baseline and 2000_capacity_baseline with similar values, whereas the 1000_capacity_baseline could not reach system stability within the simulated time range. As the arrival rate increases, max_capacity_arrival_+20% stabilizes at a slightly higher level than its baseline scenario. In contrast, the average number-in-system, L, starts to surge for 2000_capacity_arrival_+20%, with a 2522 increase (from 5016.01 to 7538.15), and it rises much more sharply under the 1000_capacity scenarios without stabilization.
a max_capacity_baseline. b max_capacity_arrival_+20%. c 2000_capacity_baseline. d 2000_capacity_arrival_+20%. e 1000_capacity_baseline. f 1000_capacity_arrival_+20%. This figure shows the order numbers in the system with a 1-month rolling average trend across six simulated scenarios. Resource capacity limitations are applied to the Reserved → InProgress (left) and InProgress → Completed (right) transitions with 2000 or 1000 order concurrent processing capacities.
The detailed system metrics for the various scenarios are summarized in Table 2. The state transition ratio is defined as (total occurrences of leaving the state) / (total order number). Notably, the transition ratio can be greater than 1 with the loop-back transition. The Completed transition ratio declines for the both capacity-limited scenarios when the arrival rate increases, which leads to a reduced throughput. The mean waiting time of the InProgress and the Completed queues suggests that although the overall completion remains dominant, the higher-order arrival rate increases the likelihood of delays and incomplete processing. Collectively, these results could help identify a system’s tipping point for demand- and resource-allocation planning.
Transition dynamics
Because the transition probabilities and processing times are the main input, the following simulations focus on examining how the order trajectory path will change if the probability or service time on certain paths is altered. A scenario with modified transition probabilities and capped service times is compared against the 2000_capacity_baseline configuration.
-
2000_capacity_prob_trans_change: In this scenario, transition probabilities are dynamically adjusted partway through the simulation period. Specifically, at the end of the sixth month, the probability of transitioning directly from Ready to Completed was reduced from 0.354 to 0.104, and the probability of moving from Ready to Reserved was increased from 0.530 to 0.780. This adjustment represents a deliberate restriction of the shortcut path to completion. Additionally, to model the enforcement of a service level guarantee, the upper bound of processing time on the InProgress → Completed path is capped at 720 hours starting at the end of the first simulation year.
For the comparison, Fig. 6 shows the 30-day rolling average of orders leaving the Completed state over time. Under the 2000_capacity_baseline scenario, order throughput remains relatively stable following an initial ramp-up period. In contrast, the scenario with altered transition probabilities exhibits a marked decline in throughput shortly after the probability adjustments applied at the end of the sixth month. The throughput then stabilizes at a lower level before rising again after the 720 h service time cap was imposed on the InProgress → Completed path at the end of first simulation year. The turning points illustrate how dynamic changes in transition logic and processing can significantly impact system performance over time.
Two scenarios are simulated: 2000_capacity_baseline, in which transition probabilities and service times remained constant, and 2000_capacity_prob_trans_change, where the probability of transitioning from Ready directly to Completed is reduced to restrict shortcut to completion after the sixth simulation month. At the end of the first simulation year, the maximum processing time on the InProgress → Completed transition is capped at 720 hours to simulate a service level guarantee.
To further understand the order path changes, the top-15 path variants for the 2000_capacity_baseline scenario are listed in Table 3, and the top-30 path variants for the 2000_capacity_prob_trans_change scenario are visualized in Fig. 7. Notably, after the transition probability change is applied in the 2000_capacity_prob_trans_change scenario, the most frequent order path in the baseline containing the shortcut Ready → Completed became the second-most frequent path. It is replaced by the path progressing through Ready → Reserved → InProgress → Completed, which reflects the intended shift in process flow dynamics.
Straight lines denote adjacent transitions, whereas curved arcs indicate non-adjacent (jumped) transitions. Numbers on loop-back arcs show the sequence of repeated loops if more than 1. For example, for path variant V23, 502 (0.4% of) orders follow the nine-step sequence with two recirculations through InProgress → Reserved. This is the path of V14 shown in Table 3. For path variant V28, the orders go through recirculation InProgress → Reserved and then Reserved → Ready. After modifying transition probabilities, the most common shortcut path (Ready → Completed) becomes less frequent and is replaced by longer paths with intermediate states.
As an overview, Fig. 8 shows the distribution of the unique states versus the total number of transitions along the order paths. The scenario with modified transition probabilities exhibits longer and more complex transitions with a greater proportion of orders accumulating higher transition counts and extended durations. For example, some orders pass through as many as 21 transitions spanning 7 unique states. This demonstrates that modest changes in transition logic can produce substantial order recirculation with a high time-in-system, and simulation-based approaches can provide a valuable tool for assessing the effects of such routing adjustments or system glitches on process performance and throughput.
Network analysis
In addition to the process simulation results, network-based analysis of a transition network was performed by using several centrality measures derived from the simulation results from the 2000_capacity_baseline scenario. This analysis quantifies the relative importance of each state and transition within the workflow graph and provides a structural perspective on how process dynamics emerge from the state transition topology. Multiple centrality metrics have been computed and summarized below35,36, and Fig. 9 shows the results:
-
Degree Centrality (unweighted): Counts the total number of direct incoming and outgoing edges attached to each node, regardless of transition frequencies or durations. Operational implication: Nodes with higher value is broadly connected and may act as a key interaction point. Stronger coordination may be required to manage diverse entry and exit pathways effectively.
-
Betweenness Centrality (weighted): Captures the extent to which a state lies on the shortest paths between all pairs of states. Higher betweenness implies greater control over information or order flows, which may define the state as a bridge or bottleneck. Operational implication: Higher betweenness implies greater control over information or order flows, which may define the state as a bridge or bottleneck. Delays here can cascade downstream, amplifying system wide congestion. Improve operational efficiency in such nodes can reduce systematic latency.
-
Closeness Centrality (weighted): Quantifies how close a node is, on average, to all other nodes in the network based on the sum of the shortest path lengths. Nodes with high closeness can quickly reach (or be reached by) all others. Operational implication: Nodes with higher value are centrally located in the network structure. It would be ideal for implementing real time monitoring or priority triage mechanisms.
-
Eigenvector Centrality (weighted): Measures not only the number of connections but also their quality by assigning higher scores to nodes linked to other high scoring nodes. This metric highlights influential hubs embedded in well-connected neighborhoods. Operational implication: Nodes with higher score reflect strategic influence within the work flow where they may benefit from enhanced audit logging or targeted quality assurance.
-
PageRank Centrality (weighted): Estimates the relative importance of nodes based on the idea that connections from important states contribute more weight, which tends to capture recursive influence within the process flow. Operational implication: High PageRank score suggest area where small issues could have outsized effects on other important nodes. These nodes may warrant periodic risk assessment.
a Degree Centrality (Unweighted). b Betweenness Centrality (Weight: Duration/Count). c Closeness Centrality (Weight: Duration/Count). d Eigenvector Centrality (Weight: Count). e PageRank Centrality (Weight: Count). f Edge Betweenness Centrality. Degree, Betweenness, Closeness, Eigenvector, and PageRank centralities were computed to characterize each state’s structural importance. Formal definitions follow established formulations: Betweenness Centrality of node v as CB(v) = ∑s≠v≠tσst(v)/σst, where σst denotes the total number of shortest paths between nodes s and t, and σst(v) counts how many of those paths pass through node v. In weighted graphs, the shortest paths are computed based on cumulative edge weights. The Betweenness Centrality of an edge follows the same principle. PageRank Centrality is computed as \(PR(v)=\frac{1-d}{N}+d{\sum }_{u\in {\rm{In}}(v)}PR(u)/L(u)\), where d is a damping factor (commonly set to 0.85), N is the total number of nodes, In(v) represents the nodes linking to v, and L(u) is the out degree of node u.
For shortest path based measures (Betweenness and Closeness centralities), weights on the edges are defined as the average transition duration divided by the order count. This simultaneously accounts for time performance and order transition frequency. For Eigenvector and PageRank centralities, order count is used as the edge weight to emphasize the strength of connections. Among all nodes, Reserved, Completed, and InProgress continuously emerge as the most structurally significant states within the workflow. Specifically, Reserved exhibited the highest score in Betweenness Centrality, which indicates that it lies on many of the shortest weighted paths and functions as a critical bridge between other stages. Completed consistently dominated in PageRank and Eigenvector centralities, which indicates its strong connectivity and importance as the most frequent termination point. InProgress also showed overall high scores across centrality metrics, thereby echoing that it serves as a major processing hub. In contrast, the abnormal states Failed, Exited, and Error maintain lower centrality scores. The Failed state has a higher score in PageRank Centrality because it is connected to the most influential state node, Completed, and commonly appears as its subsequent state. The centrality graph for Betweenness and PageRank are shown in Fig. 10.
a Betweenness Centrality (Weight: Duration/Count)–Node and Edge. b PageRank Centrality (Weight: Count). Each node represents a workflow state, and the edge thickness reflects connection strength. Betweenness centrality quantifies how frequently a state lies on the shortest weighted paths between all pairs of states. PageRank centrality captures recursive influence based on the importance of connected nodes.
At the edge level, Betweenness Centrality further clarifies which transitions serve as critical conduits. The transition Completed → InProgress has the highest edge betweenness, followed by InProgress → Reserved and Reserved → Ready, which marks them as key structural links that facilitate recirculation in the network. In summary, these results illustrate and support the logic that Completed anchors most workflows, Reserved plays an outsized role in routing and rework, and InProgress serves as a major transit state. Together, they form the backbone of the order transition network.
Discussion
This study developed and evaluated a DES framework to model EHR order flows. By reconstructing state transitions from real-world EHR data, the simulation replicated observed operational patterns, including time-in-system distributions and site-specific variability across facilities. Scenario analyses demonstrated that increased orders and limited resource capacity led to pronounced system congestion. Dynamic adjustments in transition probabilities and processing times further revealed that even modest changes in workflow behavior can drive substantial recirculation and process delays. Network-based centrality analysis has identified Reserved, InProgress, and Completed as structurally important states, which act as key hubs and potential bottlenecks within the order workflow. All the findings highlight the effectiveness and validity of simulation-based approaches in elucidating the interplay among routing logic, capacity constraints, and overall system performance.
This work provides several actionable insights for healthcare operations leaders and HIT professionals, directly grounded in the simulation and what-if scenario results:
-
Use simulation for safe policy testing and workflow design: The demonstrated what-if scenarios show that the simulation framework can serve as a virtual controllable testing environment for evaluating proposed changes to workflow, policy, or resource allocation, which enables informed decisions prior to real-world implementation.
-
Plan for scalable surge response: Simulation results under order surge conditions highlight the importance of contingency strategies such as adaptive prioritization protocols, staffing flexibility and elastic processing capacity are critical to sustain desired performance under stress.
-
Monitor critical states to detect emerging congestion: Centrality analysis identified critical states such as Reserved, InProgress, and Completed. Monitoring metrics like active order volume or dwell time in these states (e.g., using rolling average) can enable earlier detection of congestion and support timely intervention.
-
Prevent rework cycles by identifying looping transitions: In transition dynamics analysis, looping behaviors and extended process durations emerged. Automatic flagging excessive recirculation or order re-entries can help detect inefficiencies and trigger corrective workflows or alerts.
-
Leverage transition dynamics as early warning indicators: Shifts in transition duration or path probabilities can serve as leading indicators of workflow drift. Tracking these metrics over time supports proactive operational adjustments to scheduling, workload distribution, or policy enforcement.
Several limitations should be acknowledged and explored in future work. In this study, the order states were aggregated at a high level, which potentially obscures finer-grained transitions that could further clarify more process dynamics. The simulation outputs, including delays, rework loops, and order path variants also provide clinically interpretable signals that can guide workflow optimization in practice. Although the developed framework does not model detailed task level clinical activities, it serves as a system level diagnostic tool that highlights where operational bottlenecks arise and where more detailed clinical investigation may be warranted. The detailed process within each state should be investigated more thoroughly to enable effective root cause analysis. Transition probabilities and durations were assumed to remain stationary over time, which may not fully capture evolving practice patterns or system updates. Additionally, factors such as user behavior, organizational culture, and policy constraints were beyond the scope of this analysis. For future studies, the simulation can be expanded to model a more comprehensive healthcare network by incorporating additional healthcare providers and simulating referral pathways. Another promising direction is to integrate online predictive models with adaptive simulation to support real-time monitoring and decision support.
Overall, this study demonstrates that DES offers a robust quantitative framework to analyze EHR order processing performance, identify structural bottlenecks, and evaluate the impact of operational changes before deployment. By enabling data-driven scenario analysis and highlighting actionable insights, this approach supports evidence-based decision-making to enhance the efficiency of electronic order workflows in complex healthcare environments.
While the current framework is implemented using data from VHA EHR system, its core components are designed to be broadly adaptable to other healthcare settings. The DES engine and network based analysis pipeline are data-driven and modular, which can be transferable to systems given timestamped order level event logs. Key elements that can be generalizable include the structure of the simulation model and scenarios design, order path variation discovery and the network centrality analysis methods. However, two main areas need further adaptation: 1) the mapping of raw order events. The used OASIS-derived states mapping needs to be reconsidered since order process states may vary by institutions or HIT vendors; 2) transition probabilities and durations may needs different calibration approach based on local data characteristics and practice guidelines. With these customization, the framework can support EHR order workflow evaluation in a variety of healthcare environments.
Methods
EHR order flow simulation framework
Fully representing EHR systems would require modeling numerous detailed clinical tasks, workflows, staff assignment and operational constraints. The intention here is to develop a framework that can provides system level abstraction that first localizes where delays, rework loops or anomalies occur which serves as actionable entry points for investigators to conduct more detailed root cause analysis.
In the proposed discrete event–based EHR order flow simulation framework (Fig. 11), patient order information is stored in data warehouses. These EHR order records are preprocessed through denoising, mapping, and segmentation steps. Key process patterns are then extracted by using process mining to construct the order event handling workflows. Demand prediction and resources required at each workflow state are also incorporated into the simulation as configuration settings. Institutional policies serve as the underlying logic that guides order progression in the network.
Historical EHR-derived event sequences serve as the primary input to construct the order process flow. Demand and resources data provide additional configuration settings. After calibration and validation, the simulation output can be used to support decision-making in planning, surveillance, and diagnostic applications.
After the simulation model is calibrated and validated with domain experts, statistical indicators derived from simulation results can be used to evaluate policies and plan related resources. In rolling-window simulations, deviations in key metrics can be used as alerts for potential change points or anomalies in the system. Additionally, when abnormal delays or deviations are detected in the order pathways, the simulation can be easily rerun under counterfactual conditions to isolate the components or processes that are most likely responsible for the disruption. Compared with raw EHR-order, data-based retrospective detection methods, this simulation capability enhances and encourages proactively identifying and mitigating systemic inefficiencies and failures.
With the integration of real-time resource data streams and predictive capabilities, the framework has the potential to evolve into a healthcare digital twin, which could continuously mirror the operational state of the system to support real-time decision-making in an operational environment.
Raw data mapping of EHR order
To simulate the electronic order flow, discrete events of the clinical orders must be accurately and uniformly reconstructed. Taking the VA’s CDW as an example, it contains extensive clinical data, yet the event sequences for patients or computerized provider order entries are not readily available. The primary challenges in extracting meaningful sequences stem from the relevant information typically being dispersed across disparate database tables. This information can sometimes only be discerned through examination and denormalization of the data, making it an arduous and laborious task to effectively consolidate and interpret event sequences. After identifying all timestamp records that are assumed to represent a transaction on an order, all tables that have some form of activity or status update associated with that order are recorded, and then the raw data is obtained as a time-sorted event sequence 37. This data mining process can be applied to the orders in different domains (e.g., Consults, Radiology, Labs, outpatient prescriptions) to obtain the order event sequence 38. For example, a sample Consult order event sequence is shown in Fig. 12. The real date and time are masked for privacy protection.
To enhance the interpretability of event logs, this study adopted the state transition diagram from the OASIS Web Services human task standard. This approach maps various process-related activities throughout their life cycles into OASIS human task state transitions. In collaboration with subject matter experts working with the VistA EHR system and the VHA CDW, this study established mapping rules for Consult orders (summarized in Table 4). Specific DateTime columns and consult activities are aligned with corresponding OASIS states, as depicted in Fig. 13.
The diagram illustrates the standardized life cycle of an order, including common paths from Created through Completed as well as less frequent transitions to Failed, Exited, or Error. States and transitions were defined in collaboration with subject matter experts to reflect operational realities in the VHA EHR system.
This mapped state transition diagram effectively reduces dataset complexity and provides actionable insights. Domain experts can utilize the diagram to identify irregular and rare transitions by analyzing transition path statistics. The dataset analysis reveals that most orders follow a standard transition pattern: Created, Ready, Reserved, InProgress, and Completed. However, irregular and infrequent transitions to states such as Failed, Exited, and Error may indicate potential issues. A sampled order case from the VHA CDW that concluded with a Failed state is presented in Table 5.
A more detailed explanation of the data extraction and mapping process can be found in previous studies 37,38,39,40. This study focuses specifically on EHR orders within the Consults domain.
Simulation setup
By mirroring the mapped OASIS state transition diagram, DES is designed to simulate the order flow of selected facilities and assess their system performance. This approach enables a dynamic evaluation of process efficiency to enable the identification of bottlenecks and irregular transition patterns within the Consults workflow. Based on the availability of order data in the Consults domain, facilities at different complexity levels were selected. The selected date range spans from January 2017 to June 2017, and the monthly distribution of created orders is shown in Fig. 14.
The Clinical Complexity Index classifies healthcare facilities as a range of Level 1a (highest complexity) to Level 3 (lowest complexity) based on patient volume and risk, scope of clinical services, and extent of research and teaching activities42. The facility classifications represented above include Level 1a (Facility A, 50,398 orders), Level 1b (Facility B, 45,972 orders), Level 1c (Facility C, 46,634 orders), Level 2 (Facility D, 24,694 orders), and Level 3 (Facility E, 16,221 orders). The observation period spanned from January to June of 2017. Higher-complexity facilities generally processed larger volumes of Consult orders each month.
For a clearer illustration of the state transitions in the OASIS mapping of raw data, Facility D’s frequency–directly follows graph is shown in Fig. 15. This graph was generated by using the PM4PY process mining package41. For example, the raw order data in Facility D shows that, among the total 24,694 orders, 24,683 orders transitioned from the Created state to the Ready state with a mean transition time of 1 h. Additionally, 4 orders transitioned directly to the Completed state with an average duration of 3 months, and 4 orders transitioned to the Error state with an average duration of 5 months. Notably, the total number of outgoing transitions from a given state may be lower than the total number of orders in that state because some orders may still be active during data capture.
Each node represents an order state, and the number inside the node is the total number of orders that entered that state during the observation period. Directed edges between nodes represent observed transitions between states, with the order number labeled on that edge. Edge thickness is proportional to transition frequency to emphasize high volume transitions. This frequency-based visualization aids in identifying dominant pathways and rare transitions within the Consult workflow for a selected facility. Time performance (e.g., average transition time) can also be added to the edge.
Data availability
The EHR order data used in this study was derived from VHA systems and is not publicly available owing to patient privacy and data use restrictions. Access may be granted upon request and approval by the VA.
Code availability
The simulation framework and analysis scripts developed for this study are available from the corresponding author upon reasonable request.
References
U.S. Department of Veterans Affairs. Veterans Health Administration. https://www.va.gov/health/aboutvha.asp (2025).
U.S. Department of Veterans Affairs. History of IT at VA. https://www.oit.va.gov/about/history.cfm? (2025).
U.S. Department of Veterans Affairs.Va health systems research. https://www.hsrd.research.va.gov/for_researchers/cdw.cfm (2025).
Landi, H. BREAKING: VA Will Replace VistA with Cerner EHR and Intends to Create Integrated Platform. https://www.hcinnovationgroup.com/cybersecurity/article/13028723/breaking-va-will-replace-vista-with-cerner-ehr-and-intends-to-create-integrated-platform (2017).
U.S. Department of Veterans Affairs. Transforming Health Care for Veterans, Revolutionizing Health Care for All https://digital.va.gov/ehr-modernization/ (2025).
Ahlness, E. A. et al. Impacts of an electronic health record transition on veterans health administration health professions trainee experience. J. Gen. Intern. Med. 38, 1031–1039 (2023).
Cogan, A. M. et al. Using research to transform electronic health record modernization: advancing a VA partnered research agenda to increase research impacts. J. Gen. Intern. Med. 38, 965–973 (2023).
Laster, R. J. Improving the Department of Veterans Affairs Medical Centers electronic health record implementation. Thesis, College of Management and Human Potential (2024).
De Roock, E. & Martin, N. Process mining in healthcare—an updated perspective on the state of the art. J. Biomed. Inform. 127, 103995 (2022).
Mielczarek, B. Review of modelling approaches for healthcare simulation. Oper. Res. Decis. 26, 55–72 (2016).
Günal, M. M. & Pidd, M. Discrete event simulation for performance modelling in health care: a review of the literature. J. Simul. 4, 42–51 (2010).
Kalton, A. et al. Multi-agent-based simulation of a complex ecosystem of mental health care. J. Med. Syst. 40, 39 (2016).
Vahdat, V., Griffin, J. A., Stahl, J. E. & Yang, F. C. Analysis of the effects of EHR implementation on timeliness of care in a dermatology clinic: a simulation study. J. Am. Med. Inform. Assoc. 25, 827–832 (2018).
Simwita, Y. W. & Helgheim, B. I. Simulation analysis of resource flexibility on healthcare processes. J. Multidiscip. Healthc. 9, 519–528 (2016).
Vazquez-Serrano, J. I., Peimbert-Garcia, R. E. & Cardenas-Barron, L. E. Discrete-event simulation modeling in healthcare: a comprehensive review. Int. J. Environ. Res. Public Health 18, 12262 (2021).
Djanatliev, A. & German, R. Prospective healthcare decision-making by combined system dynamics, discrete-event and agent-based simulation. 2013 Winter Simulation Conference (WSC) 270–281 (2013).
dos Santos, V. H., Kotiadis, K. & Scaparra, M. P. A review of hybrid simulation in healthcare. 2020 Winter Simulation Conference (WSC) 1004–1015 (2020).
Lame, G. & Dixon-Woods, M. Using clinical simulation to study how to improve quality and safety in healthcare. BMJ Simul. Technol. Enhanc. Learn. 6, 87–94 (2020).
Salas, E., Arias, M., Aguirre, S. & Rojas, E. Combining process mining and process simulation in healthcare: a literature review. IEEE Access 12, 172562–172580 (2024).
Baril, C., Gascon, V., Miller, J. & Côté, N. Use of a discrete-event simulation in a kaizen event: a case study in healthcare. Eur. J. Operat. Res. 249, 327–339 (2016).
Kovalchuk, S. V., Funkner, A. A., Metsker, O. G. & Yakovlev, A. N. Simulation of patient flow in multiple healthcare units using process and data mining techniques for model identification. J. Biomed. Inform. 82, 128–142 (2018).
Tako, A. A. & Kotiadis, K. Partisim: a multi-methodology framework to support facilitated simulation modelling in healthcare. Eur. J. Operat. Res. 244, 555–564 (2015).
Kuruppu Appuhamilage, G. D. K., Hussain, M., Zaman, M. & Ali Khan, W. A health digital twin framework for discrete event simulation based optimised critical care workflows. npj Digit. Med. 8, 376 (2025).
van Hulzen, G., Martin, N., Depaire, B. & Souverijns, G. Supporting capacity management decisions in healthcare using data-driven process simulation. J. Biomed. Inform. 129, 104060 (2022).
Sittig, D. F. et al. Current challenges in health information technology-related patient safety. Health Inform. J. 26, 181–189 (2020).
Wang, Y. et al. Measuring the effects of computer downtime on hospital pathology processes. J. Biomed. Inform. 59, 308–15 (2016).
Ancker, J. S. et al. Effects of workload, work complexity, and repeated alerts on alert fatigue in a clinical decision support system. BMC Med. Inform. Decis. Mak. 17, 36 (2017).
Olakotan, O. O. & Mohd Yusof, M. The appropriateness of clinical decision support systems alerts in supporting clinical workflows: a systematic review. Health Inform. J. 27, 14604582211007536 (2021).
Singh, H. & Sittig, D. F. A sociotechnical framework for safety-related electronic health record research reporting: The safer reporting framework. Ann. Intern. Med. 172, S92–S100 (2020).
Denham, C. R. et al. Safe use of electronic health records and health information technology systems: trust but verify. J. Patient Saf. 9, 177–89 (2013).
OASIS. Web Services—Human Task (WS-HumanTask) Specification Version 1.1 https://docs.oasis-open.org/bpel4people/ws-humantask-1.1-spec-cs-01.pdf (2010).
Viehmann, T. Numerically more stable computation of the p-values for the two-sample Kolmogorov-Smirnov test. arXiv preprint arXiv:2102.08037 (2021).
Conover, W. Practical nonparametric statistics (John Wiley & Sons, 1999).
Gustafson, J. L.Little’s Law, 1038–1041 https://doi.org/10.1007/978-0-387-09766-4_79 (Springer, 2011).
Peng, S. et al. Influence analysis in social networks: a survey. J. Network Comput. Appl. 106, 17–32 (2018).
Center, S. H. Data Mining and Machine Learning: The NETWORK Procedure. https://documentation.sas.com/doc/en/pgmsascdc/9.4_3.2/casmlnetwork/procnetwork_network_details16.htm (2025).
Ozmen, O. et al. Feature engineering and process mining to enable hazard detection in health information technology. AMIA Joint Summits on Translational Science Proceedings 469–476 (2020).
Klasky, H. B. & Ozmen, O. Process mining in healthcare-a case study for the corporate data warehouse of the Veterans Affairs Office. Report (2019).
Omitaomu, O. A. et al. A new methodological framework for hazard detection models in health information technology systems. J. Biomed. Inform. 124, 103937 (2021).
Niu, H. et al. Detecting anomalous sequences in electronic health records using higher-order tensor networks. J. Biomed. Inform. 135, 104219 (2022).
Berti, A., van Zelst, S. & Schuster, D. PM4py: A process mining library for Python. Softw. Impacts 17, 100556 (2023).
National Academies of Sciences, Engineering, and Medicine. and Education; Committee on Facilities Staffing Requirements for Veterans Health. Appendix C: Nature of Veterans Health Administration Facilities Management (Engineering) Tasks and Staffing (National Academies Press, 2019).
Acknowledgements
This quality improvement initiative is sponsored by the VA, is supported by and utilizes VA-funded resources from the Knowledge Discovery Infrastructure at Oak Ridge National Laboratory and is backed by the US Department of Energy’s (DOE’s) Office of Science. The manuscript has been authored by UT-Battelle LLC under contract DE-AC05-00OR22725 with the DOE. The US government retains and the publisher, by accepting the article for publication, acknowledges that the US government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this manuscript, or allow others to do so, for US government purposes. DOE will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (https://www.energy.gov/doe-public-access-plan). The initiative aimed to improve care for Veterans within the VHA and involved models specifically tailored to the VHA system. Therefore, it is not generalizable outside of the VHA and is officially considered non-research by the VHA. Oak Ridge National Laboratory received Institutional Review Board approval due to the secondary use of patient data, which complies with internal policies and standards. The authors acknowledge the broader partnership and express gratitude to the veterans receiving care at VA.
Author information
Authors and Affiliations
Contributions
Y.C. led the simulation development and drafted the manuscript. H.N. provided technical guidance and contributed to the methodological design. O.A.O. conceived the original idea, supervised the research activities, and managed the overall project. S.B. reviewed the manuscript and provided additional feedback. A.L., A.T., V.S., and J.R.N. served as domain experts on the VHA EHR data and sponsored the project. All authors reviewed and approved the final version of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Chen, Y., Niu, H., Omitaomu, O.A. et al. A simulation framework for evaluating electronic order workflows in integrated health records. npj Health Syst. 3, 14 (2026). https://doi.org/10.1038/s44401-026-00067-2
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s44401-026-00067-2

















