Abstract
As a well-known Bayesian network structure learning algorithm in equivalence class space (E-space), Greedy equivalence search (GES) is used in many fields. However, it encounters high complexity when searching for information from an empty graph. If the initial graph of GES is an equivalence class containing the strongest dependencies instead of an empty graph, its performance will be significantly improved. In this study, we propose a three-phase algorithm to establish the initial graph. First, we design a measure based on relative entropy to evaluate the relation between variables. Then, the variables are connected based on the previously designed metrics and the resulting graph is transformed into E-space. Finally, the resulting graph is used as the initial graph of GES for E-space optimization. We compare the proposed algorithm with GES in efficiency and accuracy, and the results show that our algorithm improves the efficiency and accuracy of GES. Furthermore, extensive comparisons are designed to compare our method with other state-of-the-art methods on benchmarks and real data about COVID-19 pandemic in the UK.
Introduction
Discovering the causal mechanisms behind data is a fundamental challenge in the field of artificial intelligence. Bayesian networks (BNs)1, as a type of probabilistic graphical model, provide an intuitive representation of the causal relationships among random variables. Hence, BNs have been applied in a wide variety of domains in the past decades, such as prediction of protein-protein2, explainability of chemical sciences3, time series forecasting4, Earth systems5, and telecom-fraud risk warnings6.
BNs encode the variables and independent relationships between variables as nodes and edges between nodes, respectively, which forms a directed acyclic graph (DAG) that becomes the graphical structure of a BN. The process of recovering the DAG corresponding to a BN is called BN structure learning (BNSL). Existing BNSL methods can be divided into function-based, constraint-based (CB), score-and-search (S&S), and hybrid algorithms. The function-based methods describe the relationship between variables using the given function, such as the linear non-Gaussian acyclic model (LiNGAM)7, post-nonlinear (PNL)8, or additive noise models9, among others. These methods usually perform well in the case of continuous variables that satisfy their given assumptions. However, when the precise function does not apply in the case of discrete variables, identifying the direction of causality becomes intractable10. CB approaches generally use a series of hypothesis tests to determine the conditional independence relationships between variables. The most popular CB methods include PC11, IAMB12, and their improved versions or variants13,14,15. S&S approaches rely on well-defined scoring functions such as BIC/MDL16,17, BDeu18, and so on, to transform the BNSL problem into a task of searching for the DAG with the highest score. S&S methods can be further divided into two classes of algorithms, according to their different search strategies: exact search algorithms and local search algorithms. Although the former can find globally optimal solutions, such as GOBNILP19,20, A*21, and their variants22,23, they hardly scale beyond 60 nodes; thus, local search algorithms are becoming increasingly popular. BNSL methods using local search typically search for optimal structures in three spaces: DAG space (D-space), equivalent class (EC, introduced in “Equivalence class” section) space (E-space), and ordering space (O-space). D-space methods operate on individual edges in the DAG until a locally optimal score is reached, such as Hill-climbing (HC)24 and its more complex version25,26. However, the number of candidate structures in D-space grows super-exponentially as the number of variables increases. The number of ECs is generally less than the number of DAGs, and some of the inefficient operations of D-space can be avoided in E-space, so E-space is much simpler than D-space (For more details, refer to Section 2.3 in27.)27,28. The most representative algorithm in E-space is greedy equivalence search (GES)29, which updates the current EC to another EC with a higher score. Order-based search (OBS)30 provides the third search space for BNSL, and it uses a topological order to denote a DAG. Although some excellent O-space methods have been proposed31,32, O-space is rooted in consistency rules for node order, which provides easier optimization but also limits its performance31,32. Naturally, there are approaches to combine CB and S&S algorithms, and the most representative form is Max-min HC (MMHC)33. However, these algorithms do not perform better than CB and S&S methods34. Therefore, this study focuses on E-space.
Since GES was proposed, researchers have studied how to improve its performance in the past two decades. Chickering designed two clever operators and refined the structure of GES to make it a baseline algorithm for E-space27,35, upon which subsequent E-space algorithms have been proposed. Nielsen et al. slightly improved the performance of GES by making a tradeoff between randomness and greediness36. Chen et al. combined ordering and EC to apply the genetic algorithm to E-space37. Zhang et al. used the maximal information coefficient to determine the draft and modified the draft to obtain a higher score38. Alonso et al. restricted the search space of the first phase of GES to improve efficiency39. Ramsey et al. refined the score cache and parallelized the search phases, significantly speeding up GES40. Alonso et al. modified the changes in scoring at each step to incorporate some excellent metaheuristic algorithms, improving accuracy41. Nandy et al. proposed a hybrid E-space algorithm by combining GES and CB methods for high-dimensional continuous data42. Liu et al. recently designed two frameworks to transfer the local optimum of E-space to other spaces, which improves the accuracy of GES43. Liu et al. proposed a hybrid algorithm based on mutual information for discrete data44. Laborda et al. used GES as a local solver for distributed parallel optimization45. Chen et al. developed an algorithm to find the k-best EC using dynamic programming and model averaging46. Chen et al. used A* to search for the globally optimal EC on networks with fewer than 30 variables47.
Zheng et al. recently transformed the BNSL problem into a continuous optimization process48, which serves as a novel way to discover the DAG. Many works have used different deep learning methods to improve performance from Zheng’s perspective49,50,51,52. However, some comparison results show that these methods generally perform poorly on discrete data25,49.
In this study, we focus on improving the performance of GES in terms of both efficiency and accuracy on discrete data. When GES starts searching from an empty graph, the search space is very complex. Naturally, a high-level intuition is that the efficiency will be improved significantly if GES searches for the optimal EC from an excellent initial graph. Therefore, we first design a metric to measure the degree of dependence between variables based on relative entropy. Then, a framework containing connection rules according to the quantified dependency strengths is proposed. Finally, following these linking rules, the variables are connected and converted into EC as the initial graph of GES for optimization. Our framework can also simultaneously improve the local optimum of the forward phase of GES. We conduct extensive comparisons using the proposed method, GES, and other state-of-the-art algorithms to reveal the performance of the proposed method.
The remainder of this paper is organized as follows: “Preliminaries” section provides the information about graph theory, BN, and EC. “Methods” section describes how GES works, and proposes our novel algorithm with a new metric and introduces it in detail. “Results” section compares the proposed method with GES and other state-of-the-art algorithms on well-known benchmarks and a real dataset. “Discussion” section summarizes the paper and gives future directions for our research.
Preliminaries
Graph theory
A graph \(G=(\textbf{V},\textbf{E})\) is composed of sets of edges \(\textbf{E}\) and nodes \(\textbf{V}\). If two arbitrary nodes \(X,Y\in \textbf{V}\) are connected by an undirected edge \(X-Y\), they are neighbors, and the set of the neighbors of X is \(\textbf{N}{{\textbf{e}}_{X}}\). We say that Y is a parent of X if there is a directed edge pointing from Y to X, such as \(Y\rightarrow X\). All parents of X form a parent set, expressed as \(\textbf{P}{{\textbf{a}}_{X}}\), and the maximum value of \(\left| \textbf{P}{{\textbf{a}}_{X}} \right|\) in G is called the maximum in-degree \({{d}_{G}}\). Regardless of whether the edges between X and Y are directed or undirected, we consider X and Y to be adjacent, and all nodes adjacent to X are denoted as \(\textbf{A}{{\textbf{d}}_{X}}\). If some nodes are pairwise adjacent, they form a clique. A V-structure is a subgraph formed by three nodes, where two non-adjacent nodes are parents of the third node, that is, three nodes are connected as \(Y\rightarrow X\leftarrow Z\).
\(X\mapsto Y\) is a path from X along some directed or undirected edges that end up at Y. If the edges of the path are all undirected edges, then \(X\mapsto Y\) is an undirected path. Similarly, \(X\mapsto Y\) is a directed path if these edges are directed from the previous node to the next node in the path. Otherwise, \(X\mapsto Y\) is considered a semi-directed path if there is at least one directed edge and one undirected edge in the path. In a directed path \(X\mapsto Y\), all nodes following X are descendants of X. A graph G containing a cycle means that there is a path \(X\mapsto Y\) in G and X that is adjacent to Y. G is a DAG if there is no cycle in G and every edge in G is directed. Similarly, if there is at least one semi-directed path in G that does not contain any cycle, G is a partially DAG (PDAG).
Bayesian networks
A BN is a probabilistic graphical model that can be expressed as \(BN=(G,\mathbb {P})\), where G is a DAG and \(\mathbb {P}\) is its joint probability distribution. Let \(\textbf{N}{{\textbf{d}}_{X}}\) denote all descendants of X. G encodes the conditional independence assumptions as follows:
Based on the conditional independences and chain rule of probability, the joint probability P over \(\textbf{V}\) can be factorized as:
where \(\left| \textbf{V} \right|\) is the number of nodes, and Eq. (2) obviously simplifies the complexity of computing joint probability.
S&S methods allow the BNSL problem to be represented as follows:
where \(\textbf{G}\) is all candidate structures, \({{G}^{*}}\) is the optimal structure, and \(\textbf{D}\) is the data to be studied. In this paper, we use BIC, one of the most used scores, to estimate the learned structure. BIC has a fundamental property: it is decomposable, which means the BIC of a BN can be decomposed into the sum of the scores of each node with its parents, that is,
where
For discrete observational data:
In this equation, \({{\Omega }_{X}}\) and \({{\Omega }_{\textbf{P}{{\textbf{a}}_{X}}}}\) represent the state spaces of X and \(\textbf{P}{{\textbf{a}}_{X}}\), respectively; \({{N}_{x,pa}}\) denotes is the number of instances with \(X=x\wedge \textbf{P}{{\textbf{a}}_{X}}=pa\) in \(\textbf{D}\), so we omit \(\textbf{D}\) in the expansion of the equation; \(\log {{\tilde{\theta }}_{x|pa}}\) is the log-likelihood estimation of the conditional probability \(P(X=x|\textbf{P}{{\textbf{a}}_{X}}=pa)\).
Then, we introduce the relationship between the graph G and distribution \(\mathbb {P}\). Let I(G) and \(I(\mathbb {P})\) denote the sets of conditional independence assertions entailed by G and \(\mathbb {P}\), respectively. We say that G is a perfect map (p-map) of \(\mathbb {P}\) if and only if \(I(G) = I(\mathbb {P})\). Let \(\varepsilon (G)\) denote the equivalence class (CPDAG) of a DAG G.
Equivalence class
If the independence relationships and distributions of G and \(G'\) from two DAGs are identical, they are equivalent to each other. The skeleton of a DAG is an undirected graph that transforms the directed edges of the DAG into undirected edges. Based on skeleton and v-structures, Verma and Pearl prove the following theorem:
Theorem 1
53 Two DAGs are equivalent iff they share the same skeleton and v-structures.
If a PDAG \({{G}_{P}}\) has the same skeleton and v-structures as a DAG G, and all directed edges in \({{G}_{P}}\) also exist in G, we say that the PDAG admits a consistent extension and the DAG is a consistent extension of the PDAG.
All equivalent DAGs form their equivalent classes (EC), denoted by \(\varepsilon\). Obviously, PDAG is not a precise representation of EC because we cannot ensure that every DAG in \(\varepsilon\) is a consistent extension of the PDAG corresponding to one DAG in \(\varepsilon\). To get a suitable representation of \(\varepsilon\) shown in Definition 2, we distinguish the edges of the DAGs in \(\varepsilon\) as in Definition 1.
Definition 1
If a directed edge exists in every DAG in \(\varepsilon\), the edge is compelled. Otherwise, the edge is reversible, that is, this edge exists in at least two of the DAGs of \(\varepsilon\) in the opposite direction.
Definition 2
A PDAG is a completed PDAG (CPDAG) if every directed edge of the PDAG is compelled and all undirected edges are reversible.
Based on Definition 2, any DAG in \(\varepsilon\) is a consistent extension of the CPDAG corresponding to \(\varepsilon\), so we also use \(\varepsilon\) to denote the CPDAG corresponding to \(\varepsilon\) in the rest of this paper. Figure 1 shows a simple DAG, its skeleton, the CPDAG corresponding to the EC to which it belongs, and another consistent extension of the CPDAG.
With these definitions, we introduce the notion of a consistent score function. Let \(G^{*}\) denote a p-map of \(\mathbb {P}\), and let G be an arbitrary DAG that is not a p-map. A score function is consistent if, in the limit as \(|\textbf{D} |\rightarrow \infty\), the following hold:
-
\(G^{*}\) attains the highest score;
-
The score of any non-p-map DAG G is strictly less than that of any DAG in \(\varepsilon (G^{*})\), i.e.,
$$\begin{aligned} score(G, \textbf{D}) < score(H, \textbf{D}), \quad \forall H \in \varepsilon (G^{*}). \end{aligned}$$
Theorem 2
35 BIC is a consistent score function.
Methods
The original greedy equivalent search
GES searches for the EC with the highest score in two phases (The consistency of GES relates to the nature of the perfect map and the consistency of the score (see54 for a detailed introduction), for detailed proof, see Sections 4.2 and 4.3 of35.) In the first phase, called forward equivalent search (FES), iterative updates to the current MEC occur by adding one edge using a well-designed operator \(Insert(X,Y,\textbf{T})\). The second phase is called Backward Equivalent Search (BES), and it deletes one edge in one operation via the operator \(Delete(X,Y,\textbf{H})\). The operations \(Insert(X,Y,\textbf{T})\) and \(Delete(X,Y,\textbf{H})\) are shown in Definition 3 and Definition 4.
Definition 3
For any non-adjacent variables X and Y, and any subset \(\textbf{T}\) of the set \(\textbf{Nn}{{\textbf{A}}_{Y,X}}\) (\(\textbf{Nn}{{\textbf{A}}_{Y,X}}=\textbf{N}{{\textbf{e}}_{Y}}\bigcap \{\textbf{V}\backslash \textbf{A}{{\textbf{d}}_{X}}\}\)), \(Insert(X,Y,\textbf{T})\) updates the current \(\varepsilon\) in the following ways:
-
Adding the edge \(X\rightarrow Y\);
-
Changing the undirected edges between Y and any \(T\in \textbf{T}\) into directed edges \(T\rightarrow Y\).
Definition 4
For any adjacent variables X and Y, and any subset \(\textbf{H}\) of the set \(\textbf{N}{{\textbf{A}}_{Y,X}}\) (\(\textbf{N}{{\textbf{A}}_{Y,X}}=\textbf{N}{{\textbf{e}}_{Y}}\bigcap \textbf{A}{{\textbf{d}}_{X}}\)), \(Delete(X,Y,\textbf{H})\) operates the current \(\varepsilon\) in the following ways:
-
Deleting the edge \(X\rightarrow Y\) or \(X-Y\);
-
Changing the original undirected edges \(Y-H\) and \(X-H\) for any \(H\in \textbf{H}\) into the directed edges \(Y\rightarrow H\) and \(X\rightarrow H\).
Algorithm 1 shows the outline of FES. Since the results of \(Insert(X,Y,\textbf{T})\) are not necessarily CPDAG, the function \(PDAG\_To\_CPDAG\) in Line 7 transforms the PDAG into the corresponding CPDAG. It is combined by two functions, \(PDAG\_To\_DAG\)55 and \(DAG\_To\_CPDAG\)56. The former extracts a consistent extension from the PDAG, and the latter transforms a DAG into the CPDAG to which it belongs. The operation of BES is similar to FES. We can obtain BES by replacing the validity tests (in Line 15), local score calculation criterion (in Line 16), and the operation (in Line 18) of \(Insert(X,Y,\textbf{T})\) in Algorithm 1 with those of \(Delete(X,Y,\textbf{H})\). The validity tests of \(Insert(X,Y,\textbf{T})\) and \(Delete(X,Y,\textbf{H})\) test how to determine whether or not an \(Insert(X,Y,\textbf{T})\) or \(Delete(X,Y,\textbf{H})\) operator is valid (more details and proof refer to Theorems 15 and 17 in35.) The local score changes of two operators are introduced in Theorems 3 and 4.
Theorem 3
The changes in BIC that result from applying a valid operator \(Insert(X,Y,\textbf{T})\) to a CPDAG \(\varepsilon\) is
Theorem 4
The increments in BIC that result from applying a valid operator \(Delete(X,Y,\textbf{H})\) to a CPDAG \(\varepsilon\) is
Based on the above, GES can be simply expressed as \(GES:\varepsilon *\leftarrow BES(FES(\textbf{D},{{\varepsilon }_{empty}})),\) where \({{\varepsilon }_{empty}}\) is an empty graph because the CPDAG of the EC to which the empty graph belongs is still an empty graph based on Definition 2.
Theorem 5
35 If the distribution \(\mathbb {P}\) admits a p-map and \(\varepsilon ^{*}\) denotes the CPDAG found by GES with a consistent score function, then in the limit of large \(|\textbf{D} |\), \(\varepsilon ^{*} \text { is a p-map of } \mathbb {P}.\)
The proposed greedy equivalent search
When GES starts from an empty graph, it is inefficient as the search space is very complex. Naturally, if GES searches for the optimal EC from an excellent seed EC with some strong dependency edges, the search space will be restricted, and the efficiency will be improved. Therefore, for the proposed method, a metric to evaluate the relationships between variables is first designed. Then, we connect the most correlated variables to form a preliminary undirected graph. After determining the v-structure, we obtain an excellent seed graph of GES.
Considering the prior probability P(X) of a variable X and the posterior probability P(X|Y) of that variable given another variable Y, we can use P(X) and P(X|Y) to measure the relationship between variables X and Y. Recall the meaning of relative entropy (RE): the difference between two probability distributions R and Q, which is formulated as follows:
Based on RE, we design a metric, prior-posterior relative entropy (PPRE), to quantify the strength of dependencies between X and Y, as follows:
We analyze why PPRE can determine variable dependencies in terms of both meaning and formula. On one hand, PPRE reflects the difference between P(X) and P(X|Y). That is, the greater the PPRE, the greater the effect of Y on X, indicating a stronger relationship between the variables Y and X. On the other hand, the posterior probability is substituted into Eq. (11) in the form of joint and marginal distribution, that is,
Equation (12) represents that PPRE(X, Y) is mainly affected by the product of the marginal distributions P(X)P(Y) and their joint distribution P(X, Y), which indicates the extent to which the uncertainty about X is reduced given that Y is known. To clarify, the proposed PPRE metric is defined as RE between the prior distribution P(X) and the posterior distribution P(X|Y). This direct comparison of prior and posterior distributions naturally captures the effect of Y on X, reflecting the conditional dependency between variables. Unlike general-purpose divergence measures that compare arbitrary distributions, PPRE is specifically designed to quantify variable dependencies in Bayesian network structure learning. PPRE is derived from RE, so it has three properties similar to RE.
Property 1
If X and Y are independent, then \(PPRE(X,Y)=0\).
Proof
If X and Y are independent, then \(P(X|Y)=P(X)\). Therefore, \(PPRE(X,Y)=0\) based on Eq. (11). \(\square\)
Let \(\textbf{I}{{\textbf{d}}_{X}}\) denote the set of variables independent of X, and then we get Property 2.
Property 2
For \(\forall Y\in \textbf{I}{{\textbf{d}}_{X}}\), iff the distributions of X and Y are identical, \(PPRE(X,Y)=PPRE(Y,X)\).
Proof
According to Eq. (12),
where \(Q(X,Y)=\left( P(X)P(Y)-P(X,Y) \right) \log \left( P(X)P(Y)-P(X,Y) \right)\). Since Y is independent of X, Q(X, Y) must be nonzero. Therefore, iff the distributions of X and Y are identical, Eq. (13) is zero, and the property is proven. \(\square\)
Property 3
\(PPRE(X,Y)\ge 0\)
Proof
The property naturally follows from the nonnegativity of the RE. \(\square\)
PPRE compares P(X) and \(P(X\mid Y)\) via a symmetrized RE for a fixed target X. Hence \(PPRE(X,Y)\ge 0\) and \(PPRE(X,Y)=0\) if \(P(X\mid Y)=P(X)\) (e.g., when \(X\perp Y\)). Note that PPRE(X, Y) and PPRE(Y, X) are generally different, since they quantify how Y affects X and how X affects Y, respectively. This directional asymmetry is used in the proposed algorithm through the strongest dependent sets to determine edge orientation.
For the evaluation criteria for dependencies, we designed a PPRE-based GES algorithm, called PPGES, the pseudo-code of which is given in Algorithm 2. The PPGES will be introduced in detail in stages. Before describing the pseudo-code, we clarify several notations used in Algorithm 2. \(\varepsilon *\) denotes the current partially directed acyclic graph (PDAG) constructed during the algorithm. It is initialized as the empty graph on the variable set \(\textbf{V}\), and will be updated iteratively until it becomes a local optimal CPDAG. \(\textbf{S}{{\textbf{D}}_{X}}\) denotes the strongest dependent set of variable X, which is a subset of candidate parent variables for X. It is obtained by ranking all variables \(Y \ne X\) according to PPRE(X, Y). cache is a temporary data structure storing the pairwise PPRE(X, Y) values between variable X and the remaining variables. For each X, cache is sorted in descending order of PPRE values to facilitate the selection of \(\textbf{S}{{\textbf{D}}_{X}}\).
Obviously, PPGES is a three-phase algorithm. The inputs of PPGES include the observation data \(\textbf{D}\), maximum in-degree of G \({{d}_{G}}\), and the threshold t. In the first phase (Lines 2-10), we compute the PPRE to determine the strongest dependent sets (expressed as \(\textbf{S}{{\textbf{D}}_{X}}\)) for each variable. Since subsequent operations may query the PPRE multiple times, the PPRE of each variable and the remaining variables is calculated and stored in the cache for efficiency, which has a complexity of \(O\left( \left| \textbf{D} \right| \cdot \left( \left| \textbf{V} \right| -1 \right) \right)\). In order to facilitate subsequent operation, PPGES sorts \(PPRE(X,\textbf{V}\text \!\!\backslash \!\!\text X),X\in \textbf{V}\) in the cache, the worst complexity of which is \(O\left( \left( \left| \textbf{V} \right| -1 \right) \cdot \log \left( \left| \textbf{V} \right| -1 \right) \right)\). After sorting, we select the first \({{d}_{G}}\) variables from the cache as the \(\textbf{S}{{\textbf{D}}_{X}}\) of the variable X. Finally, the variables with a PPRE greater than the threshold t were filtered out as the final \(\textbf{S}{{\textbf{D}}_{X}}\). As for the threshold t, we will obtain it statistically based on the PPRE of the standard network, and the determination of it will be discussed in “Results” section. The complexity of sorting and filtering is \(O\left( 2{{d}_{G}} \right)\). Therefore, for each variable \(X\in \textbf{V}\), the worst-case complexity is \(O\left( \left| \textbf{D} \right| \cdot \left| \textbf{V} \right| +\left| \textbf{V} \right| \cdot \log \left( \left| \textbf{V} \right| -1 \right) -\log \left( \left| \textbf{V} \right| -1 \right) \right)\). Naturally, the worst complexity of the first phase is \(O\left( {{\left| \textbf{V} \right| }^{2}} \right)\).
In the second phase (Lines 11-23), PPGES obtains a CPDAG with the strongest dependencies. Based on the set \(\textbf{S}{{\textbf{D}}_{X}}\), we conduct connect rules for the variables:
-
Rule 1. If \(X\in \textbf{S}{{\textbf{D}}_{Y}}\wedge Y\notin \textbf{S}{{\textbf{D}}_{X}}\), the algorithm applies \(AddEdge(X\rightarrow Y)\);
-
Rule 2. If \(X\notin \textbf{S}{{\textbf{D}}_{Y}}\wedge Y\in \textbf{S}{{\textbf{D}}_{X}}\), the algorithm applies \(AddEdge(X\leftarrow Y)\);
-
Rule 3. If \(X\in \textbf{S}{{\textbf{D}}_{Y}}\wedge Y\in \textbf{S}{{\textbf{D}}_{X}}\), the algorithm applies \(AddEdge(X-Y)\).
These connection rules are based on the conditional dependence information captured by PPRE. For Rules 1 and 2, if X strongly depends on Y but not vice versa, we direct the edge as \(X\leftarrow Y\), reflecting that knowing Y reduces the uncertainty of X more than the reverse. Rule 3 handles the case where both variables are strongly dependent on each other; since the direction cannot be reliably determined from PPRE alone, we add an undirected edge \(X-Y\), consistent with the PDAG representation in Bayesian networks. Overall, these rules ensure the resulting PDAG respects acyclicity and can serve as a sound seed graph for the subsequent GES optimization. While heuristic in implementation, they are theoretically motivated by conditional dependence and BN structure constraints. In order to prevent the repeated addition of edges, X is removed from \(\textbf{S}{{\textbf{D}}_{Y}}\) after connecting them. Since the operation of the first phase, for \(\forall X\in \textbf{V},\left| \textbf{S}{{\textbf{D}}_{X}} \right| \le {{d}_{G}}\), the worst case of the above operations’ complexity is \(O\left( d_{G}^{2}\cdot \left| \textbf{V} \right| \right)\). The function \(AddEdge(\cdot )\) in Lines 15 and 18 denotes the operation of adding an edge into the current graph. In our implementation, it is realized by functions such as set.edge() and set.arc() in the bnlearn package. After connecting their strongest dependent variables for each variable, the resulting graph is a PDAG. However, the required input for GES is CPDAG. Thus, the final step of the second phase transforms the resulting graph into a CPDAG using \(PDAG\_to\_CPDAG\), whose worst complexity is \(O\left( d_{G}^{2}\cdot \left| \textbf{V} \right| +d_{G}^{2}\cdot \left| \textbf{E} \right| +d_{G}^{2} \right)\), where \(\left| \textbf{E} \right|\) represents the number of edges in G. Therefore, the upper bound of the second phase’s complexity is \(O\left( \left| \textbf{V} \right| \right)\).
The final phase (Line 24) of PPGES uses the CPDAG obtained in the second phase as input to optimize the final CPDAG further using GES. Figure 2 shows an example of PPGES.
Results
In this section, we first introduce the backgrounds of method comparisons in “Backgrounds of comparisons” section, such as benchmarks, comparison algorithms, and hardware and software conditions. Then, “Comparisons on GES and PPGES” section compares the performance differences between GES and PPGES with different parameters. “Comparisons on PPGES and other algorithms” section shows the accuracy of the results of PPGES and other competitive algorithms. Finally, “Comparisons on real data” section applies the proposed algorithm and the state-of-the-art solver to a real dataset.
Backgrounds of comparisons
All benchmarks in “Comparisons on GES and PPGES” and “Comparisons on PPGES and other algorithms” sections are sampled from the ground-truth models, which can be downloaded from the Bayesian Network Repository (https://www.bnlearn.com/bnrepository/). We select two small networks (Asia, Sachs), four medium networks (Alarm, Child, Insurance, Water), and three large networks (Hailfinder, Hepar2, Win95pts), as well as the classification criteria are also from the Bayesian Network Repository. The characteristics of these real networks are given in Table 1.
All the state-of-the-art or well-established algorithms used for the comparison are briefly described as follows:
-
PC.stable13: The competitive CB baseline algorithm;
-
GOBNILP20: The state-of-the-art exact search solver;
-
GES35: The most representative E-space local search algorithm;
-
ILGES41: The state-of-the-art E-space local search algorithm;
-
ARGES42: The hybrid algorithm in E-space;
-
MMHC: The hybrid baseline algorithm in D-space;
-
MMPC+GES: The initial graph for GES is provided by MMPC;
-
MAHC26: A D-space local search algorithm using model averaging;
-
DAG-GNN49: The competitive continuous optimization baseline algorithm.
-
GOLEM51: A continuous optimization baseline algorithm.
-
DAGMA52: The competitive continuous optimization baseline algorithm.
PC.stable and MMHC are programmed using the R package bnlearn57 (https://www.bnlearn.com/). GOBNILP runs in C based on the source code (https://www.cs.york.ac.uk/aig/sw/gobnilp/). GES, ILGES, and the proposed PPGES are implemented in R based on bnlearn and pcalg. ARGES runs based on the R package pcalg58. MAHC runs in Java using the source code (https://bayesian-ai.eecs.qmul.ac.uk/bayesys/). ESTOBS runs in C++, leveraging the source code from its authors. DAG-GNN (https://github.comfi/shmoon1234/DAG-GNN), and DAGMA (https://github.com/kevinsbello/dagma) run in Python based on their source code. GOLEM is provided by the Python toolbox gCastle59. It is noted that GOBNILP plays a role in searching for the global optimum in the comparisons, which provides a reference for other algorithms. MMPC+GES combines the MMPC phase of MMHC with GES according to the operations of the authors of ARGES. The parameters of all the above algorithms use the default values from their papers. All comparisons are made on a computer with an Intel Core i7-10700f (2.9GHz) processor and 32GB of RAM.
In “Comparisons on GES and PPGES” section, the criteria of comparisons are BIC and run time. In “Comparisons on PPGES and other algorithms” section, the criteria are BIC and structural hamming distances (SHD). SHD is defined by the authors of MMHC to compare the differences between the learned CPDAG \({{\varepsilon }_{learned}}\) and the true CPDAG of BN \({{\varepsilon }_{true}}\), which can be computed as follows:
That is, SHD is the sum of the number of missing edges, extra edges, and edges with the opposite direction in the learned CPDAG compared to the true CPDAG.
Sensitivity analysis of PPGES parameters: comparisons with GES
To investigate how the parameters of PPGES affect its performance, we conduct a sensitivity analysis by comparing PPGES with GES under different parameter settings. Specifically, for the maximum in-degree, we consider three alternatives \({d_G} = \{1,3,6\}\). For the threshold parameter, we select the mean, median, and upper-hinge (the latter two are values in Tukey’s five-number summary60) of PPRE, which corresponds to the strongest variable dependence among all ground-truth models. Note that these threshold values are computed globally across all variables rather than individually. This yields nine combinations of parameters for PPGES, which are evaluated in this section.
We sample 10,000 instances for each benchmark network as this section’s data sets. On the one hand, the differences between GES and PPGES run time, \({\Delta _{time}} = {T_{PPGES}} - {T_{GES}}\), are used to show that PPGES improves the efficiency of GES with different parameters. On the other hand, \({\Delta _{BIC}} = {BIC_{PPGES}} - {BIC_{GES}}\) shows the effect of PPGES on the scores of GES under different parameters. The results are given in Figs. 3 and 4. The combination of parameters is represented as “in-degree=X,threshold=X”. For example, “d=3,t=upper” indicates the result when the maximum degree is set to 3, and the threshold is set to upper-hinge. The lighter bars in Figs. 3 and 4 indicate that PPGES improves the performance of GES, while the darker bars imply that PPGES does not perform as well as GES.
In each graph of Fig. 3, the results of PPGES with different parameters are lighter bars, which means that PPGES with different parameters achieve better efficiency. To the contrary, some combinations of parameters obtain worse scores than GES. In summary, the parameters we choose for PPGES are \({d_G} = 3\) and \(t = upper{\text { - }}hinge\). PPGES with the combination outperformed GES in 7/9 networks, and of those seven networks, four were the most accurate results. Meanwhile, PPGES with the combination performs well in terms of efficiency as shown in Fig. 3.
With the parameters, we compared the ratio in which PPGES improved efficiency with different amounts of data. Meanwhile, in this subsection, we compare the efficiency of MMPC+GES and PPGES, and their accuracy will be compared in “Comparisons on PPGES and other algorithms” section. In Fig. 5, the x-axis represents the amount of data, and the y-axis represents \(\frac{{{T}_{GES}}-{{T}_{algorithm}}}{{{T}_{GES}}}\).
Obviously, both MMPC+GES and PPGES can significantly improve the efficiency of GES. As a well-established CB method, MMPC is more efficient than the proposed algorithm. MMPC+GES runs faster than PPGES in 29/36 cases, and the differences between their results are less than 0.25 in 16/29 cases.
Comparisons on PPGES and other algorithms
In this section, we sample 1000, 5000, 10,000, and 100,000 instances for each ground-truth model. The BIC scores obtained by all the comparison algorithms are listed separately in Tables 2, 3, 4, and 5. The results of SHD are shown in Tables 6, 7, 8, and 9. The best score for each of the benchmarks in the table is written in bold, and two decimal places are reserved for BIC. If the algorithm goes wrong for some reason, the results will be represented as “–”.
In Table 2, although GOBNILP finds a global optimum on Asia, Sachs, and Child, it fails when the number of nodes increases. MAHC, DAG-GNN, GOLEM and DAGMA can stably output results, but their scores are far from those of other algorithms, especially in the Hailfinder model. PC.stable wins 1 of 9 models, but its performance is limited by the conditional independence tests. PPGES and ILGES obtain higher scores than GES in most models. In detail, ILGES outputs the Hepar2 model with the highest score, whereas PPGES outperforms other competitive solvers in 6/9 models. Moreover, although PPGES is not the best method for the Sachs and Hepar2 models, it also gets excellent scores that are very close to the highest score.
The results of 5000 instances are similar to those of 1000 instances, as shown in Table 3. ARGES successfully outputs a CPDAG in the Insurance model, but it fails in Table 2. It also maintains excellent performance in the Asia model but still obtains a poor result on the Alarm model. Obviously, PPGES still outperforms other algorithms on 6/9 models. Although PPGES is not the best algorithm on models Asia, Sachs, and Child models, its score is only 1.37%, 2.53%, and 0.43% away from the optimal score, respectively.
As the amount of data increases, algorithms perform better in all models, such as DAG-GNN, which scored only 0.47% worse than the highest score in the Water model. Obviously, PPGES stably outputs better CPDAG for 7/9 models with 10,000 instances. In these seven networks, PPGES improves from 0.30% to 2.11% over the suboptimal algorithm. For the Asia and Child models, PPGES is only 1.05% and 0.11% less than the global optimal, respectively.
In the comparison results of 100,000 instances, seven comparative algorithms search for the global optimal in the Sachs model, as shown in Table 5. PPGES performs better than the results in Tables 2, 3, and 4. In the results of the Child model, the proposed method outputs the same score as GOBNILP. Although MMPC+GES obtains the same result as GES for the Asia and Sachs models, it performs worse than GES for 6/9 models.
As for SHD, PPGES does not perform as well as in comparisons of BIC, especially in Table 6. That is because GES is an S&S algorithm, and the search for the best results is based on the score function. Similarly, the structures with global optimal scores found by GONBNILP do not necessarily have the lowest SHD. As the amount of data increased, the trend of BIC and SHD gradually matched, and PPGES performed better and better. In the results of Table 9, PPGES found the CPDAG on 6/9 models that were closest to the real structure. In contrast to the BIC comparison results, MAHC performed better in the SHD comparison, possibly because the method used the strategy of model averaging. Similar to BIC, MMPC+GES is still not better than GES in terms of any amount of data.
As for the comparisons between PPGES and MMPC+GES, the proposed algorithm can improve GES in terms of both efficiency and accuracy, while MMPC+GES only enhances efficiency and even worsens accuracy in many cases.
It is also worth noting that a better BIC score does not necessarily imply a lower SHD, because BIC and SHD evaluate different aspects of model quality. As a score-equivalent and consistent scoring function, BIC is guaranteed to asymptotically identify the p-map of the underlying distribution. In contrast, SHD is a purely combinatorial metric that counts edge-level differences between graphs, which makes SHD more sensitive to local structural variations, sample fluctuations, and characteristic of benchmark networks. Consequently, it is possible for BIC to show clear improvement while SHD appears unstable in certain cases. Nevertheless, in line with the asymptotic property of BIC (see Theorem 2), our experiments show that the SHD of PPGES improves with increasing sample size across most benchmark networks (see Tables 6–9).
Comparisons on real data
Since it is much more difficult to obtain the true causal structure from real datasets, in this section we only select PPGES and MAHC, which performed best in SHD in the previous comparisons, and apply them to a real dataset. This dataset is derived from a research61 that recorded the development of the 2019 coronavirus disease (COVID-19) in the UK over 866 days from January 2020 to June 2022, forming raw data with 866 rows and 18 columns. The raw data, processed data, processing methods, ground-truth graph, and data set descriptions are available from the literature61.
The scoring performances of PPGES and MAHC are listed in Table 10. Undoubtedly, PPGES has more advantages compared with MAHC on this real dataset, whether in terms of structure score or graphical evaluation. Then, to observe the causal structures obtained by the two methods more intuitively, the subgraphs of the causal structures they output are shown in Fig. 6. In Fig. 6a and b, the correct edges, the extra edges (the edges that do not exist in the ground-truth graph but are output by the algorithm), and the incorrect edges (the missing edges compared to the ground-truth graph and the edges in the opposite direction) are marked in black, blue dash lines, and red, respectively.
In detail, MAHC recovers five correct edges, all of which are also correctly discovered by PPGES. PPGES correctly determine the causal relationships, \(New\_infections \rightarrow Positive\_tests\) and \(Hospital\_admissions \rightarrow Patients\_in\_hospital\), while MAHC fails in these causality. Moreover, PPGES discovers the causal relationships between the policy node Lockdown and the mobility node \(Transportation\_activity\), and between the two mobility nodes \(Work\_and\_school\_activity\) and \(Transportation\_activity\). However, MAHC do not correctly search the edges \(Lockdown \rightarrow Transportation\_activity\) and \(Work\_and\_school\_activity \rightarrow Transportation\_activity\). PPGES incorrectly finds an undirected edge \(Positive\_tests - Majority\_COVID\_19\_variant\), while MAHC does not make the same mistake.
Discussion
In this study, we proposed an improved E-space algorithm called PPGES to improve the performance of GES. First, the strongest dependence relations were determined. In order to find the relations behind the data, we designed a metric called PPRE based on the relative entropy of the prior and posterior distributions. Then, for the proposed metric, we constructed a framework containing well-designed connection rules to obtain an excellent initial graph of GES. The connected graph is converted into E-space and further optimized using GES to obtain the final CPDAG. The experimental results reveal that the proposed algorithm improves the performance of GES in terms of time and scores. Moreover, PPGES outperformed other solvers in most models with different instances. Finally, PPGES also output a more accurate structure for a real dataset about COVID-19.
Our future works will focus on parallelizing the algorithm to further improve its efficiency and studying how to achieve the same excellent results under weak faithfulness conditions.
Data availability
All data generated or analysed during this study are included in this paper (including, but not limited to, repository name, author, and URL).
References
Pearl, J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference (Elsevier, 2014).
Jansen, R. et al. A Bayesian networks approach for predicting protein-protein interactions from genomic data. Science 302, 449–453 (2003).
Feng, J., Lansford, J. L., Katsoulakis, M. A. & Vlachos, D. G. Explainable and trustworthy artificial intelligence for correctable modeling in chemical sciences. Sci. Adv. 6, eabc3204 (2020).
Wang, B. & Liu, X. Fuzzy-probabilistic time series forecasting combining Bayesian network and fuzzy time series model. Symmetry 17, 275 (2025).
Phan, T. D., Smart, J. C., Capon, S. J., Hadwen, W. L. & Sahin, O. Applications of Bayesian belief networks in water resource management: A systematic review. Environ. Model. Softw. 85, 98–111 (2016).
Hu, M., Li, X., Li, M., Zhu, R. & Si, B. A framework for analyzing fraud risk warning and interference effects by fusing multivariate heterogeneous data: A Bayesian belief network. Entropy 25, 892 (2023).
Shimizu, S., Hoyer, P. O., Hyvärinen, A., Kerminen, A. & Jordan, M. A linear non-Gaussian acyclic model for causal discovery. J. Mach. Learn. Res. 7 (2006).
ZHANG, K. On the identifiability of the post-nonlinear causal model. In Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence (UAI), Vol. 647 (AUAI Press, 2009).
Hoyer, P., Janzing, D., Mooij, J. M., Peters, J. & Schölkopf, B. Nonlinear causal discovery with additive noise models. Adv. Neural Inf. Process. Syst. 21 (2008).
Glymour, C., Zhang, K. & Spirtes, P. Review of causal discovery methods based on graphical models. Front. Genet. 10, 524 (2019).
Spirtes, P. & Glymour, C. An algorithm for fast recovery of sparse causal graphs. Soc. Sci. Comput. Rev. 9, 62–72 (1991).
Tsamardinos, I., Aliferis, C. F., Statnikov, A. R. & Statnikov, E. Algorithms for large scale Markov blanket discovery. In FLAIRS, Vol. 2, 376–381 (2003).
Colombo, D. et al. Order-independent constraint-based causal structure learning. J. Mach. Learn. Res. 15, 3741–3782 (2014).
Wu, X. et al. Accurate Markov boundary discovery for causal feature selection. IEEE Trans. Cybern. 50, 4983–4996 (2019).
Jaber, A., Kocaoglu, M., Shanmugam, K. & Bareinboim, E. Causal discovery from soft interventions with unknown targets: Characterization and learning. Adv. Neural. Inf. Process. Syst. 33, 9551–9561 (2020).
Schwarz, G. Estimating the dimension of a model. Ann. Stat. 461–464 (1978).
Suzuki, J. A construction of bayesian networks from databases based on an mdl principle. In Uncertainty in Artificial Intelligence, 266–273 (Elsevier, 1993).
Buntine, W. Theory refinement on Bayesian networks. In Uncertainty Proceedings 1991, 52–60 (Elsevier, 1991).
Cussens, J. Bayesian network learning with cutting planes. In 27th Conference on Uncertainty in Artificial Intelligence, 153–160 (AUAI Press, 2011).
Bartlett, M. & Cussens, J. Integer linear programming for the Bayesian network structure learning problem. Artif. Intell. 244, 258–271 (2017).
Yuan, C., Malone, B. & Wu, X. Learning optimal Bayesian networks using a* search. In Twenty-second International Joint Conference on Artificial Intelligence (2011).
Wang, Z., Gao, X., Tan, X. & Liu, X. Learning Bayesian networks using a* search with ancestral constraints. Neurocomputing 451, 107–124 (2021).
Liao, Z. A., Sharma, C., Cussens, J. & van Beek, P. Finding all bayesian network structures within a factor of optimal. In Proceedings of the AAAI Conference on Artificial Intelligence Vol. 33, 7892–7899 (2019).
Heckerman, D., Geiger, D. & Chickering, D. M. Learning Bayesian networks: The combination of knowledge and statistical data. Mach. Learn. 20, 197–243 (1995).
Liu, X., Gao, X., Wang, Z., Ru, X. & Zhang, Q. A metaheuristic causal discovery method in directed acyclic graphs space. Knowl.-Based Syst. 276, 110749 (2023).
Constantinou, A. C., Liu, Y., Kitson, N. K., Chobtham, K. & Guo, Z. Effective and efficient structure learning with pruning and model averaging strategies. Int. J. Approx. Reason. 151, 292–321 (2022).
Chickering, D. M. Learning equivalence classes of Bayesian-network structures. J. Mach. Learn. Res. 2, 445–498 (2002).
Gillispie, S. B. & Perlman, M. D. Enumerating markov equivalence classes of acyclic digraph dels. In Proceedings of the Seventeenth Conference on Uncertainty in Artificial Intelligence, 171–177 (2001).
Meek, C. Graphical Models: Selecting Causal and Statistical Models. Ph.D. thesis, (Carnegie Mellon University, 1997).
Teyssier, M. & Koller, D. Ordering-based search: a simple and effective algorithm for learning Bayesian networks. In Proceedings of the Twenty-First Conference on Uncertainty in Artificial Intelligence, 584–590 (2005).
Wang, Z., Gao, X., Tan, X. & Liu, X. Determining the direction of the local search in topological ordering space for Bayesian network structure learning. Knowl.-Based Syst. 234, 107566 (2021).
Scanagatta, M., de Campos, C. P., Corani, G. & Zaffalon, M. Learning bayesian networks with thousands of variables. Adv. Neural Inf. Process. Syst.28 (2015).
Tsamardinos, I., Brown, L. E. & Aliferis, C. F. The max-min hill-climbing Bayesian network structure learning algorithm. Mach. Learn. 65, 31–78 (2006).
Scutari, M., Graafland, C. E. & Gutiérrez, J. M. Who learns better Bayesian network structures: Accuracy and speed of structure learning algorithms. Int. J. Approx. Reason. 115, 235–253 (2019).
Chickering, D. M. Optimal structure identification with greedy search. J. Mach. Learn. Res. 3, 507–554 (2002).
Nielsen, J. D., Kočka, T. & Peña, J. M. On local optima in learning Bayesian networks. In Proceedings of the Nineteenth Conference on Uncertainty in Artificial Intelligence, 435–442 (2002).
Chen, F., Wang, X. & Rao, Y. Learning Bayesian networks using genetic algorithm. J. Syst. Eng. Electron. 18, 142–147 (2007).
Zhang, Y., Zhang, W. & Xie, Y. Improved heuristic equivalent search algorithm based on maximal information coefficient for Bayesian network structure learning. Neurocomputing 117, 186–195 (2013).
Alonso-Barba, J. I. et al. Scaling up the greedy equivalence search algorithm by constraining the search space of equivalence classes. Int. J. Approx. Reason. 54, 429–451 (2013).
Ramsey, J., Glymour, M., Sanchez-Romero, R. & Glymour, C. A million variables and more: The fast greedy equivalence search algorithm for learning high-dimensional graphical causal models, with an application to functional magnetic resonance images. Int. J. Data Sci. Anal. 3, 121–129 (2017).
Alonso, J. I., Ossa, L., Gamez, J. A. & Puerta, J. M. On the use of local search heuristics to improve GES-based Bayesian network learning. Appl. Soft Comput. 64, 366–376 (2018).
Nandy, P., Hauser, A. & Maathuis, M. H. High-dimensional consistency in score-based and hybrid structure learning. Ann. Stat. 46, 3151–3183 (2018).
Liu, X., Gao, X., Ru, X., Tan, X. & Wang, Z. Improving greedy local search methods by switching the search space. Appl. Intell. 53, 22143–22160 (2023).
Liu, X., Gao, X., Ru, X. & Wang, Z. A hybrid Bayesian network structure learning algorithm in equivalence class space. In 2023 8th International Conference on Control and Robotics Engineering (ICCRE), 1–4 (IEEE, 2023).
Laborda, J. D., Torrijos, P., Puerta, J. M. & Gámez, J. A. Parallel structural learning of Bayesian networks: Iterative divide and conquer algorithm based on structural fusion. Knowl.-Based Syst. 296, 111840 (2024).
Chen, Y. & Tian, J. Finding the k-best equivalence classes of Bayesian network structures for model averaging. In Proceedings of the AAAI Conference on Artificial Intelligence, vol. 28 (2014).
Chen, E. Y.-J., Choi, A. C. & Darwiche, A. Enumerating equivalence classes of Bayesian networks using EC graphs. In Artificial Intelligence and Statistics, 591–599 (PMLR, 2016).
Zheng, X., Aragam, B., Ravikumar, P. K. & Xing, E. P. Dags with no tears: Continuous optimization for structure learning. Adv. Neural Inf. Processing Syst. 31 (2018).
Yu, Y., Chen, J., Gao, T. & Yu, M. DAG-GNN: DAG structure learning with graph neural networks. In International Conference on Machine Learning, 7154–7163 (PMLR, 2019).
Yu, Y., Gao, T., Yin, N. & Ji, Q. DAGS with no curl: An efficient DAG structure learning approach. In International Conference on Machine Learning, 12156–12166 (PMLR, 2021).
Ng, I., Ghassami, A. & Zhang, K. On the role of sparsity and DAG constraints for learning linear DAGS. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual (eds Larochelle, H. et al.) (2020).
Bello, K., Aragam, B. & Ravikumar, P. DAGMA: learning DAGS via m-matrices and a log-determinant acyclicity characterization. In Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022 (eds Koyejo, S. et al.) (2022).
VERMA, T. Equivalence and synthesis of causal models. In Proceedings of the Sixth Conference on Uncertainty in Artificial Intelligence, 1991 (Elsevier, 1991).
Koller, D. & Friedman, N. Probabilistic Graphical Models: Principles and Techniques (MIT Press, 2009).
Dor, D. & Tarsi, M. A simple algorithm to construct a consistent extension of a partially oriented graph. Technicial Report R-185, Cognitive Systems Laboratory, UCLA 45 (1992).
Chickering, D. M. A transformational characterization of equivalent bayesian network structures. In Proceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence, 87–98 (1995).
Scutari, M. Learning Bayesian networks with the bnlearn r package. J. Stat. Softw. 35 (2010).
Kalisch, M., Mächler, M., Colombo, D., Maathuis, M. H. & Bühlmann, P. Causal inference using graphical models with the r package pcalg. J. Stat. Softw. 47, 1–26 (2012).
Zhang, K. et al. gcastle: A python toolbox for causal discovery (2021). arXiv:2111.15155.
Tukey, J. W. Exploratory Data Analysis (Reading/Addison-Wesley, 1977).
Constantinou, A. et al. Open problems in causal structure learning: A case study of covid-19 in the UK. Expert Syst. Appl. 234, 121069 (2023).
Acknowledgements
This work was supported by the National Natural Science Foundation of China (61573285, 62262016, 52402453). This work was also sponsored by the Innovation Foundation for Doctor Dissertation of Northwestern Polytechnical University (CX2022047). This work was also supported by the Fundamental Research Funds for the Central Universities (G2022KY0602). This work was also supported by the Open Project of the State Key Laboratory of Intelligent Game(Grant No. ZBKF-23-05). This work was also supported by the Key Research and Development Program of Shaanxi Province (Grant No.2023-GHZD-33).
Author information
Authors and Affiliations
Contributions
Conceptualization, Xiaohan Liu; methodology, Xiaohan Liu and Ziyi Yang; software, Xiaohan Liu; validation, Xiaohan Liu; formal analysis, Xiaohan Liu and Qi Feng; investigation, Xiaohan Liu; resources, Xiaohan Liu and Ziyi Yang; data curation, Xiaohan Liu and Ziyi Yang; writing???original draft preparation, Xiaohan Liu and Shuying Wu; writing???review and editing, Xiaohan Liu, Shuying Wu, and Yuqing Yang; visualization, Xiaohan Liu and Chuchao He; supervision, Xiaoguang Gao; project administration, Xiaohan Liu, Qi Feng, Ziyi Yang, Shuying Wu, Yuqing Yang, Jia Ren and Chuchao He; funding acquisition, Xiaoguang Gao.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Liu, X., Feng, Q., Yang, Z. et al. An improved greedy equivalent search method based on relative entropy. Sci Rep 15, 37250 (2025). https://doi.org/10.1038/s41598-025-21219-8
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-21219-8







