Abstract
In recent years, mixed Bayesian networks have received increasing attention across various fields for probabilistic reasoning. Though many studies have been devoted to propagation computation on strong junction trees for mixed Bayesian networks, few have addressed the construction of appropriate strong junction trees. In this work, we establish a connection between the minimal strong triangulation for marked graphs and the minimal triangulation for star graphs. We further propose a minimal strong triangulation method for the moral graph of mixed Bayesian networks and develop a polynomial-time algorithm to derive a strong junction tree from this minimal strong triangulation. Moreover, we also focus on the propagation computation of all posteriors on this derived strong junction tree. We conducted multiple numerical experiments to evaluate the performance of our proposed method, demonstrating significant improvements in computational efficiency compared to existing approaches. Experimental results indicate that our minimal strong triangulation approach provides a robust framework for efficient probabilistic inference in mixed Bayesian networks.
MSC:
62H22; 62F15
1. Introduction
Originating from artificial intelligence [1], Bayesian networks offer intuitive graphical representations for joint probability distributions and efficient algorithms for probabilistic reasoning. These networks are powerful tools for managing uncertainty and are applied across various fields, including bioinformatics [2,3], computational biology [4,5,6], and neuroscience [7], among others.
Over the past two decades, there has been a growing interest in Bayesian networks that incorporate both continuous and discrete variables [8]. This paper focuses on probabilistic models of such mixed networks, specifically those constrained to conditional linear Gaussian (CLG) distributions. Lauritzen [9] introduced a method for the exact local computation of means and variances in mixed Bayesian networks, but it suffered from numerical stability issues. Subsequently, Lauritzen and Jensen [10] developed an alternative, stable local computation scheme for these conditional Gaussian networks. However, this scheme is intricate, involving evaluations of matrix generalized inverses and recursive combinations of potentials. Cowell [11] proposed an alternative architecture utilizing exchange operations for variable elimination within an elimination tree. These exchange operations avoid complex matrix computations. Building upon the push and exchange operations, Madsen [12] extended lazy propagation, leveraging independence relations induced by evidence to enhance inference efficiency. The push operation is employed to obtain the posterior of continuous variables, while the exchange operation ensures the maintenance of a valid directed acyclic graph (DAG) structure during the belief update process.
Previous works have significantly improved the propagation computation for mixed Bayesian networks, primarily relying on strong junction trees. However, to our best knowledge, few works discuss how to construct appropriate strong junction trees. In this paper, we try to construct a strong junction tree under a minimal strong triangulation framework, which employs a polynomial-time minimal triangulation algorithm to minimize the number of fill-in edges. Additionally, we focus on the propagation computation of all posteriors. The process primarily involves three steps: finding triangulation and junction trees, performing inward and outward message passing, and computing posteriors for all non-evidence variables. This propagation computation based on junction trees facilitates the sharing of computations [13,14] among different posteriors. As suggested by Madsen [12], we applied exchange operations to eliminate both continuous and discrete variables during message passing. We conducted multiple simulation experiments, comparing our method with two existing methods: CGBayesNets [15] and BayesNetBP [16]. The results indicate that our method outperforms the existing methods in computational speed.
The primary contribution of this paper is the construction of a strong junction tree within a minimal strong triangulation framework. Specifically, we establish a connection between the minimal strong triangulation for marked graphs and the minimal triangulation for star graphs. We further propose a minimal strong triangulation method for the moral graph of mixed Bayesian networks and develop a polynomial-time algorithm to derive a strong junction tree from this minimal strong triangulation. Additionally, we present two theorems that guarantee our algorithm can successfully identify strong junction trees for mixed Bayesian networks.
This paper is organized as follows. In Section 2, the notation and definitions used in mixed Bayesian networks are reviewed. In Section 3, we discuss the minimal strong triangulation for marked graphs. Section 4 describes our method to find strong junction trees for mixed Bayesian networks by using minimal strong triangulations. In Section 5, we give a detailed analysis on exchange operation. Section 6 presents message passing on strong junction trees and posterior computation. Some numerical experimental results are shown in Section 7. Finally, Section 8 concludes this paper.
2. Basic Notions on Decomposition and Triangulation for Marked Graphs
Considering two graphs and , we call a subgraph of G if and . Given a subset A of vertices in V, it induces a subgraph where . A graph is complete if every pair of distinct vertices is adjacent. A clique is a maximal vertex set inducing a complete graph. Given a graph , a vertex subset S is called a separator for two disjoint vertex subsets if every path in G between some and contains a vertex in S. We also say that S separates A and B in G. If and , , , we call as a partition of V.
Definition 1.
For a graph , a partition of V is a decomposition of G if are separated by S in G, and S is complete.
Let be an undirected graph with two types of vertices, and , and this graph G is called a marked graph. Moral graphs of directed acyclic graphs (DAGs) of mixed Bayesian networks are marked graphs and those graph operations introduced below can be applied to moral graphs.
Definition 2.
A partition of is a strong decomposition of G if is a decomposition of G and any of the three conditions , or holds.
Leimer introduced the concept of marked graphs and their strong decomposition [17]. We specifically adopt strong decomposition in Definition 2 for marked graphs because it preserves the secondary structure [18] of the marked graph, as indicated by the closure [19] of marginal distributions during inward message passing.
A graph G is called triangulated if G is complete or there is a decomposition of G such that and are triangulated. Similarly, a marked graph is called strongly triangulated if G is complete or there is a strong decomposition of G such that and are strongly triangulated. Given a marked graph G, we can construct a star graph from G by adding a vertex ⋆ to its vertex set and connecting ⋆ with every discrete variable. This star graph is denoted as , where . For any subset , the star graph of the induced graph is just the induced subgraph of the star graph . The following Proposition 1 [20] characterizes the equivalent relationship between the strongly triangulated marked graph and its star graph.
Proposition 1
([20]). A marked graph G is strongly triangulated if and only if is triangulated.
3. Minimal Strong Triangulation for Marked Graphs
A graph is called a triangulation of if H is triangulated. Any edge is referred to as a fill edge. A triangulation of is minimal if is not triangulated for any proper subset of F. For a marked graph , a graph is called a strong triangulation of G if H is strongly triangulated. Let be a set of all the strong triangulations of G. A strong triangulation of a marked graph is minimal if is not strongly triangulated for any proper subset of F. The following theorem establishes the relationship between minimal strong triangulations and minimal triangulations.
Theorem 1.
For a marked graph , is a minimal strong triangulation of G if and only if is a minimal triangulation of .
Proof.
(⇒) If is a minimal strong triangulation of G, then H is strongly triangulated and is triangulated. Denote as the edge set . Thus, is a triangulation of . If is also a triangulation of and , then is a strong triangulation of G by Proposition 1. Since H is a minimal strong triangulation of G, it is a contradiction. Then is a minimal triangulation of .
(⇐) If is a minimal triangulation of , then H is strongly triangulated and H is a strong triangulation of G. If is also a strong triangulation of G and , then is triangulated by Proposition 1 and is a triangulation of . Since is a minimal triangulation of , it is a contradiction. Thus, is a minimal strong triangulation of G. □
Theorem 1 indicates that a minimal strong triangulation for a marked graph G can be obtained from a minimal triangulation for the star graph , which is an undirected graph. Specifically, we can apply the minimal triangulation algorithm MSC-M [21] to , and subsequently obtain a minimal strong triangulation for G by dropping the star vertex.
Notice that the weight of a marked graph G can be defined as the minimum sum of the weights of all the cliques over all possible strong triangulations of G, that is,
where is the set of all the cliques in the triangulation H. The weight of a clique C is calculated as [22], where s represents the product of the number of states of the discrete nodes in C, and m denotes the number of continuous nodes in C. In , s has an exponential weight on the number of discrete nodes in C, while acts as a polynomial weight on the number of continuous nodes. Unlike the optimal triangulation computation for with an NP-hard computational complexity [23], our approach in this paper employs a polynomial-time minimal triangulation strategy to minimize the number of fill-in edges. Specifically, the time cost [21] of a minimal triangulation is , where and are the number of vertices and edges, respectively.
In this paper, we focus on directed acyclic graphs (DAGs) , where V is marked into two groups, and . The vertices in represent discrete variables, while those in represent continuous variables. Since moral graphs of those DAGs with two types of variables are marked graphs, then we deduce the following corollary from Theorem 1:
Corollary 1.
Let be the moral graph of a DAG G with two types of variables; H is a minimal strong triangulation of if and only if is a minimal triangulation of .
4. Construction of Strong Junction Trees for Moral Graphs
Given a DAG G with two types of variables, its moral graph is a marked graph. Utilizing the MCS-M algorithm [21], we can derive a minimal triangulation graph K of . In this algorithm, the star vertex ⋆ can be designated as the first numbered vertex. Consequently, the resulting graph K manifests as a star graph of some graph H, as ⋆ remains unconnected to any continuous vertex. Therefore, H is a minimal strong triangulation of according to Corollary 1. The time complexity of the MCS-M algorithm for is , where and denote the set of vertices and edges of , respectively.
The SMCS algorithm [17] shares a similar definition with the MCS algorithm [24], but always chooses discrete vertices if there is some discrete vertex with the maximum weight. Utilizing the SMCS algorithm enables us to acquire a D-numbering [25] denoted as for H. Consequently, we obtain a D-ordered clique sequence for H. Specifically, , where with , , and so forth, until . For any index k, forms a clique in H. The positions of can be determined by the equivalent condition that .
Theorem 2.
The D-ordered clique sequence obtained by the SMCS algorithm is also strong, i.e., or for , where and .
Proof.
Without loss of generalization, we assumes that , , and . Since H is strongly triangulated and the partition is a decomposition of H, this partition is also a strong decomposition of H. Otherwise, there are , , and . Furthermore, there exists a path such that , and other vertices are continuous. From Corollary 2.7 in [20], H is not a strong triangulated graph since is not adjacent to . It is a contradiction. So the partition is also a strong decomposition of H.
If , the result is trivial. If and , then since is a strong decomposition of H. Thus, . Moreover, we have that . Otherwise, means there is only one clique in H, and the result is trivial. Since is the first vertex numbered by the SMCS algorithm, the cardinality of the discrete vertex is the same as that of the continuous vertex in . Thus, the discrete vertex is always numbered earlier than the continuous vertex by the SMCS. The number of discrete vertex in is always bigger than that of the continuous vertex in . It is a contradiction. So, if . Furthermore, since the subgraph is also strongly triangulated, the conclusion is obtained by recursive decompositions. □
As demonstrated above, the SMCS algorithm can be employed to generate a strong D-ordered clique sequence for a strong triangulation H, enabling the subsequent construction of a strong junction tree using this sequence. The time complexity for utilizing the SMCS algorithm [17] to find a strong D-ordered clique sequence and construct a strong junction tree is , where and denote the sets of vertices and edges of H, respectively.
From Corollary 1 and Theorem 2, we can devise the following Algorithm 1 with polynomial time complexity to find strong junction trees for mixed Bayesian networks:
| Algorithm 1: Construct a Strong Junction Tree |
| Input: A DAG G with two types of variables Output: A strong junction tree T on G 1 Construct the Star Graph: Create the star graph from the moral graph of the DAG G; 2 Minimal Triangulation: Apply the MCS-M algorithm to the star graph with the star chosen as the first numbered vertex to derive its minimal triangulation graph K; 3 Obtain Strong Triangulation: Remove the star vertex ⋆ from K to yield a minimal strong triangulation graph H of ; 4 Construct Strong Junction Tree: Apply the SMCS algorithm to the minimal strong triangulation graph H for generating a strong junction tree T. |
5. The Exchange Operation for Mixed Bayesian Networks
In this paper, our focus lies on directed acyclic graphs (DAGs) denoted as . Here, the vertices of represent discrete variables, while those of represent continuous variables. To exploit conditional linear Gaussian (CLG) distributions for modeling, we make an additional assumption that no continuous variables have discrete children in G. A mixed Bayesian network with CLG distributions comprises a DAG G, a set of conditional probability distributions , and a set of CLG density functions , where is the set of the parents of Z in G. A mixed Bayesian network has a joint distribution of the form: . Let Y be a continuous variable and and . Given variables and , the conditional distribution of Y is , where is a table of real numbers, is a table of -dimensional vectors, and is a table of non-negative values.
To compute all the marginal posteriors given evidence in mixed Besian networks, we employ propagation computation based on message passing within a strong junction tree structure. During message passing, barren variables [26] in the local structure can be directly eliminated without computation. And exchange operations [11] can avoid complex matrix operations, and make for finding barren variables [27]. In this section, we will discuss the usage of exchange operations in mixed Bayesian networks in detail.
Assume that are continuous variables and there is a pair of conditional Gaussian distributions:
The exchange operation, equivalent to Theorem 1 of the work [28], converts (1) and (2) into another pair of distributions:
where . These formulae from Cowell [11] are essentially from the Bayes’ theorem. (3) and (4) keep the same joint distribution as (1) and (2).
Let us consider exchange operations in mixed Bayesian networks . Assume that , and there are no other directed paths from Y to Z in G. The parent set of Y is partitioned into and , where and . Similarly, the parent set of Z is partitioned into and , where and . The conditional density of Y and of Z have the following forms:
Since there are no other directed paths from Y to Z except , variables in are non-descendants of Y. By conditional independence relations in Bayesian networks, we have that
Similarly, we also have that
since any parent vertex of Y is non-descendant of Z. Denote as , and as . Thus from Equations (5)–(8), we have that:
For any , thus or . If for some , thus, . Similarly, if for some , thus . The exchange operation can convert Equations (5) and (6) into the following pair of conditional densities:
where .
Following the exchange operation, we reconstruct the mixed Bayesian network G into based on the altered conditional distributions of Y and Z. Specifically, the new network is the same as the original G except that and in for any and . And since and no other directed path exists from Y to Z in G, it can be verified that the new network is also a DAG. Thus, is also a mixed Bayesian network with the same joint distribution with , where . The process from G to is called arc-reversal. By using arc-reversals step by step, any continuous variable can become a node with no children. Thus, the exchange operation provides us a method to eliminate continuous variables and avoids matrix operations [11].
Now, we consider handling discrete variables by using the exchange operation. Let with parent set and let with parent set . and are the corresponding probabilities of Q and X, respectively. The exchange operation converts the above pair of conditional probabilities into and , which maintain the same joint probability distribution of the original pair:
From (13) and (14), the exchange operation converts and into the following two conditional probabilities :
by the assumption that there are no other directed paths from Q to X in G.
After the exchange operation, similar to the case of continuous variables, the mixed Bayesian network G can be reconstructed into according to the changed conditional distributions of Q and X. is the same as G except that and in for any and , and is also a DAG. Thus, is also a mixed Bayesian network with the same joint distribution with , where . The process from G to is also called arc-reversal.
In this paper, arc-reversals for discrete variables with no continuous children are only used in message passing for further finding barren variables, and the lazy propagation [12,27,29] is also utilized to improve the computational efficiency. Following message passing, direct variable eliminations [30] are used to eliminate discrete variables in the marginal posterior computation.
6. Message Passing and Posterior Computing
Let denote the head variable of and the tail variables of . For instance, if , then and . Let denote the domain of p. and are similarly defined for . A potential is represented as , where denotes the set of conditional probability and denotes that related to conditional density. The combination of two potentials, and , is defined as .
Let be a strong junction tree of H, which is a minimal strong triangulation of . For any , choose a clique such that and allocate p to C. For any , choose a clique such that and allocate f to C. For any clique , A has an original potential , where p is allocated to A} and f is allocated to A}.
Assume that two cliques are adjacent on T and the current potential of A is . We define as the union of domains of all variable in and , and , where . Let be a DAG induced by the current potential . The information potential from A to B is computed by the following operations:
(1) Find barren variables N [12] with respect to set S, evidence and DAG . Remove conditional probabilities and densities from , whose head variable is contained in N, and the remaining potential is denoted by . Remove N from and the remaining graph is denoted by . Let X be the set d-connected to S about in . Remove conditional probabilities and densities from , whose domain has no variables in . The resulting potential is denoted by . The set X can be partitioned into and , which denote discrete and continuous non-evidence variables, respectively.
(2) Eliminate variables in by exchange operations. Based on these operations, conditional densities of are changed, and those densities whose head variable is contained in are removed from . The remaining potential is denoted by .
(3) Eliminate variables in with no continuous children by exchange operations. Let denote . If for some proper , exchange operations for eliminating x are delayed. Based on these operations, conditional probabilities of are changed, and those probabilities whose head variable x is contained in and are removed from .
By these operations, the remaining potential is denoted by . Thus, . When considering the information from A to B, the current potential is the combination of the original potential of A with all the information from the clique to A. Then, .
, there is a clique A containing X. To compute the posterior distribution of X,
For Equation (20), exchange operations are performed on continuous variables, and the sum-out procedure [14] is used to eliminate discrete variables. Thus we can obtain the following Algorithm 2 for computing all the posteriors:
| Algorithm 2: Propagation Computation for All the Posteriors. |
| Input: A mixed Bayesian network with given evidence variable set Output: All the posteriors for non-evidence variables 1 Construct a strong junction tree T on G by Algorithm 1; 2 Choose a clique in the strong junction tree T as the root R; 3 Perform inward message passing to the root such that root R receives information from its neighbors; 4 Perform outward message passing from the root such that each clique in T receives information from its neighbors; 5 Compute the posterior for each non-evidence variable X by Equations (17)–(20). |
7. Numerical Experiment
Our experiments were conducted on a collection of randomly generated CLG sparse Bayesian networks consisting of n variables and edges. Referring to the work [12], we set the number of variables n to {50, 75, 100} and the ratio of continuous variables to {0, 0.25, 0.5, 0.75, 1}. With this simulation setting, we can evaluate the performances of our method across various sizes of mixed Bayesian networks and different proportions of continuous variables. For each network size, each proportion of continuous variables, and each size of the evidence set, one hundred networks were randomly generated, with each discrete variable having two states. These experiments were conducted using C++ on a PC equipped with an Intel Core i9-9900X CPU running at 3.5GHz. We mainly utilized three standard libraries in C++: <vector>, <random>, and <algorithm>. The <vector> library is a versatile container for dynamic arrays but often allocates more memory than necessary, leading to potential overhead. The <random> library offers a comprehensive set of tools for generating random numbers but may produce inconsistent sequences across different platforms due to varying implementations. The <algorithm> library provides a rich set of functions for operations on data sequences, like sorting and searching, but does not directly support complex data structures such as graphs and trees.
In our experiments, the average time required to find minimal strong triangulations and strong junction trees for networks with 50, 75, and 100 variables is approximately 0.0005, 0.0014, and 0.0031 s, respectively. These results demonstrate that the algorithm outlined in Section 4 is efficient in producing strong junction trees for mixed Bayesian networks. Subsequent belief updates based on these strong junction trees are also carried out in our experiments.
Figure 1, Figure 2 and Figure 3 depict the average time costs of the propagation computation of our method, including both message passing and the posterior computation for all non-evidence variables, across networks with 50, 75, and 100 variables. Meanwhile, Figure 4, Figure 5 and Figure 6 display the median time costs of the propagation computation. As shown in Figure 1, Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6, the speed of belief updates in networks only containing continuous variables remains consistently fast, whether considering the average or median time costs. Moreover, as shown in Figure 5 and Figure 6, at low numbers of instantiations, a low proportion of continuous variables tends to increase the median time costs for belief updates in Bayesian networks. This is due to the state space size of the largest clique potentially being larger in networks with a lower proportion of continuous variables compared to networks with a higher proportion, under similar conditions. However, once the number of instantiations exceeds half of the variables, the discrepancy among networks with varying proportions of continuous variables becomes small. This reduction occurs because the average largest discrete domain size decreases as the size of the evidence set increases [12].
Figure 1.
Average time in seconds for belief update in networks with 50 variables.
Figure 2.
Average time in seconds for belief update in networks with 75 variables.
Figure 3.
Average time in seconds for belief update in networks with 100 variables.
Figure 4.
Median time in seconds for belief update in networks with 50 variables.
Figure 5.
Median time in seconds for belief update in networks with 75 variables.
Figure 6.
Median time in seconds for belief update in networks with 100 variables.
A notable observation when comparing the average time costs depicted in Figure 3 with the median time costs shown in Figure 6 is the significant disparity between the two measures. This disparity arises primarily from the considerable variance in time costs resulting from the random sampling of Bayesian networks. For instance, in networks with 100 variables, a continuous variable proportion of 0.5, and 10 pieces of evidence, the average time is 54.3024 s with a standard deviation of 538.1081, while the median time is 0.1355 s. It is evident that this huge average time is caused by certain outliers in the time cost for 100 randomly generated networks. The reason for this abnormal result may lie in the generation of large-scale cliques with discrete variables under certain random seeds when the size of the evidence set is ten.
For a further analysis of the efficiency of our method, we conducted a comparative study with two existing packages: CGBayesNets (a MATLAB package [15]) and BayesNetBP (an R package [16]). Figure 7 illustrates the average time costs on propagation computation of three methods for 100 randomly generated networks, each comprising 50 variables. In the left panel, the size of the evidence set is fixed at five, while in the right panel, the proportion of continuous variables is set to 0.5. As shown in both panels, it is evident that our method outperforms the other two methods in terms of computational speed.
Figure 7.
Time costs of propagation computation for our method, CGBayesNets, and BayesNetBP.
8. Conclusions
The Bayesian network is a prominent predictive modeling framework, where its nodes often represent genes or proteins interconnected by edges denoting relationships. These edges, informed by experimental data, impart a predictive capability to Bayesian computational methods [15,31], capturing both the strength and direction of connections. Despite over two decades of development in mixed Bayesian networks, combining continuous and discrete variables [8], the field still lacks efficient algorithms for identifying appropriate strong junction trees in their moral graphs. In our study, we address this gap by investigating the minimal strong triangulation of moral graphs and presenting a polynomial-time algorithm for deriving strong junction trees based on this approach. Moreover, we also consider the message passing computation of all posteriors on this derived strong junction tree. Our experimental findings affirm the superior efficiency of our algorithm compared to existing methods for message passing in mixed Bayesian networks. The proposed approach not only demonstrates remarkable improvements in computational speed but also provides a robust framework for efficient probabilistic inference in mixed Bayesian networks, making it a valuable tool for diverse applications in bioinformatics, computational biology, and neuroscience.
Author Contributions
Methodology, Y.L. and X.W.; Software, S.W. and C.Z.; Formal analysis, X.W. and S.W.; Writing—original draft, Y.L. and X.W. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the National Natural Foundation of China (No. 12171076, 12226003, 11971097), the Natural Science Foundation of Jiangsu (No. BK20230439).
Data Availability Statement
All the codes can be found in https://github.com/KLASMOE/BPMBN.
Acknowledgments
Authors acknowledge valuable suggestions from the reviewers for this work.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Pearl, J. Probabilistic Inference in Intelligent Sysetms; Morgan Kaufmann: Burlington, MA, USA, 1988. [Google Scholar]
- Laccourreye, P.; Bielza, C.; Larrañaga, P. Explainable Machine Learning for Longitudinal Multi-Omic Microbiome. Mathematics 2022, 10, 1994. [Google Scholar] [CrossRef]
- Sazal, M.; Mathee, K.; Ruiz-Perez, D.; Cickovski, T.; Narasimhan, G. Inferring directional relationships in microbial communities using signed Bayesian networks. BMC Genom. 2020, 21 (Suppl. S6), 663. [Google Scholar] [CrossRef] [PubMed]
- Bernaola, N.; Michiels, M.; Larrañaga, P.; Bielza, C. Learning massive interpretable gene regulatory networks of the human brain by merging Bayesian networks. PLoS Comput. Biol. 2023, 19, e1011443. [Google Scholar] [CrossRef]
- Joshi, A.; Rienks, M.; Theofilatos, K.; Mayr, M. Systems biology in cardiovascular disease: A multiomics approach. Nat. Rev. Cardiol. 2021, 18, 313–330. [Google Scholar] [CrossRef]
- Liu, F.; Zhang, S.W.; Guo, W.F.; Wei, Z.G.; Chen, L. Inference of Gene Regulatory Network Based on Local Bayesian Networks. PLoS Comput. Biol 2016, 12, e1005024. [Google Scholar] [CrossRef]
- Michiels, M.; Larrañaga, P.; Bielza, C. BayeSuites: An open web framework for massive Bayesian networks focused on neuroscience. Neurocomputing 2021, 428, 166–181. [Google Scholar] [CrossRef]
- Salmerón, A.; Rumí, R.; Langseth, H.; Nielsen, T.D.; Madsen, A.L. A review of inference algorithms for hybrid Bayesian networks. J. Artif. Intell. Res. 2018, 62, 799–828. [Google Scholar] [CrossRef]
- Lauritzen, S.L. Propagation of probabilities, means and variances in mixed graphical association models. J. Am. Stat. Assoc. Stat. 1992, 87, 1098–1108. [Google Scholar] [CrossRef]
- Lauritzen, S.L.; Jensen, F. Stable local computation with mixed Gaussian distributions. Stat. Computing 2001, 11, 191–203. [Google Scholar] [CrossRef]
- Cowell, R.G. Local propagation in conditional Gaussian Bayesian networks. J. Mach. Learn. Res. 2005, 6, 1517–1550. [Google Scholar]
- Madsen, A.L. Belief update in CLG Bayesian networks with lazy propagation. Int. J. Approx. 2008, 49, 503–521. [Google Scholar] [CrossRef]
- Lauritzen, S.L.; Spiegelhalter, D.J. Local computations with probabilities on graphical structures and their application to expert systems. J. R. Stat. Soc. Ser. 1988, 50, 157–224. [Google Scholar] [CrossRef]
- Zhang, N.L. Computational properties of two exact algorithms for Bayesian networks. J. Appl. Intell. 1998, 9, 173–184. [Google Scholar] [CrossRef]
- McGeachie, M.J.; Chang, H.H.; Weiss, S.T. CGBayesNets: Conditional Gaussian Bayesian network learning and inference with mixed discrete and continuous data. PLoS Comput. Biol. 2014, 10, e1003676. [Google Scholar] [CrossRef]
- Yu, H.; Moharil, J.; Blair, R.H. BayesNetBP: An R Package for Probabilistic Reasoning in Bayesian Networks. J. Stat. Softw. 2020, 94, 1–31. [Google Scholar] [CrossRef]
- Leimer, H.G. Triangulated graphs with marked vertices. Ann. Discret. Math. 1989, 41, 311–324. [Google Scholar]
- Wang, X.F.; Guo, J.H. The tree structure of various graphs in statistics. Stat. Its Interface 2013, 6, 151–164. [Google Scholar]
- Wang, X.F.; Guo, J.H.; He, X. Finding the minimal set for collapsible graphical models. Proc. Am. Math. Soc. 2011, 139, 361–373. [Google Scholar] [CrossRef]
- Lauritzen, S.L. Graphical Models; Clarendon Press: Oxford, UK, 1996. [Google Scholar]
- Berry, A.; Blair, J.R.S.; Heggernes, P.; Peyton, B.W. Maximum cardinality search for computing minimal triangulations of graphs. Algorithmica 2004, 39, 287–298. [Google Scholar] [CrossRef]
- Jensen, F. HUGIN API Referene Manual. 2024. Available online: https://download.hugin.com/webdocs/manuals/api-manual.pdf (accessed on 19 June 2024).
- Wen, W.X. Optimal decomposition of belief networks. In Proceedings of the Sixth Workshop on Uncertainty in Artificial Intelligence, Cambridge, MA, USA, 27–29 July 1990; pp. 245–256. [Google Scholar]
- Blair, J.R.S.; Peyton, B.W. An introduction to chordal graphs and clique trees. In Sparse Matrix Computations: Graph Theory Issues and Algorithms; George, J.A., Liu, J.W.H., Eds.; IMA Volumes in Mathematics and its Applications; Springer: Berlin/Heidelberg, Germany, 1993; Volume 56, pp. 1–30. [Google Scholar]
- Leimer, H.G. Optimal decomposition by clique separators. Discret. Math. 1993, 113, 99–123. [Google Scholar] [CrossRef]
- Shachter, R.D. Evaluating Influence Diagrams. Oper. Res. 1986, 34, 871–882. [Google Scholar] [CrossRef]
- Madsen, A.L. Improvements to message computation in lazy propagation. Int. J. Approx. 2010, 51, 499–514. [Google Scholar] [CrossRef]
- Shachter, R.D.; Kenley, C.R. Gaussian influence diagrams. Manag. Sci. 1989, 35, 527–549. [Google Scholar] [CrossRef]
- Madsen, A.L.; Jensen, F.V. Lazy propagation: A junction tree inference algorithm based on lazy evaluation. Artif. Intell. 1999, 113, 203–245. [Google Scholar] [CrossRef]
- Butz, C.J.; Konkel, K.; Lingras, P. Join tree propagation utilizing both arc reversal and variable elimination. Int. J. Approx. Reason. 2011, 52, 948–959. [Google Scholar] [CrossRef][Green Version]
- Needham, C.J.; Bradford, J.R.; Bulpitt, A.J.; Westhead, D.R. A primer on learning in Bayesian networks for computational biology. PLoS Comput. Biol. 2007, 3, e129. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).






