# Bayesian probabilistic network modeling from multiple independent replicates

- 1.5k Downloads

## Abstract

Often protein (or gene) time-course data are collected for multiple replicates. Each replicate generally has sparse data with the number of time points being less than the number of proteins. Usually each replicate is modeled separately. However, here all the information in each of the replicates is used to make a composite inference about signal networks. The composite inference comes from combining well structured Bayesian probabilistic modeling with a multi-faceted Markov Chain Monte Carlo algorithm. Based on simulations which investigate many different types of network interactions and experimental variabilities, the composite examination uncovers many important relationships within the networks. In particular, when the edge's partial correlation between two proteins is at least moderate, then the composite's posterior probability is large.

### Keywords

Posterior Probability Receiver Operating Characteristic Curve Partial Correlation Lasso High Posterior Probability### List of abbreviations

- BIC
Bayesian Information Criterion

- DAG
directed acyclic graph

- DFP
double block false positive

- DLFP
double block log odds based false positive

- FPR
false positive rate

- QFP
quadruple block false positive

- QLFP
quadruple block log odds based false positive

- rep
replicate

- ROC
receiver operating characteristic

- TFP
triple block false positive

- TLFP
triple block log odds based false positive

- TPR
true positive rate.

## Introduction

Often the laboratory collection of protein phosphorylation time-course data results not in a single set of time-course data, but in multiple sets of time-course data. Typically the data are sparse: the number *t* of time points is significantly less than the number *k* of proteins. Even though there are differences between these data sets, the underlying biochemical interactions (signal) are reflected in each of these data sets. Many times these individual sets of time course protein data are modeled individually. The discussion in this paper focuses on protein measurements. However, it equally applies to sparse time course measurements obtained through gene microarrays.

The methods in this paper incorporate data from multiple replications of a systems biology investigation to determine composite posterior probabilities of network relationships. These methods are motivated by the desire to predict interactions between proteins based on probabilistically incorporating all of the data from several independent investigations. Utilizing underlying Gaussian-based regression likelihood, low informative empirical priors and Bayesian model averaging, closed form (up to a proportional constant) posterior probabilities are computed of networks, each of which is a directed acyclic graph (DAG). Extensive searching through the space of DAGs is performed with a multi structured Metropolis-Hastings Markov Chain Monte Carlo based algorithm. These model DAG posterior probabilities are combined with Bayesian model averaging [1] to produce posterior probabilities for relationships between the proteins.

Since the combined likelihood from our *m* independent replications has approximately *m* times more information, whether Akaike, Bayesian, Dirichlet information criteria or Fisher information, than that of a single replication, the combined technique tends to yield more precise estimates for posterior probabilities [2, 3]. Also in this paper, simulations demonstrate that this combined analysis captures more of a network's signal.

In a previous paper [4], an approximate Bayesian posterior analysis for a single sparse replication was developed. As with the current paper, it used multiple regression to model cotemporal associations between the proteins' measurements, where each sampled time provided insight into the proteins' relationships. Diagnostics to test the suitability of this [4] method for a particular data set were presented. These diagnostics easily can be employed on each separate replication of a multiple replication study in order to test the current paper's suitability for a particular multiple replication study. Furthermore, many of the theoretical justifications from [4] carry over to the multiple replication setting. In particular, the previous and current methodology strongly relates to low-order (small number of predictors) dependency networks [5, 6, 7]. The use of DAGs, with the proteins being the nodes and directed edges signifying the relations, allows the splitting of a replicates' likelihood into conditionally independent parts [6, 8, 9, 10]. While in [4] approximate posterior probabilities were obtained for a single replication through the use of Bayesian information based scores, in the current paper an exact Bayesian posterior probability is obtained for the combined *m* replications.

Recent papers have examined combining data from multiple studies. In [11], a score function based on the expected number of associations was developed, and its results were weighted in a Bayesian fashion with supplementary information from gene ontology and protein structures. In a frequentist, non-Bayesian manner, the authors in [12] weighted different studies so as to maximize the statistical power (chance of claiming a true positive). They also obtained integrated p-value estimates. In [13], linear programming was used to find the subnetworks which are most consistent from one replication to the next.

## Methods

Edge probabilities are computed for an undirected graph where the nodes represent individual proteins and where an edge between two nodes represents a relationship between the two corresponding proteins. These edge probabilities are based on an algorithmic search through the space of all models (DAGs) guided by the posterior probability of the DAGs. This DAG posterior probability takes into account all of the data sets. Verification of the effectiveness of this technique requires simulations. A simulation consists of generating multiple sets of data from the same underlying signal. The following discussions focus on each of the following important ideas: DAG posterior probability, algorithmic search, testing data sets and analysis techniques.

### Posterior probability

*parents-child*linear associations between the proteins. In particular, the set of parents (predictors) of a particular child (response) is the set of vertices which have directed edges going from the parent to the child. In order to give equal consideration to each child and each potential parent, each protein's time course values within a data set are standardized, using its average and standard deviation. The number of parents for any particular child is restricted to be less than or equal to

*t -*2, guaranteeing valid regression settings [14]. Acyclic refers to not allowing cycles in the graph, i.e. not allowing a protein to be a direct descendent of itself. In this paper, we present a theoretically strong probabilistic method which comprehensively incorporates multiple data sets. For convenience, the data sets are referred to as

*reps*, even though the reps may have different signal parameters and may differ in their variances about the underlying signal network. Separately for each DAG, rep and child combination, we utilize independent unit-informative empirical g-priors for the slope parameters of the parents-child linear regression relations that are specified by the DAG [15]. As well, independent unit informative inverse gamma priors are independently placed on the residual variances. Thus, due to the prior structure, the reps' data sets are independent from one replicate to the next. In addition, due to the DAG structure, each child's conditional likelihood is independent from that of another child [8]. Therefore a particular DAG's Bayes factor, is

*Y*represents the (standardized) data, and ${Y}_{i}^{\left(j\right)}$is the data for child

*i*in rep

*j*. Due to the conjugacy nature of the priors, the child

*i*rep

*j*Bayes factor, ${P}_{r}\left({Y}_{i}^{\left(j\right)}|DAG\right)$, has a closed form expression [15, Chapter 9]. Specifically, this Bayes factor is given by

*P*(

*i*) is the number of parents of child

*i*for the DAG, ${X}_{i}^{\left(j\right)}$is child

*i*'s parents data matrix for rep

*j*for the DAG, ${s}_{ij}^{2}$ is the corresponding residual regression (error) variance, and

The simple value of ${P}_{r}\left({Y}_{i}^{\left(j\right)}|DAG\right)$when *i* has no parents is a consequence of the standardized data.

From Equations (1) and (2), a closed form is easily found for the DAG's overall Bayes factor. As is common, we assume that the prior probability of one model (DAG) is the same as that of another. This yields that the posterior probability of a particular DAG, given all of the data, is proportional to the DAG's (overall) Bayes factor. Since for even a moderate number of proteins the DAG space is too large for a census, an intelligent search algorithm must be used.

### Algorithmic search

In a Markov chain manner, the Metropolis-Hastings algorithm moves through the DAG space. From equations (1) and (2), for any element, a DAG, its Bayes factor (proportional posterior probability) can be computed. Given a current element in the search space, the algorithm generates a candidate element from the current one. If the probability of the candidate is greater than the probability of the current element, then the candidate replaces the current element. If the probability of the candidate is not an improvement over the probability of the current element, then the candidate replaces the current element with a probability of $\frac{candidateBayesProb}{currentBayesProb}$. For each of 10 runs there are 50 million iterations, and the highest 200 scoring DAGs, along with their Bayes factors, are collected. These 10 lists of 200 are amalgamated into one list, *TopD*. From the list, *TopD*, probabilities for undirected graphs (*TopU*), protein-to-protein edge posterior probabilities, and other network feature probabilities are computed. Details of the Metropolis-Hastings algorithm are found in [16].

The Metropolis-Hastings algorithm used in this research is a variation of the one presented in [4, 17, 18]. This paper's single rep algorithm is a pure Bayes posterior modification of the previous Bayesian information criterion (BIC) based approximation algorithm [4]. The multiple rep algorithm has similarities to the single rep algorithm, but it moves through the DAG space based on Bayes posterior probabilities after incorporating multiple reps, and it allows multiple edge insertions and deletions.

The use of a single move, a single insertion or deletion of an edge, in a Metropolis-Hastings search is common and is motivated by the Metropolis requirement that all moves be reversible with equal probability of a move and its inverse [16]. In the multiple rep algorithm, single, double, or triple reversible moves are allowed. Each vertex in the directed acyclic graphs has bounded in-degree (a maximum number of parents of a given child), typically 3. This condition must be enforced as well. The implementation of multiple moves is straightforward. First, the number of changes (1, 2 or 3) is chosen uniformly. Second, using this chosen number, either edge insertions or deletions are selected and applied yielding a candidate directed acyclic graph. Third, if the candidate directed acyclic graph is found to be infeasible then the process of choosing a candidate starts over.

### Testing data

In order to assess the quality of the models found by the multiple rep algorithm, it is necessary to engineer replicate data where the underlying signal is known. For this paper, five studies of multiple replicate data are generated from known underlying signals.

*t*time points and

*k*proteins, we draw

*t*independent samples from a

*k*-dimensional multivariate normal distribution which has a mean vector of zeros and a selected generating covariance matrix, which provides the selected network signal. We use covariance matrices that are block diagonal with first-order autocorrelations [14, page 414] within the blocks (and zero correlations between blocks). Specifically, if proteins

*p*

_{1},

*p*

_{2}, and

*p*

_{3}constitute a block of 3 correlated proteins, then the covariance (equivalent to signal Pearson correlation) block corresponding to them is of the form:

where *ρ* is the Pearson correlation between *adjacent* proteins within the block. We say that the (triple) block of three vertices is correlated with intensity *ρ*.

For a block of 4 proteins, the partial correlations between *p*_{1} and *p*_{2} and between *p*_{3} and *p*_{4} equal $\frac{\rho}{\sqrt{{\rho}^{2}+1}}$, while the partial correlation between *p*_{2} and *p*_{3} is $\frac{\rho}{{\rho}^{2}+1}$. All other pairs of the four proteins have a partial correlation of zero.

### Analysis of data and models

Our overall strategy is to conduct five illustrative simulation studies where each study consists of a set of three reps, each generated from a specific signal. Each of the five studies are designed to examine potentially different characteristics of biological networks. The three reps in a study mimic biological replicates. For each study, the multiple rep Bayesian Metropolis-Hastings algorithm is applied to all three replicates, giving the composite results. For comparison purposes, the single rep algorithm is applied to each of the three individual reps. Separately, for the composite and for each of the individual executions of the modeling algorithms, the matrix of protein-protein edge posterior probabilities and the vector of within-block connectivity probabilities are obtained.

*connected*if given any pair of the nodes in the block, there exists a sequence of edges from the first node in the pair to the second node in the pair, where each edge is incident only with nodes in the block. The probability that a block of nodes, representing our proteins, is connected is estimated by the sum of the probabilities of the top undirected graphs in which those nodes are connected. Mathematically, this is:

where the characteristic function χ(*v*_{1}*, . . . , v*_{ n }*, G*) has the value 1 if and only if the vertices *v*_{1}*, . . . , v*_{ n } are connected in the undirected graph *G*. The computation of the characteristic function χ () for 3 and 4 nodes is straightforward.

## Results

Specific results of the five simulation studies are presented. The discussion of the first study is more detailed than that of the remaining four since some of the details of all five studies are quite similar. Following the discussion of the individual studies, a further analysis of the posterior probabilities is presented.

### Individual studies

The first of five simulation studies is a set of three reps, ${R}_{1}^{\left(1\right)}$, ${R}_{2}^{\left(1\right)}$ and ${R}_{3}^{\left(1\right)}$. Each of these reps reflects *t* = 10 simulated measurements of twelve proteins, ${p}_{1}^{\left(1\right)},\dots ,{p}_{12}^{\left(1\right)}$. Furthermore, the underlying generating covariance matrix has assigned high correlation intensity, *ρ* = 0.94, to each of the four blocks $\left\{{p}_{1}^{\left(1\right)},\dots ,{p}_{3}^{\left(1\right)}\right\}$, $\left\{{p}_{4}^{\left(1\right)},\dots ,{p}_{6}^{\left(1\right)}\right\}$, $\left\{{p}_{7}^{\left(1\right)},\dots ,{p}_{8}^{\left(1\right)}\right\}$, and $\left\{{p}_{10}^{\left(1\right)},\dots ,{p}_{12}^{\left(1\right)}\right\}$, as described in the Methods. The observed Pearson correlations in the reps are close to the Pearson correlations of their generator.

Study 1 posterior probabilities

C | ${\mathbf{R}}_{\mathbf{1}}^{\left(\mathbf{1}\right)}$ | ${\mathbf{R}}_{\mathbf{2}}^{\left(\mathbf{1}\right)}$ | ${\mathbf{R}}_{\mathbf{3}}^{\left(\mathbf{1}\right)}$ | |
---|---|---|---|---|

${p}_{1}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{2}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{3}^{\left(1\right)}$ | 1.0 | 1.0 | 0.9744 | 1.0 |

${p}_{4}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{5}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{6}^{\left(1\right)}$ | 1.0 | 1.0 | 1.0 | 1.0 |

${p}_{7}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{8}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{9}^{\left(1\right)}$ | 1.0 | 1.0 | 1.0 | 1.0 |

${p}_{10}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{11}^{\left(1\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{12}^{\left(1\right)}$ | 1.0 | 1.0 | 0.993 | 1.0 |

TFP Avg | 0.0376 | 0.0246 | 0.0641 | 0.0346 |

TLFP Avg | 0.0008 | 0.0008 | 0.0047 | 0.0017 |

Consider a particular one of the generating 3 *×* 3 blocks in this first simulation, say based on the ordered proteins: *A*, *B*, and *C*. It has high moderate Pearson correlations between all of its protein pairs. There is also substantial partial correlation between two adjacency proteins, namely between *A* and *B*, and also between *B* and *C*. However, there is zero partial correlation between the non-adjacent proteins *A* and *C*. In other words, for a fixed value of *B*, there is no correlation between *A* and *C*. It is informative to compare this setting to the biological setting where a parent protein *A*^{ * } has a causal influence on a child protein *B*^{ * } which has a causal influence on a grandchild protein *C*^{ * }. However, for a fixed result for *B*^{ * }, *A*^{ * } has no influence on *C*^{ * }. Hence, there is no partial causal influence between *A*^{ * } and *C*^{ * }.

Lasso and adaptive Lasso partial correlation estimates for ${R}_{1}^{\left(1\right)}$

${p}_{1}^{\left(1\right)}$ | ${p}_{2}^{\left(1\right)}$ | ${p}_{3}^{\left(1\right)}$ | ${p}_{4}^{\left(1\right)}$ | ${p}_{5}^{\left(1\right)}$ | ${p}_{6}^{\left(1\right)}$ | ${p}_{7}^{\left(1\right)}$ | ${p}_{8}^{\left(1\right)}$ | ${p}_{9}^{\left(1\right)}$ | ${p}_{10}^{\left(1\right)}$ | ${p}_{11}^{\left(1\right)}$ | ${p}_{12}^{\left(1\right)}$ | |
---|---|---|---|---|---|---|---|---|---|---|---|---|

${p}_{1}^{\left(1\right)}$ | | 0.77 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{2}^{\left(1\right)}$ | 0.84 | | 0.56 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{3}^{\left(1\right)}$ | 0.00 | 0.48 | | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{4}^{\left(1\right)}$ | 0.00 | 0.00 | 0.00 | | 0.76 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{5}^{\left(1\right)}$ | 0.00 | 0.00 | 0.00 | 0.58 | | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{6}^{\left(1\right)}$ | 0.00 | 0.00 | 0.00 | 0.00 | 0.42 | | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{7}^{\left(1\right)}$ | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | | 0.96 | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{8}^{\left(1\right)}$ | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.70 | | 0.00 | 0.00 | 0.00 | 0.00 |

${p}_{9}^{\left(1\right)}$ | 0.00 | 0.00 | 0.15 | 0.00 | 0.00 | 0.00 | 0.32 | 0.28 | | 0.00 | 0.00 | 0.00 |

${p}_{10}^{\left(1\right)}$ | 0.00 | -0.03 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | | 0.89 | 0.27 |

${p}_{11}^{\left(1\right)}$ | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 | 0.01 | 0.00 | 0.00 | 0.63 | | 0.00 |

${p}_{12}^{\left(1\right)}$ | 0.00 | 0.00 | 0.06 | 0.00 | 0.00 | 0.00 | 0.01 | 0.00 | 0.00 | 0.50 | 0.28 | |

Ridge partial correlation estimates for ${R}_{1}^{\left(1\right)}$

${p}_{1}^{\left(1\right)}$ | ${p}_{2}^{\left(1\right)}$ | ${p}_{3}^{\left(1\right)}$ | ${p}_{4}^{\left(1\right)}$ | ${p}_{5}^{\left(1\right)}$ | ${p}_{6}^{\left(1\right)}$ | ${p}_{7}^{\left(1\right)}$ | ${p}_{8}^{\left(1\right)}$ | ${p}_{9}^{\left(1\right)}$ | ${p}_{10}^{\left(1\right)}$ | ${p}_{11}^{\left(1\right)}$ | |
---|---|---|---|---|---|---|---|---|---|---|---|

${p}_{2}^{\left(1\right)}$ | 0.75 | | |||||||||

${p}_{3}^{\left(1\right)}$ | 0.10 | 0.30 | | ||||||||

${p}_{4}^{\left(1\right)}$ | 0.49 | -0.49 | 0.00 | | |||||||

${p}_{5}^{\left(1\right)}$ | 0.00 | 0.06 | 0.13 | 0.54 | | ||||||

${p}_{6}^{\left(1\right)}$ | -0.16 | 0.14 | 0.06 | 0.31 | 0.31 | | |||||

${p}_{7}^{\left(1\right)}$ | 0.14 | 0.05 | 0.00 | 0.02 | -0.07 | 0.00 | | ||||

${p}_{8}^{\left(1\right)}$ | -0.16 | 0.08 | 0.07 | 0.11 | -0.05 | -0.08 | 0.49 | | |||

${p}_{9}^{\left(1\right)}$ | 0.00 | -0.04 | 0.16 | 0.13 | -0.07 | 0.00 | 0.43 | 0.39 | | ||

${p}_{10}^{\left(1\right)}$ | 0.07 | -0.09 | 0.00 | 0.18 | 0.00 | 0.00 | -0.20 | -0.05 | -0.04 | | |

${p}_{11}^{\left(1\right)}$ | 0.21 | 0.03 | -0.08 | -0.16 | 0.00 | 0.06 | 0.38 | -0.00 | -0.09 | 0.50 | |

${p}_{12}^{\left(1\right)}$ | -0.21 | -0.05 | 0.09 | -0.22 | 0.11 | 0.05 | -0.05 | 0.00 | 0.19 | 0.54 | 0.40 |

Pearson correlations for ${R}_{1}^{\left(1\right)}$

${p}_{1}^{\left(1\right)}$ | ${p}_{2}^{\left(1\right)}$ | ${p}_{3}^{\left(1\right)}$ | ${p}_{4}^{\left(1\right)}$ | ${p}_{5}^{\left(1\right)}$ | ${p}_{6}^{\left(1\right)}$ | ${p}_{7}^{\left(1\right)}$ | ${p}_{8}^{\left(1\right)}$ | ${p}_{9}^{\left(1\right)}$ | ${p}_{10}^{\left(1\right)}$ | ${p}_{11}^{\left(1\right)}$ | |
---|---|---|---|---|---|---|---|---|---|---|---|

${p}_{1}^{\left(1\right)}$ | 0.94 | | |||||||||

${p}_{2}^{\left(1\right)}$ | 0.82 | 0.92 | | ||||||||

${p}_{3}^{\left(1\right)}$ | 0.16 | 0.14 | 0.31 | | |||||||

${p}_{4}^{\left(1\right)}$ | 0.18 | 0.22 | 0.39 | 0.90 | | ||||||

${p}_{5}^{\left(1\right)}$ | 0.07 | 0.19 | 0.31 | 0.78 | 0.86 | | |||||

${p}_{6}^{\left(1\right)}$ | 0.55 | 0.64 | 0.61 | -0.06 | -0.18 | -0.09 | | ||||

${p}_{7}^{\left(1\right)}$ | 0.51 | 0.61 | 0.62 | -0.05 | -0.16 | -0.12 | 0.98 | | |||

${p}_{8}^{\left(1\right)}$ | 0.54 | 0.64 | 0.69 | 0.03 | -0.07 | -0.01 | 0.97 | 0.97 | | ||

${p}_{9}^{\left(1\right)}$ | -0.20 | -0.18 | -0.08 | -0.09 | 0.00 | 0.12 | 0.02 | -0.02 | 0.01 | | |

${p}_{10}^{\left(1\right)}$ | -0.11 | 0.08 | -0.02 | -0.16 | -0.06 | 0.08 | 0.17 | 0.10 | 0.13 | 0.97 | |

${p}_{11}^{\left(1\right)}$ | -0.06 | 0.01 | 0.15 | -0.01 | 0.09 | 0.22 | 0.26 | 0.21 | 0.27 | 0.95 | 0.95 |

The second simulation study consists of *t* = 5 measurements of 9 proteins, ${p}_{1}^{\left(2\right)},\dots ,{p}_{9}^{\left(2\right)}$. The replicate data is generated from a signal giving correlation intensity *ρ* = 0.9 to each of the two triple blocks, $\left\{{p}_{1}^{\left(2\right)},{p}_{2}^{\left(2\right)},{p}_{3}^{\left(2\right)}\right\}$ and $\left\{{p}_{7}^{\left(2\right)},{p}_{8}^{\left(2\right)},{p}_{9}^{\left(2\right)}\right\}$, as well as the block of two proteins $\left\{{p}_{4}^{\left(2\right)},{p}_{5}^{\left(2\right)}\right\}$. All other pairs of proteins are assigned zero correlation. Note that protein ${p}_{6}^{\left(2\right)}$has zero signal correlation to all other proteins.

*C*

^{(2)}as well as the individual rep analysis of ${R}_{1}^{\left(2\right)}$ are successful in recognizing the generating signal, whereas individual rep analyses of ${R}_{2}^{\left(2\right)}$ and ${R}_{3}^{\left(2\right)}$ do not fare as well. Many of these ${R}_{2}^{\left(2\right)}$ and ${R}_{3}^{\left(2\right)}$ deviations from signal are associated with corresponding deviations of their sample partial correlations from the signal, which is most likely caused by the small sample size of

*t*= 5.

Study 2 posterior probabilities

C | ${\mathbf{R}}_{\mathbf{1}}^{\left(\mathbf{1}\right)}$ | ${\mathbf{R}}_{\mathbf{2}}^{\left(\mathbf{1}\right)}$ | ${\mathbf{R}}_{\mathbf{3}}^{\left(\mathbf{1}\right)}$ | |
---|---|---|---|---|

${p}_{1}^{\left(2\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{2}^{\left(2\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{3}^{\left(2\right)}$ | 1.0 | 0.9902 | 0.0416 | 0.7189 |

${p}_{7}^{\left(2\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{8}^{\left(2\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{9}^{\left(2\right)}$ | 1.0 | 1.0 | 0.9397 | 1.0 |

TEP Avg | 0.1482 | 0.1053 | 0.1680 | 0.1440 |

TLFP Avg | 0.0185 | 0.0212 | 0.0821 | 0.0377 |

${p}_{4}^{\left(2\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{5}^{\left(2\right)}\phantom{\rule{0.3em}{0ex}}$ | 0.8624 | 0.9974 | 0.0659 | 0.3481 |

DFP Avg | 0.1943 | 0.1493 | 0.2612 | 0.2174 |

DLEP Avg | 0.0488 | 0.0536 | 0.1692 | 0.0996 |

The third study contains 10 measurements of 10 proteins, ${p}_{1}^{\left(3\right)},\dots ,{p}_{10}^{\left(3\right)}$. An underlying signal is generated with correlation intensity *ρ* = 0.94 for the two quadruple blocks of vertices $\left\{{p}_{1}^{\left(3\right)},\dots ,{p}_{4}^{\left(3\right)}\right\}$ and $\left\{{p}_{5}^{\left(3\right)},\dots ,{p}_{8}^{\left(3\right)}\right\}$ as well as the double block $\left\{{p}_{9}^{\left(3\right)},{p}_{10}^{\left(3\right)}\right\}$.

High signal correlations between proteins yield high sample Pearson correlations (ranging from 0.73 to 0.99) within the reps. Additionally, the zero signal correlation pairs yield low sample Pearson correlations, ranging from -0.13 to 0.40.

Study 3 posterior probabilities

C | ${\mathbf{R}}_{\mathbf{1}}^{\left(\mathbf{3}\right)}$ | ${\mathbf{R}}_{\mathbf{2}}^{\left(\mathbf{3}\right)}$ | ${\mathbf{R}}_{\mathbf{3}}^{\left(\mathbf{3}\right)}$ | |
---|---|---|---|---|

${p}_{1}^{\left(3\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{2}^{\left(3\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{3}^{\left(3\right)}-{p}_{4}^{\left(3\right)}$ | 1.0 | 1.0 | 1.0 | 1.0 |

${p}_{5}^{\left(3\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{6}^{\left(3\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{7}^{\left(3\right)}-{p}_{8}^{\left(3\right)}$ | 1.0 | 1.0 | 1.0 | 1.0 |

QFP Avg | 0.0611 | 0.0683 | 0.0809 | 0.0459 |

QLFP Avg | 0.0035 | 0.0038 | 0.0059 | 0.0008 |

${p}_{9}^{\left(3\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{10}^{\left(3\right)}\phantom{\rule{0.3em}{0ex}}$ | 1.0 | 1.0 | 1.0 | 1.0 |

DFP Avg | 0.094 | 0.1203 | 0.1044 | 0.0751 |

DLEP Avg | 0.0232 | 0.0094 | 0.0144 | 0.0045 |

*ρ*values to the four blocks of three proteins are 0.7, 0.6, 0.6 and 0.5, respectively. As in the previous three studies, average sample correlations between proteins in different blocks remain low throughout the reps, ranging from -0.0329 to 0.1159. The sample correlations between proteins that are signal correlated within blocks are representative of the signal correlations. The exception occurs in ${R}_{2}^{\left(4\right)}$, where ${R}_{2}^{\left(4\right)}$ has relatively low within block sample correlations ranging from

*-*0.1141 to 0.7817. This influenced its signal inconsistencies in edge probabilities (refer to Figure 4). In general, triples associated with lower generating

*ρ*values receive lower and more inconsistent correlations, which also speak to some edge probability inconsistencies in Figure 4.

Study 4 posterior probabilities

C | ${\mathbf{R}}_{\mathbf{1}}^{\left(\mathbf{4}\right)}$ | ${\mathbf{R}}_{\mathbf{2}}^{\left(\mathbf{4}\right)}$ | ${\mathbf{R}}_{\mathbf{3}}^{\left(\mathbf{4}\right)}$ | |
---|---|---|---|---|

${p}_{1}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{2}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{3}^{\left(4\right)}$ | 1.0 | 1.0 | 0.0 | 1.0 |

${p}_{4}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{5}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{6}^{\left(4\right)}$ | 0.9874 | 0.3109 | 0.6007 | 0.8037 |

${p}_{7}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{8}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{9}^{\left(4\right)}$ | 0.0455 | 0.0021 | 0.0095 | 0.0106 |

${p}_{10}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{11}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{12}^{\left(4\right)}$ | 0.3655 | 0.0038 | 0.1190 | 0.0117 |

TFP Avg | 0.0853 | 0.0467 | 0.0692 | 0.0700 |

TLFP Avg | 0.0043 | 0.0018 | 0.0046 | 0.0068 |

*ρ*= 0.9, the second replicate has

*ρ*= 0.82 and the third replicate has

*ρ*= 0.7. In this study the generating signal is not as strong as the signals in the first three studies. The block correlations for each replicate are derived from the

*ρ*value assigned to each rep. As expected, ${R}_{1}^{\left(5\right)}$ has the highest sample correlations among triple signal correlated proteins, followed by ${R}_{2}^{\left(5\right)}$ and lastly ${R}_{3}^{\left(5\right)}$. All have small sample correlation averages among signal zero correlated proteins. The edge probability diagrams (Figure 5) are symbolic of these results.

*C*

^{(5)}and ${R}_{1}^{\left(5\right)}$ show triple connectivity probabilities of 1.0 for all blocks, while the single rep analysis for ${R}_{2}^{\left(5\right)}$ does not recognize signal in $\left\{{p}_{1}^{\left(5\right)},{p}_{2}^{\left(5\right)},{p}_{3}^{\left(5\right)}\right\}$ (see Figure 5). The analysis for ${R}_{3}^{\left(5\right)}$, which was generated with the lowest correlation intensity, does not recognize the signal as well as those of ${R}_{1}^{\left(5\right)}$ and

*C*

^{(5)}.

Study 5 posterior probabilities

C | ${\mathbf{R}}_{\mathbf{1}}^{\left(\mathbf{5}\right)}$ | ${\mathbf{R}}_{\mathbf{2}}^{\left(\mathbf{5}\right)}$ | ${\mathbf{R}}_{\mathbf{3}}^{\left(\mathbf{5}\right)}$ | |
---|---|---|---|---|

${p}_{1}^{\left(5\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{2}^{\left(5\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{3}^{\left(5\right)}$ | 1.0 | 1.0 | 0.0016 | 1.0 |

${p}_{4}^{\left(5\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{5}^{\left(5\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{6}^{\left(5\right)}$ | 1.0 | 1.0 | 0.9874 | 0.9819 |

${p}_{7}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{8}^{\left(4\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{9}^{\left(4\right)}$ | 1.0 | 1.0 | 1.0 | 1.0 |

${p}_{10}^{\left(5\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{11}^{\left(5\right)}\phantom{\rule{0.3em}{0ex}}-{p}_{12}^{\left(5\right)}$ | 1.0 | 1.0 | 1.0 | 0.3477 |

TFP Avg | 0.0643 | 0.0418 | 0.0854 | 0.0613 |

TLFP Avg | 0.0019 | 0.0023 | 0.0017 | 0.0017 |

### ROC analysis of posterior probabilities

*x*,

*y*) coordinates are generated by the decreasing sequence of edge posterior probability

*cutoffs*(i.e. lower limits for classifying positive edges). The

*y*-coordinate, the true positive rate (TPR), is the fraction of signal edges that are classified as positive edges. The

*x*-coordinate, the false positive rate (FPR), is the fraction of signal non-edges that are classified as positive edges. In each of Figures 6(a)-6(d), comparisons are made between ROC curves whose signal edges are determined by non-zero Pearson correlation (ranging from 0.250 to 0.940) versus those whose signal edges are determined by non-zero partial correlation (ranging from 0.447 to 0.685). The signal partial correlation ROC curves tend to be above and to the left of the signal Pearson correlation ROC curves. This represents the algorithm's ability to identify, with higher posterior probability, signal partial correlation edges over signal Pearson correlation edges.

As can be seen in covariance block (3) and partial correlation matrix (4), there are more signal Pearson correlation edges than signal partial correlation edges. For the ROC curves based on partial correlation edges, there are only two edges within a triple block, and each has a relatively high posterior probability. This leads to their ROC curves increasing at a faster rate than those based on Pearson correlation edges. Overall, the signal non-zero Pearson correlation edges have lower posterior probabilities than do the signal non-zero partial correlation edges. In addition, the composite ROC curves tend to be to the upper left of their corresponding individual replication ROC curves. This corresponds to higher posterior probabilities for the true signal edges under composite analysis than under most individual rep analyses.

## Conclusions

Structured Bayesian posterior probabilities are developed for network features based on multiple sparse time-course data sets. This methodology allows for the incorporation of data sets with varying degree of experimental variability. For our simulations, the multiple rep composite method performs well in uncovering strong network signals. The composite method does better than a single rep method in uncovering moderate network signals. The composite method assigns high posterior probability to edges with at least moderate network partial correlation, while it assigns moderate to small posterior probabilities to edges with 0.0 network partial correlation.

Composite ROC curves based on system non-zero partial correlation (solid lines in Figure 6(a)) have small area above them which signifies that our composite method provides excellent detection of edges having partial correlation.

The five simulation studies span a range of network situations. The first three studies examine networks consisting of block subnetworks with high signal correlations within blocks. These blocks are of varying sizes. The composite method is more successful in identifying blocks of three or four proteins, rather than smaller blocks. For study four, blocks with moderate, rather than high, signal correlation within the blocks are examined. The composite method does not perform as well for these blocks but it does outperform the single rep method for all study four's subnetworks. In the fifth study, the different reps have varying degree of experimental variability. Still, the composite method recognizes the network's signals with high posterior probabilities.

The multiple rep method utilizes independent empirical priors acting on independent reps. Thus, as suggested by the fifth study, this method can be used even if there are major fixed, non-random differences between the reps. Each rep still contributes information about the network structure. This likelihood based methodology automatically weights the reps in the sense that reps having more experimental variability will receive less weight in determining the subnetworks which have highest posterior probability.

The computation of posterior probabilities lends itself towards the identification of various network features. These features can correspond to connected subgraphs in the interaction network. With experimental data, where the goal is to discover the generating signal, searching for high probability features is quite valuable.

If there are strictly random differences between the reps, it may be useful to employ a hierarchical structure (e.g. assuming that parent-child regression slopes for one rep come from the same distributions as do those from another rep). This approach would involve substantially more complex Bayes factors, and thus would be more computationally intensive. We are currently developing methodology for this setting.

## Notes

### Acknowledgements

The authors thank the Center for Molecular Communication and Signaling of Wake Forest University for their support through a grant. The authors thank our colleague, Jacquelyn Fetrow, who suggested the ROC analysis after one of our research presentations. The authors also thank Steven Wicker for his valuable assistance with the manipulation of graphical images.

This article has been published as part of *BMC Bioinformatics* Volume 13 Supplement 9, 2012: Selected articles from the IEEE International Conference on Bioinformatics and Biomedicine 2011: Bioinformatics. The full contents of the supplement are available online at http://www.biomedcentral.com/bmcbioinformatics/supplements/13/S9.

### References

- 1.Hoeting JA, Madigan D, Raftery AE, Volinsky CT: Bayesian Model Averaging: a tutorial (with comments by M. Clyde, David Draper and E.I. George, and a rejoinder by the authors). Statistical Science. 1999, 14 (4): 382-417. 10.1214/ss/1009212519.CrossRefGoogle Scholar
- 2.Burnham K, Anderson D: Model Selection and Multimodel Inference: a Practical Information-Theoretic Approach. 2002, New York: Springer, 2Google Scholar
- 3.DeGroot MH, Schervish MJ: Probability and Statistics. 2002, Addison-WesleyGoogle Scholar
- 4.John DJ, Fetrow JS, Norris JL: Continuous cotemporal probabilistic modeling of systems biology networks from sparse data. IEEE/ACM Trans Comput Biol Bioinform. 2011, 8 (5): 1208-1222.PubMedCentralCrossRefPubMedGoogle Scholar
- 5.Heckerman D, Chickering DM, Meek C, Rounthwaite R, Kadie C: Dependency networks for inference, collaborative filtering, and data visualization. J Mach Learn Res. 2000, 1: 49-75.Google Scholar
- 6.Wille A, Zimmermann P, Vranová E, Fürholz A, Laule O, Bleuler S, Hennig L, Prelic A, von Rohr P, Thiele L, Zitzler E, Gruissem W, Bühlmann P: Sparse graphical Gaussian modeling of the isoprenoid gene network in
*Arabidopsis thaliana*. Genome Biol. 2004, 5: R92-10.1186/gb-2004-5-11-r92.PubMedCentralCrossRefPubMedGoogle Scholar - 7.Wille A, Bühlmann P: Low-order conditional independence graphs for inferring genetic networks. Stat Appl Genet Mol Biol. 2006, 5: Article1-PubMedGoogle Scholar
- 8.Friedman N, Nachman I, Pe'er D: Learning Bayesian network structures from massive datasets: the sparse candidate algorithm. Proceedings of Uncertainty in Artificial Intelligence. 1999, 196-205.Google Scholar
- 9.Pearl J: Probabilistic Reasoning in Intelligent Systems. 1988, Morgan KaufmannGoogle Scholar
- 10.Segal E, Pe'er D, Regev A, Koller D, Friedman N: Learning module networks. J Mach Learn Res. 2005, 6: 557-588.Google Scholar
- 11.Lee H, Deng M, Sun F, Chen T: An integrated approach to the prediction of domain-domain interactions. BMC Bioinformatics. 2006, 7: 269-10.1186/1471-2105-7-269.PubMedCentralCrossRefPubMedGoogle Scholar
- 12.Hwang D, Rust AG, Ramsey S, Smith JJ, Leslie DM, Weston AD, de Atauri P, Aitchison JD, Good L, Siegel AF, Bolouri H: A data integration methodology for systems biology. Proc Natl Acad Sci USA. 2005, 102 (48): 17296-17301. 10.1073/pnas.0508647102.PubMedCentralCrossRefPubMedGoogle Scholar
- 13.Wang Y, Joshi T, Zhang XS, Xu D, Chen L: Inferring gene regulatory networks from multiple microarray datasets. Bioinformatics. 2006, 22 (19): 2413-2420. 10.1093/bioinformatics/btl396.CrossRefPubMedGoogle Scholar
- 14.Johnson RA, Wichern DW: Applied Multivariate Statistical Analysis. 2007, Prentice-Hall, 6Google Scholar
- 15.Hoff P: A First Course in Bayesian Statistical Methods. 2009, SpringerCrossRefGoogle Scholar
- 16.Chib S, Greenberg E: Understanding the Metropolis-Hastings algorithm. Am Stat. 1995, 49 (4): 327-335.Google Scholar
- 17.Black RA, Fetrow JS, John DJ, Norris JL: Examining effects of variability on systems biology modeling algorithms. ACM-BCB 2010. Edited by: Liao L, Li G, Zhang A, Borodovsky M, Ozoyoglu G, Mikler AR. 2010, Niagara Falls, New York, USA: ACM, 374-377.Google Scholar
- 18.John DJ, Fetrow JS, Norris JL: Metropolis-Hastings algorithm and continuous regression for finding next-state models of protein modification using information scores. Proceedings of the 7th International Symposium on BioInformatics and BioEngineering, Volume I. Edited by: Yang JY, Yang MQ, Zhu MM, Zhang Y, Arabnia HR, Deng Y, Bourbakis N. 2007, IEEE, 35-41.Google Scholar
- 19.de la Fuente A, Bing N, Hoeschele I, Mendes P: Discovery of meaningful associations in genomic data using partial correlation coefficients. Bioinformatics. 2004, 20 (18): 3565-3574. 10.1093/bioinformatics/bth445.CrossRefPubMedGoogle Scholar
- 20.Dobra A, Hans C, Jones B, Nevins JR, West M: Sparse graphical models for exploring gene expression data. J Multivar Anal. 2004, 90: 196-212. 10.1016/j.jmva.2004.02.009.CrossRefGoogle Scholar
- 21.Krämer N, Schäfer J, Boulesteix AL: Regularized estimation of large-scale gene association networks using graphical Gaussian models. BMC Bioinformatics. 2009, 10: 384-10.1186/1471-2105-10-384.PubMedCentralCrossRefPubMedGoogle Scholar
- 22.Lauritzen SL: Graphical Models. 1996, Oxford Clarendon PressGoogle Scholar
- 23.Li H, Gai J: Gradient directed regularization for sparse Gaussian, concentration graphs with applications to inference of genetic networks. Biostatistics. 2006, 7 (2): 302-317.CrossRefPubMedGoogle Scholar
- 24.Magwene PM, Kim J: Estimating genome expression networks using first-order conditional independence. Genome Biol. 2004, 5: R100-10.1186/gb-2004-5-12-r100.PubMedCentralCrossRefPubMedGoogle Scholar
- 25.Schäfer J, Strimmer K: A shrinkage approach to large-scale covariance matrix estimation and implications for functional genomics. Stat Appl Genet Mol Biol. 2005, 4: Article32-PubMedGoogle Scholar
- 26.Schäfer J, Strimmer K: An empirical Bayes approach to inferring large-scale gene association networks. Bioinformatics. 2005, 21: 754-764. 10.1093/bioinformatics/bti062.CrossRefPubMedGoogle Scholar
- 27.Mason SJ, Graham NE: Areas beneath the relative operating characteristics (ROC) and relative operating levels (ROL) curves: statistical significance and interpretation. Quarterly Journal of the Royal Meterological Society. 2002, 128: 2145-2166. 10.1256/003590002320603584.CrossRefGoogle Scholar

## Copyright information

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.