- Research
- Open Access
- Published:

# Activating and inhibiting connections in biological network dynamics

*Biology Direct*
**volume 3**, Article number: 49 (2008)

## Abstract

### Background

Many studies of biochemical networks have analyzed network topology. Such work has suggested that specific types of network wiring may increase network robustness and therefore confer a selective advantage. However, knowledge of network topology does not allow one to predict network dynamical behavior – for example, whether deleting a protein from a signaling network would maintain the network's dynamical behavior, or induce oscillations or chaos.

### Results

Here we report that the balance between activating and inhibiting connections is important in determining whether network dynamics reach steady state or oscillate. We use a simple dynamical model of a network of interacting genes or proteins. Using the model, we study random networks, networks selected for robust dynamics, and examples of biological network topologies. The fraction of activating connections influences whether the network dynamics reach steady state or oscillate.

### Conclusion

The activating fraction may predispose a network to oscillate or reach steady state, and neutral evolution or selection of this parameter may affect the behavior of biological networks. This principle may unify the dynamics of a wide range of cellular networks.

### Reviewers

Reviewed by Sergei Maslov, Eugene Koonin, and Yu (Brandon) Xia (nominated by Mark Gerstein). For the full reviews, please go to the Reviewers' comments section.

## Background

Many biological processes involve networks of interacting proteins or genes. Examples include networks that control the cell cycle, transcriptional regulation, cellular signaling, and cell-fate determination in development. As more biochemical networks are mapped, detailed analysis of networks has become possible. Many researchers have analyzed the connections among nodes in the networks [1–3]. Different studies have emphasized the importance of network structure, motifs, or other properties [4–7]. While the topology of biochemical networks is informative – for example, feedback loops are necessary for oscillatory dynamics – topology does not fully describe network behavior. The dynamic response to different inputs is a key property that biological networks have evolved; perturbing the network can alter the dynamics [8], and the topological structure of the network may be a byproduct of selection for dynamical behavior [9–11]. Understanding the relationship between network topology and dynamics can give insight into the evolution of cellular oscillators and switches [12–14].

Using a simple model of biochemical network dynamics [15], we show that the balance between activating and inhibiting connections strongly influences whether network dynamics reach steady state or oscillate. A high fraction of activating connections predisposes network dynamics to reach steady state. Tuning this parameter alters the oscillation period. We found significant dependence of the dynamics on the fraction of activating connections in random networks, optimized networks, and examples of biological networks. Our work is related to previous work on the sign of interactions in transcription modules [16] and biological network subsystems [17].

Our model includes key features of a biochemical network: interactions of varying strengths, strongly nonlinear dynamics, and saturating response to inputs [15]. Variants of the model have been used to study robustness in genetic networks, with a focus on dynamics that reach steady state [18–21]. The model describes interactions among the nodes – which represent the genes, mRNA transcripts, or proteins – and the activity of each node – which represents the expression and/or activity level of the molecule. (Activity of a molecule changes if its concentration changes, or because of chemical changes such as phosphorylation.) The interaction strengths are given by the matrix *W*, where *W*_{
ij
}is the strength of the effect of node *j* on node *i* (fig. 1). Each *W*_{
ij
}can be positive (activating), negative (inhibiting), or zero (no interaction). The activating fraction *a* is the fraction of nonzero interaction strengths which are positive. Nodes can self-regulate, an effect known to be important in biochemical networks [22, 23]. The activity vector is **s**, with *s*_{
i
}the activity of node *i*. Each *s*_{
i
}is between -1 ("off") and 1 ("on"), where 0 corresponds to the basal activity of the node. This model is similar to a class of models of neuronal networks [24], where the balance of activating and inhibiting connections is also of interest [25–27].

The activity vector changes in time according to

**s**(*t* + 1) = tanh(*rW* **s**(*t*)).

The hyperbolic tangent saturates the interactions among nodes so that *s*_{
i
}∈ [-1, 1]. We use *r* = 100 unless otherwise specified. The dynamics can reach steady state (fig. 1e), oscillate (fig. 1f) or change chaotically. The dynamics are fully specified by eq. 1 and the initial condition. We randomly assigned *s*_{
i
}(0) = 1 or -1 with equal probability. We then determined whether the dynamics reach a steady state and the period of oscillation.

## Results and discussion

We first studied the dynamics of random Erdös-Renyi networks. We generated networks with specified size *N*, probability of nonzero connection between nodes *c*, activating fraction connections *a*, and random Gaussian *W*_{
ij
}.

For a wide range of conditions, the activating fraction *a* is strongly correlated with the probability that the dynamics of a random network reach steady state. For *a* near 1, nearly all runs of the network dynamics reach steady state, while for *a* near 0, few runs reach steady state; the network dynamics typically oscillate (fig. 2). When *a* = 0, ~0.01% to 10% of runs reach steady state, depending on network degree. The probability of reaching steady state is only weakly dependent on the size of the network, if the number of connections per node is fixed (fig. 2). Altering the typical magnitude of connection strengths has little effect on the dynamics, because the dynamics are highly saturated (see Methods).

The average period of oscillation varies strongly with *a*. When *a* = 1, most runs reach steady state (period 1), while for *a* = 0, period-2 oscillations are typical (the activity of each node switches between -1 and 1). The longest average period occurs for *a* = 0.5, where activating and inhibiting connections are equally likely (fig. 2). Increasing *N* or *c* tends to increase the period.

In random Erdös-Renyi networks, changing *a* has a strong effect on the dynamics. We then studied how the activating fraction affects the ability of a mutation/selection process to optimize the dynamical behavior. We numerically optimized 10-node networks to robustly reach steady state or oscillate, using the great deluge algorithm [28]. The convergence rate of the algorithm reflects the density of networks in parameter space: optimization converges more rapidly in regions of parameter space in which desirable networks are dense. We found that the number of mutations required to change a random graph into a network with robust dynamics strongly varies with *a*.

Although robustness is a key property of biological networks [11, 29, 30] there is no consistent definition of network robustness. The choice of definition is important because increasing one type of robustness can decrease another type [31]. We considered three types of robust dynamical behavior: *topological robustness* means the dynamics are robust to alterations in the network wiring; *mutational robustness* means the dynamics are robust to alterations in the interaction strengths (with no changes to which nodes are connected); and *environmental robustness* means that the dynamics are robust to changes in the initial conditions. We optimized for topologically and environmentally robust networks: we perturbed the network topology and measured how the dynamics changed, averaged over many initial conditions. To quantify robustness, we measured changes in the steady state (if a steady state is reached) or the oscillation period (see Methods).

Networks which robustly reach steady state were rapidly found by our optimization for high *a* = 0.8: the optimization required 26 iterations to converge, on average (fig. 3a). For lower *a*, the results depend on the network degree: the increase in number of iterations to converge was a factor of ~2 to 25 when *a* was decreased to 0.2. Small increases occurred in *a* when selecting for steady state behavior. The networks that were initialized with the lowest value of *a* = 0.2 showed statistically significant increases in the average *a* of 0.13–0.45.

Optimization for low-period oscillation converged most rapidly for low *a* and for long-period oscillation at intermediate *a* (fig. 3b–d). We note that robust period-2 oscillation was found with fewer iterations of the optimization procedure than robust steady state, while long-period oscillation required many more iterations. For period-2 oscillation, the convergence slowed with *a* in a way that depended on the network degree.

When we optimized for long-period oscillation, a non-optimal *a* of 0.2 or 0.8 required 1.6–2 times more iterations to converge than an optimal *a* of 0.5, for a degree of 8 connections per node (fig. 3c–d). When selecting for long-period oscillation, we observed statistically significant (see Methods) shifts in *a* of the selected networks: the activating fraction shifted toward *a* = 0.5.

Both our network-optimization and random network results suggest that high *a* promotes dynamics that reach steady state, while intermediate and low *a* promotes oscillatory dynamics. Low *a* is correlated with short-period oscillation; intermediate *a* is correlated with long-period oscillation. To understand the connection between these observations and the dynamics of biochemical networks, we studied 5 relatively small, and therefore tractable, biological network topologies. While we did not attempt to replicate biochemically realistic dynamics of these networks, we sought to understand how the network topology and activating fraction together affect the dynamics. For biochemical networks, the network topology is clearly important, because these networks have been influenced by billions of years of evolution. But are such networks therefore insensitive to changes in *a*?

We studied two circadian oscillators and three signaling networks: the 10-node *Drosophila* circadian clock [32], the 8-node *Arabidopsis* circadian clock [33], the 15-node core Notch pathway [34], the 27-node Wnt/*β*-catenin signaling pathway [35]; and the 22-node Nerve Growth Factor (NGF) pathway [36] (see Methods for details). We expect circadian clock networks to oscillate, and the signaling networks to reach steady state. Consistent with this idea, the circadian networks have an average *a* of 0.74, while the signaling networks have a higher average *a* of 0.83.

We first studied the known topology of each network. On average, the circadian clock network dynamics oscillate in > 95% of trials, while the signaling network dynamics reach steady state in 80% of trials. This result suggests that our simple model captures features of the biological network dynamics. Because numerical values for the connection strengths are not known, we averaged over randomly chosen interaction strengths, drawn from a Gaussian distribution with unit mean and variance. Therefore, the biological networks we studied show high *mutational* and *environmental robustness* because their dynamics are not sensitive to the exact values of the interaction strengths or initial conditions. Two of the signaling networks' dynamics (Wnt and Notch) always reach steady state; the networks have no feedback loops, which are required for oscillations.

We then varied *a* from its natural value by changing the sign of some interactions in the network, without changing which nodes are connected. For the circadian networks, the fraction of runs that reach steady state, while always low, increases dramatically for unnatural values of *a*. Increasing *a* to 1 by making all connections positive increases the chance of reaching steady state from ~1% to 18% for the *Drosophila* circadian network. For the *Arabidopsis* circadian network, this increase in *a* increases the chance of reaching steady state from 0 to 8% (fig. 4). The oscillation period for the circadian networks is longest near the natural *a*. Both the topology and the pattern of activating and inhibiting connections is important for circadian network dynamics.

The NGF network dynamics reach steady state much more often than do the circadian-network dynamics. For the natural *a*, ~50% of runs reach steady state. A decrease to *a* = 0.5 decreased the chance of reaching steady state to 20% (figure 4). The NGF network dynamics showed a non-monotonic dependence of the steady state fraction on *a*, for reasons not understood. The other two signaling networks have dynamics without feedback that reach steady state for all values of *a*.

We note that the results presented here normalize the baseline activation state of genes in a given organism to zero, and thus a change of the default state of genes from on to off would change the overall network properties, and hence the balance of activators to repressors. For example, in eukaryotes, most genes are turned off by default, and the ratio of positive to negative regulators in *S. cerevisiae* is about 3:1 in favor of positive regulations [37], whereas in bacteria, most genes are turned on by default, and the Regulon database [38] accordingly suggests that *E. coli* has a lower ratio, 3:2 in favor of positive regulators. Determining whether networks with different basal levels of expression but the same ratio of activators to repressors relative to baseline would be a fascinating topic for empirical studies, for example by changing the properties of the nodes of the repressilator.

## Conclusion

Our results suggest that the fraction of activating connections *a* is an important determinant of network dynamics. In the biochemical networks we studied, the specific network topology is clearly important. However, changing *a* also significantly altered the network dynamics. The circadian networks at their natural activating fraction almost always oscillate; increasing the activating fraction decreases the average oscillation period. The NGF signaling network dynamics become half as likely to reach steady state as *a* is changed from the natural value.

Much current research on biological networks focuses on finding network features that determine the behavior of the network, including network dynamics and the response of the network to stimuli or perturbations [39, 40]. Our result that *a* strongly affects network dynamics has implications for the evolution of biological clocks and switching networks. An intermediate value of *a* = 0.5 increases both the chance of long-period oscillations (in random Erdös-Renyi networks) and rate of convergence to a network with long-period oscillation (in optimized networks). This suggests that a network with equal numbers of activiating and inhibiting connections may be at an advantage in evolving a circadian clock. When *a* is high, random network dynamics are likely to reach steady state and optimized networks that reach steady state are most rapidly selected. Biochemical networks which reach steady state are more likely to be robust when the fraction of activating connections is high.

## Methods

### Model

We adapted the model of Wagner [15], as used by Siegal and Bergman [18, 19]. The network consists of nodes (representing genes, mRNA transcripts, or proteins) and edges (interactions between pairs of nodes). Each node potentially has an effect on each other node. This effect is represented by a number, *W*_{
ij
}, which describes the strength of the effect of node *j* on node *i*. An interaction strength can be positive (activating connection), negative (inhibiting connection), or zero (no connection). The interactions among the *N* nodes are collected in an *N* × *N* matrix *W*. Note that nodes can regulate themselves; these terms are the *W*_{
ii
}entries of the matrix.

Each network node has an activity level, which represents how activated or repressed the node is. The activity level could represent the expression (if the node is a gene or transcript) or activity level (if the node is a protein). The relative activity of node *i* is represented by *s*_{
i
}, which lies between -1 and 1. An activity level of 0 represents the basal activity of the node, while an activity level of 1 or -1 describes a node which is maximally activated or maximally inhibited, respectively. The activity of each node changes in time due to the effects of other nodes. The activity levels of all *N* nodes are collected in an *N*-element vector **s**. The activity vector changes in time; the dynamical rule which describes this change is the most important feature of the model. Given an activity vector **s**(*t*) at a specific time *t*, the activity vector at the next time, **s**(*t* + 1) must be determined. For node *i*, we sum over *j* the product of the activity level of node *j* times the interaction strength of node *j* on node *i*. This models the idea that node *j* has an effect on node *i* that depends both on the interaction strength and the activity of node *j*. Multiplying the interaction strength by the activity of node *j* produces the actual effect of *j* on *i* under current conditions. These effects are summed over all nodes that regulate *i*. We then transform the result by applying a nonlinear function *f*

The quantity ∑*W*_{
ij
}*s*_{
j
}may lie outside the range (-1, 1), and so therefore may not be a valid activity level for node *i*. The nonlinear function enforces the saturation of interactions among nodes: this models the idea that no matter how strong the activating (repressing) influence on a node, there is a maximum (minimum) activity level that can be attained. We use the nonlinear function *f*(*x*) = tanh(*rx*) with *r* = 100.

The dynamical rule describing the time evolution of the interacting genes is given in eq. 1.

### Network dynamics

Starting from a random initial condition, we iterate the dynamics (eq. 1) to determine the activity vector as a function of time. We then compute the difference between a given **s**(*t*) and the expression vectors at the previous time steps, **s**(*t* - 1), **s**(*t* - 2), and so on. To determine whether a steady state has been reached, we examine whether the Hamming-like distance

between the vectors at time points *t* and *t* - 1 is below some threshold *δ*. Then the vectors are considered equal, and the dynamics have an oscillation with period 1, a steady state. (Typically we use the value *δ* = 10^{-4}, although we verified that varying *δ* between 10^{-2} and 10^{-10}did not change the results.) To determine the period of oscillation, we examine whether the last *n* values of **s** are equal (within error) to the next-to-last *n* values for any value of *n*. Note that a steady state is equivalent to an oscillation with period *n* = 1, because at steady state we have **s**(*t*) = **s**(*t* - 1).

When determining the oscillation period, we iterate the dynamics a fixed number of times, typically 100. Therefore we cannot distinguish between chaotic dynamics and oscillations with period longer than 100.

### Network sampling

To develop a sample of networks with given properties, we specify the network size *N*, the probability of (nonzero) connection between nodes *c*, the fraction of activating connections *a*, and the distribution of connection strengths. Each characteristic is separately controlled. In this work we used a classical Erdös-Renyi random network model, where a connection between any two nodes is made with probability *c*. Therefore the average number of nodes that affect any given node in the network is *cN*. The activating fraction 0 ≤ *a* ≤ 1 determines the probability that two connected nodes have a positive (activating) interaction strength. The remaining nonzero connections have a negative interaction strength, which corresponds to an inhibiting interaction. We varied these parameters to study networks with *N* between 10 and 100, connection probability between 0.01 and 1 and activating fraction between 0 and 1.

The absolute value of the connection strength is randomly chosen according to a specified distribution. We used a wrapped Gaussian distribution, where the values are chosen from a Gaussian distribution but any negative values are made positive. The mean *μ* and standard deviation *σ* of the Gaussian distribution were varied in our simulations. We used distributions with mean between 0 and 1 and standard deviation between 0.03 and 1. Altering parameters which control the magnitude of connection strengths has little effect on the dynamics, on average: neither the fraction of runs which reach steady state nor the average period is sensitive to alteration in the mean or standard deviation of the connection strength distribution. The connection strength parameters have little effect because the nonlinear function used in our dynamics is nearly always saturated.

We implemented the graphs using arrays in the Python Numeric [41] package. The graph topology and the edge weights were generated separately, allowing us to vary the topology and the connection strengths and signs independently.

### Robustness measures

We consider three different types of robustness: *topological robustness*, which means network dynamics are robust to alterations in the network wiring; *mutational robustness*, which means the network dynamics are robust to alterations in connection strength; and *environmental robustness*, which means that the network dynamics are robust to changes in the initial condition.

In our model, each type of robustness requires a different perturbation to the network and measure of the change in behavior. We implement the perturbations corresponding the three types of robustness as follows: (*i*) *Topological perturbation*: we alter the network by randomly deleting one connection and then randomly adding a connection between two previously unconnected nodes, with random strength. The value of the strength of the connection is drawn from the same distribution used to originally make the network. This model preserves the total number of connections in the network. (*ii*) *Mutational perturbation*: we randomly select one (nonzero) connection and alter its strength. The strength of the connections is drawn from the same distribution used to originally make the network. (*iii*) *Environmental perturbation*: we alter one node of the vector of initial conditions.

To quantify robustness, we use different measures, depending on whether the network dynamics reach steady state or oscillate with period > 1. The robustness measure for dynamics which reach steady state is the mean phenotypic distance [18]. This measure assumes that both the original and perturbed networks reach a steady-state activity vector. We determine the steady-state activity vector **s**_{
r
}of a reference matrix and initial condition **s**_{0}. Then we alter the matrix and/or initial condition *M* times, and determine a set of *M* steady-state vectors **s**_{
p
}(*j*) that result from the perturbation (here *j* = 1, 2, ⋯ *M*). The mean phenotypic distance {\overline{d}}_{p} is the average Hamming-like distance (eq. (3)) between each perturbed activity vector and the reference activity vector

If the perturbations to the network tend to lead to the same steady state activity levels, the mean phenotypic distance will be low. In this case we say that the network is robust to the alteration. Typically we use *M* = 100.

For networks that oscillate, we measure robustness either using the mean absolute change in period or the mean change in period distribution. To determine the mean absolute change in period, we first determine the period of oscillation *p*_{
r
}of a reference matrix and initial condition. We then perturb the matrix and/or initial condition *M* times, and determine a set of *M* oscillation periods *p*_{
p
}(*j*) for each perturbed network (here *j* = 1, 2, ⋯ *M*). Then the mean absolute change in period is

For a more detailed measure of robustness of oscillatory networks, we study the mean change in period distribution. This measure allows us to study the effects of both varying initial conditions and alterations in the interaction network, and assess both steady-state and oscillating network dynamics. We consider a reference network and a set of *M* perturbed networks. For each network, we sample *k* different initial conditions. Each initial condition leads to a period of oscillation (where period 1 represents steady state). For a given network, we determine the period distribution: the number of times that the different initial conditions lead to oscillations of period 1, period 2, and so on. The vectors of counts are **v**_{
r
}for the reference matrix and **v**_{
p
}(*j*) for the perturbed networks (here *j* = 1, 2, ⋯ *M*). The mean change in period distribution is

If the alterations in the network or initial condition tend to lead to the same period of oscillation, the mean change in period distribution will be low. In this case we say that the network was robust to the alteration.

### Network optimization

We used the great deluge algorithm (GDA) [28] to test whether graphs with specific properties are more abundant in some regions of parameter space than others. At each step in the optimization, one edge of the graph was randomly changed to a new value, drawn from the same distribution used to create the graph (which was a Gaussian with unit variance "wrapped" by taking the absolute value of all values, so all draws have the same sign). The optimization was repeated for starting populations of graphs with different values of *a* and *c*, as shown in table 1.

Regions of parameter space in which desirable networks are dense converge in few iterations, whereas regions where they are sparse require more iterations for convergence. Therefore, the convergence rate of the algorithm discriminates between parameters where a desired dynamical behavior is common or uncommon. The GDA is a nonlinear optimization algorithm that performs well on a range of highly nonlinear problems and converges more rapidly than simulated annealing or genetic algorithms [28]. When used for maximization (seeking the largest value of a fitness function), the fitness surface is imagined as a landscape where higher values (peaks) are better. The GDA uses the concept of a "water level", below which changes are unacceptable. A random walk is performed on the landscape of possibilities, and the fitness is evaluated at each step. If the fitness at the new location is higher, the new location is accepted and the water level rises by a constant value based on the initial cost. (This value was selected after trial runs; we attempted to balance the convergence rate with the quality of the solution found.) If the fitness at the new location is lower but above the water level, the change is accepted. If the fitness at the new location is below the water level, the change is rejected.

The great deluge algorithm is less sensitive to local optima than is hill-climbing (always accepting the solution if it is better) or gradient methods, and is typically faster and less parameter-dependent than algorithms such as simulated annealing (for which temperature parameters and gradients must be selected) or genetic algorithms (for which a large population of solutions and a number of parameters are required). In our simulations, a random step in the fitness landscape was performed by a single perturbation to the network. The fitness function we used is the sum of the phenotypic distance and the mean change in period distribution, so F=\overline{{d}_{p}}+\overline{\left|\Delta v\right|}. An increase in fitness as a result of a perturbation means that the new graph is more robust (to both changes in the network wiring and changes in the initial condition) than the previously proposed graph. We measured the number of iterations required for the GDA to converge, and we discarded runs in which no convergence was obtained after 10,000 steps.

To test whether the connection probability and the activating fraction had changed, we compared the known means for the starting population to the sample means for the selected population using one-tailed, one-sample *t* tests (*n* was approximately 50 in each sample). Because each *t* test is an independent test of the hypothesis that the sample means had changed in the direction predicted, we used Fisher's method of combining independent tests of a hypothesis [42] to test whether the overall level of change was significant. We compared mean convergence times between each pair of parameter settings using two-sample *t* tests assuming unequal variances.

The results of the network selection are shown in detail in table 1.

### Biological network topologies

We studied five network topologies: the 10-node *Drosophila* circadian network sketched in reference [32], figures 4/6; the 8-node *Arabidopsis* circadian network sketched in reference [33], figure 4; the 15-node core Notch pathway sketched in reference [34]; the 27-node WNT/*β*-catenin signaling pathway sketched in reference [35]; and the 22-node Nerve Growth Factor (NGF) signaling pathway sketched in reference [36], figure 3. Details of the size and number of activating connections for these networks are given in table 2. The representation of network topologies in our model is given in tables 3, 4, 5, 6, 7.

In our simulations, we generated networks with topology corresponding to the biological network. Keeping this topology constant, we varied *A*, the number of activating connections, from 0 (producing a network that contains no activating connections, i.e., all connections are inhibiting connections) to *A*_{
m
}, the number of edges in that network (all connections are activating connections). Since the network has *A*_{
m
}edges, the activating fraction is *a* = *A*/*A*_{
m
}. Note that we did not change which network nodes were connected, only the sign of the interaction (activating or inhibiting). We chose to pick the inhibiting connections to match the network to the biological network as much as possible. Therefore, as the number of inhibiting connections was increased, we first randomly selected from the inhibiting connections that occur in the real biological network, then (once all of those inhibiting connections were made) we randomly selected from the remaining connections in the network. For each choice of the number of inhibiting connections, we generated *k* = 10^{3} networks with randomly chosen interaction strengths and random initial conditions. The signs of the connections of each of these networks were chosen independently. For example, if there were originally 5 inhibiting connections, each of the *k* = 10^{3} networks generated for 3 inhibiting connections was independently assigned one of the \left(\begin{array}{c}5\\ 3\end{array}\right) = 10 ways of choosing which edges remained inhibiting. For each of the *k* graphs and initial conditions we determined the period of oscillation. Summing over all *k* networks and initial conditions gave us a period distribution for the graph with *A* activating connections. The mean of this distribution is the average period for the network with *A* activating connections.

## Reviewers' comments

### Reviewer's report 1

*Reviewer 1: Sergei Maslov, Department of Condensed Matter Physics and Materials Science, Brookhaven National Laboratory*

**Reviewer's comment:** The manuscript reports an interesting study of how the balance between positive and negative regulatory interactions affects the dynamics of regulatory networks. This subject definitely deserves a careful analysis and (to the best of my knowledge) was not explored before. Authors use a simple dynamical model introduced by Andreas Wagner in 1996 and later used in a number of publications. The most appealing property of this model is its simplicity: the genes (or their protein products) exist in either "active", "on" state (+1) or "inactive", "off" state (-1). While this boolean property is somewhat relaxed in this manuscript, it still (approximately) applies to the majority of nodes. The dynamical rules of the model remind the threshold activation rules in model neural networks, where contributions of multiple inputs weighted by their connection strength *W*_{
ij
}are simply added up and compared with the activation threshold of a node. As such they cannot model an arbitrary logical function of input variables (one of the many simplifications of the model).

My main concern about this model is that it represents inactive genes/proteins by -1 (instead of simply 0). As a consequence even inactive genes keep sending negative/positive signals to their targets which they are positively/negatively regulating. This property of the model is rather artificial since in most cases genes/proteins that are not active *send no signal* instead of sending a *negative* signal. This can be easily rectified by the simultaneous change of *s*_{
i
}variables (*s*_{
i
}→ (*s*_{
i
}+ 1)/2) and activation thresholds (equal to 0 in the present model). However, as a result of this transformation the activation thresholds become non-trivially (and unrealistically) coupled to magnitudes and *signs* of regulatory interactions *W*_{
ij
}. As a consequence, I am not sure which of this study's conclusions about positive-to-negative ratio would survive in the more realistic scenario of arbitrary activation thresholds.

**Authors' response:** This simplification is noted by Wagner [15]: "For reasons of computational simplicity... ", and, as the reviewer notes, has been widely used in subsequent work (e.g. Siegal and Bergman 2002 [18]). It is precisely for the reason that the reviewer notes (that a large number of arbitrary activation thresholds must be introduced) that the variable transformation to a scale of 0 – 1 is not typically used. We would argue that our results show that in a widely used model of gene expression, we are able to identify key parameters related to network dynamics and to relate them to biological networks: there are many other restrictions of this admittedly simplified model that could be relaxed, and exploring which of these features is most influential for the results would be a fascinating topic for future work.

**Reviewer's comment:** Another simplification of the model is the synchronous update rule in which the states of all nodes are modified simultaneously with each other. In reality regulatory networks are characterized by a broad distribution of timescales of gene activation/deactivation corresponding to a more complicated (and generally asynchronous) update rule. Can authors comment on how sensitive are their results with respect to update rules?

**Authors' response:** Wagner (1996) also noted this limitation and cited the need for computational simplicity again. We did not model different update rules in this study, so we are unable to comment on the sensitivity. The simplicity of this model is sufficient to capture basic dynamics of a network and the model is further supported by its ability to represent "biological networks" as well as the repressilator. The total runtime was over 100,000 CPU hours, and the work itself is algorithmically complex. The bottleneck is thus not the hardware, and the simplifications required in 1996 are still needed today despite Moore's Law. However, again we agree that testing which model assumptions affect the result will ultimately be important for establishing the generality of the work.

**Reviewer's comment:** My final remark on this study is that in its present form it does not reveal which network property is being optimized by the empirically observed ratio of positive and negative regulations. Perhaps, authors could use the following observation: the ratio between positive and negative regulations in the genome-wide regulatory network of baker's yeast appears to be around 3:1 in favor of positive regulations (S. Maslov, K. Sneppen, Physical Biology, 2005). This is consistent with the ratio authors observed in their examples of real-life small subnetworks/pathways (except for the NGF signaling network). On the other hand, according to the data from the Regulon database, the regulatory network in *E. coli* has fewer positive regulations (the empirical ratio is only 3:2 in favor of positive regulations). This could be due to the fact that the default state of most bacterial genes is "on". That is to say, they are being expressed at a considerable rate even in the absence of any positive regulatory signals. Hence, they require fewer positive and more negative inputs than comparable networks in eukaryotes such as S. cerevisiae whose default state is typically off. Could authors comment on what are their assumptions about the default state of genes and how the change of the default state would affect the results?

**Authors' response:** This is a very interesting point, and we have incorporated this suggestion into the discussion as shown below. We thank the reviewer for the suggested references.

"We note that the results presented here normalize the baseline activation state of genes in a given organism to zero, and thus a change of the default state of genes from on to off would change the overall network properties, and hence the balance of activators to repressors. For example, in eukaryotes, most genes are turned off by default, and the ratio of positive to negative regulators in *S. cerevisiae* is about 3:1 in favor of positive regulations [37], whereas in bacteria, most genes are turned on by default, and the Regulon database [38] accordingly suggests that *E. coli* has a lower ratio, 3:2 in favor of positive regulators. Determining whether networks with different basal levels of expression but the same ratio of activators to repressors relative to baseline would be a fascinating topic for empirical studies, for example by changing the properties of the nodes of the repressilator."

### Reviewer's report 2

*Reviewer 2: Eugene V. Koonin, National Center for Biotechnology Information*

**Reviewer's comment:** The paper uses simulations to explore network dynamics and reaches a clear and interesting conclusion: that the fraction of activating (as opposed to inhibitory) connections determines the dynamical regime, that is whether the network arrives at a steady state of goes into an oscillating regime. As interesting as these findings are, I am somewhat uncomfortable with the conclusion: "The activating fraction may be a control parameter that cells use to predispose a network to oscillate or reach steady state." I am not sure that this (seemingly) adaptationist interpretation is justified or that the uncovered principle is as general as implied here. In particular, the oscillatory regime certainly is natural in the case of the circadian clock but its wider relevance remains to be proven.

**Authors' response:** We agree with the reviewer that the adaptationist language may be too strong here, and have changed the sentence to read "The activating fraction may predispose a network to oscillate or reach steady state, and neutral evolution or selection of this parameter may affect the behavior of biological networks."

### Reviewer's report 3

*Reviewer 3: Yu (Brandon) Xia, Bioinformatics Program and Department of Chemistry, Boston University (nominated by Mark Gerstein, Department of Molecular Biology and Biophysics, Yale University)*

**Reviewer's comment:**1. You correlated the simulated dynamics with biological function for five biological networks. I am wondering how much of the correlation can be teased out by simple static topological measures, such as number and percentage of feedback loops, and fraction of activation links?

**Authors' response:** We believe that the value of the present work is that we are able to show a link between even these static topological properties and the dynamics. Explaining the fraction of the variance in the behavior of real biological networks that is explained by these measures, as opposed to other properties, awaits both the development of more realistic models and the measurement of network dynamics in far more detail. However, it will be fascinating to explore these relationships in future.

**Reviewer's comment:**2. It will useful to illustrate some of these networks. For example, it will be useful to have one figure for a random network, one figure for an optimized network, and one figure for an actual biological network. It will be easier for the readers to see the similarities and differences between these networks.

**Authors' response:** We considered adding a few example networks, but in the end decided that it would be better to provide a recipe that allows interested readers to generate these networks with any arbitrary parameter settings, including those used throughout the paper. Accordingly, the following code snippet can be used to generate and display such networks:

>>>from networkx import XDiGraph, draw

>>>from numpy.random import normal

>>>from random import random

>>>from pylab import show

>>>def f(n, p):

... g = XDiGraph( )

... g.add_nodes_from(zip(range(n), [choice([-1,1]) for i in range(n)]))

... for n1 in g.nodes( ):

... for n2 in g.nodes( ):

... if random( ) < p:

... g.add_edge(n1, n2, normal(0,1))

... return g

...

>>>g = f(20,0.1) #this is just an example: any value of n and p can be used

>>>draw(g)

>>>show( )

This can be used to generate random graphs equivalent to those used in the study. The dependencies are the NetworkX, numpy, and matplotlib packages.

## References

Tanaka R, Yi TM, Doyle J: Some protein interaction data do not exhibit power law statistics. FEBS Letters. 2005, 579 (23): 5140-5144. 10.1016/j.febslet.2005.08.024.

Xia Y, Yu HY, Jansen R, Seringhaus M, Baxter S, Greenbaum D, Zhao HY, Gerstein M: Analyzing cellular biochemistry in terms of molecular networks. Ann Rev Biochem. 2004, 73: 1051-1087. 10.1146/annurev.biochem.73.011303.073950.

Barabasi AL, Oltvai ZN: Network biology: Understanding the cell's functional organization. Nature Reviews Genetics. 2004, 5 (2): 101-U15. 10.1038/nrg1272.

Batada NN, Reguly T, Breitkreutz A, Boucher L, Breitkreutz BJ, Hurst LD, Tyers M: Stratus not altocumulus: A new view of the yeast protein interaction network. PLoS Biology. 2006, 4 (10): 1720-1731. 10.1371/journal.pbio.0040317.

Albert R: Scale-free networks in cell biology. J Cell Sci. 2005, 118 (21): 4947-4957. 10.1242/jcs.02714.

Milo R, Itzkovitz S, Kashtan N, Levitt R, Shen-Orr S, Ayzenshtat I, Sheffer M, Alon U: Superfamilies of evolved and designed networks. Science. 2004, 303 (5663): 1538-1542. 10.1126/science.1089167.

Mangan S, Alon U: Structure and function of the feed-forward loop network motif. PNAS. 2003, 100 (21): 11980-11985. 10.1073/pnas.2133841100.

Li CM, Klevecz RR: A rapid genome-scale response of the transcriptional oscillator to perturbation reveals a period-doubling path to phenotypic change. PNAS. 2006, 103 (44): 16254-16259. 10.1073/pnas.0604860103.

Siegal ML, Promislow DEL, Bergman A: Functional and evolutionary inference in gene networks: does topology matter?. Genetica. 2007, 129: 83-103. 10.1007/s10709-006-0035-0.

Prill RJ, Iglesias PA, Levchenko A: Dynamic properties of network motifs contribute to biological network organization. PLoS Biol. 2005, 3 (11): e343-10.1371/journal.pbio.0030343.

Kollmann M, Lovdok L, Bartholome K, Timmer J, Sourjik V: Design principles of a bacterial signalling network. Nature. 2005, 438 (7067): 504-507. 10.1038/nature04228.

Chen LN, Wang RQ: Designing gene regulatory networks with specified functions. IEEE Trans Circ Sys I. 2006, 53 (11): 2444-2450. 10.1109/TCSI.2006.883880.

Huang S, Eichler G, Bar-Yam Y, Ingber DE: Cell fates as high-dimensional attractor states of a complex gene regulatory network. Phys Rev Lett. 2005, 94 (12): 128701-10.1103/PhysRevLett.94.128701.

Tyson JJ, Chen KC, Novak B: Sniffers, buzzers, toggles and blinkers: dynamics of regulatory and signaling pathways in the cell. Curr Opin Cell Biol. 2003, 15 (2): 221-231. 10.1016/S0955-0674(03)00017-6.

Wagner A: Does evolutionary plasticity evolve?. Evolution. 1996, 50 (3): 1008-1023. 10.2307/2410642.

Shen-Orr SS, Milo R, Mangan S, Alon U: Network motifs in the transcriptional regulation network of Escherichia coli. Nature Genetics. 2002, 31: 64-68. 10.1038/ng881.

DasGupta B, Enciso GA, Sontag E, Zhang Y: Algorithmic and complexity results for decompositions of biological networks into monotone subsystems. Biosystems. 2006, 90 (1): 161-178. 10.1016/j.biosystems.2006.08.001.

Siegal ML, Bergman A: Waddington's canalization revisited: Developmental stability and evolution. PNAS. 2002, 99 (16): 10528-10532. 10.1073/pnas.102303999.

Bergman A, Siegal ML: Evolutionary capacitance as a general feature of complex gene networks. Nature. 2003, 424 (6948): 549-552. 10.1038/nature01765.

Masel J: Genetic assimilation can occur in the absence of selection for the assimilating phenotype, suggesting a role for the canalization heuristic. J Evol Biol. 2004, 17 (5): 1106-1110. 10.1111/j.1420-9101.2004.00739.x.

Azevedo RBR, Lohaus R, Srinivasan S, Dang KK, Burch CL: Sexual reproduction selects for robustness and negative epistasis in artificial gene networks. Nature. 2006, 440 (7080): 87-90. 10.1038/nature04488.

Camas FM, Blazquez J, Poyatos JF: Autogenous and nonautogenous control of response in a genetic network. PNAS. 2006, 103 (34): 12718-12723. 10.1073/pnas.0602119103.

Verma M, Rawool S, Bhat PJ, Venkatesh KV: Biological significance of autoregulation through steady state analysis of genetic networks. Biosystems. 2006, 84: 39-48. 10.1016/j.biosystems.2005.10.001.

Hopfield JJ: Neural Networks and Physical Systems with Emergent Collective Computational Abilities. Proc Natl Acad Sci U S A. 1982, 79 (8): 2554-2558. 10.1073/pnas.79.8.2554.

Shadlen MN, Newsome WT: Noise, neural codes and cortical organization. Curr Opin Neurobiol. 1994, 4 (4): 569-79. 10.1016/0959-4388(94)90059-0.

Troyer TW, Miller KD: Physiological gain leads to high ISI variability in a simple model of a cortical regular spiking cell. Neural Computation. 1997, 9 (5): 971-983. 10.1162/neco.1997.9.5.971.

Rajan K, Abbott LF: Eigenvalue spectra of random matrices for neural networks. Physical Review Letters. 2006, 97 (18):

Dueck G: New optimization heuristics: the great deluge algorithm and the record-to-record travel. J Comp Phys. 1993, 104: 86-10.1006/jcph.1993.1010.

Kitano H: Biological robustness. Nat Rev Genet. 2004, 5 (11): 826-837. 10.1038/nrg1471.

de Visser J, Hermisson J, Wagner GP, Meyers LA, Bagheri-Chaichian H, Blanchard JL, Chao L, Cheverud JM, Elena SF, Fontana W, Gibson G, Hansen TF, Krakauer D, Lewontin RC, Ofria C, Rice SH, von Dassow G, Wagner A, Whitlock MC: Perspective: Evolution and detection of genetic robustness. Evolution. 2003, 57 (9): 1959-1972. 10.1554/02-750R.

Cooper TF, Morby AP, Gunn A, Schneider D: Effect of random and hub gene disruptions on environmental and mutational robustness in Escherichia coli. BMC Genomics. 2006, 7: 237-10.1186/1471-2164-7-237.

Bell-Pedersen D, Cassone VM, Earnest DJ, Golden SS, Hardin PE, Thomas TL, Zoran MJ: Circadian rhythms from multiple oscillators: Lessons from diverse organisms. Nature Rev Genet. 2005, 6 (7): 544-556. 10.1038/nrg1633.

Locke JCW, Southern MM, Kozma-Bognar L, Hibberd V, Brown PE, Turner MS, Millar AJ: Extension of a genetic network model by iterative experimentation and mathematical analysis. Mol Syst Biol. 2005, 1: 2005.0013-10.1038/msb4100018.

KEGG PATHWAY. 2005, http://www.genome.jp/kegg/pathway.html, [http://www.genome.jp/kegg/pathway/hsa/hsa04330.html]

BioCarta. 2000, [http://www.biocarta.com/pathfiles/h_wntPathway.asp]

Huang EJ, Reichardt LF: Trk receptors: Roles in neuronal signal transduction. Ann Rev Biochem. 2003, 72: 609-642. 10.1146/annurev.biochem.72.121801.161629.

Maslov S, Sneppen K: Computational architecture of the yeast regulatory network. Phys Biol. 2005, 2: S94-100. 10.1088/1478-3975/2/4/S03.

Gama-Castro S, Jimenez-Jacinto V, Peralta-Gil M, Santos-Zavaleta A, Penaloza-Spinola M, Contreras-Moreira B, Segura-Salazar J, Muniz-Rascado L, Martinez-Flores I, Salgado H, Bonavides-Martinez C, Abreu-Goodger C, Rodriguez-Penagos C, Miranda-Rios J, Morett E, Merino E, Huerta A, Trevino-Quintanilla L, Collado-Vides J: RegulonDB (version 6.0): gene regulation model of Escherichia coli K-12 beyond transcription, active (experimental) annotated promoters and Textpresso navigation. Nucleic Acids Res. 2008, 36: D120-124. 10.1093/nar/gkm994.

Brandman O, James E, Ferrell J, Li R, Meyer T: Interlinked Fast and Slow Positive Feedback Loops Drive Reliable Cell Decisions. Science. 2005, 310 (5747): 496-498. 10.1126/science.1113834.

Variano EA, McCoy JH, Lipson H: Networks, dynamics, and modularity. Phys Rev Lett. 2004, 92 (18):

NumPy. 2005, [http://numpy.scipy.org/]

Fisher RA: Statistical methods for research workers. 1925, London: Oliver & Loyd, 13

## Acknowledgements

We thank Natalie Ahn, Corrie Detweiler, Todd Gibson, Debra Goldberg, Micah Hamady, Xuedong Liu, and Brad Olwin for helpful discussions. We thank Matthew Woitaszek and Henry Tufo for assistance with the parallel computations and for use of the Hemisphere cluster. This work was supported in part by the Butcher Foundation, the Alfred P. Sloan Foundation and the University of Colorado UROP program.

## Author information

### Authors and Affiliations

### Corresponding author

## Additional information

### Competing interests

The authors declare that they have no competing interests.

### Authors' contributions

DM wrote the code, ran simulations and interpreted results. LW ran preliminary simulations, participated in programming and interpreted results. RK directed the coding, participated in programming, and interpreted results. MB directed the project, analyzed results, and wrote the manuscript. All authors read and approved the final manuscript.

## Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

## Rights and permissions

**Open Access**
This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License (
https://creativecommons.org/licenses/by/2.0
), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## About this article

### Cite this article

McDonald, D., Waterbury, L., Knight, R. *et al.* Activating and inhibiting connections in biological network dynamics.
*Biol Direct* **3, **49 (2008). https://doi.org/10.1186/1745-6150-3-49

Received:

Accepted:

Published:

DOI: https://doi.org/10.1186/1745-6150-3-49

### Keywords

- Nerve Growth Factor
- Network Dynamic
- Circadian Clock
- Biological Network
- Connection Strength