On the Performance of Particle Swarm Optimization for

advertisement
184
On the Performance of Particle Swarm
Optimization for Parameterizing Kinetic Models
of Cellular Networks
Natalie Berestovsky
Riya Fukui
Luay Nakhleh
Department of Computer Science
Rice University
6100 Main Street, Houston
TX 77005, USA.
Email: ny1@rice.edu
Department of Computer Science
Rice University
6100 Main Street, Houston
TX 77005, USA.
Email: rf7@rice.edu
Department of Computer Science
Rice University
6100 Main Street, Houston
TX 77005, USA.
Email: nakhleh@rice.edu
Abstract—Recent advances in high-throughput technologies
and an increased knowledge of biochemical systems have
enabled the reconstruction of cellular regulatory networks.
While these reconstructions often consist of the connectivity
patterns among the network’s components, an important task
in this area is to derive values for the kinetic parameters of
the system so that its dynamics can be elucidated and testable
hypotheses can be generated. The problem here is to reverse
engineer the model (in our case, values of kinetic parameters)
given the data that consists of a network connectivity and
time-series data. Particle swarm optimization (PSO) has been
recently used for this reverse engineering task. In this paper,
we introduce three scoring metrics for assessing the optimality
of the solution found by PSO. Using biological data sets, we
study the performance of the PSO framework under the three
scoring metrics. Our results show that PSO achieves very good
parameterization of the cellular networks we consider, regardless of the complexity of the dynamics patterns generated by the
underlying kinetics. Given the flexibility of the PSO framework,
as well as its natural amenability to parallelization and general
high-performance implementations and execution, our results
indicate PSO may be a good candidate for parameterizing
cellular networks.
Index Terms—Particle swarm optimization, cellular networks, kinetic models, time series data.
I. I NTRODUCTION
Understanding biology at the system level requires examining not only the properties of individual molecular species,
but also their organization into a tightly-connected and finetuned network of structural and dynamic interactions [1].
Advanced biotechnologies are amassing data on molecular
interactions from a variety of organisms at an exponential
rate [2]. These data typically take the form of graphs, where
nodes represent the molecules, such as genes or proteins,
and the edges represent interactions. However, the dynamic
behavior may not depend only on the wiring, but also
on the values of the kinetic parameters associated with
the strength of interactions (though views that the wiring
patterns can solely determine the dynamics exists; e.g., [3]).
Typically, inferring parameter values is done by obtaining the
978-1-4673-1191-5/12/$31.00 ©2012 IEEE
known values from the literature, and using experiments to
evalute the remaining ones [4] (using “educated guessing”
of parameter values is also common; the process is very
much ad hoc, indeed). However, the literature has sparse
information regarding the parameters and those available
depend highly on the the experimental conditions, making
them insufficient or inappropriate for analytical purposes in
general.
This triumvirate of issues—the systems biology view, the
availability of network connectivity, and the need to parameterize networks for dynamics—underlies the significance of
the problem of parameter estimation of nonlinear dynamic
biochemical pathways and motivates the emerging efforts on
solving it in a systematic way [5].
To address this problem, several global optimization techniques have been developed and applied to learning the
values of kinetic parameters in cellular networks. Moles et
al. provide a review and comparison of various global optimization techniques applied to biochemical systems [5]. The
authors showed that not all global optimization techniques
work well for parameter estimation of regulatory networks.
Nonetheless, they showed that the general evolutionarybased strategy was able to solve the problem successfully.
In this work, we examine the applicability and performance of Particle Swarm Optimization (PSO), a biologically
inspired algorithm, for parameter estimation of biochemical
pathways using time-series data. Combining aspects of genetic and evolutionary algorithms, PSO uses the concept of
fitness and utilizes random values to search the (solution)
space. Adjustment towards local and global optima by
the particles in the swarm is conceptually similar to the
crossover operation utilized by genetic algorithms [6]. We
employ PSO in conjunction with scoring metrics that we
devise for assessment of the optimality in each search iteration. We show that for some systems these metrics perform
equally well, but for others, some metrics outperform the
others. As such, we analyze the time-series data to elucidate
185
the patterns the govern and affect the performance of each
metric, so as the produce guidelines for which metric, or
metrics, to use given the observable trends in the data.
We demonstrate the validity of PSO as a method for
parameter estimation from time-series data. Further, its
flexibility (in terms of customizing and fine-tuning the
search parameter values), as well as its amenability to highperformance techniques, make it a good candidate for reverse
engineering biological systems from time-series data.
II. PARTICLE SWARM OPTIMIZATION (PSO) FOR
PARAMETER ESTIMATION
PSO is a probabilistic search algorithm that is based on a
social model of, e.g., a swarm of insects searching for a food
source. It was first introduced by Kennedy and Eberhart in
1995 [6], and since then PSO has been studied by several
authors in terms of performance and applications [7], [8].
The basic swarm optimization algorithm works as follows:
k particles are uniformly positioned within given bounds and
initialized with a velocity. At each search step, each particle
determines its new position by, first, updating its velocity
vector using the memory of its own previous experience and
the experience of the best particle in the swarm, and, then,
calculating the new position vector. The particle’s update
steps are stochastic (not deterministic). A more detailed
description of the general PSO framework is shown in
Algorithm 1.
Algorithm 1 Particle Swarm Optimization (PSO)
1) Input: A swarm of k particles: {ρ1 , . . . , ρk }, each represented by an N dimensional initial position vector pρ and initial velocity vector vρ .
2) Repeat:
a) For each particle ρ, update the velocity vector vρ for time t + 1
according to
t+1
vρ
t
t
t
= ωvρ + c1 r1 (lρ − pρ ) + c2 r2 (g − pρ )
(1)
where for particle ρ
ω
is the inertia of the particle.
t
vρ
is the velocity at time t of this particle.
t
pρ
is the position at time t of this particle.
lρ
is the best solution found by this particle thus far (local
memory).
g
is the best solution found by the swarm thus far (global
memory).
c1 , c2
are parameters representing a particle’s “trust” into itself
and the swarm, respectively.
r1 , r2
are random numbers in the range (0, 1).
b) For each particle ρ, update its position at time t + 1 according to
t+1
pρ
t
t+1
= pρ + vρ
(2)
c) For each particle ρ, score the position vector pρ in the context of the
problem being addressed and update:
lρ
if the current position produces the best local score thus far.
g
if the current position of ρ produces the best global score thus
far.
3) Output: The global best position g.
Aside from the stochasticity contributing to the individual
behavior of the particles, there are two other aspects that
influence the outcomes of the optimization. First, there are
three parameters that could be fine-tuned specifically to the
problem of interest. The inertia parameter ω, controls the
exploration properties of the algorithm with higher values
favoring global behavior and lower values focusing on the
local search. The best results can be achieved by dynamically
changing ω: increasing it if no score progress has been made
for several iterations, invoking global search, or decreasing
it for better local refinement. Shi and Eberhart suggest using
0.8 < ω < 1.4 as an optimal inertia range [9]. Other two
PSO parameters are c1 and c2 , which indicate how much
“trust” the particle has in itself and in the swarm, respectively. These values weight the importance of “local best vs.
global best”. The parameter selection for PSO can highly
depend on the type of the problem being addressed, and
many modified approaches have been proposed to address
individual problems [10], [11], [9], [12]. In this work we
use the settings 0.1 < ω < 1.1 and c1 = c2 = 1.49. This
choice of parameters has shown to be appropriate for PSO
parameter estimation for signaling networks in recent work
by Iadevaia et al. [13].
The second factor that influences the success of the swarm
is its initialization. To achieve best results the particles in the
swarm should be uniformly distributed across the solution
space. It is a common practice to use random initialization,
even though it does not always guarantee perfect coverage.
To address this, Venter et al. carried out a series of numerical
experiments showing that space filling design approach is not
significantly superior to the random initialization [8]. This is
due to the swarm’s great dynamics during the execution. In
this work we use random initialization for particle swarms.
Next, we discuss the application of PSO to parameter
estimation of cellular networks from time-series data.
A. PSO for cellular networks
The general optimization problem, to which PSO can be
easily applied, consist of some objective functions, where
the independent variables comprise the problem dimension,
and the dependent variables serve as a scoring metric. The
convergence of the algorithm is assumed if for a specified
number of consecutive iterations, the change in the objective function score is less than some pre-defined expected
change.
In the context of cellular networks, each particle in the
PSO tries to find the parameter set that best fits the timeseries data. The particle’s position vector represents a single
parameter set. At each iteration, the particles update their
position vector according to Equation 2 in Algorithm 1,
trying to find the set that results in the fitness score higher
than the current global best. The best position vector is
determined by the highest scores from a metric associated
with the PSO execution. The scoring of a particular position
vector follows a three step procedure: set, simulate, and
compare. First, the values in the position vector are set as
the model parameters; second, the model is simulated to
produce the dynamic profiles associated with this parameter
186
set; finally, the simulated time-series data are compared
with experimental time-series data using one of the metrics
proposed below. The processes is repeated until convergence
is achieved. In the next section, we propose three distinct
scoring metrics for scoring the solutions found by the
particles in each iteration.
B. Comparison metrics
For the comparison metrics, we are give two sets of
time-series data, one produced by the PSO-parameterized
model, labeled sP SO below, and the other is the experimental
(“true”) data, labeled strue below. We assume each data
set consists of n molecules, the concentration of which is
measured at m different time points.
1) Euclidean distance: One scoring metric is the average
Euclidean distance between sP SO and strue . While this
metric has the highest lower bound, it does capture the
correlation between parameter inconstancy and difference in
the curves. This score is calculated according to
n
qP
(3)
score = P
n
m−1 P SO
true )2
(s
−
s
j
j
i=1
j=0
In order to focus on the comparison of the systems’ dynamics, both times-series data sets were trimmed to exclude
steady-state tails and then were adjusted to have the same
size by using the Piecewise Aggregate Approximation (PAA)
algorithm [14].
2) Tangent slope distance: Another metric measures the
distance between the linear features of the two time-series
data sets. Assuming the concentrations were measured at
time points t0 , t1 , . . . , tm−1 , we define the vector of slopes
for each of the time-series data sets, α, using the formula
sj+1 − sj
,
(4)
αj =
tj+1 − tj
where 0 ≤ j ≤ m − 2. Then, we define the score as
n
qP
score = P
n
m−2
P SO − αtrue )2 )
j
i=1
j=0 (αj
(5)
3) Feature extraction by Discrete Fourier Transform
(DFT): A third scoring metric is based on feature extraction.
The Discrete Fourier Transform is the projection of a signal
from the time domain into the frequency domain by
n
−2πif t
1 X
cf = √
f (t)e n
(6)
n t=0
√
where f = 1, . . . , m and i = −1. Each of these complex
numbers cf , represent the amplitude and the phase of a
sinusoidal function term in a decomposed time series-curve.
DFT assumes the curve to be periodic, and the complex
numbers describe the global frequencies of the curve. The
low frequencies describe the general feature or the trend
of the curve, whereas the high frequencies describe the
details of the curve, which are often obstructed by noise.
By using only the first k complex numbers for comparison,
the metric can compare the general characteristic of the two
curves, since the first complex numbers correspond to low
frequencies of the curve [15]. The score is then calculated
according to
n
qP
(7)
score = P
n
k
P SO k − kctrue k)2
(kc
j
j
i=1
j=0
where k = m
4 . DFT was calculated using a Fast Fourier
Transform (FFT) algorithm.
III. C ASE STUDIES
In this section we examine the performance of PSO
parameter estimation for signaling pathways from timeseries data. We consider three systems of different levels of
complexity: MAPK cascade, MAPK with negative feedback,
and Jak/Stat. The input for all test cases consist of a COPASI
model [16] and a text file containing the experimental time
series data. In the first two cases, the experimental data
is synthetic and generated from the corresponding COPASI
model given the parameter sets described in their sources.
The last system, on the other hand, contains real experimental data. For it, we devise a simple mass-action model
that incorporates all experimental species with known upstream/downstream interactions and attempt to parametrize
it using PSO method. Both MAPK systems have the initial
concentrations corresponding to the ones used in generation
of synthetic “true” data, Jak/Stat has initial concentrations
set to zero for all species except the source.
For all systems, the inertia parameter is in range 0.1 <
ω < 1.1 and changes dynamically by the factor of 2 according the the improvement counter δ. Each particle keeps
track of its δ by incrementing it with each improvement
and decrementing when no improvement happens. Inertia is
incremented when δ falls below 3 to favor global search
and decremented when δ is above 5 to favor local search.
The “trust” parameter c1 and c2 both are set to 1.49 [13].
All simulation use 200 particles, random initialization, and
executed until no improvement was observed for 2000 PSO
iterations.
We show that in some cases all comparison metrics
perform equivalently well, and in others, some metrics
are more appropriate than others. In addition, we confirm
that multiple runs of the PSO converges to a plethora of
parameter estimates of varying quality of fit with respect to
the profiles [17].
A. A MAPK pathway
The mitogen-activated protein kinase (MAPK) cascade
has been heavily studied in the past two decades. This
pathway is very significant and has been shown to participate
in regulating gene expression in a larger variety of biological
processes such as cell differentiation, cell survival/apoptosis,
adaptation to drugs, etc. Ultra-sensitivity of MAPK cascade
187
Fig. 1. Time-series data of the MAPK pathway. Left: synthetic time-series data generated by the curated COPASI model [18]. Right: replication of
MAPK sensitivity experiment demonstrated in [18] using the COAPSI model.
(a) Euclidean
(b) Tangent slope
(c) DFT
Fig. 2. Convergence of PSO on the MAPK pathway using the three scoring metrics. Each colored line corresponds to a score convergence for a
single run of PSO parameter estimation. Each reaches a different optimal solution, with some processes converging within few hundred iterations, while
others taking thousands of iterations.
was first studied in great detail both experimentally and
computationally by Huang and Ferrell in 1996 [18]. Since
then, over a dozen of MAPK models arose from Huang and
Ferrell work [19].
Huang and Farrell developed a system of ordinary differential equations (ODEs) to model the MAPK pathway
using experimental data to parametrize it. Using this model,
they demonstrated how the cascade organization of the
mitogen-activated protein kinase (MAPK) greatly enhances
the sensitivity of cellular targets to external stimuli. The
curated version of this model is available in the BioModels
database [20]. Figure 1 shows the raw time-series data
generated by the curated model (left) and the replication
of the MAPK cascade sensitivity validation demonstrated in
their work (right).
We utilize this computational model to assess the performance of PSO parameter estimation. Time-series data in
Figure 1(left) is used as the synthetic “true” data. The range
for possible parameter values is restricted to [0.001, 10000].
Each of the three comparison metrics is examined and the
experiments are repeated five times.
Figure 2 shows the convergence of five experiments for
each of the three metrics. It is clear that each PSO parameter
estimation run reaches different optimal solutions, with some
converging within a few hundred iterations and others taking
thousands of iterations. Next, we examine the quality of
those solutions by visually validating the highest and lowest
scoring solutions from each metric.
Figure 3 shows the the COPASI simulation using the
parameters corresponding to the highest and the lowest
score associated with each comparison metric. All metrics
consistently generate the parameters resulting in very good
fit to the original data set. It is worth noticing that all PSOgenerated parameter sets achieve the desired dynamics in a
very short amount of time. The reason for this could be that
the PSO was given the large parameter search range, and
considered values significantly different from those in the
curated model. We also repeated the experiments considering
the range closer to that of curated model and PSO was much
less successful in finding optimal parameter set.
Furthermore, we examined whether the PSO-determined
parameters preserve the ultra-sensitivity in the MAPK cascade that has been shown by Huang and Farrell [18].
Using the parameters produced by the highest-scoring PSO
execution, we conducted the sensitivity experiment. Figure 4 shows the replication of sensitivity experiment with
PSO generated parameters. All metrics preserved the ultrasensitivity property. However, the sensitivity of the cascade
is triggered by different concentrations of MAPKKK activator compared to the original system. This is due to the
difference in the magnitude of the parameter values in the
system.
B. A MAPK pathway with negative feedback
Next, we examined the model of Kholdenko [21] on the
dynamic properties of MAPK cascade. Kholdenko demon-
188
(a) Euclidean highest
(b) Tangent slope highest
(c) DFT highest
(d) Euclidean lowest
(e) Tangent slope lowest
(f) DFT lowest
Fig. 3. Visual validation of the highest and lowest scoring solutions for the MAPK pathway. Top row shows the results for highest scoring solutions
under all three metrics (a-c), and the bottom row shows the results for the lowest scoring solutions (d-f). The color coding of the species correspond to
those in Figure 1 (left).
(a) Euclidean
(b) Tangent slope
(c) DFT
Fig. 4. MAPK ultra-sensitivity to the MAPKKK activator. The experiment is replicated using the PSO generated parameters using each of the three
metrics.
strated that a negative feedback loop combined with intrinsic
ultra-sensitivity of the MAPK cascade can bring about sustained oscillations in MAPK phosphorylation. We obtained
the curated COPASI model of the MAPK pathway of [21]
from the BioModels database [20]. The oscillatory patterns
produced by this model are shown in Figure 5. Using it as the
synthetic “true” data, we apply particle swarm optimization
to determine the parameters that best fit this system. Figure
6 shows the highest and the lowest scored solutions out of
the five runs of the algorithm.
It is clear from the results, the Euclidean distance is not
appropriate metric for this highly oscillatory system. While
tangent slope approach performed reasonably well, it is not
surprising that DFT comparison produced the best results.
At its core, DFT uses a series of sinusoidal curves to fit
the data and it is much superior to other metrics in this
case. For all metrics, PSO success is less consistent in the
case of such complex system. Additionally, the parameter
range for allowed values was restricted to [0.001, 100],
since with the larger range, the particles generated many
parameter sets that failed to converge in the simulation step
of optimality scoring. One reason for this may be the use of a
simple underlying dynamic model, with too few parameters
to capture the true dynamics.
189
Fig. 5.
[20].
A MAPK pathway with negative feedback. The time-series data correspond to the model of [21] as obtained from the BioModels database
(a) Euclidean highest
(b) Tangent slope highest
(c) DFT highest
(d) Euclidean lowest
(e) Tangent slope lowest
(f) DFT lowest
Fig. 6. Visual validation of the highest and lowest scoring solutions for the MAPK pathway with negative feedback. Top row shows the results
for highest scoring solutions under all three metrics (a-c), and the bottom row shows the results for the lowest scoring solutions (d-f). The color coding
of the species correspond to those in Figure 5.
C. A Jak/Stat pathway
For the last case study, we applied the PSO algorithm
to experimental data (as opposed to the two cases above,
where the data was generated synthetically from curated
dynamic models). We used a Jak/Stat signaling network with
negative regulatory feedback loops. The dynamics of this
pathway were explored by Bachmann et al. to determine the
roles of the two transcriptional negative feedback regulators
Socs and CIS [22]. We obtained the time-series data for this
system from [22] (shown in Figure 7). The experimental
data shows a spike in the activity of the phosphorylated
signaling components (pJak, pEpoR, and pStat) at the initial
time points, but later suppression by the inhibitory feedback
from the expressed genes Socs and CIS.
The challenge in analyzing this systems lies in the fact
that we have no curated dynamic model associated with it.
Using the structural knowledge of the Jak2/Stat5 system as
described by Bachmann et al., we constructed a small mass
action model corresponding to the high level interactions
among these five species:
pEpoR + Jak ↔ pEpoRJak
pJak + Stat ↔ pJakStat
pStat + Socs ↔ pStatSocs
pStat + CIS ↔ pStatCIS
SocsRNA + pJak ↔ SocsRNApJak
CISRNA + pStat ↔ CISRNApStat
pEpoRJak → pJak + pEpoR
pJakStat → pJak + pStat
pStatSocs → pStat + SocsRNA
pStatCIS → pStat + CISRNA
SocsRNApJak → SocsRNA + Jak
CISRNApStat → Stat + CISRNA
All initial concentration were set to 0, the source con-
190
centration (pEpoR) was set to 1, and the allowed parameter
value range was [0.001, 10000]. Using this model, we applied the PSO algorithm to determine the parameter values.
The results of the parameter estimation are shown in
Figure 8. All comparison metrics capture the expression of
two RNA components, and the activation with further deactivation of pJak. However, all systems missed the activation and suppression of pStat. Additionally, it is interesting
that DFT showed less consistency in the accuracy of its
solution than the other two metrics. This could be due to
the fact that not enough data points were given to capture
the periodic cycle needed to extract the signal frequencies.
IV. D ISCUSSION
Parameter estimation is a central problem in modeling the
dynamics of cellular networks. In this work, we analyzed
the performance of Particle Swarm Optimization (PSO) for
estimating the parameters from a given time-series data.
We showed in many cases that PSO successfully determines optimal solutions for a given system. When multiple
solutions were found, not all final parameter sets showed
good results. However, as we have shown, to achieve best
results, the characteristics of the estimated data set should
match the choice of scoring metric. Systems with simpler
dynamics might not need complex metrics like DFT, but
others with complex dynamics might only perform well
under DFT. To quantify the complexity of dynamics we
calculated the Shannon entropy for each system [23]. The
values are 2.98, 3.48, and 2.25 for the MAPK pathway, the
MAPK pathway with negative feedback, and the Jak/Stat
pathway, respectively. Based on these values, a correlation
emerges between systems with complex dynamics (higher
entropy of the time-series data) and the performance of more
detailed scoring metrics, such as DFT.
In exploring the performance of PSO, we observed that
some of the runs differed significantly in the execution time.
In Table I, we show the average times (in seconds) and their
standard deviation for a single optimization step with 200
particles. It is clear that DFT is a much more computationTABLE I
T HE AVERAGE TIME µ ( IN SECONDS ) FOR
A SINGLE 200- PARTICLE
OPTIMIZATION STEP AND ITS STANDARD DEVIATION σ.
MAPK cascade
MAPK (feedback)
Jak2/Stat5
Euclidean
µ = 18.45
σ = 0.015
µ = 43.29
σ = 1.348
µ = 1.18
σ = 0.003
Tangent slope
µ = 13.62
σ = 0.0197
µ = 60.89
σ = 0.587
µ = 1.08
σ = 0.003
DFT
µ = 32.47
σ = 0.708
µ = 103.80
σ = 0.484
µ = 9.56
σ = 0.038
ally demanding metric. In many cases the Euclidean distance
and the tangent slope distance can achieve good results in
much less time, and the use of DFT can be reserved for
systems with higher complexity in the data.
We also examined how fast PSO runs achieve convergence. We showed that multiple runs potentially lead to
multiple solutions. Sometimes, having an ability to acquire
multiple sets of parameter estimates could be beneficial since
multiple solutions could be examined by the experimentalist
for biological plausibility [17].
V. ACKNOWLEDGMENTS
This work was supported by a Seed Grant from the Gulf
Coast Center for Computational Cancer Research, funded by
John and Ann Doerr Fund for Computational Biomedicine,
as well as by the National Cancer Institute [grant number
R01CA125109]. The content is solely the responsibility of
the authors and does not necessarily represent the official
views of the National Cancer Institute or the National
Institutes of Health.
R EFERENCES
[1] H. Kitano, “Systems biology: A brief overview,” Science, vol. 295,
no. 5560, pp. 1662–1664, 2002.
[2] R. Sharan and T. Ideker, “Modeling cellular machinery through
biological network comparison,” Nature Biotechnology, vol. 24, no. 4,
pp. 427–433, 2006.
[3] R. Albert and H. G. Othmer, “...but no kinetic details needed,” SIAM:
Society for Industrial and Applied Mathematics News, vol. 36, no. 10,
2003.
[4] U. S. Bhalla and R. Iyengar, “Emergent properties of network of biological signaling pathways,” Science, vol. 283, pp. 381–387, January
1999.
[5] G. Moles, P. Mendes, and J. Banga, “Parameter estimation in biochemical pathways: A comparison of global optimization methods,”
Genome Research, vol. 13, 2003.
[6] J. Kennedy and R. Eberhard, “Particle swarm optimization,” in Proc.
IEEE International Conference on Neural Networks. IEEE Service
Center, 1995, pp. 1942–1948.
[7] Y. Shi and R. Eberhard, “A modified particle swarm optimizer,” in
Proc. IEEE International Conference on Evolutionary Computation,
1998, pp. 69–73.
[8] G. Venter and J. Sobieszczanski-Sobieski, “Particle swarm optimization,” The American Institute of Aeronautics and Astronautics (AIAA)
Journal, vol. 41, no. 8, pp. 1583– 1589, 2003.
[9] Y. Shi and R. Eberhard, “Parameter selection in particle swarm optimization,” Evolutionary Programming VII, Lecture Notes in Computer
Science, pp. 591–600, 1998.
[10] N. Iwasaki, K. Yasudo, and G. Ueno, “Particle swarm optimization:
Dynamic parameter adjustment using swarm activity,” in IEEE International Conference on Systems, Man and Cybernetics., October
2008, pp. 2634 – 2639.
[11] R. Eberhard and Y. Shi, “Comparing inertia weights and constriction
factors in particle swarm optimization,” in Proceedings of the 2000
Congress on Evolutionary Computing, 2000, pp. 84–88.
[12] M. Clerc and J. Kennedy, “The particle swarm - explosion, stability,
and convergence in a multidimensional complex space,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 1, pp. 58–73, 2002.
[13] S. Iadevaia, Y. Lu, F. C. Morales, G. B. Mills, and P. Ram, “Identification of optimal drug combinations targeting cellular networks:
Integrating phospho-proteomics and computational network analysis,”
Cancer Research, vol. 70, no. 17, September 2010.
[14] E. Keogh, K. Chakrabarti, M. Pazzani, and S. Mehrotra, “Dimensionality reduction for fast similarity search in large time series databases,”
Journal of knowledge and information systems, vol. 3, pp. 263–286,
2000.
[15] F. MoĢˆrchen, “Time series feature extraction for data mining using
DWT and DFT,” Departement of Mathematics and Computer Science
Philipps-University Marburg, Tech. Rep. 33, 2003.
191
Fig. 7.
Experimental data for five species in the Jak/Stat pathway. Data based on [22].
(a) Euclidean highest
(b) Tangent slope highest
(c) DFT highest
(d) Euclidean lowest
(e) Tangent slope lowest
(f) DFT lowest
Fig. 8. Visual validation of the highest and lowest scoring solutions for Jak/Stat experimental data. Top row shows the results for highest scoring
solutions under all three metrics (a-c), and the bottom row shows the results for the lowest scoring solutions (d-f). The color coding of the species
correspond to those in Figure 7.
[16] S. Hoops, S. Sahle, R. Gauges, C. Lee, J. Pahle, N. Simus, M. Singhal,
L. Xu, P. Mendes, and U. Kummer, “Copasi — a complex pathway
simulator,” Bioinformatics, vol. 22, pp. 3067–74, 2006.
[17] P. Naval, Jr, L. G. Sison, and E. Mendoza, “Metabolic network parameter inference using particle swarm optimization,” in International
Conference on Molecular Systems Biology, 2006.
[18] C.-Y. Huang and J. E. Ferrell, “Ultrasensitivity in the mitogenactivated protein kinase cascade,” Proceedings of the National
Academy of Sciences, vol. 93, pp. 10 078–10 083, September 1996.
[19] S. J. Vayttaden, S. M. Ajay, and U. S. Bhalla, “A spectrum of models
of signaling pathways,” ChemBioChem, vol. 5, no. 10, pp. 1365–1374,
2004.
[20] C. Li, M. Donizelli, N. Rodriguez, H. Dharuri, L. Endler, V. Chelliah,
L. Li, E. He, A. Henry, M. Stefan, J. Snoep, M. Hucka, N. L.
Novere, and C. Laibe, “Biomodels database: An enhanced, curated
and annotated resource for published quantitative kinetic models,”
BMC Systems Biology, vol. 4, no. 92, 2010.
[21] B. Kholdenko, “Negative feedback and unltrasensitivey can bring
about oscillations in the mitogen-activated protein kinase cascade,”
European Journal of Biochemistry, vol. 267, pp. 1583–1588, 2000.
[22] J. Bachmann, A. Raue, M. Schilling, M. E. Bohm, C. Kreutz,
D. Kaschek, H. Basch, N. Gretz, W. D. Lehmann, J. Timmer, and
U. Klingmuller, “Division of labor by dual feedback regulator controls
jak2/stat5 signaling over broad ligand range,” Molecular Systems
Biology, vol. 7, no. 516, 2011.
[23] C. Shannon and W. Weaver, A mathematical theory of communication.
University of Illinois Press, 1963.
Download