NEW VARIANTS OF THE DIFFERENTIAL EVOLUTION ALGORITHM:
APPLICATION FOR NEUROSCIENTISTS
L. Buhry, A. Giremus, E. Grivel, S. Sa¨ ıghi and S. Renaud
IMS Laboratory, UMR 5218 CNRS
University of Bordeaux – ENSEIRB
351, Cours de la Lib´ eration 33405 TALENCE – FRANCE
Email: laure.buhry, audrey.giremus, eric.grivel, sylvain.saighi, sylvie.renaud@ims-bordeaux.fr
ABSTRACT
When dealing with non-linear estimation issues, metaheuristics are
often used. In addition to genetic algorithms (GAs), simulating an-
nealing (SA), etc., a great deal of interest has been paid to differen-
tial evolution (DE). Although this algorithm requires less iterations
than GAs or SA to solve optimization issues, its computational cost
can still be reduced. Variants have been proposed but they do not
necessarily converge to the global minimum. In this paper, our con-
tribution is twofold: 1) we present new variants of DE. They have
the advantage of converging faster than the standard DE algorithm
while being robust to local minima. 2) To confirm the efficiency of
our variants, we test them with a benchmark of functions often con-
sidered when studying metaheuristic performance. Then, we use
them in the field of neurosciences to estimate the parameters of the
Hodgkin–Huxley neuronal activity model.
1. INTRODUCTION
In the field of signal processing, several approaches operate in two
steps: system modelling and model parameter estimation. The esti-
mation step consists in searching for the set of parameters which
minimizes a given error criterion such as the least square error,
the maximum likelihood or the minimum variance. Depending on
the constraints of the application (real-time and storage capacity),
different off-line/on-line or iterative algorithms can be considered.
Thus, one can use subspace methods, adaptive or optimal filtering,
or the expectation-maximisation (EM) algorithm, etc. Their opti-
mization steps may be based on classical techniques such as the
steepest descent or the Newton Raphson methods. Nevertheless,
when dealing with highly non-linear estimation issues, alternative
solutions, such as metaheuristics, must be considered. They may be
potentially useful, especially when locating the global optimum is
a difficult task. Although their computational costs are quite high,
they are well suited to optimization problems.
These optimization techniques are inspired by natural systems
like metallurgy for the simulated annealing (SA), biology of evolu-
tion for the genetic algorithms (GAs) or ethology for the ant colony
algorithms or theparticle swarm optimization. Thus, they have al-
ready been used in a large number of application areas, such as
biomedical applications to estimate intensity distributions for brain
magnetic resonance images [11], Rayleigh fading channel simula-
tion [5], in the field of multimedia [1], etc.
In this paper, we focus our attention on the differential evo-
lution (DE) algorithm. Invented by Price and Storn in 1995 [12],
DE belongs to the class of evolutionary algorithms. Like GAs, DE
consists of a population of individuals which evolves towards a pa-
rameter vector that minimizes a beforehand defined fitness function
F
f it
. The purpose of DE is then to explore and to evaluate new
regions of the solution space, by building new candidate solutions
from existing ones. It uses mechanisms inspired by biological evo-
lution, namely the reproduction, the recombination, and the selec-
tion. However, unlike GAs, the exploration is automatically regu-
lated since new individuals result only from the recombination of
individuals of the initial population. In addition, instead of keeping
the best individuals of the k
th
generation
1
, DE consists in compar-
ing pairwise one individual of a generation with its mutant. Thus,
in DE, every candidate of the new generation has a fitness function
F
f it
that is lower (or equal) to the fitness function of the previous
one.
Our contribution is twofold: firstly, we suggest a new variant of
the DE in order to improve its convergence speed, hence its compu-
tational cost, while avoiding local minima. Secondly, we evaluate
the resulting approach in the field of neurosciences. Indeed, in this
area, signal processing plays a role that is more and more active.
Thus, we have already studied the relevance of GAs to estimate the
parameters of the neuronal activity model proposed by Hodgkin–
Huxley [6] and widely used in neurosciences [2].
The remainder of this paper is organized as follows. In the sec-
ond part, we present the classical DE strategies. Then, in the third
part, we describe the proposed variants and compare the perfor-
mance of all the strategies on the benchmark of functions introduced
by De Jong in [7] and widely used in the optimization field. Finally,
in the fourth part, we apply and test our variant on a neuro–scientific
problem: the parameter estimation of the Hodgkin-Huxley model.
2. CLASSICAL STRATEGIES OF THE DIFFERENTIAL
EVOLUTION ALGORITHM
2.1 Classical DE
DE consists in generating a population of NP
2
individuals which are
composed of D parameters, also called “genes”. The population is
initialized by randomly choosing individuals in a uniform manner
within the boundary constraints of the model. Then, at each time
step, new trial individuals are built by means of two operations: the
so–called “differentiation” and the “recombination”. In the follow-
ing, we define X
r
k
(i) as the i
th
gene of the r
th
individual of the k
th
generation.
Differentiation: the r
th
new parameter vector, X
r
k,trial
, is gener-
ated by adding to an individual X
r
1
k
randomly chosen among the k
th
generation in a uniform manner, the weighted difference between
two other population members, X
r
2
k
and X
r
3
k
, with r
1
= r
2
= r
3
:
∀r = 1,..., NP , X
r
k,trial
= X
r
1
k
+ F.(X
r
2
k
− X
r
3
k
), (1)
where F is usually set to 0.5.
Recombination: the r
th
mutant individual, X
r
k,mut
, inherits genes
of X
r
k,trial
with a probability CR, where CR ∈ [0, 1] is usually set to
0.9. By generating u according to a uniform distribution over [0, 1]
one has:
∀i = 1,..., D, ∀r = 1, .., NP , X
r
k,mut
(i)=
X
r
k,trial
(i) if u < CR
X
r
k
(i) otherwise.
1
k
th
“generation” denotes here the population at the iteration k.
2
For the sake of clarity, we use the same notations as in the seminal
paper [12].
17th European Signal Processing Conference (EUSIPCO 2009) Glasgow, Scotland, August 24-28, 2009
© EURASIP, 2009 2352