John Reinitz1, Sergey Vakulenko2, Dmitri Grigoriev3, Andreas Weber4. 1. Departments of Statistics, Ecology and Evolution, Molecular Genetics and Cell Biology, University of Chicago, Chicago, IL, USA. 2. Saint Petersburg National Research University of Information Technologies, Mechanics and Optics, Saint Petersburg, Russian Federation. 3. CNRS, Mathématiques, Université de Lille, Villeneuve d'Ascq, France. 4. Department of Computer Science, University of Bonn, Bonn, Germany.
Abstract
We consider evolution of a large population, where fitness of each organism is defined by many phenotypical traits. These traits result from expression of many genes. Under some assumptions on fitness we prove that such model organisms are capable, to some extent, to recognize the fitness landscape. That fitness landscape learning sharply reduces the number of mutations needed for adaptation. Moreover, this learning increases phenotype robustness with respect to mutations, i.e., canalizes the phenotype. We show that learning and canalization work only when evolution is gradual. Organisms can be adapted to many constraints associated with a hard environment, if that environment becomes harder step by step. Our results explain why evolution can involve genetic changes of a relatively large effect and why the total number of changes are surprisingly small. Copyright:
We consider evolution of a large population, where fitness of each organism is defined by many phenotypical traits. These traits result from expression of many genes. Under some assumptions on fitness we prove that such model organisms are capable, to some extent, to recognize the fitness landscape. That fitness landscape learning sharply reduces the number of mutations needed for adaptation. Moreover, this learning increases phenotype robustness with respect to mutations, i.e., canalizes the phenotype. We show that learning and canalization work only when evolution is gradual. Organisms can be adapted to many constraints associated with a hard environment, if that environment becomes harder step by step. Our results explain why evolution can involve genetic changes of a relatively large effect and why the total number of changes are surprisingly small. Copyright:
A central idea of modern biology is that evolution proceeds by mutation and selection. This process may be represented as a walk in a fitness landscape leading to fitness increase and slow adaptation
[1]. According to classical ideas this walk can be considered a sequence of small random steps with small phenotypic effects. Nevertheless, there is a limited amount of experimental support for this idea
[2] and some experimental evidence that evolution can involve genetic changes of a relatively large effect and that the total number of changes are surprisingly small
[3]. Another intriguing fact is that organisms are capable of making adaptive predictions of environmental changes
[4].To explain those facts new evolutionary concepts have been suggested (see the review by
5 and references therein). The main idea is that a population can “learn” (recognize) fitness landscapes
[5–
7]. The approach developed in these works is a generalization of ideas from machine learning in which learning (regression to data) is viewed as selection and generalization (interpolation or extrapolation) is viewed as adaptation.A mathematical basis for investigation of evolution learning problems has been developed by
8. However, this work uses a simplified model, where organisms are represented as Boolean circuits seeking an “ideal answer” to environmental challenges. These circuits involve
N
Boolean variables that can be interpreted as genes, and the ideal circuit answer maximizes the fitness. A similar model was studied numerically by
7 to confirm the theory of “facilitated variation” explaining the appearance of genetic variations which can lead to large phenotypic ones. In the work by
9 a theory of the evolution of these Boolean circuits was advanced. It was shown that, under some conditions—weak selection, see
10—a polynomially large population over polynomially many generations (polynomial in
N
) will end up almost surely consisting exclusively of assignments, which satisfy all constraints. This theorem can shed light on the problem of the evolution of complex adaptations since that satisfiability problem can be considered as a rough mathematical model of adaptation to many constraints.In
6 it is shown that, in the regime of weak selection, population evolution can be described by the multiplicative weight update algorithm (MWUA), which is a powerful tool, well known in theoretical computer science and a generalization of such famous algorithms as Adaboost and others
[11]. Note that in
6 infinitely large populations are investigated whereas the results of
9 hold only for finite populations and take into account genetic drift.Evolution Computation(EC) problems are considered recently by many papers
[12–
16] mainly for artificial test fitness functions like OneMax or LeadingOnes (for an overwiew of EC problems, see
17).In this paper, we investigate adaptation and the fitness landscape learning problem for more realistic fitness function This fitness can model adaptation for insects and connected with a fundamental hard combinatorial problem:
K-SAT.The main results can be outlined as follows. We show that, in a fixed environment, genes can serve as learners in the machine learning sense. Indeed, if an organism has survived for a long period, this fact alone constitutes important information, which can be used. The biological interpretation of this fact is simple: if a population is large enough and mutations are sufficiently rare, deleterious mutations are eliminated by purifying selection. Hence, those non-neutral mutant alleles which have become fixed in natural populations will, with probability close to 1, be adaptive and cause a positive increment of fitness (see
Theorem 3.1 and
Theorem 3.2 in
Subsection 3.1 and
Subsection 3.2). We obtain mathematical results, which allows us to estimate the reduction of mutation number due to that learning landscape procedure. Learning can sharply reduce the number of mutations needed to form a phenotypic trait useful for adaptation that is consistent with experimental data mentioned above (see
3).Another important result is as follows. We estimate the accuracy of fundamental Nagylaki equations
[6,
10] for a realistic population model, where the population size is bounded and a non-zero mutation rate is taken into account (in the case of asexual reproduction). Those accuracy estimates are fulfilled for all possible values of mutation rates and population sizes.
2 Model
In this section, we describe our model and mathematical approach.
2.1 Genome
We assume that the genotype can be described by Boolean strings of length
N
, where
N
is the number of genes. Thenwhere
s
= 1 means that gene
i is activated (switched on) and
s
= 0 means that it is repressed (switched off). Correspondence between Boolean hypercube and genotypes is considered for example in
12.
2.2 Phenotypic traits
Although phenotype is controlled by genes, it is also influenced by environmental conditions and various epigenetic processes. In this paper, we suppose that phenotypic traits are controlled by genotype only. We consider levels
f
of expressions of those traits as real variables in the interval (0, 1). Then the vector
f = (
f
1, . . . ,
f
) can be considered to represent the organismal phenotype. We suppose thatwhere
f
∈ (0, 1) is a real valued function of the Boolean string
s, the genotype.Only a part of
s
is involved in
f
. Namely, for each
j we have a set of indices
K
= {
i
1,
i
2, . . .,
i
} such that
f
depends on
s
with
i ∈
K
, so thatwhere
i
∈
K
and
n
is the number of genes involved in the control of the trait expression.The representation of phenotype by the quantities
f
is suggestive of quantitative traits because the
f
are real valued. The limiting values of 0 or 1 suggest another interpretation, however, in terms of cell type. Multicellular organisms consist of cells of different types. One can suppose that the organismal phenotype is defined completely by the corresponding cell pattern. The cell type
j is determined by morphogenes, which can be identified as gene products or signaling molecules that can change cell type (or genes that code for signaling molecules that can determine cell types or cell-cell interactions and then finally the cell pattern). The morphogene activity is defined by (
2.2).We further suppose:Assumption M.
Assume activities fThe sets K
= {1, . . .
N
}We denote the total number of genes involved in regulation of all
f
by
N
, whereAssumption
M implies that the genetic control of the phenotype is organized, in a sense, randomly, and that only a portion of the full set of genes controls phenotypic traits. That modularity of gene control is well known from experimental data (see
18,
19) and for evolution computation problems it was studied, for example, in
16.Consider an example, where the assumption
M holds, where we have a saturated expression, inspired by earlier work
[20,
21]. Letwhere
j = 1, . . . ,
N
. Here
σ(
z) is a sigmoidal function of real
z such thatand
w
,
h
are some coefficients (their meaning will be explained below). As an example, we can take
σ(
S) = (1 + exp(
−bS))
−1, where
b > 0 is a sharpness parameter. Note that for large
b this sigmoidal function tends to the step function and for
b = +
∞ our model becomes a Boolean one. The parameters
h
defines thresholds for trait expression
[20]. The relation (
2.4) can be interpreted as a simple mathematical model for quantitative trait locus (QTL) action.To understand the role of
h
consider a trait
f
and suppose that for a well adapted organism
f
1. Let, for simplicity,
w
take the values 1, 0, or
−1. Then the parameter
h
defines how many genes involved in the control of the
f
expression should be activators and how many should be repressors. Let the numbers of activator and repressor genes be
, respectively.Then
f
1 ifOne can suppose that
h
describes a direct influence of environment on phenotype, such as stress, that can exert epigenetic effects. In
Section 2.6 using data from
22 we will show that the model defined by (
2.4) are capable to describe main topological characteristics of really observed fitness functions in the case of mimicry, camouflage and thermoregulation for insects.Let us introduce the matrix
W of size
N
×
N
with the entries
w
. The coefficients
w
determine the effects of terminal differentiation genes (see
23), and hence encodes the genotype-phenotype map. We assume that the coefficients
w
are random, with the probability that
w
> 0 or that
w
< 0 is
β/2
N, where
β > 0 is a parameter. This quantity
β <<
N defines a genetic redundancy, i.e., averaged numbers of genes that control a trait. Note that then large
β ≫ 1 one has
n
< 4
β with the probability Pr
, which is exponentially close to 1: Pr
> 1 – exp(−0.1
β), thus, the number
n
are bounded.
2.3 Fitness
We know little about the details of how fitness relates to the phenotype of multicellular organisms, and for that reason classic neo-Darwinian theory takes fitness to be a function of genotype. Some models which take account of epistasis have been proposed
[24]. The random field models assign fitness values to genotypes independently from a fixed probability distribution. They are close to mutation selection models introduced by
25, and can be named House of Cards (HoC) model. The best known model of this kind is the NK model introduced by Kauffman and Weinberger
[26], where each locus interacts with
K other loci. Rough Mount Fuji (RMF) models are obtained by combining a random HoC landscape with an additive landscape models
[27]. In evolution computations (EC) some artificial fitness models were used, for example OneMax and Leading Ones to test evolution algorithms, see for example
[15].In this work, we use the classical approach of R. Fisher by introducing an explicit representation of phenotype,
f, and allow it to determine fitness through interaction with an environment
b. That is, we assume that the phenotype is completely determined by the phenotype trait expression, and thus the fitness depends on the genotype
s via
f
.We express the relative fitness
F and its dependence on environment
b via an auxiliary function
W via the relationwhere
K
is a positive constant and
b = (
b
1,...,
b
) is a vector consisting of coefficients
b
, respectively. Below we refer to
W as a fitness potential, and we assume thatSometimes, if the parameter
b is fixed, we shall omit the corresponding argument in notation for
W and
F.We consider fitness as a numerical measure of interactions between the phenotype and an environment. For a fixed environment, this idea gives us the fitness of classical population genetics. A part of the fitness, however, depends on the organism developing properly and for now we represent it as independent of the environment, although we are aware that this is not always the case. Note that some coefficients
b
may be negative and others may be positive, and that the model (
2.7) can describe gene epistatic effects via dependence of
f
on
s if
f
are nonlinear in
s.The expression (
2.7) can serve as a rough approximation of the fitness function in the case of insects such as grasshoppers or fruit flies. In fact, important factors, which determine insect survival, are thermoregulation, mimicry and camouflage levels
[18,
22,
28]. All those factors depend on colour pigmentation pattern. Blackwhite pigmentation patterns can be roughly described by vectors
f = (
f
1,
f
2,...,
f
), where
f
1 and
f
0 mean that the cell
j is black, or white, respectively, Then thermoregulation depends on
The mimicry level can be approximately defined by expression
, where
f
* is a target pattern corresponding to an insect to mimic. Colour patterns can be also described by classical RGB formalism.The representation of the fitness as a sum of terms (
2.7) is of course a rough approximation; however if assumption
M holds that representation is consistent with important observed facts. First, mutations have been identified that alter one part of the pigment pattern without affecting any other. This independence of different pattern parts can be explained by the modular organization of the genetic regulation that controls pigmentation. In the course of evolution, different aspects of the pigment pattern have clearly evolved independently of each other
[18]. Second, the topology of the fitness landscapes was studied in
22 by field experiments in the case of insect mimicry. Main conclusions are as follows. A number of studies of fitness landscapes in natural populations have demonstrated low fitness of intermediate phenotypes, i.e., existence of valley in the fitness landscape. It is found
[22] that natural selection promotes genetic architecture preventing the expression of intermediate phenotypes. Close fitness peaks are separated by ridges, favouring colour pattern switches and allowing drift from local peaks.In
Section 2.6 we will show that the fitness model defined by (
2.4) and (
2.7) have those topological properties.
2.4 Population dynamics model
For simplicity, we consider populations with asexual reproduction. (Although a part of the results remain valid for sexual reproduction, as we discuss at the end of this subsection). We choose initial genotypes randomly from a gene pool and assign them to organisms. This choice is invariant with respect to the population member, i.e,. the probability to assign a given genotype
s to a member of the population does not depend on that member.In each generation, there are
N
pop(
t) individuals, the genome of each of which is denoted by
s(
t), where
t = 0, 1, 2,... stands for the evolution step number). Following the classical Wright-Fisher ideas, we suppose that generations do not overlap. In each generation (i.e., for each
t), the following three steps are performed:Each individual
s at each evolution step can mutate with probability
p
mut per gene;At evolution step
t each individual with a genotype
s produces
k progeny, where
k is a random non-negative integer, distributed according to the Poisson lawwhere
q =
F(
s) is the fitness of that individual;To take into account ecological restrictions on the population size, we introduce the maximal population size
N
popmax. If
(
t) >
N
popmax, where
(
t) is the number of progeny produced by the population at step
t, we kill randomly selected individuals in a population-dependent manner. The probability of the death of an individual is given by
p
kill(
) = 1
− (
N
popmax/
(
t)). If
(
t)
≤ N
popmax, we do nothing. We refer to this as the “massacre procedure.”Conditions 1 and 2 imply that mutations in the genotypes create a new genetic pool and then a new round of selection starts. Condition 3 expresses the fundamental ecological limitation that all environments can only support populations of a limited size. If
N
popmax ≫ 1 then by (
2.8) and the Central Limit Theorem one can show, under some additional conditions, (see
Section 4) that fluctuations of the population size are small, and thus the population is ecologically stable and
N
pop(
t) ≈
N
popmax.In the limit case of infinitely large populations we will write the discrete dynamical equation for the time evolution of the frequency
X(
s,
t) of the genotype
s in the population aswhere
(
t) is the average fitness of the population at the moment
t defined bywhere
S(
t) is the set of genotypes existing in the population at time
t (the genetic pool) and
X(
s,
t) =
N(
s,
t)
/N
pop(
t) is the frequency of the genotype
s. Here
N(
s,
t) denotes the number of the population members with the genotype
s at the step
t.The equations (
2.9) do not take mutations into account. They only describe changes in the genotype frequencies because of selection at the
t-th time step. The same equations govern evolution in the case of sexual reproduction in the limit of weak selection
[6,
10]. Note that for an evolution defined by (
2.9), the average fitness
(
t) defined by (
2.10) satisfies Fisher’s theorem, so that this function increases at each time step
t:
(
t + 1) ≥
(
t).
2.5 Gene regulation network
In this section, we follow ideas of the classic paper
[29]: the model should include a regulatory network, which evolves itself.Regulatory genes as well as environmental factors, such as temperature, can influence the trait expression. This effect can be realized via thresholds
h
(we shall describe it below), or via a regulation of coefficients
w
(see
30). In fact, these approaches are similar for sharp sigmoidal functions
σ that are close to step functions, as can be shown by ideas from
31. Consider the expressioninvolved in relation (
2.4). Suppose following
31 that
w
take the values
γ, 0 or −
γ, where
γ is a parameter, which can be regulated. Let
h
≪
γ be fixed.Assume that at an evolution step we have
S
>
d
, where
d
> 0 is a parameter, which is more than
γ. According to our Theorems this fact indicates, that for a well adapted organism the trait
f
(
s)
≈ 1 and the corresponding coefficient
b
= 1. On the contrary, if
S
< −
d
, then one can expect that
b
= −1 and
f
(
s) must be 0.In the both cases, we can regulate the trait expression by a feedback so that whenever
S
attains a critical level, i.e., a trait is well expressed, then
γ should be increased. In our numerical simulations we use an alternative model, where we change
h
. The alternative model, which is used in our numerical simulations, can be described as follows. We suppose that depending on activity of some regulatory genes or proteins (such as Hsp90), the threshold value can take three values
,
and
such thatwhere
D
> 1 is a large parameter that defines the number of genes involved in trait specification (see
Subsection 2.2). Thus,
h
can take large negative or positive values, and also a neutral value close to 0. The feedback can be described as follows: if
S
> 0 is large enough and
h
is small, then
h
= −
D
; if
S
< 0 and
is large enough and
h
is small, then
h
=
D
; otherwise, we do not change
h
.In our simulations, each Δ
T evolution steps we modify
h
from 0 to −
D
or
D
for the trait with the maximal value
.Such a regulation produces a good adaptation even when the number of genes is essentially less than the number of the traits. The plot in
Figure 3 shows a difference between an evolution without any regulation and with the regulation by (
2.12) described above.
Figure 3.
Frequency distributions of degree of gene pleiotropy for model (
2.4) with the parameters
N
= 4000,
β = 4,
h = 0,
N
= 3000.
However, the evolution of gene regulation via
h
has an advantage: it makes phenotype more robust. In fact, the traits with large
are non-sensitive with respect to mutations. The effect produced by this robustness is shown on
Figure 1.
Figure 1.
This graph illustrates a difference between the adaptation for evolution without evolution of gene regulation via threshold (the red curve) and with that evolution by 2.12 (the green curve).
The parameters are
N
= 50,
M = 300, the mutation rate
p
mut = 0.01,
γ = 1 and
K = 4. Non-zero coefficients
w
are random numbers distributed according to the standard normal law. The initial genome is a random binary string, where each value is 0 or 1 with probability 1/2. The coefficients
b
are either 1 or −1, where the probability of 1 is
p
= 0.8.
This graph illustrates a difference between the adaptation for evolution without evolution of gene regulation via threshold (the red curve) and with that evolution by 2.12 (the green curve).
The parameters are
N
= 50,
M = 300, the mutation rate
p
mut = 0.01,
γ = 1 and
K = 4. Non-zero coefficients
w
are random numbers distributed according to the standard normal law. The initial genome is a random binary string, where each value is 0 or 1 with probability 1/2. The coefficients
b
are either 1 or −1, where the probability of 1 is
p
= 0.8.This regulation is even more effective, if we consider the modular evolution, following the recent paper
[32]. Indeed, biological systems are characterized by a high degree of modularity. This modularity allows biological systems to vary only in a small subset of traits at each evolution round.
Figure 2 shows an effect of this modularity. We consider a toy example, where a system with only 10 genes should be adapted to 200 constraints. Without regulation, we have no chances to make an adaptation (only about 10 traits are correctly adapted, see the green curve). It is a consequence of a formidable pleiotropy (20 traits on 1 gene). However, if at each evolution stage an organism should be adapted to 4 traits whereas the remaining ones are made robust by high
, then already 66 traits are correctly expressed after 20000 evolution steps. Note that the learning plays a key role in the regulation. In fact, the sign of the regulation threshold
h
depends on the sign of the corresponding coefficient
b
, and the knowledge of that sign gives us important information on the fitness landscape.
Figure 2.
This graph shows that the modular evolution of gene regulation allows adaptation with a few genes, in that case
N
= 10 and the number of traits
M = 200.
This means that we have a very big pleiotropy. Evolution proceeds in 20000 steps. The green curve corresponds to random walk with fixed small
h without any evolution of gene regulation. We observe that with 10 genes 9 traits are correctly expressed. If evolution goes into 50 rounds then 66 traits are correctly adapted, and if evolution goes in 190 rounds then 129 traits are correctly adapted (the red curve). In that last case, at each step we make adaptation to at most one trait. Parameters are as follows. The mutation rate
p
mut = 0.01 and
K = 5. Nonzero coefficients
w
are random numbers distributed according to the standard normal law. The initial genome is a random binary string, where each value is 0 or 1 with probability 1/2. The coefficients
b
are either 1 or –1, where the probability of 1 is
p
= 0.8.
This graph shows that the modular evolution of gene regulation allows adaptation with a few genes, in that case
N
= 10 and the number of traits
M = 200.
This means that we have a very big pleiotropy. Evolution proceeds in 20000 steps. The green curve corresponds to random walk with fixed small
h without any evolution of gene regulation. We observe that with 10 genes 9 traits are correctly expressed. If evolution goes into 50 rounds then 66 traits are correctly adapted, and if evolution goes in 190 rounds then 129 traits are correctly adapted (the red curve). In that last case, at each step we make adaptation to at most one trait. Parameters are as follows. The mutation rate
p
mut = 0.01 and
K = 5. Nonzero coefficients
w
are random numbers distributed according to the standard normal law. The initial genome is a random binary string, where each value is 0 or 1 with probability 1/2. The coefficients
b
are either 1 or –1, where the probability of 1 is
p
= 0.8.
2.6 Adaptation as a hard combinatorial problem
Adaptation (i.e., maximization of fitness in a changing environment) is a very hard problem since over evolutionary history we observe the coevolution of many traits accompanied by changes in many genes. In its general context, this is a problem in the theory of macroevolution, which in general requires the integration of population genetics and developmental biology for its full understanding. There are two key components of this problem. First, development is itself a dynamical process operating over time. Second, there is a combinatorial component of development wherein different combinations of gene must be expressed in different cell types. This combinatorial aspect of the problem means that straightforward theoretical methods of considering the relationship between gene expression and a changing environment that have been very successful in single celled organisms
[33] cannot be applied to metazoa. In this work, for the sake of tractability, we focus on the combinatorial aspect of the problem and neglect developmental dynamics. Even at the highly simplified level of our model, adaptation is a hard computational problem, as we now demonstrate.Consider the case, where
f
are defined by relations (
2.4) and assume thatσ is the step function;b
> 0.As a consequence of the second assumption,
F attains its maximum for
f
1 = 1,
f
2 = 1,...,
f
= 1. Let us show that, even in this particular case, the problem of the fitness maximization with respect to
s is very complex. In fact, for a choice of
h
it reduces to the famous NP-complete problem, so-called
K-SAT, which has received a great deal of attention in the last few decades (see
34–
39). The
K-SAT can be formulated as follows.K
-SAT problem.
Let us consider the set V
= {
s
1,...,
s
}
of Boolean variables s
∈ {0, 1}
and a set
z
,
z
,...,
z
of sCook and Levin
[34,
35] have shown that the
K-SAT problem is NP-complete and therefore in general it is not feasible in a reasonable running time. In subsequent studies—for instance, by
36 —it was shown that
K-SAT of a random structure is feasible under the condition that
N
(
K)
N
, where
α
(
K)
≈ 2
log 2 for large
K.The set
of solutions of random
K-SAT has a nontrivial structure depending on parameter
α =
N
[37,
39]. For sufficiently small
α < α
(
K), where
α
(
K)
≈ 2
log(
K)
/K is some critical parameter, the set
forms a giant cluster, where nearest solutions are connected by a single flip and one can go from a solution to another by a sequence of single flips (pointed mutations)
[39]. For
α ∈ (
α
,
α
), where
α
(
k)
< α
is another critical value, solutions form a set of disconnected clusters. The local search algorithms do not work in the domain
α > α
.Probably, for evolution context
K-SAT was applied first in
40, where it was used for an investigation of speciation problem.To see the connection of our model with
K-SAT, consider equation (
2.4) supposing that
w
∈ {1, 0, −1} and
h
=
−C
+ 0.5, where
C
is the number of negative
w
in the sum
We set
m =
N
and
n =
N
. Under this choice of
h
, the terms
σ(
S
) can be represented as disjunctions of literals
z
. Each literal
z
equals either
s
or
, where
denotes negation of
s
. To maximize the fitness, we must assign
s
such that all disjunctions will be satisfied. If we fix the number
n
of the literals participating in each disjunction (clause) and set
n
=
K, this assignment problem is precisely the
K-SAT problem formulated above.Reduction to the
K-SAT problem is a transparent way of representing the idea that multiple constraints need to be satisfied. The number
K defines the gene redundancy and the probability of gene pleiotropy. Remind that pleiotropy occurs when one gene influences two or more seemingly unrelated phenotypic traits. The threshold
h
and
K define the number of genes which need be flipped in order to attain a high expression of the trait
f
. Note that gene pleiotropy is a fundamental characteristics
[41], which is studied for real organisms only recently (see
19). We can compare experimental observations and consequence of model (
2.4), which is a generalisation of
K-SAT (compare plot
Figure 3 and plots on Figure 1 in
19). So, we can fit our model parameters using real data. Moreover, we can check validity of our model by the following arguments.We note that, in the case of giant cluster formation, the topological properties of the solution set
, mentioned above, outline the properties of really observed fitness landscapes
[22]: existence of many peaks, valleys and ridges connecting peaks. Namely, existence of many solutions of
K-SAT, when a giant cluster exists, means that the landscape has a number of peaks separated by valleys. On the other hand, connectance of solutions within the giant cluster can be interpreted that there exist ridges that connect peaks.Note that there are important differences between
K-SAT in Theoretical Computer Science and fitness maximization problems. First, the signs of
b
are unknown for real biological situations since the fitness landscape is unknown. Second, our adaptation problem involves the threshold parameters
h
(see (
2.4)). In contrast to
K-SAT, in our case the Boolean circuit is plastic, because the
h
are not fixed.If the
b
are unknown, the adaptation (fitness maximization) problem becomes even harder because we do not know the function to optimize. Therefore, many algorithms for
K-SAT are useless for biological adaptation problems. Below we will nonetheless obtain some analytical results based on the assumption that
b
are random.
3. Main theorems
The subsequent material is organized as follows. First we formulate a result on regulation mechanism power. Furthermore, we prove two fitness landscape learning theorems.
3.1 Fitness landscape learning theorems
For simplicity, we consider asexual reproduction. To obtain similar results for sexual reproduction, one can consider a weak selection regime and use the results of
10, where eq. (
2.9) are derived.Let us introduce two sets of indices
I
+ and
I
, such that
I
+ ∪
I
= {1,...,
N
}. We refer to these sets in the sequel as positive and negative sets, respectively. We haveThe biological interpretation of that definition is transparent: the expression of the traits
f
with
j ∊
I
+ increases the fitness and for
j ∊
I
– expression of the trait decreases the fitness.Let
s and
be two genotypes. Then we denote by Diff(
s,
) the set of positions
i such that
s
≠
:The set Diff(
s,
) indicates which genes in
s should be flipped in order to obtain
.We formulate two theorems on fitness landscape learning. First we consider the case of infinitely large populations.Theorem 3.1.
Suppose that the evolution of the genotype frequencies X(
s,
t)
is determined by equations (I
for all t ∊ [
T
1,
T
1 +
T
]
, where T
1,
T
0
are integers, the population contains two genotypes s and
such that the frequencies X(
s,
t)
and X(
,
t)
satisfyII
we havefor some j. In other words, the genes s
≠
are involved in a single regulation set KIII
LetandThen, ifwe have j ∊
I
+.
If f
(
s) >
f
(
),
then j ∊
I
.Before proving this, let us make some comments. The biological meaning of the theorem is simple: for simple fitness models, where unknown parameters
b
are involved in a linear way, in the limit of infinitely large populations fitness landscape learning is possible.Moreover, note that we do not make any specific assumptions about the nature of mutation, but only that all genetic variation between
s and
are contained in a single regulatory set
K
.The assertion of
Theorem 3.1 is not valid if the set Diff(
s,
) belongs to a union of different regulation sets
K
=
j
1, . . . ,
j
with
p > 1. This effect of belonging to different sets
K
is pleiotropy in gene regulation. Note that if
N
≪
N
then the pleiotropy probability is small for large genome lengths
N
. On the contrary, if
N
≫
N
then assumption
II is invalid.Assumption
II looks natural if when we deal with point mutations. In fact, if
is obtained from
s by a single point mutation then condition (
3.4) always holds for some
j. For small mutation rates the probability of two point mutations is essentially below than the probability of a single mutation.To conclude let us note that Theorem gives a rough estimate for the learning time
T
:Proof. The main idea is simple. Negative mutations lead to elimination of mutant genotypes from the population, and the corresponding frequencies become, for large times, exponentially small.Assume that (
3.7) holds. Let
j ∊ I
, and thus
b
0. Consider the quantityAccording to assumption
IIAssumption
III entails thatRelations (
2.6) and (
3.10) implyBy (
2.9) and the last inequality we find that for
T > T
1Consider inequality (
3.11) for
T =
T
1 +
T
. Let us note that in the relation
Q(
T
1) =
X(
s,
T
1)
/X(
,
T
1) the numerator is
p
0 whereas the denominator ≤ 1. Thus,
Q(
T
1) ≥
p
0. The same arguments show that
Q(
T
1 +
T
) ≤ 1
/p
1. Therefore, by (
3.11) one obtains thatThis inequality leads to a contradiction for
T
satisfying (
3.6), thus completing the proof.
3.2 The case of finite populations
Theorem 3.1 can be extended to the case of finite populations and non-zero mutation rates. is small. To formulate this generalization, we need an additional assumption about the fitness function. Suppose thatwhere
c
,
C
0 are constants independent of
t. For example, ifthen
c
=
K
exp(–
γ) and
C
=
K
exp(
γ) and (
3.13) holds if
K
> exp(
γ).Condition (
3.13) means that each individual gives birth to at least
c
and at most
C
descendants, where those bounds do not depend on the population size and evolution step.LetNote that for simplicity in the next
Theorem 3.2 we consider point mutations (bit flipping) only. The model used here cannot represent mutations of arbitrarily small effect, but it can include insertions or deletions. In contrast to
Theorem 3.2,
Theorem 3.1 is valid for all kinds of mutations.Then we haveTheorem 3.2.
Consider the population dynamics defined by model
1
-
3
in
(
3.14)
and
M
hold, and assumptions (Then if j ∊
Iis fulfilled with the probability Prwhere for large N
→ 0Interpretation ofIt is interesting to compare
Theorem 3.1 and
Theorem 3.2. The previous one asserts that for infinite populations the probability of the event
j ∊ I
is zero whereas the second one claims that this probability becomes exponentially small as the population size increases.This theorem also shows that evolution can make a statistical test checking the hypothesis
H
that
j ∊ I
– against the hypothesis
H
+ that
j ∊ I
+. Suppose that
H
– is true. Let
V be the event that the frequency
X(
,
t) of the genotype
in the population is larger than
p
1 within a sufficiently large time
T
. According to estimate (
3.18), the probability of the event
V is so small that it is almost unbelievable. Therefore, the hypothesis
H
– should be rejected. We will refer
T
as the
checking time.Rare mutants. In this Theorem we assume that the frequencies
p
0 and
p
1 of genotypes (wild and mutant) are fixed and our estimate is valid as
p
mut → 0. I.e., we do not consider mutants with a very small frequencies (fractions). Of course, a large population always contains a small number of such mutants. In numerical simulations we assume that evolution is successful and population is perfectly adapted, if, say, 95 or 99 percents of population members have the maximal fitness.Ideas for the proof. The main idea is the same as that for the previous theorem: we compare the frequencies of the organisms with the genotype
and the organisms with the genotype
s. However, the proof includes a number of technical details connected with estimates of mutation effects and fluctuations. The formal proof can be found in
Section 4. It is based on estimates of the accuracy of the Nagylaki equations (
2.9). The main
Lemma 4.5 for the proof of
Theorem 3.2. admits a transparent interpretation. We show that fraction
X(
s,
t) of genotype
s evolves in time in such a way that the estimates.are satisfied, where
F(
s) is a fitness of genotype
s,
is average population fitness, and
r(
p
,
N
)) are small corrections, which converge to zero uniformly in
s as the mutation rate
p
mut → 0 and the population size
N
popmax → ∞. This means that in the limit
p
mut → 0,
N
popmax → ∞ we have equation (
2.9). The main problem with the application of
Theorem 3.2 is how it allows to perform fitness landscape learning. It can be done by a regulation, as is detailed in the following section.
4 Proof of theorems
Let us prove
Theorem 3.2.
4.1 Main tools and auxiliary Lemmas
Let us introduce notation and make some preliminary remarks. Remind that we denote by
N(
s,
t) the number of the population members with the genotype
s at the moment
t. Let
X(
t) be the set of all population members at the moment
t. For each
x ∊ X(
t) let us denote by
(
x,
t) the number of progeny born by the individual
x at the moment
t before the massacre (see point 3 of model from
Subsection 2.4). Let
s
(
x) be the genotype of
x. Then, according to (
2.8), the mean of
(
x,
t) iswhere
EX denotes the expected value of
X. By
(
s,
t) we denote the number of all progeny born by individuals with the genotype
s at the moment
t before the massacre. Since all progeny are produced independently and randomly, the previous relation givesOur main analytical tools are the Chernoff bounds and the Hoeffding inequalities. We also use the Markov inequality: for a positive random quantity
X and
a > 0 one hasMoreover, we use two elementary estimates. Let be an event in stochastic population dynamics. We denote by
Not the negation (complement) of
and by Pr(
|ℬ) the conditional probability of
under the condition
ℬ. For events
,
ℬ
1, . . . ,
ℬ
we haveFor two events
,
ℬ one hasLemma 4.1.
Let X
= 1, . . . ,
n. Let each XThen for all δ > 0whereSimilarly,Proof. Note that for any
λ > 0Since
X
are independent quantities, we haveThe straight forward computation shows thatTherefore, due to the Markov inequality (
4.3) and estimate (
4.8) one haswhereWe minimize
f with respect to
λ and obtain (
4.6). To derive (
4.7), we useand repeat the same arguments. The Lemma is proved.Lemma 4.2.
Let X
= 1, . . . ,
n such that X
{0, 1}
and EX
=
p. ThenwhereProof. Note that for any
λ > 0Since
X
are independent quantities, we haveNote that
E exp(
−λX
) =
p exp(
−λ) + 1
− p. LetWe take
λ = ln 2 and find that
G(ln 2,
p) =
−g(
p). Now by using the Markov inequality (
4.3) and estimate (
4.12) one obtains (
4.10). The Lemma is proved.We also use the following Chernoff-Hoeffding theorem. Let
X
be i.i.d. quantities such that
X
∊ {0, 1} and
EX
=
p, where
i = 1, . . . ,
n. Then for
one haswhere
D(
x||y) is the Kullback-Leibler divergenceMoreover, we will use the Hoeffding Theorem: if i.i.d. quantities
X
[0, 1] with the probability 1 then
4.2 Main lemmas
First we estimate the population size fluctuations.Lemma 4.3.
Let
(
t)
be the number of all progeny, born in the population at the moment t before the massacre, and ε
1 > 0
be a small number. Thenwith probabilitywhereProof. Let
(
x,
t) denote the number of progeny produced by the individual
x before the massacre at the
t-th evolution step. The number
(
t) is the sumof the mutually independent random quantities. According to (
4.2), the average
E
(
x,
t) is
F(
s
(
x)). Therefore,We setand use the
Lemma 4.1 that gives us (
4.17).Lemma 4.4.
Let
ε
2 ∊ (0, 1)
be fixed and condition (
3.13)
be fulfilled. Assume, moreover, thatwhereand c
> 1
is defined by (
(
t)
byThen one haswhereandProof. Let
ξ(
x) be random quantities defined as follows:
ξ(
x) = 1 if the individual
x is survived as a result of massacre (see point 3 of our model from
Subsection 2.4), and
ξ(
x) = 0 otherwise. Let
(
t) be the set of progeny produced by all individuals from the population. Then the number
N
sur(
t) =
N
pop(
t + 1) of finally survived progeny can be computed as follows:Note that
|
(
t)
| =
(
t). Moreover,
Eξ(
x ) =
N
popmax/
(
t) for
(
t)
≥ N
popmax. Therefore, if
(
t)
≥ N
popmax thenLet us define the eventwhere the interval
J
(
t) is defined by (
4.16) and
is defined by (
4.27). By (
4.4) we haveNow we apply the Hoeffding inequality (
4.15). For each
ε
2 > 0 we obtainIf
ℬ (
t) takes place, then
(
t)
≥ N
popmax and consequentlyTherefore,Moreover, by
Lemma 4.3Inequalities (
4.30), (
4.32) and (
4.33) prove (
4.25).The following lemma, in particular, allows us to obtain equations (
2.9) and (
2.10) in the limit of infinite populations and for small mutation probabilities.Recall that
denotes the number of non-mutated progeny generated by the individuals with the genotype
s before the massacre. Let
N
sur(
s,
t) be the number of those progeny that survived after that massacre.Lemma 4.5.
Let ε
0
be a positive number satisfying (
4.75)
andThen one haswith the probability Pr
such thatwhereSimilarly,with the probability Pr
such thatProof. Step 1, estimates of fluctuations. First let us estimate the fluctuations of the number
. For each
ε
2 > 0 let us define the eventBy
Lemma 4.1 one hasNote thatAs a result, by (
4.46) we obtainStep 2. Here we estimate the number of progeny that survived as a result of the massacre procedure (point 3 of the population dynamics model, see subsection 2.9). Let
X' (
s, t) be the set of progeny produced by individuals with the genotype
s. Then the number
N
sur(
s, t) of survived progeny
x for individuals
x belonging to the set
Z' (
s, t) iswhere
ξ(
x) are defined in the proof of the previous Lemma. For
ε
3 > 0 we consider the eventLet us estimate the probability Pr(
sur,
(
t)). According to the Hoeffding Theorem (
4.15)Note that
ξ(
x) and
ξ(
y) are independent quantities for different
x and
y, thus under the condition
>
N
popmaxtherefore,Let us define the events
ℬ
(
t) and
ℬ(
t) byThen using (
4.4) one hasWe observe that under conditions ℬ
(
t) and ℬ(
t)In that estimate let us set
ε
2 = 0.5 and
ε
1 = 1. Taking into account that
E
=
N
pop(
t) <
2C
popmax, we have thatwhereMoreover, according to (
4.47)and due to (
4.17)where
R
1,
R
2 are defined by (
4.37) and (
4.38). Finally,Step 3, estimate of the number of mutants.Let us estimate how many individuals with genotypes
s can mutate. The probability of mutation is
p
mut. Let
N
mut(
s,
t) be the number of such mutants. Let us define the event
mut,
(
t) bySince the random quantity
N
mut(
s,
t) is subject to the Bernoulli law, we can apply the Chernoff-Hoeffding inequality (
4.13). Then we obtain thatwhere, according to definition (
4.14) of
D(
x||
y), one hasand
g is defined by (
4.11).Using (
4.4) one hasAs a result, by
Lemma 4.3 one findswhere
R
4,
R
5 are defined by (
4.40) and (
4.41).To prove (
4.35), we set
ε
3 =
ε
0/2. Taking into account condition (
4.75) for
ε
0 we see that if the both events Not
mut,
(
t) and Not
sur,
(
t) take place, then inequality (
4.35) is fulfilled. Thuswhere
R
are defined by (
4.37)–(
4.41).Finally, taking into account the results of steps 1, 2 and 3 we see that estimate (
4.35) holds with the probability Pr
t,+. It completes the proof of (
4.35). The second inequality (
4.42) can be obtained in the same way.
4.3 Remaining part of the proof of
Theorem 3.2
We use the same idea that in the proof of
Theorem 3.1 but first we establish uniform bounds for the population size and other quantities involved in the proof.Step 1 Here we estimate the population size. Let us setin
Lemma 4.4. Let us consider the events
ε
(
t) defined by (
4.24) in
Lemma 4.4. If the events
ε
(
t) take place for all
t ∈ [
T
1,
T
1 +
T
] and
N
pop(0) =
N
popmax, we have thatThen conditions (
3.15), (
3.16) of
Theorem 3.2 implyThose inequalities imply the following estimates for the quantities
R
defined by (
4.37)–(
4.41):where
q
are defined bywhereandwhere
= 1 – (
κc
)
–1, andFor each
p ∈ (0, 1) let us define an auxiliary functionwhere
q
,
are defined by relations (
4.68)–(
4.72). We can find asymptotics of
ρ under natural assumptions that
p
mut → 0 and
N
popmax → ∞ while all the rest parameters are fixed. Then the leading term in the right hand side of (
4.36) is
q
4 and
As a result, we haveStep 2. Let
Q(
t) is defined by (
3.8) and, moreover, let
j ∈
I
−. We use
Lemma 4.5 inductively for genotypes
s and
. Let us setwhere
ε is defined by (
4.75). We remark that the inequalityholds with a probability Pr
> 0. Let us obtain a uniform estimate of that probability. Let
ℰ(
t) be the event that (
4.78) holds at the step
t. Using (
4.4) we havewhere, according to
Lemma 4.4, the probability of the event Not
ε
(
t) is less than
η, where
η is defined by (
4.25), andWe conclude by (
4.5) thatwhere
is defined by (
4.74). This estimate is uniform in
t
∈ [1, . . . ,
T
]. By (
4.80) we obtain then that the inequalityis satisfied with the probability Pr
such thatFor
ε
0 defined by (
4.75). one hasNow repeating the same arguments that in the end of the proof of
Theorem 3.1, and taking into account asymptotics (
4.77), we obtain the conclusion of
Theorem 3.2.
5 Discussion
In this paper, we proposed a model for fitness landscape learning, which extends earlier work by
7–
9 in two ways. First, we use hybrid circuits involving two kinds of variables. The first class of variables are real valued in the interval (0, 1) and can be interpreted as relative levels of phenotypic traits, other variables are Boolean and can be interpreted as genes. Second, we use a threshold scheme of regulation, which is inspired by ideas of the paper by
20. All variables are involved in gene regulation via thresholds.The work presented here is a major extension of a long term effort to explicitly model the effects of phenotypic buffering in evolution by considering a class of Boolean and mixed Boolean-continuous models in which the phenotype is represented explicitly and the degree of phenotypic buffering can be controlled in various ways. For example, we have demonstrated that the idea of an “evolutionary capacitor”
[42,
43] can be implemented by explicit control of phenotypic buffering in a hub-and-spokes architecture
[23] and that in a more general class of genetic architecture numerical simulations show that an intermediate level of buffering is optimal for evolution in a changing environment
[31].The results reported here are very promising, since they are consistent with the results of recent experiments by
44 and
45 on heat shock stress. The essential mechanism is that the exploration of the fitness landscape by the genetic network in such a way that future mutations are more likely to be adaptive. We have shown that, at least for some fitness landscapes, rapid evolutionary changes—perhaps instances of the “hopeful monsters” of Goldschmidt
[46]—can be created by a combination of random small mutations and epigenetic effects. The main idea is that small mutations pave the way for large epigenetic or genetic changes. The hypothetical mechanism, which we propose, can be outlined as follows (see
Figure 4,
Figure 5).
Figure 4.
This graph illustrates Goldshmidt’s leaps.
At the initial moment the trait expressions take the values
x = 0.5,
y = 0.5. According to Fisher’s ideas, random large mutations decrease the fitness
F =
K
exp(
W). (Changes of
x =
F
,
y =
f
, which are induced by mutations, are shown by red vectors.) Thus such mutations produce non-viable organisms.
Figure 5.
This plot illustrates the main ideas of evolution based on the fitness landscape learning.
At the initial time the trait expressions take the values
x = 0.5,
y = 0.5. Evolutionary changes go in two stages. First we make small random mutations (shown by red vectors), which explore the fitness landscape. If such a mutation is not eliminated from the population, this means that a correct evolution direction is found, and gene regulation system makes a big leap (shown by the green vector) in the direction of that small mutation. Such a two step model can be called clever Goldshmidt leaps. Note that evolution is gradual, and the existence of clusters of almost identical genes involved in the same QTL increases the chances to create a clever Goldschmidt hopeful monster.
This graph illustrates Goldshmidt’s leaps.
At the initial moment the trait expressions take the values
x = 0.5,
y = 0.5. According to Fisher’s ideas, random large mutations decrease the fitness
F =
K
exp(
W). (Changes of
x =
F
,
y =
f
, which are induced by mutations, are shown by red vectors.) Thus such mutations produce non-viable organisms.
This plot illustrates the main ideas of evolution based on the fitness landscape learning.
At the initial time the trait expressions take the values
x = 0.5,
y = 0.5. Evolutionary changes go in two stages. First we make small random mutations (shown by red vectors), which explore the fitness landscape. If such a mutation is not eliminated from the population, this means that a correct evolution direction is found, and gene regulation system makes a big leap (shown by the green vector) in the direction of that small mutation. Such a two step model can be called clever Goldshmidt leaps. Note that evolution is gradual, and the existence of clusters of almost identical genes involved in the same QTL increases the chances to create a clever Goldschmidt hopeful monster.This network modifies the thresholds by a simple feedback mechanism. Although we are not aware of clear experimental evidence for the existence of such a mechanism, we nevertheless think that such a mechanism can be connected with regulations via enhancers
[47,
48], where enhancer action is described by deep network models based on thermodynamics, and chemical kinetics, and those models contain threshold parameters. Alternative variants involve modifications of weights. To some extent, both mechanisms are mathematically equivalent. However, the regulation via thresholds has an important advantage: it makes phenotypes robust with respect to mutations. In this second version of the paper we included a subsection about regulation and two pictures, which show results of numerical simulations based on the “strong selection weak mutation” (SSWM) algorithm. As is the case for threshold signs, they of course should have different signs. The key question is how an individual obtains information about the fitness landscape. In our model that information is the sign of
b
. If the
b
is positive the corresponding threshold must be negative, otherwise, it should be negative. Actually, we think that the gene of individuals have that information! According to the main theorems proved in this paper, the fact that a mutated individual survives for a sufficient long period of time gives us that information. The very fact of the existence of the individual carries the most important information. Look at the mutations of its genotype and you will know where evolution should go!As is described in
Subsection 2.5, we assume that expression of genes involved in the expression of phenotypic traits depends on threshold parameters
h
, which take three values: a large negative one, a neutral value close to zero and a large positive one. First the threshold parameter
h
is small and thus the phenotypic trait is sensitive with respect to even small mutations. Those mutations play a fundamental role working as scouts exploring environments (see
Figure 4). If a mutation occurred and the corresponding mutant has survived within
T
≫ 1 generations then according to
Theorem 3.1 and
Theorem 3.2 these events mean that that mutation increases the fitness that allows the network to estimate the correct direction of evolution. Then gene regulation detects that increase to change the threshold according to simple rules. Namely, if the trait is less expressed in that mutant with respect to wild type parent, the gene regulation system decreases the threshold up to the large negative value. On the contrary, if the trait is strongly expressed in the mutant, the gene regulation system increases the threshold up to the large positive value. This simple regulation control not only sharply reduces the number of mutations needed for adaptation, but also canalizes the phenotype since for large thresholds the trait expression level becomes insensitive with respect to mutations. We suppose that these threshold modifications can be inherited.So, we propose the mechanism: small mutations serve as scouts finding the way for large epigenetic or genetic changes, which can be performed by gene regulatory system.The mechanism may also explain the results of
4 on prediction of environmental changes. In fact, let us suppose that environment varies in time. The first, perhaps relatively small, variations can trigger the threshold mechanism described above. As a result, the population will be adapted to the subsequent changes in advance.Our results show that evolution can proceed rapidly because it reduces the number of mutations required for adaptive change.The primary limitation of our results is that the representation of the evolving genetic network is limited to the network of gene controlling phenotype, represented here by the Boolean strings
s. Other model variables represent the coarse-grained activities of genes. One class is the terminal differentiation genes represented by
w
, and another are the genes or epigenetic factors controlling the thresholds
h and their associated learning rules. A more careful consideration of the relationship of these moieties to observable molecular entities is an important objective of future work. At the mathematical level, the key analytical results were obtained in a simplified context that falls short of a realistic level of pleiotropy and thus of the level of NP-hard complexity exhibited by fully pleotropic forms of our model. We believe that our analytical results can be generalized, which we plan to address in future work.
Data availability
All data underlying the results are available as part of the article and no additional source data are required.This is a throught-provoking, hypothetical paper with considerable potential, duly revised.I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.In the current work the authors study a population genetics model in which fitness is a linear function of a set of phenotypic traits, and where the genotype-to-phenotype map is given by a linear transformation composed with sigmoidal functions. Despite the seeming simplicity of the fitness function, the authors make the case that optimizing fitness is a hard, NP-complete, problem. Under this model, they study the extent to which the fitness landscape (that is, the question of which phenotypic traits contribute positively to fitness, and which contribute negatively) can be inferred from the distribution of these traits in the population after being subject to evolution for a moderate span of time. They then connect this question with that of whether learning algorithms (potentially epigenetic in nature) can help optimize and speed up evolution by learning the fitness landscape.Toward the goal of inferring the fitness landscape, the authors prove two theorems. Theorem 3.1 concerns a simplified model of an infinite population with no mutations, while Theorem 3.2 concerns a more complex model of a finite population with mutation, stochastic number of descendants and a culling process. In either case, the result is that under certain assumptions, if a mutant genotype is present in the population with high enough frequency after a long enough period of elapsed time, then we can confidently infer that any phenotypic trait differential between the wildtype and the mutant is associated with a higher fitness.While Theorem 3.1 ignores mutation, even Theorem 3.2 seems at odds with mutation-selection balance. Even a mutant with lower fitness will be present in the population at a frequency that is on the order of the rate of mutation, while the theorem seems to claim that the frequency of such a mutant will be exponentially small with high probability. The authors should resolve this apparent discrepancy. In addition, the main ideas of the proof would be more clearly communicated if the authors would include a treatment of the intermediate model of an infinite population with mutation.As for the discussion of the way learning can speed up the evolutionary process, this part of the paper remains unclear and underdeveloped. The authors discuss a two-step evolutionary process in which the first step consists of small mutations in order to explore the fitness landscape, and the next step involves changing the thresholds involved in the genotype-to-phenotype map in a way that promotes phenotypic traits associated with higher fitness. While this idea is interesting and worth exploring, a few issues arise. A conceptual issue remaining to be addressed is whether the threshold h
i is part of the genotype and what mechanism are needed to alter its value. According to the authors, the h
i’s can be modified genetically or epigenetically. If epigenetically, it is not apparent what are the environmental cues that will lead to such learning let alone the actual mechanism of modifying them. If genetically, it is similarly unclear in what way the learning of the fitness landscape is being stored, if at all, in the genotype, and what is the connection to the Theorems of chapter 3. The theorems in chapter 3 rely on observing the frequency of a genotype in the population, but such information is not stored in individual genotypes. Additionally, if the thresholds are understood to be variable and subject to selection, then the fitness-maximization problem in fact becomes easy (in contrast to the prior analysis of it as NP-complete), unless we impose restrictions on the range of the threshold. Indeed, one can set the thresholds at positive or negative infinity depending on whether the corresponding trait is positive or negative in order to effectively keep the trait on or off regardless of genotype.Finally, as this manuscript addresses the relation between learning and the rate of evolution, it would benefit from including a reference to one of the most relevant and intuitive articles written in the late 80’s by Geoffrey E. Hinton & Steven J. Nowlan, “How Learning Can Guide Evolution”
Complex Systems, 1, 495-502
[1]. In it, Hinton and Nowlan showed how learning alters the shape of the fitness landscape and thereby provides easier evolutionary paths towards sets of co-adapted alleles. Hinton and Nowlan demonstrated that this effect allows learning individuals to evolve faster than non-learners. Though the learning model presented by Hinton and Nowlan operates at “somatic” timescale, the analogy to mutations at the evolutionary timescale can be drawn.To conclude, though it can be improved by the above suggestions, this article touches upon very interesting and important issues in the field of evolutionary biology which are still only lightly investigated, and highlight what might be a fruitful path towards better understanding.We confirm that we have read this submission and believe that we have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however we have significant reservations, as outlined above.This is a new paper in a series of exciting papers about phenotypic evolution, based on models of genes genetic regulatory networks, and phenotypes.There are many technical details in the paper that most biologists will find difficult to follow – I would put at least some of these items into an Appendix. Also, I am wondering how many (even theoretical) biologists would readily grasp the meaning of the K-SAT problem as defined. A more accessible formulation (in addition to the strictly technical one) could do a lot for easier understanding (one of the papers by the authors contains a didactive figure, for example).I consider this paper a serious attempt at broadening our views of the role of learning dynamics in evolution, but I think the message could be made a lot clearer. The punch line in a way is in the Discussion that I find both intriguing and baffling.First, as it is portrayed in this paper, the effect of genes is filtered through a genetic regulatory network. It seems that, just like in Ref 40
[1], that this filter itself does not evolve, or at least the dynamics is not given. In other words, how do the weights in the network evolve? Note that in the Watson-Wagner paper in
Evolution
[2] it was the evolution of the weights that was related to Hebbian dynamics. So, how do we stand on this?Second, in this paper the genetic control of the thresholds in also kept implicit, but the evolution of threshold values does play an important role in the Goldschmidtian argument. Is the change in thresholds genetic or epigenetic? If the latter, then it ought to be part of the regulatory network. I guess this is what the statement “gene regulation detects that increase to change the threshold according to simple rules”. How these rules would be implemented in mechanistic terms remains obscure. And how would they arise in evolution? The simple rule implies feedback from expression levels to threshold values.Third, do I understand it correctly that here, in contrast to Ref 40
[1], thresholds can be negative? But a large negative threshold value would mean that expression levels will increase, since negative times negative is positive (Eq. 2.4). If the statement “take three values: a large negative one, a neutral value close to zero and a large positive one” instead refers to gene expression levels rather than threshold values (the latter would in this case be nonnegative) then I do not understand the thought example. Increasing the threshold would reduce the expression level that was selected to go up in the first place!These points seem to me crucial, so their clarification is badly needed. I would appreciate a few numerical examples along with at least hints to answers to the more conceptual questions on the dynamics.Also, according to the cited Nagylaki dynamics
[3] the population evolves almost as if it were in linkage equilibrium – which cannot hold for the asexual population considered by the authors.There are also some minor issues in the paper: in Eq. 2.4 “s” should be replaced by a sigma, and there are also examples where plurals and singulars in the same sentence do not match.I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above.