SlideShare a Scribd company logo
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 921
Evolutionary Generative Adversarial Networks
Chaoyue Wang, Chang Xu , Xin Yao , Fellow, IEEE, and Dacheng Tao , Fellow, IEEE
Abstract—Generative adversarial networks (GANs) have been
effective for learning generative models for real-world data.
However, accompanied with the generative tasks becoming more
and more challenging, existing GANs (GAN and its variants)
tend to suffer from different training problems such as insta-
bility and mode collapse. In this paper, we propose a novel
GAN framework called evolutionary GANs (E-GANs) for sta-
ble GAN training and improved generative performance. Unlike
existing GANs, which employ a predefined adversarial objective
function alternately training a generator and a discriminator,
we evolve a population of generators to play the adversar-
ial game with the discriminator. Different adversarial training
objectives are employed as mutation operations and each indi-
vidual (i.e., generator candidature) are updated based on these
mutations. Then, we devise an evaluation mechanism to mea-
sure the quality and diversity of generated samples, such
that only well-performing generator(s) are preserved and used
for further training. In this way, E-GAN overcomes the lim-
itations of an individual adversarial training objective and
always preserves the well-performing offspring, contributing to
progress in, and the success of GANs. Experiments on several
datasets demonstrate that E-GAN achieves convincing genera-
tive performance and reduces the training problems inherent in
existing GANs.
Index Terms—Deep generative models, evolutionary computa-
tion, generative adversarial networks (GANs).
I. INTRODUCTION
GENERATIVE adversarial networks (GANs) [1] are one
of the main groups of methods used to learn generative
Manuscript received August 16, 2018; revised November 29, 2018; accepted
January 15, 2019. Date of publication January 28, 2019; date of current ver-
sion November 27, 2019. This work was supported in part by the Australian
Research Council under Project FL-170100117, Project DP-180103424,
Project IH180100002, and Project DE180101438, in part by the National Key
Research and Development Program of China under Grant 2017YFC0804003,
in part by the Engineering and Physical Sciences Research Council (EPSRC)
under Grant EP/J017515/1 and Grant EP/P005578/1, in part by the Program
for Guangdong Introducing Innovative and Enterpreneurial Teams under
Grant 2017ZT07X386, in part by the Shenzhen Peacock Plan under Grant
KQTD2016112514355531, in part by the Science and Technology Innovation
Committee Foundation of Shenzhen under Grant ZDSYS201703031748284,
and in part by the Program for University Key Laboratory of Guangdong
Province under Grant 2017KSYS008. (Corresponding author: Chang Xu.)
C. Wang, C. Xu, and D. Tao are with the UBTECH Sydney Artificial
Intelligence Centre, Faculty of Engineering and Information Technologies,
University of Sydney, Darlington, NSW 2008, Australia, and also with
the School of Computer Science, University of Sydney, Darlington, NSW
2008, Australia (e-mail: chaoyue.wang@sydney.edu.au; c.xu@sydney.edu.au;
dacheng.tao@sydney.edu.au).
X. Yao is with the Shenzhen Key Laboratory of Computational Intelligence,
University Key Laboratory of Evolving Intelligent Systems of Guangdong
Province, Department of Computer Science and Engineering, Southern
University of Science and Technology, Shenzhen 518055, China, and also
with the CERCIA, School of Computer Science, University of Birmingham,
Birmingham B15 2TT, U.K. (e-mail: x.yao@cs.bham.ac.uk).
Color versions of one or more of the figures in this paper are available
online at http://ieeexplore.ieee.org.
Digital Object Identifier 10.1109/TEVC.2019.2895748
models from complicated real-world data. As well as using
a generator to synthesize semantically meaningful data from
standard signal distributions, GANs (GAN and its variants)
train a discriminator to distinguish real samples in the train-
ing dataset from fake samples synthesized by the generator.
As the confronter, the generator aims to deceive the discrim-
inator by producing ever more realistic samples. The training
procedure continues until the generator wins the adversarial
game; that is, the discriminator cannot make a better decision
than randomly guessing whether a particular sample is fake or
real. Compared with other existing generative models, GANs
provide a concise and efficient framework for learning genera-
tive models. Therefore, GANs have recently been successfully
applied to image generation [2]–[6], image editing [7]–[10],
video prediction [11]–[13], and many other tasks [14]–[18].
Comparing with most existing discriminative tasks (e.g.,
classification and clustering), GANs perform a challenging
generative process, which projects a standard distribution to
a much more complex high-dimensional real-world data dis-
tribution. First, although GANs have been utilized to model
large-scale real-world datasets, such as CelebA, LSUN, and
ImageNet, they easily suffer from the mode collapse problem,
i.e., the generator can only learn some limited patterns from
the large-scale target datasets, or assigns all of its probability
mass to a small region in the space [19]. Meanwhile, if the
generated distribution and the target data distribution do not
substantially overlap (usually at the beginning of training), the
generator gradients can point to more or less random direc-
tions or even result in the vanishing gradient issue. In addition,
to generate high-resolution and high-quality samples, both the
generator and discriminator are asked to be deeper and larger.
Under the vulnerable adversarial framework, it is hard to bal-
ance and optimize such large-scale deep networks. Thus, in
most of the existing works, appropriate hyper-parameters (e.g.,
learning rate and updating steps) and network architectures
are critical configurations. Unsuitable settings reduce GAN’s
performance or even fail to produce any reasonable results.
Overall, although GANs already produce visually appealing
samples in various applications, they are still facing many
large-scale optimization problems.
Many recent efforts on GANs have focused on over-
coming these optimization difficulties by developing vari-
ous adversarial training objectives. Typically, assuming the
optimal discriminator for the given generator is learned, dif-
ferent objective functions of the generator aim to measure
the distance between the generated distribution and the tar-
get data distribution under different metrics. The original
GAN uses Jensen–Shannon divergence (JSD) as the met-
ric. A number of metrics have been introduced to improve
GAN’s performance, such as least squares [20], absolute
1089-778X c
 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
922 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019
deviation [21], Kullback–Leibler (KL) divergence [22], and
Wasserstein distance [23]. However, according to both theo-
retical analyses and experimental results, minimizing each dis-
tance has its own pros and cons. For example, although mea-
suring KL divergence largely eliminates the vanishing gradient
issue, it easily results in mode collapse [22], [24]. Likewise,
Wasserstein distance greatly improves training stability but can
have nonconvergent limit cycles near equilibrium [25].
Through observation, we find that most existing GAN
methods are limited by the specified adversarial optimization
strategy. Since the training strategy is fixed, it is hard to adjust
the balance between the generator and discriminator during the
training process. Meanwhile, as aforementioned, each exist-
ing adversarial training strategy has its own pros and cons in
training GAN models.
In this paper, we build an evolutionary GAN (E-GAN),
which treats the adversarial training procedure as an evo-
lutionary problem. Specifically, a discriminator acts as the
environment (i.e., provides adaptive loss functions) and a pop-
ulation of generators evolve in response to the environment.
During each adversarial (or evolutionary) iteration, the dis-
criminator is still trained to recognize real and fake samples.
However, in our method, acting as parents, generators undergo
different mutations to produce offspring to adapt to the envi-
ronment. Different adversarial objective functions aim to min-
imize different distances between the generated distribution
and the data distribution, leading to the different mutations.
Meanwhile, given the current optimal discriminator, we mea-
sure the quality and diversity of samples generated by the
updated offspring. Finally, according to the principle of “sur-
vival of the fittest,” poorly performing offspring are removed
and the remaining well-performing offspring (i.e., generators)
are preserved and used for further training. Based on the evo-
lutionary paradigm to optimize GANs, the proposed E-GAN
overcomes the inherent limitations in the individual adversarial
training objectives and always preserves the well-performing
offspring produced by different training objectives (i.e., muta-
tions). In this way, we contribute to progress in and the success
of the large-scale optimization of GANs. Following vanilla
GAN [1], we evaluate the new algorithm in image generation
tasks. Overall, the proposed evolutionary strategy is largely
orthogonal to the existing GAN models. Through applying the
evolutionary framework to different GAN models, it is possi-
ble to perform different kinds of generation tasks. For example,
GAN objectives were devised for generating text. Considering
them as the mutation operations, the proposed evolutionary
framework can be applied to solve text generation tasks.
Meanwhile, the proposed E-GAN framework also provides
a novel direction to apply evolutionary learning paradigm
on solving deep learning problems. Recent years, although
deep learning algorithms have achieved promising perfor-
mances on a variety of applications, they are still facing
many challenges in solving real-world problems. Evolutionary
computation, as a powerful approach to complex real-world
problems [26]–[29], has been utilized to solve many deep
learning challenges. Among them, Real et al. [30] devised an
evolutionary algorithm to automatically search the architecture
and hyper-parameters of deep networks. Moreover, evolution
strategies have been utilized as an alternative to MDP-based
techniques to optimize reinforcement learning models [31].
In this paper, we attempt to combine the back propagation
algorithm and the evolutionary algorithm for optimizing deep
generative models. The parameters updated by different learn-
ing objectives are regarded as variation results during the
evolutionary process. By introducing suitable evaluation and
selection mechanisms, the whole training process can be more
efficient and stable. We hope the proposed evolutionary learn-
ing framework can be generalized to more deep learning
problems, such as reinforcement learning.
In summary, we make the following contributions in this
paper.
1) In order to stabilize GAN’s training process, we devised
a simple yet efficient evolutionary algorithm for opti-
mizing generators within GANs framework. To the best
of our knowledge, it is the first work that introduces
the evolutionary learning paradigm into learning GAN
models.
2) Through analyzing the training process of E-GAN,
some properties of existing GANs objectives are further
explored and discussed.
3) Experiments evaluated on several large-scale datasets are
performed, and demonstrate that convincing results can
be achieved by the proposed E-GAN framework.
The rest of this paper is organized as follows. After a brief
summary of the previous related works in Section II, we illus-
trate the proposed E-GAN together with its training process
in Section III. Then, we exhibit the experimental validation
of the whole method in Section IV. Finally, we conclude this
paper with some future directions in Section V.
II. BACKGROUND
In this section, we first review some previous GANs devoted
to reducing training instability and improving the genera-
tive performance. We then briefly summarize some of the
evolutionary algorithms on deep neural networks.
A. Generative Adversarial Networks
GANs provides an excellent framework for learning deep
generative models, which aim to capture the probability dis-
tributions over the given data. Compared to other generative
models, GAN is easily trained by alternately updating a gener-
ator and a discriminator using the back propagation algorithm.
In many generative tasks, GANs (GAN and its variants)
produce better samples than other generative models [32].
Besides image generation tasks, GANs have been introduced
to more and more tasks, such as video generation [11], [33],
visual tracking [34]–[36], domain adaption [37], hashing
coding [38]–[40], and feature learning [41], [42]. In these
tasks, the adversarial training strategy also achieved promising
performances.
However, some problems still exist in the GANs training
process. In the original GAN, training the generator was equal
to minimizing the JSD between the data distribution and the
generated distribution, which easily resulted in the vanishing
gradient problem. To solve this issue, a nonsaturating heuristic
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
WANG et al.: E-GANs 923
objective (i.e., “− log D trick”) replaced the minimax objective
function to penalize the generator [1]. Then, Radford et al. [22]
and Salimans et al. [43] designed specified network (DCGAN)
architectures and proposed several heuristic tricks (e.g., feature
matching, one-side label smoothing, and virtual batch normal-
ization) to improve training stability. Meanwhile, energy-based
GAN [21] and least-squares GAN [20] improved training sta-
bility by employing different training objectives. Although
these methods partly enhanced training stability, in practice,
the network architectures, and training procedure still required
careful design to maintain the discriminator-generator balance.
More recently, Wasserstein GAN (WGAN) [23] and its vari-
ant WGAN-GP (with gradient penalty) [44] were proposed
to minimize the Wasserstein-1 distance between the gener-
ated and data distributions. Since the Wasserstein-1 distance
is continuous everywhere and differentiable almost everywhere
under only minimal assumptions [23], these two methods con-
vincingly reduce training instability. However, to measure the
Wasserstein-1 distance between the generated distribution and
the data distribution, they are asked to enforce the Lipschitz
constraint on the discriminator (also known as critic), which
may result in some optimization problems [44].
Besides devising different objective functions, some works
attempt to stable GAN training and improve generative
performance by introducing multiple generators or discrimina-
tors into an adversarial training framework. Nguyen et al. [45]
proposed a dual discriminator GAN, which combines the KL
and reverse KL divergences into a unified objective func-
tion through employing two discriminators. Multidiscriminator
GAN frameworks [46], [47] are devised and utilized for provid-
ing stable gradients to the generator and further stabilizing the
adversarial training process. Moreover, Tolstikhin et al. [48]
applied boosting techniques to train a mixture of genera-
tors by continually adding new generators to the mixture.
Ghosh et al. [49] trained many generators by using a mul-
ticlass discriminator that predicts which generator produces
the sample. Mixture GAN [50] is proposed to overcome the
mode collapse problem by training multiple generators to spe-
cialize in different data modes. Overall, these multigenerator
GANs aim to learn a set of generators, and the mixture of their
learned distributions would approximate the data distribution,
i.e., different generators are encouraged to capture different data
modes. Although the proposed E-GAN also creates multiple
generators during training, we always keep the well-performing
candidates through survival of the fittest, which helps the final
learned generator achieving better performance. Note that, in
our framework, only one generator was learned to represent
the whole target distribution at the end of the training.
B. Evolutionary Computation
Over the last 20 years, evolutionary algorithms have
achieved considerable success across a wide range of
computational tasks, including modeling, optimization, and
design [51]–[55]. Inspired by natural evolution, the essence
of an evolutionary algorithm is to equate possible solutions to
individuals in a population, produce offspring through varia-
tions, and select appropriate solutions according to fitness [56].
Recently, evolutionary algorithms have been introduced
to solve deep learning problems. To minimize human
participation in designing deep algorithms and automati-
cally discover such configurations, there have been many
attempts to optimize deep learning hyper-parameters and
design deep network architectures through an evolutionary
search [57], [58]. Among them, Real et al. [30] proposed a
large-scale evolutionary algorithm to design a whole deep
classifier automatically. Meanwhile, different from widely
employed gradient-based learning algorithms (e.g., backprop-
agation), evolutionary algorithms have also demonstrated their
capacity to optimize neural networks. EPNet [59] was devised
for evolving and training neural networks using evolution-
ary programming. In [60], EvoAE was proposed to speed
up the training of autoencoders for constructing deep neu-
ral networks. Moreover, Salimans et al. [31] proposed a
novel evolutionary strategy as an alternative to the popu-
lar MDP-based reinforcement learning techniques, achieving
strong performance on reinforcement learning benchmarks. In
addition, an evolutionary algorithm was proposed to compress
deep learning models by automatically eliminating redundant
convolution filters [61]. Last but not the least, the evolution-
ary learning paradigm has been utilized to solve a number of
deep/machine tasks, such as automatic machine learning [62],
multiobjective optimization [63], etc. However, to the best of
our knowledge, there is still no work attempt to optimize deep
generative models with the evolutionary learning algorithms.
III. METHODS
In this section, we first review the original GAN formu-
lation. Then, we introduce the proposed E-GAN algorithm.
By illustrating E-GAN’s mutations and evaluation mechanism,
we further discuss the advantage of the proposed framework.
Finally, we conclude with the entire E-GAN training process.
A. Generative Adversarial Networks
GAN, first proposed in [1], studies a two-player minimax
game between a discriminative network D and a generative
network G. Taking noisy sample z ∼ p(z) (sampled from a
uniform or normal distribution) as the input, the generative
network G outputs new data G(z), whose distribution pg is
supposed to be close to that of the data distribution pdata.
Meanwhile, the discriminative network D is employed to dis-
tinguish the true data sample x ∼ pdata(x) and the generated
sample G(z) ∼ pg(G(z)). In the original GAN, this adversarial
training process was formulated as
min
G
max
D
Ex∼pdata

log D(x)

+ Ez∼pz

log(1 − D(G(z)))

. (1)
The adversarial procedure is illustrated in Fig. 1(a). Most exist-
ing GANs perform a similar adversarial procedure in different
adversarial objective functions.
B. Evolutionary Algorithm
In contrast to conventional GANs, which alternately update
a generator and a discriminator, we devise an evolutionary
algorithm that evolves a population of generator(s) {G} in a
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
924 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019
(a) (b)
Fig. 1. (a) Original GAN framework. A generator G and a discriminator D play a two-player adversarial game. The updating gradients of the generator G
are received from the adaptive objective, which depends on discriminator D. (b) Proposed E-GAN framework. A population of generators {Gθ } evolves in a
dynamic environment, the discriminator D. Each evolutionary step consists of three substages: variation, evaluation, and selection. The best offspring are kept.
given environment (i.e., the discriminator D). In this popu-
lation, each individual represents a possible solution in the
parameter space of the generative network G. During the evo-
lutionary process, we expect that the population gradually
adapts to its environment, which means that the evolved gener-
ator(s) can generate ever more realistic samples and eventually
learn the real-world data distribution. As shown in Fig. 1(b),
during evolution, each step consists of three substages.
1) Variation: Given an individual Gθ in the population,
we utilize the variation operators to produce its off-
spring {Gθ1 , Gθ2 , . . .}. Specifically, several copies of
each individual—or parent—are created, each of which
are modified by different mutations. Then, each modified
copy is regarded as one child.
2) Evaluation: For each child, its performance—or indi-
vidual’s quality—is evaluated by a fitness function F(·)
that depends on the current environment (i.e., discrimi-
nator D).
3) Selection: All children will be selected according to their
fitness value, and the worst part is removed—that is,
they are killed. The rest remain alive (i.e., free to act as
parents) and evolve to the next iteration.
After each evolutionary step, the discriminative network
D (i.e., the environment) is updated to further distinguish
real samples x and fake samples y generated by the evolved
generator(s), i.e.,
LD = −Ex∼pdata

log D(x)

− Ey∼pg

log(1 − D(y))

. (2)
Thus, the discriminative network D (i.e., the environment) can
continually provide the adaptive losses to drive the popula-
tion of generator(s) evolving to produce better solutions. Next,
we illustrate and discuss the proposed variation (or mutation),
evaluation, and selection operators in detail.
C. Variation
We employ asexual reproduction with different mutations
to produce the next generation’s individuals (i.e., children).
Specifically, these mutation operators correspond to differ-
ent training objectives, which attempt to narrow the distances
between the generated distribution and the data distribution
from different perspectives. In this section, we introduce
the mutations used in this paper.1 To analyze the corre-
sponding properties of these mutations, we suppose that, for
each evolutionary step, the optimal discriminator D∗(x) =
([pdata(x)]/[pdata(x) + pg(x)]), according to (2), has already
been learned [1].
1) Minimax Mutation: The minimax mutation corresponds
to the minimax objective function in the original GAN
Mminimax
G =
1
2
Ez∼pz

log(1 − D(G(z))

. (3)
According to the theoretical analysis in [1], given the optimal
discriminator D∗, the minimax mutation aims to minimize the
JSD between the data distribution and the generated distri-
bution. Although the minimax game is easy to explain and
theoretically analyze, its performance in practice is disap-
pointing, a primary problem being the generator’s vanishing
gradient. If the support of two distributions lies in two man-
ifolds, the JSD will be a constant, leading to the vanishing
gradient [24]. This problem is also illustrated in Fig. 2. When
the discriminator rejects generated samples with high confi-
dence [i.e., D(G(z)) → 0], the gradient tends to vanishing.
However, if the generated distribution overlaps with the data
distribution, meaning that the discriminator cannot completely
distinguish real from fake samples, the minimax mutation
provides effective gradients and continually narrows the gap
between the data distribution and the generated distribution.
2) Heuristic Mutation: Unlike the minimax mutation,
which minimizes the log probability of the discriminator being
correct, the heuristic mutation aims to maximize the log
1Although more mutation operations can be included in our framework,
according to the theoretical analysis below, we adopt three interpretable
and complementary objectives as our mutations. Meanwhile, we have tested
more mutation operations, yet the mutations described in this paper already
delivered a convincing performance.
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
WANG et al.: E-GANs 925
Fig. 2. Mutation (or objective) functions that the generator G receives given
the discriminator D.
probability of the discriminator being mistaken, i.e.,
Mheuristic
G = −
1
2
Ez∼pz

log(D(G(z))

. (4)
Compared to the minimax mutation, the heuristic muta-
tion will not saturate when the discriminator rejects the
generated samples. Thus, the heuristic mutation avoids van-
ishing gradient and provides useful generator updates (Fig. 2).
However, according to [24], given the optimal discriminator
D∗, minimizing the heuristic mutation is equal to minimiz-
ing [KL(pg||pdata) − 2JSD(pg||pdata)], i.e., inverted KL minus
two JSDs. Intuitively, the JSD sign is negative, which means
pushing these two distributions away from each other. In prac-
tice, this may lead to training instability and generative quality
fluctuations [44].
3) Least-Squares Mutation: The least-squares mutation is
inspired by LSGAN [20], where the least-squares objec-
tives are utilized to penalize its generator to deceive the
discriminator. In this paper, we formulate the least-squares
mutation as
M
least-square
G = Ez∼pz

(D(G(z)) − 1)2

. (5)
As shown in Fig. 2, the least-squares mutation is nonsaturat-
ing when the discriminator can recognize the generated sample
[i.e., D(G(z)) → 0]. When the discriminator output grows, the
least-squares mutation saturates, eventually approaching zero.
Therefore, similar to the heuristic mutation, the least-squares
mutation can avoid vanishing gradient when the discrimina-
tor has a significant advantage over the generator. Meanwhile,
compared to the heuristic mutation, although the least-squares
mutation will not assign an extremely high cost to generate
fake samples, it will also not assign an extremely low cost to
mode dropping,2 which partly avoids mode collapse [20].
2Arjovsky and Bottou [24] demonstrated that the heuristic objective suffers
from mode collapse since KL(pg||pdata) assigns a high cost to generating fake
samples but an extremely low cost to mode dropping.
Note that, different from GAN-minimax and GAN-heuristic,
LSGAN employs a different objective (“least squares”) to
optimize the discriminator, i.e.,
LLSGAN
D =
1
2
Ex∼pdata

(D(x) − 1)2

+
1
2
Ez∼pz

D(G(z))2

=

x
1
2

pdata(x)(D(x) − 1)2
+ pg(x)D(x)2

dx. (6)
Yet, with respect to D(x), the function LLSGAN
D achieves its
minimum in [0, 1] at ([pdata(x)]/[pdata(x) + pg(x)]), which is
equivalent to ours [i.e., (2)].
Therefore, although we employ only one discriminator as
the environment to distinguish real and generated samples, it is
sufficient to provide adaptive losses for all mutations described
above.
D. Evaluation
In an evolutionary algorithm, evaluation is the operation of
measuring the quality of individuals. To determine the evo-
lutionary direction (i.e., individuals’ selection), we devise an
evaluation (or fitness) function to measure the performance of
evolved individuals (i.e., children). Typically, we mainly focus
on two properties: 1) the quality and 2) the diversity of gener-
ated samples. Quality is measured for each generated sample.
If a generated image could be realistic enough, it will fool
the discriminator. On the other hand, the diversity measures
whether the generator could spread the generated samples out
enough, which could largely avoid mode collapse.
First, we simply feed generator produced images into the
discriminator D and observe the average value of the output,
which we name the quality fitness score
Fq = Ez[D(G(z))]. (7)
Note that discriminator D is constantly upgraded to be optimal
during the training process, reflecting the quality of genera-
tors at each evolutionary (or adversarial) step. If a generator
obtains a relatively high-quality score, its generated samples
can deceive the discriminator and the generated distribution is
further approximate to the data distribution.
Besides generative quality, we also pay attention to the
diversity of generated samples and attempt to overcome
the mode collapse issue in GAN optimization. Recently,
Nagarajan and Kolter [25] proposed a gradient-based regular-
ization term to stabilize the GAN optimization and suppress
mode collapse. When the generator collapses to a small region,
the discriminator will subsequently label collapsed points as
fake with obvious countermeasure (i.e., big gradients). In con-
trast, if the generator is capable of spreading generated data out
enough, the discriminator will not be much confident to label
generated samples as fake data (i.e., updated with small gra-
dients). Other techniques, e.g., exploiting conditioning latent
vector [64] or subspace [65], can also be applied to purse diver-
sity. But this issue does not fall within the scope of this paper.
We employ a similar principle to evaluate generator
optimization stability and generative diversity. Here, since the
gradient-norm of the discriminator could vary largely during
training, we employed an logarithm to shrink its fluctuation.
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
926 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019
Specifically, the minus log-gradient-norm of optimizing D is
utilized to measure the diversity of generated samples. If an
evolved generator obtains a relatively high value, which corre-
sponds to small discriminator gradients, its generated samples
tend to spread out enough, to avoid the discriminator from hav-
ing obvious countermeasures. Thus, the mode collapse issue
can be suppressed and the discriminator will change smoothly,
which helps to improve the training stability. Formally, the
diversity fitness score is defined as
Fd = − log ∇D − Ex

log D(x)

− Ez

log(1 − D(G(z)))

.
(8)
Based on the aforementioned two fitness scores, we can
finally give the evaluation (or fitness) function of the proposed
evolutionary algorithm
F = Fq + γ Fd (9)
where γ ≥ 0 balances two measurements: 1) generative qual-
ity and 2) diversity. Overall, a relatively high fitness score
F, leads to higher training efficiency and better generative
performance.
It is interesting to note that, the proposed fitness score
can also be regarded as an objective function for genera-
tor G. However, as demonstrated in our experiments, without
the proposed evolutionary strategy, this fitness objective is
still hard to achieve a convincing performance. Meanwhile,
similar with most existing objective functions of GANs, the
fitness function will continuously fluctuate during the dynamic
adversarial training process. Specifically, in each iteration, the
fitness score of generators are evaluated based on the current
discriminator. Comparing them, we are capable of selecting
well-performing ones. However, since the discriminator will
be updated according to different generators, it is hard to dis-
cuss the connection between the fitness scores over different
iterations.
E. Selection
In an evolutionary algorithm, the counterpart of the muta-
tion operators is the selection. In the proposed E-GAN, we
employ a simple yet useful survivor selection strategy to deter-
mine the next generation based on the fitness score of existing
individuals.
First of all, we should notice that both the generators (i.e.,
population) and the discriminator (i.e., environment) are opti-
mized alternately in a dynamic procedure. Thus, the fitness
function is not fixed and the fitness score of generators can
only be evaluated by the corresponding discriminator in the
same evolutionary generation, which means fitness scores eval-
uated in different generations cannot compare with each other.
In addition, due to the mutation operators of the proposed
E-GAN actually correspond to different adversarial training
objectives, selecting desired offspring is equivalent to select-
ing the effective adversarial strategies. During the adversarial
process, we hope that generator(s) can do it best to fool the
discriminator (i.e., implement the optimal adversarial strategy).
Considering these two points, we utilize the comma selec-
tion, i.e., (μ, λ)-selection [66] as the selection mechanism of
Algorithm 1 E-GANs. Default Values α = 0.0002, β1 = 0.5,
β2 = 0.99, nD = 3, nm = 3, and m = 32
Require: the batch size m. the discriminator’s updating steps
per iteration nD. the number of parents μ. the number
of mutations nm. Adam hyper-parameters α, β1, β2, the
hyper-parameter γ of evaluation function.
Require: initial discriminator’s parameters w0. initial genera-
tors’ parameters {θ1
0 , θ2
0 , . . . , θ
μ
0 }.
1: for number of training iterations do
2: for k = 0, . . . , nD do
3: Sample a batch of {x(i)}m
i=1 ∼ pdata (training data),
and a batch of {z(i)}m
i=1 ∼ pz (noise samples).
4: gw ← ∇w[ 1
m
m
i=1 log Dw(x(i))
5: + 1
m
μ
j=1
m/μ
i=1 log(1 − Dw(Gθj (z(i))))]
6: w ← Adam(gw, w, α, β1, β2)
7: end for
8: for j = 0, . . . , μ do
9: for h = 0, . . . , nm do
10: Sample a batch of {z(i)}m
i=1 ∼ pz
11: gθj,h ← ∇θj Mh
G({z(i)}m
i=1, θj)
12: θ
j,h
child ← Adam(gθj,h , θj, α, β1, β2)
13: Fj,h ← F
j,h
q + γ F
j,h
d
14: end for
15: end for
16: {Fj1,h1 , Fj2,h2 , . . . } ← sort({Fj,h})
17: θ1, θ2, . . . , θμ ← θ
j1,h1
child , θ
j2,h2
child , . . . , θ
jμ,hμ
child
18: end for
E-GAN. Specifically, after sorting the current offspring popu-
lation {xi}λ
i=1 according to their fitness scores Fi, the μ-best
individuals are selected to form the next generation.
F. E-GAN
Having introduced the proposed evolutionary algorithm and
corresponding mutation operations, evaluation criteria and
selection strategy, the complete E-GAN training process is
concluded in Algorithm 1. Overall, in E-GAN, generators {G}
are regarded as an evolutionary population and discriminator D
acts as an environment. For each evolutionary step, generators
are updated with different mutations (or objectives) to accom-
modate the current environment. According to the principle of
survival of the fittest, only well-performing children will sur-
vive and participate in future adversarial training. Unlike the
two-player game with a fixed and static adversarial training
objective in conventional GANs, E-GAN allows the algorithm
to integrate the merits of different adversarial objectives and
generate the most competitive solution. Thus, during training,
the evolutionary algorithm not only largely suppresses the lim-
itations (vanishing gradient, mode collapse, etc.) of individual
adversarial objectives, but it also harnesses their advantages to
search for a better solution.
IV. EXPERIMENTS
To evaluate the proposed E-GAN, we run experiments on
serval generative tasks and present the experimental results
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
WANG et al.: E-GANs 927
TABLE I
ARCHITECTURES OF THE GENERATIVE AND DISCRIMINATIVE NETWORKS
in this section. Compared with some previous GAN methods,
we show that the proposed E-GAN can achieve impressive
generative performance on large-scale image datasets.
A. Implementation Details
We evaluate E-GAN on two synthetic datasets and three
image datasets: 1) CIFAR-10 [67]; 2) LSUN bedroom [68];
and 3) CelebA [69]. For fair comparisons, we adopted the
same network architectures with existing works [22], [44].
In addition, to achieve better performance on generating
128 × 128 images, we slightly modified both the generative
network and the discriminator network based on the DCGAN
architecture. Specifically, the batch norm layers are removed
from the generator, and more features channels are applied
to each convolutional layers. The detailed networks are listed
in Table I, note that the network architectures of the other
comparison experiments can be easily found in the referenced
works.
We use the default hyper-parameter values listed in
Algorithm 1 for all experiments. Note that the hyper-parameter
γ is utilized to balance the measurements of samples quality
(i.e., Fq) and diversity (i.e., Fd). Usually, the quality fit-
ness score Fq lies in [0, 1], while the diversity fitness score
Fd measures the log-gradient-norm of the discriminator D,
which can vary largely according to D’s scale. Therefore,
we first determine γ ’s range based on the selected discrim-
inator D. Then, we run grid search to find its value. In
practice, we choose γ = 0.5 for the synthetic datasets, and
γ = 0.001 for real-world data. In addition, recently, some
works [44], [70] proposed the gradient penalty (GP) term to
regularize the discriminator to provide precise gradients for
updating the generator. Within the adversarial training frame-
work, our contributions are largely orthogonal to the GP term.
In our experiments, through the setting without GP term, we
demonstrated the efficiency of the proposed method. Then,
after introducing the GP term, the generative performance was
further improved, which demonstrated our framework could
also benefit from the regularization technique for the discrim-
inator. Furthermore, all experiments were trained on Nvidia
Tesla V100 GPUs. To train a model for 64 × 64 images using
the DCGAN architecture cost around 20 h on a single GPU.
B. Evaluation Metrics
Besides directly reported generated samples of the learned
generative networks, we choose the maximum mean discrep-
ancy (MMD) [71], [72], the inception score (IS) [43], and
the Fréchet Inception distance (FID) [73] as quantitative met-
rics. Among them, the MMD can be utilized to measure
the discrepancy between the generated distribution and the
target distribution for synthetic Gaussian mixture datasets.
However, the MMD is difficult to directly apply to high-
dimensional image datasets. Therefore, through applying the
pretrained Inception v3 network [74] to generated images, the
IS computes the KL divergence between the conditional class
distribution and the marginal class distribution. Usually, this
score correlates well with the human scoring of the realism
of generated images from the CIFAR-10 dataset, and a higher
value indicates better image quality. However, some recent
works [6], [75] revealed serious limitations of the IS, e.g., the
target data distribution (i.e., the training data) has not been
considered. In our experiments, we utilize the IS to measure
the E-GAN’s performance on the CIFAR-10, and compare our
results to previous works. Moreover, FID is a more principled
and reliable metric and has demonstrated better correlations
with human evaluation for other datasets. Specifically, FID
calculates the Wasserstein-2 distance between the generated
images and the real-world images in the high-level feature
space of the pretrained Inception v3 network. Note that lower
FID means closer distances between the generated distribu-
tion and the real-world data distribution. In all experiments,
we randomly generated 50k samples to calculate the MMD,
IS, and FID.
C. Synthetic Datasets and Mode Collapse
In the first experiment, we adopt the experimental design
proposed in [76], which trains GANs on 2-D Gaussian mix-
ture distributions. The mode collapse issue can be accurately
measured on these synthetic datasets, since we can clearly
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
928 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019
Fig. 3. KDE plots of the target data and generated data from different GANs trained on mixtures of Gaussians. In the first row, a mixture of 8 Gaussians
arranged in a circle. In the second row, a mixture of 25 Gaussians arranged in a grid.
TABLE II
MMD (×10−2) WITH MIXED GAUSSIAN DISTRIBUTIONS ON OUR TOY
DATASETS. WE RAN EACH METHOD FOR TEN TIMES, AND REPORT
THEIR AVERAGE AND BEST RESULTS. THE METHOD WITH LOWER
MMD VALUE IMPLIES THE GENERATED DISTRIBUTION IS
CLOSER TO THE TARGET ONE
observe and measure the generated distribution and the target
data distribution. As shown in Fig. 3, we employ two challeng-
ing distributions to evaluate E-GAN, a mixture of 8 Gaussians
arranged in a circle and a mixture of 25 Gaussians arranged
in a grid.3 Here, to evaluate if the proposed diversity fitness
score can reduce the mode collapse, we did not introduce the
GP norm and set the survived parents number μ as 1, i.e.,
during each evolutionary step, only the best candidature are
kept.
First, we utilize existing individual adversarial objectives
(i.e., conventional GANs) to perform the adversarial training
process. We train each method 50K iterations and report the
Kernel density estimation (KDE) plots in Fig. 3. Meanwhile,
the average and the best MMD values which running each
method ten times are reported in Table II. The results show that
all of the individual adversarial objectives suffer from mode
collapse to a greater or lesser degree. Then, we set hyper-
parameter γ as 0 and test the proposed E-GAN, which means
the diversity fitness score was not considered during the train-
ing process. The results show that the evolutionary framework
still has problems with the mode collapse. However, when the
3We obtain both 2-D distributions and network architectures from the code
provided in [44].
TABLE III
INCEPTION SCORES AND FIDS WITH UNSUPERVISED IMAGE
GENERATION ON CIFAR-10. THE METHOD WITH HIGHER IS OR
LOWER FID IMPLIES THE GENERATED DISTRIBUTION
IS CLOSER TO THE TARGET ONE. † [22], ‡ [6]
diversity fitness score is considered in the selection stage (in
this experiment, we set γ = 0.5), the mode collapse issue is
largely suppressed and the trained generator can more accu-
rately fit the target distributions. This demonstrates that our
diversity fitness score has the capability of measuring the sam-
ple diversity of updated generators and further suppress the
mode collapse problem.
D. CIFAR-10 and Training Stability
In the proposed E-GAN, we utilize the evolutionary algo-
rithm with different mutations (i.e., different updating strate-
gies) to optimize generator(s) {G}. To demonstrate the advan-
tages of the proposed evolutionary algorithm over existing
two-player adversarial training strategies (i.e., updating gen-
erator with a single objective), we train these methods on
CIFAR-10 and plot inception scores [43] over the training
process. For a fair comparison, we did not introduce the
GP norm into our E-GAN and set the parents number μ
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
WANG et al.: E-GANs 929
Fig. 4. Experiments on the CIFAR-10 dataset. CIFAR-10 IS over generator iterations (left), over wall-clock time (middle), and the graph of selected mutations
in the E-GAN training process (right).
TABLE IV
FIDS WITH UNSUPERVISED IMAGE GENERATION ON LSUN BEDROOM DATASET. THE METHOD WITH LOWER FID IMPLIES
THE GENERATED DISTRIBUTION IS CLOSER TO THE TARGET ONE
as 1. Moreover, the same network architecture is used for
all methods.
As shown in Fig. 4(left), E-GAN can get higher IS with
less training steps. Meanwhile, E-GAN also shows comparable
stability when it goes to convergence. By comparison, conven-
tional GAN objectives expose their different limitations, such
as instability at convergence (GAN-Heuristic), slow conver-
gence (GAN-Least square), and invalid (GAN-minimax). In
addition, we employ the proposed fitness function [i.e., (9)]
as generator’s objective function, and find its performance
is also inferior to E-GAN. This experiment further demon-
strates the advantages of the proposed evolutionary framework.
Through creating and selecting from multiple candidates, the
evolutionary framework can leverage strengths of different
objective functions (i.e., different distances) to accelerate the
training process and improve the generative performance.
Based on the evolutionary framework, E-GAN not only over-
come the inherent limitations of these individual adversarial
objectives, but it also outperforms other GANs (the WGAN
and its improved variant WGAN-GP). Furthermore, when we
only keep one parent during each evolutionary step, E-GAN
achieves comparable convergence speed in terms of wall-
clock time [Fig. 4(middle)]. During training E-GAN, we
recorded the selected objective in each step [Fig. 4(right)].
At the beginning of training, the heuristic objective and the
least-square objective are selected more frequently than the
minimax objective. It may be due to the fact that the minimax
objective is hard to provide effective gradients (i.e., vanish-
ing gradient) when the discriminator can easily recognize
generated samples. Along with the generator approaching con-
vergence (after 20K steps), ever more minimax objectives are
employed, yet the number of selected heuristic objectives is
falling. As aforementioned, the minus JSDs of the heuristic
objective may tend to push the generated distribution away
from target data distribution and lead to training instability.
However, in E-GAN, beyond the heuristic objective, we have
other options of objective, which improves the stability at
convergence.
Furthermore, we discussed the relationship between the sur-
vived parents’ number and generative performance. As shown
in Table III, both the IS and FID are utilized to evaluate
the generative performance of learned generators. First, com-
pared with the basic E-GAN (i.e., E-GAN, μ = 1, without
GP), adding the GP norm to optimize the discriminator indeed
improves generative performance. Then, we preserved multiple
parents during the E-GAN training process and measured their
scores at the end of training. We can easily observe that
the generative performance becomes better accompanied with
keeping more parents during the training. This further demon-
strates the proposed evolutionary learning paradigm could
suppress the unstable and large-scale optimization problems
of GANs.
Theoretically, if we regard updating and evaluating a
child G as an operation and define mutations number as
n, keeping p parents will cost O(np) operations in each
iteration. Comparing with traditional GANs, our evolution-
ary framework would cost more time in each iteration.
However, since the evolutionary strategy always preserves
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
930 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019
Fig. 5. Experiments of architecture robustness. Different GAN architectures, which correspond to different training challenges, trained with six different
GAN methods (or settings). The proposed E-GAN achieved promising performance under all architecture settings.
Fig. 6. Generated bedroom images on 128 × 128 LSUN bedrooms.
the well-performing off-spring, to achieve the same gen-
erative performance, E-GAN usually spends less training
steps [Fig. 4(left)]. Overall, keeping one parent during each
Fig. 7. Generated human face images on 128 × 128 CelebA dataset.
evolutionary step will only slightly reduce the time-efficiency
but with better performance [Fig. 4(middle)]. Yet, accompa-
nied by increasing p, although the generative performance can
be further improved, it will also cost more time on training
the E-GAN model. Here, if we regard the parents number
p as a hyper-parameter of our algorithm, we found setting
its value less than or equal to 4 is a preferable choice.
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
WANG et al.: E-GANs 931
Fig. 8. Interpolating in latent space. For selected pairs of generated images from a well-trained E-GAN model, we record their latent vectors z1 and z2.
Then, samples between them are generated by linear interpolation between these two vectors.
Within this interval, we can easily improve the generative
performance by sacrificing affordable computation cost. If we
continually increase the number of survivors, the generative
performance can only be improved mildly yet largely reduce
the training efficiency. In practice, we need to further bal-
ance the algorithm efficiency and performance according to
different situations.
E. LSUN and Architecture Robustness
The architecture robustness is another advantage of E-GAN.
To demonstrate the training stability of our method, we
train different network architectures on the LSUN bedroom
dataset [68] and compare with several existing works. In
addition to the baseline DCGAN architecture, we choose
three additional architectures corresponding to different train-
ing challenges: 1) limiting the recognition capability of the
discriminator D, i.e., 2-Conv-1-FC LeakyReLU discriminator
(abbreviated as weak D); 2) limiting the expression capability
of the generator G, i.e., no batchnorm and a constant num-
ber of filters in the generator (weak G); and 3) reducing the
network capability of the generator and discriminator together,
i.e., remove the BN in both the generator G and discrimina-
tor D (weak both). For each architecture, we test six different
methods (or settings): 1) DCGAN; 2) LSGAN; 3) original
WGAN (with weight clipping); 4) WGAN-GP; 5) our E-GAN
(μ = 1); and 6) E-GAN (μ = 4). For each method, we
used the default configurations recommended in the respec-
tive studies (these methods are summarized in [44]) and train
each model for 100K iterations. Some generated samples are
reported in Fig. 5, and the quantitative results (i.e., FID) are
listed in Table IV. Through observation, we find that all of
these GAN methods achieved promising performance with the
baseline architecture. For DCGAN and LSGAN, when the
balance between the generator and discriminator is broken
(i.e., only one of them is limited), these two methods have
difficulty generating any reasonable samples. Meanwhile, we
find the performance of the standard WGAN (with weight
clipping) is mostly decided by the generator G. When we
limit G’s capability, the generative performance is largely
reduced. Regarding the WGAN-GP, we find that the gener-
ative performance may mainly depend on the discriminator
(or critic). Our E-GAN achieved promising results under all
architecture settings. Moreover, we again demonstrated that the
model performance is growing with the number of survived
parents.
Furthermore, we trained E-GAN to generate higher resolu-
tion (128×128) bedroom images (Fig. 6). Observing generated
images, we demonstrate that E-GAN can be trained to gen-
erate diversity and high-quality images from the target data
distribution.
F. CelebA and Space Continuity
Besides the LSUN bedroom dataset, we also train our
E-GAN using the aligned faces from CelebA dataset.
Since humans excel at identifying facial flaws, generating
high-quality human face images is challenging. Similar to
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
932 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019
generating bedrooms, we employ the same architectures to
generate 128 × 128 RGB human face images (Fig. 7). In
addition, given a well-trained generator, we evaluate the
performance of the embedding in the latent space of noisy
vectors z. In Fig. 8, we first select pairs of generated faces
and record their corresponding latent vectors z1 and z2. The
two images in one pair have different attributes, such as gen-
der, expression, hairstyle, and age. Then, we generate novel
samples by linear interpolating between these pairs (i.e., corre-
sponding noisy vectors). We find that these generated samples
can seamlessly change between these semantically meaningful
face attributes. This experiment demonstrates that genera-
tor training does not merely memorize training samples but
learns a meaningful projection from latent noisy space to face
images. Meanwhile, it also shows that the generator trained by
E-GAN does not suffer from mode collapse, and shows great
space continuity. Overall, during the GAN training process, the
training stability is easily influenced by “bad” updating, which
could lead the generated samples to low quality or lacking
diversity, while the proposed evolutionary mechanism largely
avoids undesired updating and promotes the training to an ideal
direction.
V. CONCLUSION
In this paper, we presented an E-GAN framework for train-
ing deep generative models. To reduce training difficulties and
improve generative performance, we devised an evolutionary
algorithm to evolve a population of generators to adapt to the
dynamic environment (i.e., the discriminator D). In contrast
to conventional GANs, the evolutionary paradigm allows the
proposed E-GAN to overcome the limitations of individual
adversarial objectives and preserve the well-performing off-
spring after each iteration. Experiments showed that E-GAN
improves the training stability of GAN models and achieves
convincing performance in several image generation tasks.
In this paper, we mainly contribute to improving the image
generation performance. More generation tasks will be con-
sidered in future works, such as video generation [33] and
text generation [77].
REFERENCES
[1] I. Goodfellow et al., “Generative adversarial nets,” in Proc. Adv. Neural
Inf. Process. Syst. (NIPS), 2014, pp. 2672–2680.
[2] X. Chen et al., “InfoGAN: Interpretable representation learning by
information maximizing generative adversarial nets,” in Proc. Adv.
Neural Inf. Process. Syst. (NIPS), 2016, pp. 2172–2180.
[3] Z. Gan et al., “Triangle generative adversarial networks,” in Proc. Adv.
Neural Inf. Process. Syst. (NIPS), 2017, pp. 5253–5262.
[4] H. Zhang et al., “StackGAN: Text to photo-realistic image synthesis
with stacked generative adversarial networks,” in Proc. IEEE Int. Conf.
Comput. Vis. (ICCV), 2017, pp. 5908–5916.
[5] T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive growing of
GANs for improved quality, stability, and variation,” in Proc. Int. Conf.
Learn. Represent. (ICLR), 2018, pp. 1–26.
[6] T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida, “Spectral normal-
ization for generative adversarial networks,” in Proc. Int. Conf. Learn.
Represent. (ICLR), 2018, pp. 1–26.
[7] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation
with conditional adversarial networks,” in Proc. IEEE Conf. Comput. Vis.
Pattern Recognit. (CVPR), 2017, pp. 5967–5976.
[8] C. Wang, C. Wang, C. Xu, and D. Tao, “Tag disentangled generative
adversarial networks for object image re-rendering,” in Proc. 26th Int.
Joint Conf. Artif. Intell. (IJCAI), 2017, pp. 2901–2907.
[9] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image
translation using cycle-consistent adversarial networks,” in Proc. IEEE
Int. Conf. Comput. Vis. (ICCV), 2017, pp. 2242–2251.
[10] T.-C. Wang et al., “High-resolution image synthesis and semantic manip-
ulation with conditional GANs,” in Proc. IEEE Conf. Comput. Vis.
Pattern Recognit. (CVPR), 2018, pp. 8798–8807.
[11] C. Vondrick, H. Pirsiavash, and A. Torralba, “Generating videos with
scene dynamics,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2016,
pp. 613–621.
[12] C. Finn, I. Goodfellow, and S. Levine, “Unsupervised learning for phys-
ical interaction through video prediction,” in Proc. Adv. Neural Inf.
Process. Syst. (NIPS), 2016, pp. 64–72.
[13] C. Vondrick and A. Torralba, “Generating the future with adversar-
ial transformers,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.
(CVPR), Honolulu, HI, USA, 2017, pp. 2992–3000.
[14] Y. Zhang, Z. Gan, and L. Carin, “Generating text via adversarial
training,” in Proc. NIPS Workshop Adversarial Training, 2016, pp. 1–6.
[15] W. Fedus, I. Goodfellow, and A. M. Dai, “MaskGAN: Better text gener-
ation via filling in the___,” in Proc. Int. Conf. Learn. Represent. (ICLR),
2018, pp. 1–17.
[16] J. Lu, A. Kannan, J. Yang, D. Parikh, and D. Batra, “Best of both
worlds: Transferring knowledge from discriminative learning to a gen-
erative visual dialog model,” in Proc. Adv. Neural Inf. Process. Syst.
(NIPS), 2017, pp. 313–323.
[17] X. Lan, A. J. Ma, and P. C. Yuen, “Multi-cue visual tracking using
robust feature-level fusion based on joint sparse representation,” in
Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2014,
pp. 1194–1201.
[18] X. Chen, C. Xu, X. Yang, L. Song, and D. Tao, “Gated-
GAN: Adversarial gated networks for multi-collection style trans-
fer,” IEEE Trans. Image Process., vol. 28, no. 2, pp. 546–560,
Feb. 2019.
[19] S. Arora, R. Ge, Y. Liang, T. Ma, and Y. Zhang, “Generalization and
equilibrium in generative adversarial nets (GANs),” in Proc. 34th Int.
Conf. Mach. Learn. (ICML), 2017, pp. 224–232.
[20] X. Mao et al., “Least squares generative adversarial networks,” in Proc.
IEEE Int. Conf. Comput. Vis. (ICCV), 2017, pp. 2813–2821.
[21] J. Zhao, M. Mathieu, and Y. LeCun, “Energy-based generative
adversarial network,” in Proc. Int. Conf. Learn. Represent. (ICLR),
2017, pp. 1–17.
[22] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation
learning with deep convolutional generative adversarial networks,” in
Proc. Int. Conf. Learn. Represent. (ICLR), 2016, pp. 1–16.
[23] M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein generative adver-
sarial networks,” in Proc. 34th Int. Conf. Mach. Learn. (ICML), 2017,
pp. 214–223.
[24] M. Arjovsky and L. Bottou, “Towards principled methods for training
generative adversarial networks,” in Proc. Int. Conf. Learn. Represent.
(ICLR), 2017, pp. 1–17.
[25] V. Nagarajan and J. Z. Kolter, “Gradient descent GAN optimization is
locally stable,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017,
pp. 5591–5600.
[26] C. Qian, J.-C. Shi, K. Tang, and Z.-H. Zhou, “Constrained monotone
k-submodular function maximization using multiobjective evolutionary
algorithms with theoretical guarantee,” IEEE Trans. Evol. Comput.,
vol. 22, no. 4, pp. 595–608, Aug. 2018.
[27] S. He et al., “Cooperative co-evolutionary module identification with
application to cancer disease module discovery,” IEEE Trans. Evol.
Comput., vol. 20, no. 6, pp. 874–891, Dec. 2016.
[28] Y. Du, M.-H. Hsieh, T. Liu, and D. Tao. (2018). The Expressive
Power of Parameterized Quantum Circuits. [Online]. Available:
https://arxiv.org/abs/1810.11922
[29] L. M. Antonio and C. A. Coello Coello, “Coevolutionary multiobjective
evolutionary algorithms: A survey of the state-of-the-art,” IEEE Trans.
Evol. Comput., vol. 22, no. 6, pp. 851–865, Dec. 2018.
[30] E. Real et al., “Large-scale evolution of image classifiers,” in Proc. 34th
Int. Conf. Mach. Learn. (ICML), 2017, pp. 214–223.
[31] T. Salimans, J. Ho, X. Chen, and I. Sutskever. (2017). Evolution
Strategies as a Scalable Alternative to Reinforcement Learning. [Online].
Available: https://arxiv.org/abs/1703.03864
[32] I. Goodfellow. (2016). NIPS 2016 Tutorial: Generative Adversarial
Networks. [Online]. Available: https://arxiv.org/abs/1701.00160
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
WANG et al.: E-GANs 933
[33] S. Tulyakov, M.-Y. Liu, X. Yang, and J. Kautz, “MoCoGAN:
Decomposing motion and content for video generation,” in Proc.
IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2018,
pp. 1526–1535.
[34] Y. Song et al., “VITAL: Visual tracking via adversarial learning,” in
Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2018,
pp. 8990–8999.
[35] X. Lan, S. Zhang, P. C. Yuen, and R. Chellappa, “Learning common and
feature-specific patterns: A novel multiple-sparse-representation-based
tracker,” IEEE Trans. Image Process., vol. 27, no. 4, pp. 2022–2037,
Apr. 2018.
[36] X. Lan, A. J. Ma, P. C. Yuen, and R. Chellappa, “Joint sparse
representation and robust feature-level fusion for multi-cue visual track-
ing,” IEEE Trans. Image Process., vol. 24, no. 12, pp. 5826–5841,
Dec. 2015.
[37] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial dis-
criminative domain adaptation,” in Proc. Comput. Vis. Pattern Recognit.
(CVPR), vol. 1, 2017, p. 4.
[38] Z. Qiu, Y. Pan, T. Yao, and T. Mei, “Deep semantic hashing with gen-
erative adversarial networks,” in Proc. 40th Int. ACM SIGIR Conf. Res.
Dev. Inf. Retrieval, Tokyo, Japan, 2017, pp. 225–234.
[39] E. Yang, T. Liu, C. Deng, and D. Tao, “Adversarial examples
for hamming space search,” IEEE Trans. Cybern., to be published.
doi: 10.1109/TCYB.2018.2882908.
[40] E. Yang et al., “Pairwise relationship guided deep hashing for cross-
modal retrieval,” in Proc. AAAI, 2017, pp. 1618–1625.
[41] J. Donahue, P. Krähenbühl, and T. Darrell, “Adversarial feature learn-
ing,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–18.
[42] V. Dumoulin et al., “Adversarially learned inference,” in Proc. Int. Conf.
Learn. Represent. (ICLR), 2017, pp. 1–18.
[43] T. Salimans et al., “Improved techniques for training GANs,” in Proc.
Adv. Neural Inf. Process. Syst. (NIPS), 2016, pp. 2234–2242.
[44] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. Courville,
“Improved training of Wasserstein GANs,” in Proc. Adv. Neural Inf.
Process. Syst. (NIPS), 2017, pp. 5769–5779.
[45] T. D. Nguyen, T. Le, H. Vu, and D. Phung, “Dual discriminator gener-
ative adversarial nets,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS),
2017, pp. 2667–2677.
[46] I. Durugkar, I. Gemp, and S. Mahadevan, “Generative multi-adversarial
networks,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017,
pp. 1–14.
[47] B. Neyshabur, S. Bhojanapalli, and A. Chakrabarti. (2017). Stabilizing
GAN Training With Multiple Random Projections. [Online]. Available:
https://arxiv.org/abs/1705.07831
[48] I. O. Tolstikhin, S. Gelly, O. Bousquet, C.-J. Simon-Gabriel, and
B. Schölkopf, “AdaGAN: Boosting generative models,” in Proc. Adv.
Neural Inf. Process. Syst. (NIPS), 2017, pp. 5424–5433.
[49] A. Ghosh, V. Kulharia, V. P. Namboodiri, P. H. S. Torr, and
P. K. Dokania, “Multi-agent diverse generative adversarial networks,”
in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2018,
pp. 8513–8521.
[50] Q. Hoang, T. D. Nguyen, T. Le, and D. Phung, “MGAN: Training gen-
erative adversarial nets with multiple generators,” in Proc. Int. Conf.
Learn. Represent. (ICLR), 2018, pp. 1–24.
[51] K. A. De Jong, Evolutionary Computation: A Unified Approach.
Cambridge, MA, USA: MIT Press, 2006.
[52] H.-L. Liu, L. Chen, Q. Zhang, and K. Deb, “Adaptively allocat-
ing search effort in challenging many-objective optimization prob-
lems,” IEEE Trans. Evol. Comput., vol. 22, no. 3, pp. 433–448,
Jun. 2018.
[53] Y. Wang, M. Zhou, X. Song, M. Gu, and J. Sun, “Constructing
cost-aware functional test-suites using nested differential evolution algo-
rithm,” IEEE Trans. Evol. Comput., vol. 22, no. 3, pp. 334–346,
Jun. 2018.
[54] P. Yang, K. Tang, and X. Yao, “Turning high-dimensional optimization
into computationally expensive optimization,” IEEE Trans. Evol.
Comput., vol. 22, no. 1, pp. 143–156, Feb. 2018.
[55] D.-C. Dang et al., “Escaping local optima using crossover with emer-
gent diversity,” IEEE Trans. Evol. Comput., vol. 22, no. 3, pp. 484–497,
Jun. 2018.
[56] A. E. Eiben and J. Smith, “From evolutionary computation to
the evolution of things,” Nature, vol. 521, no. 7553, p. 476,
May 2015.
[57] R. Miikkulainen et al. (2017). Evolving Deep Neural Networks. [Online].
Available: https://arxiv.org/abs/1703.00548
[58] S. R. Young, D. C. Rose, T. P. Karnowski, S.-H. Lim, and R. M. Patton,
“Optimizing deep learning hyper-parameters through an evolutionary
algorithm,” in Proc. Workshop Mach. Learn. High Perform. Comput.
Environ., 2015, p. 4.
[59] X. Yao, “Evolving artificial neural networks,” Proc. IEEE, vol. 87, no. 9,
pp. 1423–1447, Sep. 1999.
[60] S. Lander and Y. Shang, “EvoAE—A new evolutionary method for
training autoencoders for deep learning networks,” in Proc. IEEE
39th Annu. Comput. Softw. Appl. Conf. (COMPSAC), vol. 2, 2015,
pp. 790–795.
[61] Y. Wang, C. Xu, J. Qiu, C. Xu, and D. Tao, “Towards evolutionary
compression,” in Proc. 24th Int. Conf. Knowl. Disc. Data Min., 2018,
pp. 2476–2485.
[62] R. S. Olson, N. Bartley, R. J. Urbanowicz, and J. H. Moore, “Evaluation
of a tree-based pipeline optimization tool for automating data science,”
in Proc. Genet. Evol. Comput. Conf., 2016, pp. 485–492.
[63] A. Rosales-Pérez, S. García, J. A. Gonzalez, C. A. Coello Coello, and
F. Herrera, “An evolutionary multiobjective model and instance selection
for support vector machines with Pareto-based ensembles,” IEEE Trans.
Evol. Comput., vol. 21, no. 6, pp. 863–877, Dec. 2017.
[64] M. Mirza and S. Osindero. (2014). Conditional Generative Adversarial
Nets. [Online]. Available: https://arxiv.org/abs/1411.1784
[65] J. Liang, J. Yang, H.-Y. Lee, K. Wang, and M.-H. Yang, “Sub-GAN:
An unsupervised generative model via subspaces,” in Proc. Eur. Conf.
Comput. Vis. (ECCV), 2018, pp. 698–714.
[66] O. Kramer, Machine Learning for Evolution Strategies, vol. 20. Cham,
Switzerland: Springer, 2016.
[67] A. Krizhevsky, “Learning multiple layers of features from tiny images,”
Dept. Comput. Sci., Univ. Toronto, Toronto, ON, Canada, Rep. TR-2009,
2009.
[68] F. Yu et al. (2015). LSUN: Construction of a Large-Scale Image Dataset
Using Deep Learning With Humans in the Loop. [Online]. Available:
https://arxiv.org/abs/1506.03365
[69] Z. Liu, P. Luo, X. Wang, and X. Tang, “Deep learning face attributes
in the wild,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2015,
pp. 3730–3738.
[70] W. Fedus et al., “Many paths to equilibrium: GANs do not need to
decrease adivergence at every step,” in Proc. Int. Conf. Learn. Represent.
(ICLR), 2018, pp. 1–21.
[71] A. Smola, A. Gretton, L. Song, and B. Schölkopf, “A Hilbert space
embedding for distributions,” in Proc. Int. Conf. Algorithmic Learn.
Theory, 2007, pp. 13–31.
[72] A. Gretton, K. M. Borgwardt, M. J. Rasch, B. Schölkopf, and A. Smola,
“A kernel two-sample test,” J. Mach. Learn. Res., vol. 13, pp. 723–773,
Mar. 2012.
[73] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter,
“GANs trained by a two time-scale update rule converge to a local
Nash equilibrium,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017,
pp. 6629–6640.
[74] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking
the inception architecture for computer vision,” in Proc. IEEE Conf.
Comput. Vis. Pattern Recognit., 2016, pp. 2818–2826.
[75] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena. (2018).
Self-Attention Generative Adversarial Networks. [Online]. Available:
https://arxiv.org/abs/1805.08318
[76] L. Metz, B. Poole, D. Pfau, and J. Sohl-Dickstein, “Unrolled genera-
tive adversarial networks,” in Proc. Int. Conf. Learn. Represent. (ICLR),
2017, pp. 1–25.
[77] N. Duan, D. Tang, P. Chen, and M. Zhou, “Question generation for
question answering,” in Proc. Conf. Empir. Methods Nat. Lang. Process.
(EMNLP), 2017, pp. 866–874.
Chaoyue Wang received the bachelor’s degree from
Tianjin University, Tianjin, China, and the Ph.D.
degree from the University of Technology Sydney,
Ultimo, NSW, Australia.
He is a Research Associate of machine learning
and computer vision with the School of Computer
Science, University of Sydney, Darlington, NSW,
Australia. His current research interests include
machine learning, deep learning, and generative
models.
Dr. Wang was a recipient of the Distinguished
Student Paper Award in the International Joint Conference on Artificial
Intelligence in 2017.
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
934 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019
Chang Xu received the Bachelor of Engineering
degree from Tianjin University, Tianjin, China, and
the Ph.D. degree from Peking University, Beijing,
China.
He is a Lecturer of machine learning and com-
puter vision with the School of Computer Science,
University of Sydney, Darlington, NSW, Australia.
He received fellowships from IBM, Armonk, NY,
USA, and Baidu, Beijing. His current research
interests include machine learning, data mining algo-
rithms, and related applications in artificial intelli-
gence and computer vision, including multiview learning, multilabel learning,
visual search, and face recognition. He has published in prestigious journals
and top-tier conferences in the above areas.
Xin Yao (F’03) received the B.Sc. degree from
the University of Science and Technology of China
(USTC), Hefei, China, in 1982, the M.Sc. degree
from the North China Institute of Computing
Technologies, Beijing, China, in 1985, and the Ph.D.
degree from USTC in 1990.
He is a Chair Professor of computer science with
the Southern University of Science and Technology,
Shenzhen, China, and a part-time Professor of com-
puter science with the University of Birmingham,
Birmingham, U.K. His current research interests
include evolutionary computation, ensemble learning, and their applications
to software engineering.
Dr. Yao was a recipient of the prestigious Royal Society Wolfson
Research Merit Award in 2012, the IEEE Computational Intelligence Society
(CIS) Evolutionary Computation Pioneer Award in 2013, the 2001 IEEE
Donald G. Fink Prize Paper Award for his paper on evolving artificial
neural networks, the 2010, 2016, and 2017 IEEE TRANSACTIONS ON
EVOLUTIONARY COMPUTATION Outstanding Paper Awards, the 2011 IEEE
TRANSACTIONS ON NEURAL NETWORKS Outstanding Paper Award, and
many other best paper awards. He was the President of IEEE CIS from 2014 to
2015 and the Editor-in-Chief of IEEE TRANSACTIONS ON EVOLUTIONARY
COMPUTATION from 2003 to 2008. He was a Distinguished Lecturer of
IEEE CIS.
Dacheng Tao (F’15) received the B.Eng. degree
from the University of Science and Technology of
China, Hefei, China, the M.Phil. degree from the
Chinese University of Hong Kong, Hong Kong,
and the Ph.D. degree from Birkbeck, University of
London, London, U.K.
He is a Professor of computer science and
an ARC Laureate Fellow with the School of
Computer Science and the Faculty of Engineering
and Information Technologies, and the Inaugural
Director of the UBTECH Sydney Artificial
Intelligence Centre, University of Sydney, Darlington, NSW, Australia. He
mainly applies statistics and mathematics to artificial intelligence and data
science. His research results have expounded in one monograph and over
200 publications in prestigious journals and prominent conferences, such
as the IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE
INTELLIGENCE, the IEEE TRANSACTIONS ON IMAGE PROCESSING, the
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,
the International Journal of Computer Vision, the Journal of Machine
Learning Research, Neural Information Processing Systems, International
Conference on Machine Learning, IEEE Conference on Computer Vision and
Pattern Recognition, IEEE International Conference on Computer Vision,
European Conference on Computer Vision, IEEE International Conference
on Data Mining, Association for Computing Machinery’s Special Interest
Group on Knowledge Discovery and Data Mining.
Mr. Tao was a recipient of several best paper awards, such as the Best
Theory/Algorithm Paper Runner Up Award in IEEE ICDM’07, the Best
Student Paper Award in IEEE ICDM’13, the Distinguished Paper Award
in the 2018 IJCAI, the 2014 ICDM 10-Year Highest-Impact Paper Award,
and the 2017 IEEE Signal Processing Society Best Paper Award, the
2015 Australian Scopus-Eureka Prize, and the 2018 IEEE ICDM Research
Contributions Award. He is a fellow of the Australian Academy of Science,
American Association for the Advancement of Science, International
Association for Pattern Recognition, Optical Society of America, Society of
Photo-Optical Instrumentation Engineers.
Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.

More Related Content

PDF
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Codiax
 
PDF
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Universitat Politècnica de Catalunya
 
PDF
gans_copy.pdfhjsjsisidkskskkskwkduydjekedj
fahid32446
 
PDF
Introduction to Generative Adversarial Network
vaidehimadaan041
 
PDF
Exploring The Potential of Generative Adversarial Network: A Comparative Stud...
IRJET Journal
 
PPTX
ICASSP 2018 Tutorial: Generative Adversarial Network and its Applications to ...
宏毅 李
 
PPTX
GAN Deep Learning Approaches to Image Processing Applications (1).pptx
RMDAcademicCoordinat
 
PPTX
GAN Generative Adversarial Networks.pptx
ssuser2624f71
 
Jakub Langr (University of Oxford) - Overview of Generative Adversarial Netwo...
Codiax
 
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Universitat Politècnica de Catalunya
 
gans_copy.pdfhjsjsisidkskskkskwkduydjekedj
fahid32446
 
Introduction to Generative Adversarial Network
vaidehimadaan041
 
Exploring The Potential of Generative Adversarial Network: A Comparative Stud...
IRJET Journal
 
ICASSP 2018 Tutorial: Generative Adversarial Network and its Applications to ...
宏毅 李
 
GAN Deep Learning Approaches to Image Processing Applications (1).pptx
RMDAcademicCoordinat
 
GAN Generative Adversarial Networks.pptx
ssuser2624f71
 

Similar to 6666666666666666666666666666666666666.pdf (20)

PDF
Generative Adversarial Networks GAN - Santiago Pascual - UPC Barcelona 2018
Universitat Politècnica de Catalunya
 
PPTX
Gans - Generative Adversarial Nets
SajalRastogi8
 
PDF
gan.pdf
Dr.rukmani Devi
 
PDF
A Short Introduction to Generative Adversarial Networks
Jong Wook Kim
 
PDF
Deep Generative Models II (DLAI D10L1 2017 UPC Deep Learning for Artificial I...
Universitat Politècnica de Catalunya
 
PDF
Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018
Universitat Politècnica de Catalunya
 
PPTX
GDC2019 - SEED - Towards Deep Generative Models in Game Development
Electronic Arts / DICE
 
PDF
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
Universitat Politècnica de Catalunya
 
PDF
11_gan.pdf
Ankush84837
 
PDF
Generative Adversarial Networks 2
Alireza Shafaei
 
PDF
Generative Adversarial Networks and Their Applications
Artifacia
 
PPT
GNA 13552928 deep learning for GAN a.ppt
ManiMaran230751
 
PDF
Generative Adversarial Networks (GANs) at the Data Science Meetup Luxembourg ...
Chris Hammerschmidt
 
PDF
مدل آموزش داده مصنوعی مبتنی بر شبکه GAN برای شبکه های عصبی CNN سبک
javascriptsali
 
PDF
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Universitat Politècnica de Catalunya
 
PDF
Generative Adversarial Networks
Mustafa Yagmur
 
PPTX
GANs Presentation.pptx
MAHMOUD729246
 
PPTX
Generative adversarial networks
Ding Li
 
PDF
What is a GAN Generative Adversarial Networks Guide.pdf
imoliviabennett
 
PDF
PROGRESSIVE GROWING OF GAN S FOR I MPROVED QUALITY , STABILITY , AND VARIATION
Willy Marroquin (WillyDevNET)
 
Generative Adversarial Networks GAN - Santiago Pascual - UPC Barcelona 2018
Universitat Politècnica de Catalunya
 
Gans - Generative Adversarial Nets
SajalRastogi8
 
A Short Introduction to Generative Adversarial Networks
Jong Wook Kim
 
Deep Generative Models II (DLAI D10L1 2017 UPC Deep Learning for Artificial I...
Universitat Politècnica de Catalunya
 
Deep Generative Models - Kevin McGuinness - UPC Barcelona 2018
Universitat Politècnica de Catalunya
 
GDC2019 - SEED - Towards Deep Generative Models in Game Development
Electronic Arts / DICE
 
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
Universitat Politècnica de Catalunya
 
11_gan.pdf
Ankush84837
 
Generative Adversarial Networks 2
Alireza Shafaei
 
Generative Adversarial Networks and Their Applications
Artifacia
 
GNA 13552928 deep learning for GAN a.ppt
ManiMaran230751
 
Generative Adversarial Networks (GANs) at the Data Science Meetup Luxembourg ...
Chris Hammerschmidt
 
مدل آموزش داده مصنوعی مبتنی بر شبکه GAN برای شبکه های عصبی CNN سبک
javascriptsali
 
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Universitat Politècnica de Catalunya
 
Generative Adversarial Networks
Mustafa Yagmur
 
GANs Presentation.pptx
MAHMOUD729246
 
Generative adversarial networks
Ding Li
 
What is a GAN Generative Adversarial Networks Guide.pdf
imoliviabennett
 
PROGRESSIVE GROWING OF GAN S FOR I MPROVED QUALITY , STABILITY , AND VARIATION
Willy Marroquin (WillyDevNET)
 
Ad

Recently uploaded (20)

PDF
An Uncut Conversation With Grok | PDF Document
Mike Hydes
 
PPTX
The whitetiger novel review for collegeassignment.pptx
DhruvPatel754154
 
PPTX
Introduction to Data Analytics and Data Science
KavithaCIT
 
PPTX
World-population.pptx fire bunberbpeople
umutunsalnsl4402
 
PDF
SUMMER INTERNSHIP REPORT[1] (AutoRecovered) (6) (1).pdf
pandeydiksha814
 
PPTX
INFO8116 - Week 10 - Slides.pptx big data architecture
guddipatel10
 
PPTX
Future_of_AI_Presentation for everyone.pptx
boranamanju07
 
PDF
Technical Writing Module-I Complete Notes.pdf
VedprakashArya13
 
PPTX
INFO8116 - Week 10 - Slides.pptx data analutics
guddipatel10
 
PDF
Research about a FoodFolio app for personalized dietary tracking and health o...
AustinLiamAndres
 
PDF
202501214233242351219 QASS Session 2.pdf
lauramejiamillan
 
PPTX
White Blue Simple Modern Enhancing Sales Strategy Presentation_20250724_21093...
RamNeymarjr
 
PPTX
Introduction to Biostatistics Presentation.pptx
AtemJoshua
 
PPTX
Fuzzy_Membership_Functions_Presentation.pptx
pythoncrazy2024
 
PDF
blockchain123456789012345678901234567890
tanvikhunt1003
 
PPTX
Web dev -ppt that helps us understand web technology
shubhragoyal12
 
PPTX
Data-Driven Machine Learning for Rail Infrastructure Health Monitoring
Sione Palu
 
PPTX
Employee Salary Presentation.l based on data science collection of data
barridevakumari2004
 
PPTX
short term internship project on Data visualization
JMJCollegeComputerde
 
PPTX
lecture 13 mind test academy it skills.pptx
ggesjmrasoolpark
 
An Uncut Conversation With Grok | PDF Document
Mike Hydes
 
The whitetiger novel review for collegeassignment.pptx
DhruvPatel754154
 
Introduction to Data Analytics and Data Science
KavithaCIT
 
World-population.pptx fire bunberbpeople
umutunsalnsl4402
 
SUMMER INTERNSHIP REPORT[1] (AutoRecovered) (6) (1).pdf
pandeydiksha814
 
INFO8116 - Week 10 - Slides.pptx big data architecture
guddipatel10
 
Future_of_AI_Presentation for everyone.pptx
boranamanju07
 
Technical Writing Module-I Complete Notes.pdf
VedprakashArya13
 
INFO8116 - Week 10 - Slides.pptx data analutics
guddipatel10
 
Research about a FoodFolio app for personalized dietary tracking and health o...
AustinLiamAndres
 
202501214233242351219 QASS Session 2.pdf
lauramejiamillan
 
White Blue Simple Modern Enhancing Sales Strategy Presentation_20250724_21093...
RamNeymarjr
 
Introduction to Biostatistics Presentation.pptx
AtemJoshua
 
Fuzzy_Membership_Functions_Presentation.pptx
pythoncrazy2024
 
blockchain123456789012345678901234567890
tanvikhunt1003
 
Web dev -ppt that helps us understand web technology
shubhragoyal12
 
Data-Driven Machine Learning for Rail Infrastructure Health Monitoring
Sione Palu
 
Employee Salary Presentation.l based on data science collection of data
barridevakumari2004
 
short term internship project on Data visualization
JMJCollegeComputerde
 
lecture 13 mind test academy it skills.pptx
ggesjmrasoolpark
 
Ad

6666666666666666666666666666666666666.pdf

  • 1. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 921 Evolutionary Generative Adversarial Networks Chaoyue Wang, Chang Xu , Xin Yao , Fellow, IEEE, and Dacheng Tao , Fellow, IEEE Abstract—Generative adversarial networks (GANs) have been effective for learning generative models for real-world data. However, accompanied with the generative tasks becoming more and more challenging, existing GANs (GAN and its variants) tend to suffer from different training problems such as insta- bility and mode collapse. In this paper, we propose a novel GAN framework called evolutionary GANs (E-GANs) for sta- ble GAN training and improved generative performance. Unlike existing GANs, which employ a predefined adversarial objective function alternately training a generator and a discriminator, we evolve a population of generators to play the adversar- ial game with the discriminator. Different adversarial training objectives are employed as mutation operations and each indi- vidual (i.e., generator candidature) are updated based on these mutations. Then, we devise an evaluation mechanism to mea- sure the quality and diversity of generated samples, such that only well-performing generator(s) are preserved and used for further training. In this way, E-GAN overcomes the lim- itations of an individual adversarial training objective and always preserves the well-performing offspring, contributing to progress in, and the success of GANs. Experiments on several datasets demonstrate that E-GAN achieves convincing genera- tive performance and reduces the training problems inherent in existing GANs. Index Terms—Deep generative models, evolutionary computa- tion, generative adversarial networks (GANs). I. INTRODUCTION GENERATIVE adversarial networks (GANs) [1] are one of the main groups of methods used to learn generative Manuscript received August 16, 2018; revised November 29, 2018; accepted January 15, 2019. Date of publication January 28, 2019; date of current ver- sion November 27, 2019. This work was supported in part by the Australian Research Council under Project FL-170100117, Project DP-180103424, Project IH180100002, and Project DE180101438, in part by the National Key Research and Development Program of China under Grant 2017YFC0804003, in part by the Engineering and Physical Sciences Research Council (EPSRC) under Grant EP/J017515/1 and Grant EP/P005578/1, in part by the Program for Guangdong Introducing Innovative and Enterpreneurial Teams under Grant 2017ZT07X386, in part by the Shenzhen Peacock Plan under Grant KQTD2016112514355531, in part by the Science and Technology Innovation Committee Foundation of Shenzhen under Grant ZDSYS201703031748284, and in part by the Program for University Key Laboratory of Guangdong Province under Grant 2017KSYS008. (Corresponding author: Chang Xu.) C. Wang, C. Xu, and D. Tao are with the UBTECH Sydney Artificial Intelligence Centre, Faculty of Engineering and Information Technologies, University of Sydney, Darlington, NSW 2008, Australia, and also with the School of Computer Science, University of Sydney, Darlington, NSW 2008, Australia (e-mail: [email protected]; [email protected]; [email protected]). X. Yao is with the Shenzhen Key Laboratory of Computational Intelligence, University Key Laboratory of Evolving Intelligent Systems of Guangdong Province, Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China, and also with the CERCIA, School of Computer Science, University of Birmingham, Birmingham B15 2TT, U.K. (e-mail: [email protected]). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TEVC.2019.2895748 models from complicated real-world data. As well as using a generator to synthesize semantically meaningful data from standard signal distributions, GANs (GAN and its variants) train a discriminator to distinguish real samples in the train- ing dataset from fake samples synthesized by the generator. As the confronter, the generator aims to deceive the discrim- inator by producing ever more realistic samples. The training procedure continues until the generator wins the adversarial game; that is, the discriminator cannot make a better decision than randomly guessing whether a particular sample is fake or real. Compared with other existing generative models, GANs provide a concise and efficient framework for learning genera- tive models. Therefore, GANs have recently been successfully applied to image generation [2]–[6], image editing [7]–[10], video prediction [11]–[13], and many other tasks [14]–[18]. Comparing with most existing discriminative tasks (e.g., classification and clustering), GANs perform a challenging generative process, which projects a standard distribution to a much more complex high-dimensional real-world data dis- tribution. First, although GANs have been utilized to model large-scale real-world datasets, such as CelebA, LSUN, and ImageNet, they easily suffer from the mode collapse problem, i.e., the generator can only learn some limited patterns from the large-scale target datasets, or assigns all of its probability mass to a small region in the space [19]. Meanwhile, if the generated distribution and the target data distribution do not substantially overlap (usually at the beginning of training), the generator gradients can point to more or less random direc- tions or even result in the vanishing gradient issue. In addition, to generate high-resolution and high-quality samples, both the generator and discriminator are asked to be deeper and larger. Under the vulnerable adversarial framework, it is hard to bal- ance and optimize such large-scale deep networks. Thus, in most of the existing works, appropriate hyper-parameters (e.g., learning rate and updating steps) and network architectures are critical configurations. Unsuitable settings reduce GAN’s performance or even fail to produce any reasonable results. Overall, although GANs already produce visually appealing samples in various applications, they are still facing many large-scale optimization problems. Many recent efforts on GANs have focused on over- coming these optimization difficulties by developing vari- ous adversarial training objectives. Typically, assuming the optimal discriminator for the given generator is learned, dif- ferent objective functions of the generator aim to measure the distance between the generated distribution and the tar- get data distribution under different metrics. The original GAN uses Jensen–Shannon divergence (JSD) as the met- ric. A number of metrics have been introduced to improve GAN’s performance, such as least squares [20], absolute 1089-778X c 2019 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information. Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 2. 922 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 deviation [21], Kullback–Leibler (KL) divergence [22], and Wasserstein distance [23]. However, according to both theo- retical analyses and experimental results, minimizing each dis- tance has its own pros and cons. For example, although mea- suring KL divergence largely eliminates the vanishing gradient issue, it easily results in mode collapse [22], [24]. Likewise, Wasserstein distance greatly improves training stability but can have nonconvergent limit cycles near equilibrium [25]. Through observation, we find that most existing GAN methods are limited by the specified adversarial optimization strategy. Since the training strategy is fixed, it is hard to adjust the balance between the generator and discriminator during the training process. Meanwhile, as aforementioned, each exist- ing adversarial training strategy has its own pros and cons in training GAN models. In this paper, we build an evolutionary GAN (E-GAN), which treats the adversarial training procedure as an evo- lutionary problem. Specifically, a discriminator acts as the environment (i.e., provides adaptive loss functions) and a pop- ulation of generators evolve in response to the environment. During each adversarial (or evolutionary) iteration, the dis- criminator is still trained to recognize real and fake samples. However, in our method, acting as parents, generators undergo different mutations to produce offspring to adapt to the envi- ronment. Different adversarial objective functions aim to min- imize different distances between the generated distribution and the data distribution, leading to the different mutations. Meanwhile, given the current optimal discriminator, we mea- sure the quality and diversity of samples generated by the updated offspring. Finally, according to the principle of “sur- vival of the fittest,” poorly performing offspring are removed and the remaining well-performing offspring (i.e., generators) are preserved and used for further training. Based on the evo- lutionary paradigm to optimize GANs, the proposed E-GAN overcomes the inherent limitations in the individual adversarial training objectives and always preserves the well-performing offspring produced by different training objectives (i.e., muta- tions). In this way, we contribute to progress in and the success of the large-scale optimization of GANs. Following vanilla GAN [1], we evaluate the new algorithm in image generation tasks. Overall, the proposed evolutionary strategy is largely orthogonal to the existing GAN models. Through applying the evolutionary framework to different GAN models, it is possi- ble to perform different kinds of generation tasks. For example, GAN objectives were devised for generating text. Considering them as the mutation operations, the proposed evolutionary framework can be applied to solve text generation tasks. Meanwhile, the proposed E-GAN framework also provides a novel direction to apply evolutionary learning paradigm on solving deep learning problems. Recent years, although deep learning algorithms have achieved promising perfor- mances on a variety of applications, they are still facing many challenges in solving real-world problems. Evolutionary computation, as a powerful approach to complex real-world problems [26]–[29], has been utilized to solve many deep learning challenges. Among them, Real et al. [30] devised an evolutionary algorithm to automatically search the architecture and hyper-parameters of deep networks. Moreover, evolution strategies have been utilized as an alternative to MDP-based techniques to optimize reinforcement learning models [31]. In this paper, we attempt to combine the back propagation algorithm and the evolutionary algorithm for optimizing deep generative models. The parameters updated by different learn- ing objectives are regarded as variation results during the evolutionary process. By introducing suitable evaluation and selection mechanisms, the whole training process can be more efficient and stable. We hope the proposed evolutionary learn- ing framework can be generalized to more deep learning problems, such as reinforcement learning. In summary, we make the following contributions in this paper. 1) In order to stabilize GAN’s training process, we devised a simple yet efficient evolutionary algorithm for opti- mizing generators within GANs framework. To the best of our knowledge, it is the first work that introduces the evolutionary learning paradigm into learning GAN models. 2) Through analyzing the training process of E-GAN, some properties of existing GANs objectives are further explored and discussed. 3) Experiments evaluated on several large-scale datasets are performed, and demonstrate that convincing results can be achieved by the proposed E-GAN framework. The rest of this paper is organized as follows. After a brief summary of the previous related works in Section II, we illus- trate the proposed E-GAN together with its training process in Section III. Then, we exhibit the experimental validation of the whole method in Section IV. Finally, we conclude this paper with some future directions in Section V. II. BACKGROUND In this section, we first review some previous GANs devoted to reducing training instability and improving the genera- tive performance. We then briefly summarize some of the evolutionary algorithms on deep neural networks. A. Generative Adversarial Networks GANs provides an excellent framework for learning deep generative models, which aim to capture the probability dis- tributions over the given data. Compared to other generative models, GAN is easily trained by alternately updating a gener- ator and a discriminator using the back propagation algorithm. In many generative tasks, GANs (GAN and its variants) produce better samples than other generative models [32]. Besides image generation tasks, GANs have been introduced to more and more tasks, such as video generation [11], [33], visual tracking [34]–[36], domain adaption [37], hashing coding [38]–[40], and feature learning [41], [42]. In these tasks, the adversarial training strategy also achieved promising performances. However, some problems still exist in the GANs training process. In the original GAN, training the generator was equal to minimizing the JSD between the data distribution and the generated distribution, which easily resulted in the vanishing gradient problem. To solve this issue, a nonsaturating heuristic Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 3. WANG et al.: E-GANs 923 objective (i.e., “− log D trick”) replaced the minimax objective function to penalize the generator [1]. Then, Radford et al. [22] and Salimans et al. [43] designed specified network (DCGAN) architectures and proposed several heuristic tricks (e.g., feature matching, one-side label smoothing, and virtual batch normal- ization) to improve training stability. Meanwhile, energy-based GAN [21] and least-squares GAN [20] improved training sta- bility by employing different training objectives. Although these methods partly enhanced training stability, in practice, the network architectures, and training procedure still required careful design to maintain the discriminator-generator balance. More recently, Wasserstein GAN (WGAN) [23] and its vari- ant WGAN-GP (with gradient penalty) [44] were proposed to minimize the Wasserstein-1 distance between the gener- ated and data distributions. Since the Wasserstein-1 distance is continuous everywhere and differentiable almost everywhere under only minimal assumptions [23], these two methods con- vincingly reduce training instability. However, to measure the Wasserstein-1 distance between the generated distribution and the data distribution, they are asked to enforce the Lipschitz constraint on the discriminator (also known as critic), which may result in some optimization problems [44]. Besides devising different objective functions, some works attempt to stable GAN training and improve generative performance by introducing multiple generators or discrimina- tors into an adversarial training framework. Nguyen et al. [45] proposed a dual discriminator GAN, which combines the KL and reverse KL divergences into a unified objective func- tion through employing two discriminators. Multidiscriminator GAN frameworks [46], [47] are devised and utilized for provid- ing stable gradients to the generator and further stabilizing the adversarial training process. Moreover, Tolstikhin et al. [48] applied boosting techniques to train a mixture of genera- tors by continually adding new generators to the mixture. Ghosh et al. [49] trained many generators by using a mul- ticlass discriminator that predicts which generator produces the sample. Mixture GAN [50] is proposed to overcome the mode collapse problem by training multiple generators to spe- cialize in different data modes. Overall, these multigenerator GANs aim to learn a set of generators, and the mixture of their learned distributions would approximate the data distribution, i.e., different generators are encouraged to capture different data modes. Although the proposed E-GAN also creates multiple generators during training, we always keep the well-performing candidates through survival of the fittest, which helps the final learned generator achieving better performance. Note that, in our framework, only one generator was learned to represent the whole target distribution at the end of the training. B. Evolutionary Computation Over the last 20 years, evolutionary algorithms have achieved considerable success across a wide range of computational tasks, including modeling, optimization, and design [51]–[55]. Inspired by natural evolution, the essence of an evolutionary algorithm is to equate possible solutions to individuals in a population, produce offspring through varia- tions, and select appropriate solutions according to fitness [56]. Recently, evolutionary algorithms have been introduced to solve deep learning problems. To minimize human participation in designing deep algorithms and automati- cally discover such configurations, there have been many attempts to optimize deep learning hyper-parameters and design deep network architectures through an evolutionary search [57], [58]. Among them, Real et al. [30] proposed a large-scale evolutionary algorithm to design a whole deep classifier automatically. Meanwhile, different from widely employed gradient-based learning algorithms (e.g., backprop- agation), evolutionary algorithms have also demonstrated their capacity to optimize neural networks. EPNet [59] was devised for evolving and training neural networks using evolution- ary programming. In [60], EvoAE was proposed to speed up the training of autoencoders for constructing deep neu- ral networks. Moreover, Salimans et al. [31] proposed a novel evolutionary strategy as an alternative to the popu- lar MDP-based reinforcement learning techniques, achieving strong performance on reinforcement learning benchmarks. In addition, an evolutionary algorithm was proposed to compress deep learning models by automatically eliminating redundant convolution filters [61]. Last but not the least, the evolution- ary learning paradigm has been utilized to solve a number of deep/machine tasks, such as automatic machine learning [62], multiobjective optimization [63], etc. However, to the best of our knowledge, there is still no work attempt to optimize deep generative models with the evolutionary learning algorithms. III. METHODS In this section, we first review the original GAN formu- lation. Then, we introduce the proposed E-GAN algorithm. By illustrating E-GAN’s mutations and evaluation mechanism, we further discuss the advantage of the proposed framework. Finally, we conclude with the entire E-GAN training process. A. Generative Adversarial Networks GAN, first proposed in [1], studies a two-player minimax game between a discriminative network D and a generative network G. Taking noisy sample z ∼ p(z) (sampled from a uniform or normal distribution) as the input, the generative network G outputs new data G(z), whose distribution pg is supposed to be close to that of the data distribution pdata. Meanwhile, the discriminative network D is employed to dis- tinguish the true data sample x ∼ pdata(x) and the generated sample G(z) ∼ pg(G(z)). In the original GAN, this adversarial training process was formulated as min G max D Ex∼pdata log D(x) + Ez∼pz log(1 − D(G(z))) . (1) The adversarial procedure is illustrated in Fig. 1(a). Most exist- ing GANs perform a similar adversarial procedure in different adversarial objective functions. B. Evolutionary Algorithm In contrast to conventional GANs, which alternately update a generator and a discriminator, we devise an evolutionary algorithm that evolves a population of generator(s) {G} in a Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 4. 924 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 (a) (b) Fig. 1. (a) Original GAN framework. A generator G and a discriminator D play a two-player adversarial game. The updating gradients of the generator G are received from the adaptive objective, which depends on discriminator D. (b) Proposed E-GAN framework. A population of generators {Gθ } evolves in a dynamic environment, the discriminator D. Each evolutionary step consists of three substages: variation, evaluation, and selection. The best offspring are kept. given environment (i.e., the discriminator D). In this popu- lation, each individual represents a possible solution in the parameter space of the generative network G. During the evo- lutionary process, we expect that the population gradually adapts to its environment, which means that the evolved gener- ator(s) can generate ever more realistic samples and eventually learn the real-world data distribution. As shown in Fig. 1(b), during evolution, each step consists of three substages. 1) Variation: Given an individual Gθ in the population, we utilize the variation operators to produce its off- spring {Gθ1 , Gθ2 , . . .}. Specifically, several copies of each individual—or parent—are created, each of which are modified by different mutations. Then, each modified copy is regarded as one child. 2) Evaluation: For each child, its performance—or indi- vidual’s quality—is evaluated by a fitness function F(·) that depends on the current environment (i.e., discrimi- nator D). 3) Selection: All children will be selected according to their fitness value, and the worst part is removed—that is, they are killed. The rest remain alive (i.e., free to act as parents) and evolve to the next iteration. After each evolutionary step, the discriminative network D (i.e., the environment) is updated to further distinguish real samples x and fake samples y generated by the evolved generator(s), i.e., LD = −Ex∼pdata log D(x) − Ey∼pg log(1 − D(y)) . (2) Thus, the discriminative network D (i.e., the environment) can continually provide the adaptive losses to drive the popula- tion of generator(s) evolving to produce better solutions. Next, we illustrate and discuss the proposed variation (or mutation), evaluation, and selection operators in detail. C. Variation We employ asexual reproduction with different mutations to produce the next generation’s individuals (i.e., children). Specifically, these mutation operators correspond to differ- ent training objectives, which attempt to narrow the distances between the generated distribution and the data distribution from different perspectives. In this section, we introduce the mutations used in this paper.1 To analyze the corre- sponding properties of these mutations, we suppose that, for each evolutionary step, the optimal discriminator D∗(x) = ([pdata(x)]/[pdata(x) + pg(x)]), according to (2), has already been learned [1]. 1) Minimax Mutation: The minimax mutation corresponds to the minimax objective function in the original GAN Mminimax G = 1 2 Ez∼pz log(1 − D(G(z)) . (3) According to the theoretical analysis in [1], given the optimal discriminator D∗, the minimax mutation aims to minimize the JSD between the data distribution and the generated distri- bution. Although the minimax game is easy to explain and theoretically analyze, its performance in practice is disap- pointing, a primary problem being the generator’s vanishing gradient. If the support of two distributions lies in two man- ifolds, the JSD will be a constant, leading to the vanishing gradient [24]. This problem is also illustrated in Fig. 2. When the discriminator rejects generated samples with high confi- dence [i.e., D(G(z)) → 0], the gradient tends to vanishing. However, if the generated distribution overlaps with the data distribution, meaning that the discriminator cannot completely distinguish real from fake samples, the minimax mutation provides effective gradients and continually narrows the gap between the data distribution and the generated distribution. 2) Heuristic Mutation: Unlike the minimax mutation, which minimizes the log probability of the discriminator being correct, the heuristic mutation aims to maximize the log 1Although more mutation operations can be included in our framework, according to the theoretical analysis below, we adopt three interpretable and complementary objectives as our mutations. Meanwhile, we have tested more mutation operations, yet the mutations described in this paper already delivered a convincing performance. Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 5. WANG et al.: E-GANs 925 Fig. 2. Mutation (or objective) functions that the generator G receives given the discriminator D. probability of the discriminator being mistaken, i.e., Mheuristic G = − 1 2 Ez∼pz log(D(G(z)) . (4) Compared to the minimax mutation, the heuristic muta- tion will not saturate when the discriminator rejects the generated samples. Thus, the heuristic mutation avoids van- ishing gradient and provides useful generator updates (Fig. 2). However, according to [24], given the optimal discriminator D∗, minimizing the heuristic mutation is equal to minimiz- ing [KL(pg||pdata) − 2JSD(pg||pdata)], i.e., inverted KL minus two JSDs. Intuitively, the JSD sign is negative, which means pushing these two distributions away from each other. In prac- tice, this may lead to training instability and generative quality fluctuations [44]. 3) Least-Squares Mutation: The least-squares mutation is inspired by LSGAN [20], where the least-squares objec- tives are utilized to penalize its generator to deceive the discriminator. In this paper, we formulate the least-squares mutation as M least-square G = Ez∼pz (D(G(z)) − 1)2 . (5) As shown in Fig. 2, the least-squares mutation is nonsaturat- ing when the discriminator can recognize the generated sample [i.e., D(G(z)) → 0]. When the discriminator output grows, the least-squares mutation saturates, eventually approaching zero. Therefore, similar to the heuristic mutation, the least-squares mutation can avoid vanishing gradient when the discrimina- tor has a significant advantage over the generator. Meanwhile, compared to the heuristic mutation, although the least-squares mutation will not assign an extremely high cost to generate fake samples, it will also not assign an extremely low cost to mode dropping,2 which partly avoids mode collapse [20]. 2Arjovsky and Bottou [24] demonstrated that the heuristic objective suffers from mode collapse since KL(pg||pdata) assigns a high cost to generating fake samples but an extremely low cost to mode dropping. Note that, different from GAN-minimax and GAN-heuristic, LSGAN employs a different objective (“least squares”) to optimize the discriminator, i.e., LLSGAN D = 1 2 Ex∼pdata (D(x) − 1)2 + 1 2 Ez∼pz D(G(z))2 = x 1 2 pdata(x)(D(x) − 1)2 + pg(x)D(x)2 dx. (6) Yet, with respect to D(x), the function LLSGAN D achieves its minimum in [0, 1] at ([pdata(x)]/[pdata(x) + pg(x)]), which is equivalent to ours [i.e., (2)]. Therefore, although we employ only one discriminator as the environment to distinguish real and generated samples, it is sufficient to provide adaptive losses for all mutations described above. D. Evaluation In an evolutionary algorithm, evaluation is the operation of measuring the quality of individuals. To determine the evo- lutionary direction (i.e., individuals’ selection), we devise an evaluation (or fitness) function to measure the performance of evolved individuals (i.e., children). Typically, we mainly focus on two properties: 1) the quality and 2) the diversity of gener- ated samples. Quality is measured for each generated sample. If a generated image could be realistic enough, it will fool the discriminator. On the other hand, the diversity measures whether the generator could spread the generated samples out enough, which could largely avoid mode collapse. First, we simply feed generator produced images into the discriminator D and observe the average value of the output, which we name the quality fitness score Fq = Ez[D(G(z))]. (7) Note that discriminator D is constantly upgraded to be optimal during the training process, reflecting the quality of genera- tors at each evolutionary (or adversarial) step. If a generator obtains a relatively high-quality score, its generated samples can deceive the discriminator and the generated distribution is further approximate to the data distribution. Besides generative quality, we also pay attention to the diversity of generated samples and attempt to overcome the mode collapse issue in GAN optimization. Recently, Nagarajan and Kolter [25] proposed a gradient-based regular- ization term to stabilize the GAN optimization and suppress mode collapse. When the generator collapses to a small region, the discriminator will subsequently label collapsed points as fake with obvious countermeasure (i.e., big gradients). In con- trast, if the generator is capable of spreading generated data out enough, the discriminator will not be much confident to label generated samples as fake data (i.e., updated with small gra- dients). Other techniques, e.g., exploiting conditioning latent vector [64] or subspace [65], can also be applied to purse diver- sity. But this issue does not fall within the scope of this paper. We employ a similar principle to evaluate generator optimization stability and generative diversity. Here, since the gradient-norm of the discriminator could vary largely during training, we employed an logarithm to shrink its fluctuation. Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 6. 926 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 Specifically, the minus log-gradient-norm of optimizing D is utilized to measure the diversity of generated samples. If an evolved generator obtains a relatively high value, which corre- sponds to small discriminator gradients, its generated samples tend to spread out enough, to avoid the discriminator from hav- ing obvious countermeasures. Thus, the mode collapse issue can be suppressed and the discriminator will change smoothly, which helps to improve the training stability. Formally, the diversity fitness score is defined as Fd = − log ∇D − Ex log D(x) − Ez log(1 − D(G(z))) . (8) Based on the aforementioned two fitness scores, we can finally give the evaluation (or fitness) function of the proposed evolutionary algorithm F = Fq + γ Fd (9) where γ ≥ 0 balances two measurements: 1) generative qual- ity and 2) diversity. Overall, a relatively high fitness score F, leads to higher training efficiency and better generative performance. It is interesting to note that, the proposed fitness score can also be regarded as an objective function for genera- tor G. However, as demonstrated in our experiments, without the proposed evolutionary strategy, this fitness objective is still hard to achieve a convincing performance. Meanwhile, similar with most existing objective functions of GANs, the fitness function will continuously fluctuate during the dynamic adversarial training process. Specifically, in each iteration, the fitness score of generators are evaluated based on the current discriminator. Comparing them, we are capable of selecting well-performing ones. However, since the discriminator will be updated according to different generators, it is hard to dis- cuss the connection between the fitness scores over different iterations. E. Selection In an evolutionary algorithm, the counterpart of the muta- tion operators is the selection. In the proposed E-GAN, we employ a simple yet useful survivor selection strategy to deter- mine the next generation based on the fitness score of existing individuals. First of all, we should notice that both the generators (i.e., population) and the discriminator (i.e., environment) are opti- mized alternately in a dynamic procedure. Thus, the fitness function is not fixed and the fitness score of generators can only be evaluated by the corresponding discriminator in the same evolutionary generation, which means fitness scores eval- uated in different generations cannot compare with each other. In addition, due to the mutation operators of the proposed E-GAN actually correspond to different adversarial training objectives, selecting desired offspring is equivalent to select- ing the effective adversarial strategies. During the adversarial process, we hope that generator(s) can do it best to fool the discriminator (i.e., implement the optimal adversarial strategy). Considering these two points, we utilize the comma selec- tion, i.e., (μ, λ)-selection [66] as the selection mechanism of Algorithm 1 E-GANs. Default Values α = 0.0002, β1 = 0.5, β2 = 0.99, nD = 3, nm = 3, and m = 32 Require: the batch size m. the discriminator’s updating steps per iteration nD. the number of parents μ. the number of mutations nm. Adam hyper-parameters α, β1, β2, the hyper-parameter γ of evaluation function. Require: initial discriminator’s parameters w0. initial genera- tors’ parameters {θ1 0 , θ2 0 , . . . , θ μ 0 }. 1: for number of training iterations do 2: for k = 0, . . . , nD do 3: Sample a batch of {x(i)}m i=1 ∼ pdata (training data), and a batch of {z(i)}m i=1 ∼ pz (noise samples). 4: gw ← ∇w[ 1 m m i=1 log Dw(x(i)) 5: + 1 m μ j=1 m/μ i=1 log(1 − Dw(Gθj (z(i))))] 6: w ← Adam(gw, w, α, β1, β2) 7: end for 8: for j = 0, . . . , μ do 9: for h = 0, . . . , nm do 10: Sample a batch of {z(i)}m i=1 ∼ pz 11: gθj,h ← ∇θj Mh G({z(i)}m i=1, θj) 12: θ j,h child ← Adam(gθj,h , θj, α, β1, β2) 13: Fj,h ← F j,h q + γ F j,h d 14: end for 15: end for 16: {Fj1,h1 , Fj2,h2 , . . . } ← sort({Fj,h}) 17: θ1, θ2, . . . , θμ ← θ j1,h1 child , θ j2,h2 child , . . . , θ jμ,hμ child 18: end for E-GAN. Specifically, after sorting the current offspring popu- lation {xi}λ i=1 according to their fitness scores Fi, the μ-best individuals are selected to form the next generation. F. E-GAN Having introduced the proposed evolutionary algorithm and corresponding mutation operations, evaluation criteria and selection strategy, the complete E-GAN training process is concluded in Algorithm 1. Overall, in E-GAN, generators {G} are regarded as an evolutionary population and discriminator D acts as an environment. For each evolutionary step, generators are updated with different mutations (or objectives) to accom- modate the current environment. According to the principle of survival of the fittest, only well-performing children will sur- vive and participate in future adversarial training. Unlike the two-player game with a fixed and static adversarial training objective in conventional GANs, E-GAN allows the algorithm to integrate the merits of different adversarial objectives and generate the most competitive solution. Thus, during training, the evolutionary algorithm not only largely suppresses the lim- itations (vanishing gradient, mode collapse, etc.) of individual adversarial objectives, but it also harnesses their advantages to search for a better solution. IV. EXPERIMENTS To evaluate the proposed E-GAN, we run experiments on serval generative tasks and present the experimental results Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 7. WANG et al.: E-GANs 927 TABLE I ARCHITECTURES OF THE GENERATIVE AND DISCRIMINATIVE NETWORKS in this section. Compared with some previous GAN methods, we show that the proposed E-GAN can achieve impressive generative performance on large-scale image datasets. A. Implementation Details We evaluate E-GAN on two synthetic datasets and three image datasets: 1) CIFAR-10 [67]; 2) LSUN bedroom [68]; and 3) CelebA [69]. For fair comparisons, we adopted the same network architectures with existing works [22], [44]. In addition, to achieve better performance on generating 128 × 128 images, we slightly modified both the generative network and the discriminator network based on the DCGAN architecture. Specifically, the batch norm layers are removed from the generator, and more features channels are applied to each convolutional layers. The detailed networks are listed in Table I, note that the network architectures of the other comparison experiments can be easily found in the referenced works. We use the default hyper-parameter values listed in Algorithm 1 for all experiments. Note that the hyper-parameter γ is utilized to balance the measurements of samples quality (i.e., Fq) and diversity (i.e., Fd). Usually, the quality fit- ness score Fq lies in [0, 1], while the diversity fitness score Fd measures the log-gradient-norm of the discriminator D, which can vary largely according to D’s scale. Therefore, we first determine γ ’s range based on the selected discrim- inator D. Then, we run grid search to find its value. In practice, we choose γ = 0.5 for the synthetic datasets, and γ = 0.001 for real-world data. In addition, recently, some works [44], [70] proposed the gradient penalty (GP) term to regularize the discriminator to provide precise gradients for updating the generator. Within the adversarial training frame- work, our contributions are largely orthogonal to the GP term. In our experiments, through the setting without GP term, we demonstrated the efficiency of the proposed method. Then, after introducing the GP term, the generative performance was further improved, which demonstrated our framework could also benefit from the regularization technique for the discrim- inator. Furthermore, all experiments were trained on Nvidia Tesla V100 GPUs. To train a model for 64 × 64 images using the DCGAN architecture cost around 20 h on a single GPU. B. Evaluation Metrics Besides directly reported generated samples of the learned generative networks, we choose the maximum mean discrep- ancy (MMD) [71], [72], the inception score (IS) [43], and the Fréchet Inception distance (FID) [73] as quantitative met- rics. Among them, the MMD can be utilized to measure the discrepancy between the generated distribution and the target distribution for synthetic Gaussian mixture datasets. However, the MMD is difficult to directly apply to high- dimensional image datasets. Therefore, through applying the pretrained Inception v3 network [74] to generated images, the IS computes the KL divergence between the conditional class distribution and the marginal class distribution. Usually, this score correlates well with the human scoring of the realism of generated images from the CIFAR-10 dataset, and a higher value indicates better image quality. However, some recent works [6], [75] revealed serious limitations of the IS, e.g., the target data distribution (i.e., the training data) has not been considered. In our experiments, we utilize the IS to measure the E-GAN’s performance on the CIFAR-10, and compare our results to previous works. Moreover, FID is a more principled and reliable metric and has demonstrated better correlations with human evaluation for other datasets. Specifically, FID calculates the Wasserstein-2 distance between the generated images and the real-world images in the high-level feature space of the pretrained Inception v3 network. Note that lower FID means closer distances between the generated distribu- tion and the real-world data distribution. In all experiments, we randomly generated 50k samples to calculate the MMD, IS, and FID. C. Synthetic Datasets and Mode Collapse In the first experiment, we adopt the experimental design proposed in [76], which trains GANs on 2-D Gaussian mix- ture distributions. The mode collapse issue can be accurately measured on these synthetic datasets, since we can clearly Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 8. 928 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 Fig. 3. KDE plots of the target data and generated data from different GANs trained on mixtures of Gaussians. In the first row, a mixture of 8 Gaussians arranged in a circle. In the second row, a mixture of 25 Gaussians arranged in a grid. TABLE II MMD (×10−2) WITH MIXED GAUSSIAN DISTRIBUTIONS ON OUR TOY DATASETS. WE RAN EACH METHOD FOR TEN TIMES, AND REPORT THEIR AVERAGE AND BEST RESULTS. THE METHOD WITH LOWER MMD VALUE IMPLIES THE GENERATED DISTRIBUTION IS CLOSER TO THE TARGET ONE observe and measure the generated distribution and the target data distribution. As shown in Fig. 3, we employ two challeng- ing distributions to evaluate E-GAN, a mixture of 8 Gaussians arranged in a circle and a mixture of 25 Gaussians arranged in a grid.3 Here, to evaluate if the proposed diversity fitness score can reduce the mode collapse, we did not introduce the GP norm and set the survived parents number μ as 1, i.e., during each evolutionary step, only the best candidature are kept. First, we utilize existing individual adversarial objectives (i.e., conventional GANs) to perform the adversarial training process. We train each method 50K iterations and report the Kernel density estimation (KDE) plots in Fig. 3. Meanwhile, the average and the best MMD values which running each method ten times are reported in Table II. The results show that all of the individual adversarial objectives suffer from mode collapse to a greater or lesser degree. Then, we set hyper- parameter γ as 0 and test the proposed E-GAN, which means the diversity fitness score was not considered during the train- ing process. The results show that the evolutionary framework still has problems with the mode collapse. However, when the 3We obtain both 2-D distributions and network architectures from the code provided in [44]. TABLE III INCEPTION SCORES AND FIDS WITH UNSUPERVISED IMAGE GENERATION ON CIFAR-10. THE METHOD WITH HIGHER IS OR LOWER FID IMPLIES THE GENERATED DISTRIBUTION IS CLOSER TO THE TARGET ONE. † [22], ‡ [6] diversity fitness score is considered in the selection stage (in this experiment, we set γ = 0.5), the mode collapse issue is largely suppressed and the trained generator can more accu- rately fit the target distributions. This demonstrates that our diversity fitness score has the capability of measuring the sam- ple diversity of updated generators and further suppress the mode collapse problem. D. CIFAR-10 and Training Stability In the proposed E-GAN, we utilize the evolutionary algo- rithm with different mutations (i.e., different updating strate- gies) to optimize generator(s) {G}. To demonstrate the advan- tages of the proposed evolutionary algorithm over existing two-player adversarial training strategies (i.e., updating gen- erator with a single objective), we train these methods on CIFAR-10 and plot inception scores [43] over the training process. For a fair comparison, we did not introduce the GP norm into our E-GAN and set the parents number μ Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 9. WANG et al.: E-GANs 929 Fig. 4. Experiments on the CIFAR-10 dataset. CIFAR-10 IS over generator iterations (left), over wall-clock time (middle), and the graph of selected mutations in the E-GAN training process (right). TABLE IV FIDS WITH UNSUPERVISED IMAGE GENERATION ON LSUN BEDROOM DATASET. THE METHOD WITH LOWER FID IMPLIES THE GENERATED DISTRIBUTION IS CLOSER TO THE TARGET ONE as 1. Moreover, the same network architecture is used for all methods. As shown in Fig. 4(left), E-GAN can get higher IS with less training steps. Meanwhile, E-GAN also shows comparable stability when it goes to convergence. By comparison, conven- tional GAN objectives expose their different limitations, such as instability at convergence (GAN-Heuristic), slow conver- gence (GAN-Least square), and invalid (GAN-minimax). In addition, we employ the proposed fitness function [i.e., (9)] as generator’s objective function, and find its performance is also inferior to E-GAN. This experiment further demon- strates the advantages of the proposed evolutionary framework. Through creating and selecting from multiple candidates, the evolutionary framework can leverage strengths of different objective functions (i.e., different distances) to accelerate the training process and improve the generative performance. Based on the evolutionary framework, E-GAN not only over- come the inherent limitations of these individual adversarial objectives, but it also outperforms other GANs (the WGAN and its improved variant WGAN-GP). Furthermore, when we only keep one parent during each evolutionary step, E-GAN achieves comparable convergence speed in terms of wall- clock time [Fig. 4(middle)]. During training E-GAN, we recorded the selected objective in each step [Fig. 4(right)]. At the beginning of training, the heuristic objective and the least-square objective are selected more frequently than the minimax objective. It may be due to the fact that the minimax objective is hard to provide effective gradients (i.e., vanish- ing gradient) when the discriminator can easily recognize generated samples. Along with the generator approaching con- vergence (after 20K steps), ever more minimax objectives are employed, yet the number of selected heuristic objectives is falling. As aforementioned, the minus JSDs of the heuristic objective may tend to push the generated distribution away from target data distribution and lead to training instability. However, in E-GAN, beyond the heuristic objective, we have other options of objective, which improves the stability at convergence. Furthermore, we discussed the relationship between the sur- vived parents’ number and generative performance. As shown in Table III, both the IS and FID are utilized to evaluate the generative performance of learned generators. First, com- pared with the basic E-GAN (i.e., E-GAN, μ = 1, without GP), adding the GP norm to optimize the discriminator indeed improves generative performance. Then, we preserved multiple parents during the E-GAN training process and measured their scores at the end of training. We can easily observe that the generative performance becomes better accompanied with keeping more parents during the training. This further demon- strates the proposed evolutionary learning paradigm could suppress the unstable and large-scale optimization problems of GANs. Theoretically, if we regard updating and evaluating a child G as an operation and define mutations number as n, keeping p parents will cost O(np) operations in each iteration. Comparing with traditional GANs, our evolution- ary framework would cost more time in each iteration. However, since the evolutionary strategy always preserves Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 10. 930 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 Fig. 5. Experiments of architecture robustness. Different GAN architectures, which correspond to different training challenges, trained with six different GAN methods (or settings). The proposed E-GAN achieved promising performance under all architecture settings. Fig. 6. Generated bedroom images on 128 × 128 LSUN bedrooms. the well-performing off-spring, to achieve the same gen- erative performance, E-GAN usually spends less training steps [Fig. 4(left)]. Overall, keeping one parent during each Fig. 7. Generated human face images on 128 × 128 CelebA dataset. evolutionary step will only slightly reduce the time-efficiency but with better performance [Fig. 4(middle)]. Yet, accompa- nied by increasing p, although the generative performance can be further improved, it will also cost more time on training the E-GAN model. Here, if we regard the parents number p as a hyper-parameter of our algorithm, we found setting its value less than or equal to 4 is a preferable choice. Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 11. WANG et al.: E-GANs 931 Fig. 8. Interpolating in latent space. For selected pairs of generated images from a well-trained E-GAN model, we record their latent vectors z1 and z2. Then, samples between them are generated by linear interpolation between these two vectors. Within this interval, we can easily improve the generative performance by sacrificing affordable computation cost. If we continually increase the number of survivors, the generative performance can only be improved mildly yet largely reduce the training efficiency. In practice, we need to further bal- ance the algorithm efficiency and performance according to different situations. E. LSUN and Architecture Robustness The architecture robustness is another advantage of E-GAN. To demonstrate the training stability of our method, we train different network architectures on the LSUN bedroom dataset [68] and compare with several existing works. In addition to the baseline DCGAN architecture, we choose three additional architectures corresponding to different train- ing challenges: 1) limiting the recognition capability of the discriminator D, i.e., 2-Conv-1-FC LeakyReLU discriminator (abbreviated as weak D); 2) limiting the expression capability of the generator G, i.e., no batchnorm and a constant num- ber of filters in the generator (weak G); and 3) reducing the network capability of the generator and discriminator together, i.e., remove the BN in both the generator G and discrimina- tor D (weak both). For each architecture, we test six different methods (or settings): 1) DCGAN; 2) LSGAN; 3) original WGAN (with weight clipping); 4) WGAN-GP; 5) our E-GAN (μ = 1); and 6) E-GAN (μ = 4). For each method, we used the default configurations recommended in the respec- tive studies (these methods are summarized in [44]) and train each model for 100K iterations. Some generated samples are reported in Fig. 5, and the quantitative results (i.e., FID) are listed in Table IV. Through observation, we find that all of these GAN methods achieved promising performance with the baseline architecture. For DCGAN and LSGAN, when the balance between the generator and discriminator is broken (i.e., only one of them is limited), these two methods have difficulty generating any reasonable samples. Meanwhile, we find the performance of the standard WGAN (with weight clipping) is mostly decided by the generator G. When we limit G’s capability, the generative performance is largely reduced. Regarding the WGAN-GP, we find that the gener- ative performance may mainly depend on the discriminator (or critic). Our E-GAN achieved promising results under all architecture settings. Moreover, we again demonstrated that the model performance is growing with the number of survived parents. Furthermore, we trained E-GAN to generate higher resolu- tion (128×128) bedroom images (Fig. 6). Observing generated images, we demonstrate that E-GAN can be trained to gen- erate diversity and high-quality images from the target data distribution. F. CelebA and Space Continuity Besides the LSUN bedroom dataset, we also train our E-GAN using the aligned faces from CelebA dataset. Since humans excel at identifying facial flaws, generating high-quality human face images is challenging. Similar to Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 12. 932 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 generating bedrooms, we employ the same architectures to generate 128 × 128 RGB human face images (Fig. 7). In addition, given a well-trained generator, we evaluate the performance of the embedding in the latent space of noisy vectors z. In Fig. 8, we first select pairs of generated faces and record their corresponding latent vectors z1 and z2. The two images in one pair have different attributes, such as gen- der, expression, hairstyle, and age. Then, we generate novel samples by linear interpolating between these pairs (i.e., corre- sponding noisy vectors). We find that these generated samples can seamlessly change between these semantically meaningful face attributes. This experiment demonstrates that genera- tor training does not merely memorize training samples but learns a meaningful projection from latent noisy space to face images. Meanwhile, it also shows that the generator trained by E-GAN does not suffer from mode collapse, and shows great space continuity. Overall, during the GAN training process, the training stability is easily influenced by “bad” updating, which could lead the generated samples to low quality or lacking diversity, while the proposed evolutionary mechanism largely avoids undesired updating and promotes the training to an ideal direction. V. CONCLUSION In this paper, we presented an E-GAN framework for train- ing deep generative models. To reduce training difficulties and improve generative performance, we devised an evolutionary algorithm to evolve a population of generators to adapt to the dynamic environment (i.e., the discriminator D). In contrast to conventional GANs, the evolutionary paradigm allows the proposed E-GAN to overcome the limitations of individual adversarial objectives and preserve the well-performing off- spring after each iteration. Experiments showed that E-GAN improves the training stability of GAN models and achieves convincing performance in several image generation tasks. In this paper, we mainly contribute to improving the image generation performance. More generation tasks will be con- sidered in future works, such as video generation [33] and text generation [77]. REFERENCES [1] I. Goodfellow et al., “Generative adversarial nets,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2014, pp. 2672–2680. [2] X. Chen et al., “InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2016, pp. 2172–2180. [3] Z. Gan et al., “Triangle generative adversarial networks,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017, pp. 5253–5262. [4] H. Zhang et al., “StackGAN: Text to photo-realistic image synthesis with stacked generative adversarial networks,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2017, pp. 5908–5916. [5] T. Karras, T. Aila, S. Laine, and J. Lehtinen, “Progressive growing of GANs for improved quality, stability, and variation,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2018, pp. 1–26. [6] T. Miyato, T. Kataoka, M. Koyama, and Y. Yoshida, “Spectral normal- ization for generative adversarial networks,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2018, pp. 1–26. [7] P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros, “Image-to-image translation with conditional adversarial networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2017, pp. 5967–5976. [8] C. Wang, C. Wang, C. Xu, and D. Tao, “Tag disentangled generative adversarial networks for object image re-rendering,” in Proc. 26th Int. Joint Conf. Artif. Intell. (IJCAI), 2017, pp. 2901–2907. [9] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2017, pp. 2242–2251. [10] T.-C. Wang et al., “High-resolution image synthesis and semantic manip- ulation with conditional GANs,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2018, pp. 8798–8807. [11] C. Vondrick, H. Pirsiavash, and A. Torralba, “Generating videos with scene dynamics,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2016, pp. 613–621. [12] C. Finn, I. Goodfellow, and S. Levine, “Unsupervised learning for phys- ical interaction through video prediction,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2016, pp. 64–72. [13] C. Vondrick and A. Torralba, “Generating the future with adversar- ial transformers,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Honolulu, HI, USA, 2017, pp. 2992–3000. [14] Y. Zhang, Z. Gan, and L. Carin, “Generating text via adversarial training,” in Proc. NIPS Workshop Adversarial Training, 2016, pp. 1–6. [15] W. Fedus, I. Goodfellow, and A. M. Dai, “MaskGAN: Better text gener- ation via filling in the___,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2018, pp. 1–17. [16] J. Lu, A. Kannan, J. Yang, D. Parikh, and D. Batra, “Best of both worlds: Transferring knowledge from discriminative learning to a gen- erative visual dialog model,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017, pp. 313–323. [17] X. Lan, A. J. Ma, and P. C. Yuen, “Multi-cue visual tracking using robust feature-level fusion based on joint sparse representation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2014, pp. 1194–1201. [18] X. Chen, C. Xu, X. Yang, L. Song, and D. Tao, “Gated- GAN: Adversarial gated networks for multi-collection style trans- fer,” IEEE Trans. Image Process., vol. 28, no. 2, pp. 546–560, Feb. 2019. [19] S. Arora, R. Ge, Y. Liang, T. Ma, and Y. Zhang, “Generalization and equilibrium in generative adversarial nets (GANs),” in Proc. 34th Int. Conf. Mach. Learn. (ICML), 2017, pp. 224–232. [20] X. Mao et al., “Least squares generative adversarial networks,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2017, pp. 2813–2821. [21] J. Zhao, M. Mathieu, and Y. LeCun, “Energy-based generative adversarial network,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–17. [22] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2016, pp. 1–16. [23] M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein generative adver- sarial networks,” in Proc. 34th Int. Conf. Mach. Learn. (ICML), 2017, pp. 214–223. [24] M. Arjovsky and L. Bottou, “Towards principled methods for training generative adversarial networks,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–17. [25] V. Nagarajan and J. Z. Kolter, “Gradient descent GAN optimization is locally stable,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017, pp. 5591–5600. [26] C. Qian, J.-C. Shi, K. Tang, and Z.-H. Zhou, “Constrained monotone k-submodular function maximization using multiobjective evolutionary algorithms with theoretical guarantee,” IEEE Trans. Evol. Comput., vol. 22, no. 4, pp. 595–608, Aug. 2018. [27] S. He et al., “Cooperative co-evolutionary module identification with application to cancer disease module discovery,” IEEE Trans. Evol. Comput., vol. 20, no. 6, pp. 874–891, Dec. 2016. [28] Y. Du, M.-H. Hsieh, T. Liu, and D. Tao. (2018). The Expressive Power of Parameterized Quantum Circuits. [Online]. Available: https://arxiv.org/abs/1810.11922 [29] L. M. Antonio and C. A. Coello Coello, “Coevolutionary multiobjective evolutionary algorithms: A survey of the state-of-the-art,” IEEE Trans. Evol. Comput., vol. 22, no. 6, pp. 851–865, Dec. 2018. [30] E. Real et al., “Large-scale evolution of image classifiers,” in Proc. 34th Int. Conf. Mach. Learn. (ICML), 2017, pp. 214–223. [31] T. Salimans, J. Ho, X. Chen, and I. Sutskever. (2017). Evolution Strategies as a Scalable Alternative to Reinforcement Learning. [Online]. Available: https://arxiv.org/abs/1703.03864 [32] I. Goodfellow. (2016). NIPS 2016 Tutorial: Generative Adversarial Networks. [Online]. Available: https://arxiv.org/abs/1701.00160 Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 13. WANG et al.: E-GANs 933 [33] S. Tulyakov, M.-Y. Liu, X. Yang, and J. Kautz, “MoCoGAN: Decomposing motion and content for video generation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2018, pp. 1526–1535. [34] Y. Song et al., “VITAL: Visual tracking via adversarial learning,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2018, pp. 8990–8999. [35] X. Lan, S. Zhang, P. C. Yuen, and R. Chellappa, “Learning common and feature-specific patterns: A novel multiple-sparse-representation-based tracker,” IEEE Trans. Image Process., vol. 27, no. 4, pp. 2022–2037, Apr. 2018. [36] X. Lan, A. J. Ma, P. C. Yuen, and R. Chellappa, “Joint sparse representation and robust feature-level fusion for multi-cue visual track- ing,” IEEE Trans. Image Process., vol. 24, no. 12, pp. 5826–5841, Dec. 2015. [37] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial dis- criminative domain adaptation,” in Proc. Comput. Vis. Pattern Recognit. (CVPR), vol. 1, 2017, p. 4. [38] Z. Qiu, Y. Pan, T. Yao, and T. Mei, “Deep semantic hashing with gen- erative adversarial networks,” in Proc. 40th Int. ACM SIGIR Conf. Res. Dev. Inf. Retrieval, Tokyo, Japan, 2017, pp. 225–234. [39] E. Yang, T. Liu, C. Deng, and D. Tao, “Adversarial examples for hamming space search,” IEEE Trans. Cybern., to be published. doi: 10.1109/TCYB.2018.2882908. [40] E. Yang et al., “Pairwise relationship guided deep hashing for cross- modal retrieval,” in Proc. AAAI, 2017, pp. 1618–1625. [41] J. Donahue, P. Krähenbühl, and T. Darrell, “Adversarial feature learn- ing,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–18. [42] V. Dumoulin et al., “Adversarially learned inference,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–18. [43] T. Salimans et al., “Improved techniques for training GANs,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2016, pp. 2234–2242. [44] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. Courville, “Improved training of Wasserstein GANs,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017, pp. 5769–5779. [45] T. D. Nguyen, T. Le, H. Vu, and D. Phung, “Dual discriminator gener- ative adversarial nets,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017, pp. 2667–2677. [46] I. Durugkar, I. Gemp, and S. Mahadevan, “Generative multi-adversarial networks,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–14. [47] B. Neyshabur, S. Bhojanapalli, and A. Chakrabarti. (2017). Stabilizing GAN Training With Multiple Random Projections. [Online]. Available: https://arxiv.org/abs/1705.07831 [48] I. O. Tolstikhin, S. Gelly, O. Bousquet, C.-J. Simon-Gabriel, and B. Schölkopf, “AdaGAN: Boosting generative models,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017, pp. 5424–5433. [49] A. Ghosh, V. Kulharia, V. P. Namboodiri, P. H. S. Torr, and P. K. Dokania, “Multi-agent diverse generative adversarial networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2018, pp. 8513–8521. [50] Q. Hoang, T. D. Nguyen, T. Le, and D. Phung, “MGAN: Training gen- erative adversarial nets with multiple generators,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2018, pp. 1–24. [51] K. A. De Jong, Evolutionary Computation: A Unified Approach. Cambridge, MA, USA: MIT Press, 2006. [52] H.-L. Liu, L. Chen, Q. Zhang, and K. Deb, “Adaptively allocat- ing search effort in challenging many-objective optimization prob- lems,” IEEE Trans. Evol. Comput., vol. 22, no. 3, pp. 433–448, Jun. 2018. [53] Y. Wang, M. Zhou, X. Song, M. Gu, and J. Sun, “Constructing cost-aware functional test-suites using nested differential evolution algo- rithm,” IEEE Trans. Evol. Comput., vol. 22, no. 3, pp. 334–346, Jun. 2018. [54] P. Yang, K. Tang, and X. Yao, “Turning high-dimensional optimization into computationally expensive optimization,” IEEE Trans. Evol. Comput., vol. 22, no. 1, pp. 143–156, Feb. 2018. [55] D.-C. Dang et al., “Escaping local optima using crossover with emer- gent diversity,” IEEE Trans. Evol. Comput., vol. 22, no. 3, pp. 484–497, Jun. 2018. [56] A. E. Eiben and J. Smith, “From evolutionary computation to the evolution of things,” Nature, vol. 521, no. 7553, p. 476, May 2015. [57] R. Miikkulainen et al. (2017). Evolving Deep Neural Networks. [Online]. Available: https://arxiv.org/abs/1703.00548 [58] S. R. Young, D. C. Rose, T. P. Karnowski, S.-H. Lim, and R. M. Patton, “Optimizing deep learning hyper-parameters through an evolutionary algorithm,” in Proc. Workshop Mach. Learn. High Perform. Comput. Environ., 2015, p. 4. [59] X. Yao, “Evolving artificial neural networks,” Proc. IEEE, vol. 87, no. 9, pp. 1423–1447, Sep. 1999. [60] S. Lander and Y. Shang, “EvoAE—A new evolutionary method for training autoencoders for deep learning networks,” in Proc. IEEE 39th Annu. Comput. Softw. Appl. Conf. (COMPSAC), vol. 2, 2015, pp. 790–795. [61] Y. Wang, C. Xu, J. Qiu, C. Xu, and D. Tao, “Towards evolutionary compression,” in Proc. 24th Int. Conf. Knowl. Disc. Data Min., 2018, pp. 2476–2485. [62] R. S. Olson, N. Bartley, R. J. Urbanowicz, and J. H. Moore, “Evaluation of a tree-based pipeline optimization tool for automating data science,” in Proc. Genet. Evol. Comput. Conf., 2016, pp. 485–492. [63] A. Rosales-Pérez, S. García, J. A. Gonzalez, C. A. Coello Coello, and F. Herrera, “An evolutionary multiobjective model and instance selection for support vector machines with Pareto-based ensembles,” IEEE Trans. Evol. Comput., vol. 21, no. 6, pp. 863–877, Dec. 2017. [64] M. Mirza and S. Osindero. (2014). Conditional Generative Adversarial Nets. [Online]. Available: https://arxiv.org/abs/1411.1784 [65] J. Liang, J. Yang, H.-Y. Lee, K. Wang, and M.-H. Yang, “Sub-GAN: An unsupervised generative model via subspaces,” in Proc. Eur. Conf. Comput. Vis. (ECCV), 2018, pp. 698–714. [66] O. Kramer, Machine Learning for Evolution Strategies, vol. 20. Cham, Switzerland: Springer, 2016. [67] A. Krizhevsky, “Learning multiple layers of features from tiny images,” Dept. Comput. Sci., Univ. Toronto, Toronto, ON, Canada, Rep. TR-2009, 2009. [68] F. Yu et al. (2015). LSUN: Construction of a Large-Scale Image Dataset Using Deep Learning With Humans in the Loop. [Online]. Available: https://arxiv.org/abs/1506.03365 [69] Z. Liu, P. Luo, X. Wang, and X. Tang, “Deep learning face attributes in the wild,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2015, pp. 3730–3738. [70] W. Fedus et al., “Many paths to equilibrium: GANs do not need to decrease adivergence at every step,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2018, pp. 1–21. [71] A. Smola, A. Gretton, L. Song, and B. Schölkopf, “A Hilbert space embedding for distributions,” in Proc. Int. Conf. Algorithmic Learn. Theory, 2007, pp. 13–31. [72] A. Gretton, K. M. Borgwardt, M. J. Rasch, B. Schölkopf, and A. Smola, “A kernel two-sample test,” J. Mach. Learn. Res., vol. 13, pp. 723–773, Mar. 2012. [73] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “GANs trained by a two time-scale update rule converge to a local Nash equilibrium,” in Proc. Adv. Neural Inf. Process. Syst. (NIPS), 2017, pp. 6629–6640. [74] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 2818–2826. [75] H. Zhang, I. Goodfellow, D. Metaxas, and A. Odena. (2018). Self-Attention Generative Adversarial Networks. [Online]. Available: https://arxiv.org/abs/1805.08318 [76] L. Metz, B. Poole, D. Pfau, and J. Sohl-Dickstein, “Unrolled genera- tive adversarial networks,” in Proc. Int. Conf. Learn. Represent. (ICLR), 2017, pp. 1–25. [77] N. Duan, D. Tang, P. Chen, and M. Zhou, “Question generation for question answering,” in Proc. Conf. Empir. Methods Nat. Lang. Process. (EMNLP), 2017, pp. 866–874. Chaoyue Wang received the bachelor’s degree from Tianjin University, Tianjin, China, and the Ph.D. degree from the University of Technology Sydney, Ultimo, NSW, Australia. He is a Research Associate of machine learning and computer vision with the School of Computer Science, University of Sydney, Darlington, NSW, Australia. His current research interests include machine learning, deep learning, and generative models. Dr. Wang was a recipient of the Distinguished Student Paper Award in the International Joint Conference on Artificial Intelligence in 2017. Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.
  • 14. 934 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 23, NO. 6, DECEMBER 2019 Chang Xu received the Bachelor of Engineering degree from Tianjin University, Tianjin, China, and the Ph.D. degree from Peking University, Beijing, China. He is a Lecturer of machine learning and com- puter vision with the School of Computer Science, University of Sydney, Darlington, NSW, Australia. He received fellowships from IBM, Armonk, NY, USA, and Baidu, Beijing. His current research interests include machine learning, data mining algo- rithms, and related applications in artificial intelli- gence and computer vision, including multiview learning, multilabel learning, visual search, and face recognition. He has published in prestigious journals and top-tier conferences in the above areas. Xin Yao (F’03) received the B.Sc. degree from the University of Science and Technology of China (USTC), Hefei, China, in 1982, the M.Sc. degree from the North China Institute of Computing Technologies, Beijing, China, in 1985, and the Ph.D. degree from USTC in 1990. He is a Chair Professor of computer science with the Southern University of Science and Technology, Shenzhen, China, and a part-time Professor of com- puter science with the University of Birmingham, Birmingham, U.K. His current research interests include evolutionary computation, ensemble learning, and their applications to software engineering. Dr. Yao was a recipient of the prestigious Royal Society Wolfson Research Merit Award in 2012, the IEEE Computational Intelligence Society (CIS) Evolutionary Computation Pioneer Award in 2013, the 2001 IEEE Donald G. Fink Prize Paper Award for his paper on evolving artificial neural networks, the 2010, 2016, and 2017 IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION Outstanding Paper Awards, the 2011 IEEE TRANSACTIONS ON NEURAL NETWORKS Outstanding Paper Award, and many other best paper awards. He was the President of IEEE CIS from 2014 to 2015 and the Editor-in-Chief of IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION from 2003 to 2008. He was a Distinguished Lecturer of IEEE CIS. Dacheng Tao (F’15) received the B.Eng. degree from the University of Science and Technology of China, Hefei, China, the M.Phil. degree from the Chinese University of Hong Kong, Hong Kong, and the Ph.D. degree from Birkbeck, University of London, London, U.K. He is a Professor of computer science and an ARC Laureate Fellow with the School of Computer Science and the Faculty of Engineering and Information Technologies, and the Inaugural Director of the UBTECH Sydney Artificial Intelligence Centre, University of Sydney, Darlington, NSW, Australia. He mainly applies statistics and mathematics to artificial intelligence and data science. His research results have expounded in one monograph and over 200 publications in prestigious journals and prominent conferences, such as the IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, the IEEE TRANSACTIONS ON IMAGE PROCESSING, the IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, the International Journal of Computer Vision, the Journal of Machine Learning Research, Neural Information Processing Systems, International Conference on Machine Learning, IEEE Conference on Computer Vision and Pattern Recognition, IEEE International Conference on Computer Vision, European Conference on Computer Vision, IEEE International Conference on Data Mining, Association for Computing Machinery’s Special Interest Group on Knowledge Discovery and Data Mining. Mr. Tao was a recipient of several best paper awards, such as the Best Theory/Algorithm Paper Runner Up Award in IEEE ICDM’07, the Best Student Paper Award in IEEE ICDM’13, the Distinguished Paper Award in the 2018 IJCAI, the 2014 ICDM 10-Year Highest-Impact Paper Award, and the 2017 IEEE Signal Processing Society Best Paper Award, the 2015 Australian Scopus-Eureka Prize, and the 2018 IEEE ICDM Research Contributions Award. He is a fellow of the Australian Academy of Science, American Association for the Advancement of Science, International Association for Pattern Recognition, Optical Society of America, Society of Photo-Optical Instrumentation Engineers. Authorized licensed use limited to: Trial User - University of Central Punjab. Downloaded on April 12,2023 at 08:10:42 UTC from IEEE Xplore. Restrictions apply.