
Image translation with dual‐directional generative adversarial networks
Author(s) -
Ruan Congcong,
Yuan Liuchun,
Hu Haifeng,
Chen Dihu
Publication year - 2021
Publication title -
iet computer vision
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.38
H-Index - 37
eISSN - 1751-9640
pISSN - 1751-9632
DOI - 10.1049/cvi2.12011
Subject(s) - computer science , generator (circuit theory) , hallucinating , artificial intelligence , translation (biology) , encoder , image translation , adversarial system , dual (grammatical number) , image (mathematics) , representation (politics) , feature (linguistics) , class (philosophy) , generative grammar , pattern recognition (psychology) , computer vision , power (physics) , art , biochemistry , physics , chemistry , literature , linguistics , philosophy , quantum mechanics , politics , messenger rna , political science , law , gene , operating system
Image‐to‐image translation is a class of vision and graphics problems where the goal is to learn the mapping between input images and output images. However, due to the unstable training and limited training samples, many existing GAN‐based works have difficulty in producing photo‐realistic images. Herein, dual‐directional generative adversarial networks are proposed, which consist of four adversarial networks, to produce images of high perceptual quality. In this framework, self‐reconstruction strategy is used to construct auxiliary sub‐networks, which impose more effective constraints on encoder‐generator pairs. Using this idea, this model can increase the use ratio of paired data conditioned on the same dataset and obtain well‐trained encoder‐generator pairs with the help of the proposed cross‐network skip connections. Moreover, the proposed framework not only produces realistic images but also addresses the problem where condition GAN produces sharp images containing many small, hallucinated objects. Training on multiple supervised datasets, convincing evidences are shown to prove that this model can achieve compelling results by latently learning a common feature representation. Qualitative and quantitative comparisons against other methods, demonstrate the effectiveness and superiority of the method.