Open Access
ESA‐CycleGAN: Edge feature and self‐attention based cycle‐consistent generative adversarial network for style transfer
Author(s) -
Wang Li,
Wang Lidan,
Chen Shubai
Publication year - 2022
Publication title -
iet image processing
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.401
H-Index - 45
eISSN - 1751-9667
pISSN - 1751-9659
DOI - 10.1049/ipr2.12342
Subject(s) - discriminator , computer science , artificial intelligence , generative adversarial network , generator (circuit theory) , feature extraction , feature (linguistics) , enhanced data rates for gsm evolution , rendering (computer graphics) , generative grammar , pattern recognition (psychology) , image (mathematics) , computer vision , power (physics) , telecommunications , linguistics , physics , philosophy , quantum mechanics , detector
Abstract Nowadays, style transfer is used in a wide range of commercial applications, such as image beautification, film rendering etc. However, many existing methods of style transfer suffer from loss of details and poor overall visual effect. To address these problems, an edge feature and self‐attention based cycle‐consistent generative adversarial network (ESA‐CycleGAN) is proposed. The model architecture consists of a generator, a discriminator, and an edge feature extraction network. Both the generator and the discriminator contain a self‐attention module to capture global features of the image. The edge feature extraction network extracts the edge of the original image and feeds it into the network together with the original image, thereby allowing better processing of details. Besides, a perceptual loss term is added to optimize the network, resulting in better perceptual results. ESA‐CycleGAN is applied on four datasets, respectively. The experimental results show that the authors’ computed final IS and FID values have good results compared to the results of several other existing models, indicating the superiority of the model in style transfer, which can better preserve the details of the original images with better image quality.