
Regional attention generative adversarial network
Author(s) -
Wang Wei,
Hu Haifeng,
Huang Yi,
Ruan Chongchong,
Chen Dihu
Publication year - 2019
Publication title -
electronics letters
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.375
H-Index - 146
eISSN - 1350-911X
pISSN - 0013-5194
DOI - 10.1049/el.2018.8251
Subject(s) - generative grammar , computer science , feature (linguistics) , generative adversarial network , artificial intelligence , adversarial system , mechanism (biology) , quality (philosophy) , machine learning , generative model , pattern recognition (psychology) , image (mathematics) , data mining , philosophy , linguistics , epistemology
In this Letter, the authors propose a novel attention mechanism combined with a classical generative adversarial network (GAN) model to improve the visual quality of generated samples. This novel attention model is named regional attention GAN. The proposed mechanism can build dependencies between the high‐level representations extracted from attention regions of real images and corresponding feature maps of the generative network. By modelling these dependencies, the generative network can be facilitated to learn feature mapping and fit the distribution of real data. They conduct extensive experiments on widely used datasets CIFAR‐10, STL‐10, and CelebA. The quantitative and qualitative performance improvement over state‐of‐the‐art methods demonstrates the validity of the proposed attention mechanism in improving the quality of generated images.