Abstract
Generating images using generative adversarial networks (GAN) is one of the research hotspots. The traditional convolutional GAN treats spatial and channel-wise features equally, which causes the lack of flexibility in extracting features from the discriminator and the generator. To address the issue, we propose generative adversarial networks based on convolutional block attention module (CBAM-GAN) in this paper. CBAM-GAN adds the convolutional block attention module after some convolution operators to adaptively rescale spatial and channel-wise features, which can enhance salient regions and extract more detail features. We apply the network framework of CBAM-GAN to popular GAN models and do an empirical study on MNIST and CIFAR-10 datasets. Experiments show that our model can significantly improve the quality of generated images compared with the traditional convolutional GAN.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Goodfellow, I., et al.: Generative adversarial nets. In: Proceedings of Neural Information Processing Systems (NIPS), pp. 2672–2680 (2014)
Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)
Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever, I., Abbeel., P.: InfoGAN: interpretable representation learning by information maximizing generative adversarial nets. In: Proceedings of Neural Information Processing Systems (NIPS), pp. 2172–2180 (2016)
Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of Computer Vision and Pattern Recognition (CVPR), pp. 4681–4690 (2017)
Zhu, J.-Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of International Conference on Computer Vision (ICCV), pp. 2242–2251 (2017)
Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X., Metaxas, D.: StackGAN: text to photo-realistic image synthesis with stacked generative adversarial networks. In: Proceedings of International Conference on Computer Vision (ICCV), pp. 5908–5916 (2017)
Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)
Zhang, H., Goodfellow, I., Metaxas, D., Odena, A.: Self-attention generative adversarial networks. arXiv preprint arXiv:1805.08318 (2018)
Zhang, Y., Li, K., Li, K., Wang, L., Zhong B., Fu, Y.: Image super-resolution using very deep residual channel attention networks. arXiv preprint arXiv:1807.02758v2 (2018)
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)
Xu, K., et al.: Show, attend and tell: neural image caption generation with visual attention. In: Proceedings of International Conference on Machine Learning (ICML), pp. 2048–2057 (2015)
Yang, Z., He, X., Gao, J., Deng, L., Smola, A.J.: Stacked attention networks for image question answering. In: Proceedings of Computer Vision and Pattern Recognition (CVPR), pp. 21–29 (2016)
Gregor, K., Danihelka, I., Graves, A., Rezende, D.J., Wierstra, D.: Draw: a recurrent neural network for image generation. In: Proceedings of International Conference on Machine Learning (ICML), pp. 1462–1471 (2015)
Cheng, J., Dong, L., Lapata, M.: Long short-term memory-networks for machine reading. arXiv preprint arXiv:1601.06733 (2016)
Parikh, A.P., Täckström, O., Das, D., Uszkoreit, J.: A decomposable attention model for natural language inference. arXiv preprint arXiv:1606.01933 (2016)
Chen, Y.-P., Kalantidis, Y., Li, J.-S., Yan, S.-C., Feng, J.-S.: A2-Nets: double attention networks. arXiv preprint arXiv:1810.11579v1 (2018)
Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1
Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein GAN. arXiv preprint arXiv:1701.07875 (2017)
Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.C.: Improved training of Wasserstein GANs. In: Advances in Proceedings of Neural Information Processing Systems (NIPS), pp. 5767–5777 (2017)
Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Smolley, S.P.: Least squares generative adversarial networks. In: Proceedings of International Conference on Computer Vision (ICCV), pp. 2813–2821 (2017)
The mnist database of handwritten digits. http://yann.lecun.com/exdb/mnist. Accessed 28 Oct 2018
The CIFAR-10 dataset. http://www.cs.toronto.edu/~kriz/cifar.html. Accessed 28 Oct 2018
Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved techniques for training GANs. In: Proceedings of Neural Information Processing Systems (NIPS), pp. 2234–2242 (2016)
Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: Proceedings of Neural Information Processing Systems (NIPS), pp. 6626–6637 (2017)
Huang, G., et al.: An empirical study on evaluation metrics of generative adversarial networks. arXiv preprint arXiv:1806.07755v2 (2018)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)
Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096 (2018)
IMAGENET. http://www.image-net.org. Accessed 28 Oct 2018
Large-scale CelebFaces Attributes (CelebA) Dataset. http://mmlab.ie.cuhk.edu.hk/projects/CelebA.html. Accessed 28 Oct 2018
Li, C.-L., Jiang, Y.-M., Cheslyar, M.: Embedding image through generated intermediate medium using deep convolutional generative adversarial network. CMC: Comput. Mater. Continua 56(2), 313–324 (2018)
Fang, W., Zhang, F., Sheng, V.S., Ding, Y.-W.: A method for improving CNN-based image recognition using DCGAN. CMC: Comput. Mater. Continua 57(1), 167–178 (2018)
Acknowledgments
This research study is supported by the National Natural Science Foundation of China (No. 61672108).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Ma, B., Wang, X., Zhang, H., Li, F., Dan, J. (2019). CBAM-GAN: Generative Adversarial Networks Based on Convolutional Block Attention Module. In: Sun, X., Pan, Z., Bertino, E. (eds) Artificial Intelligence and Security. ICAIS 2019. Lecture Notes in Computer Science(), vol 11632. Springer, Cham. https://doi.org/10.1007/978-3-030-24274-9_20
Download citation
DOI: https://doi.org/10.1007/978-3-030-24274-9_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-24273-2
Online ISBN: 978-3-030-24274-9
eBook Packages: Computer ScienceComputer Science (R0)