Abstract
The surging popularity of generative adversarial networks (GANs) has ignited a wave of innovation in the realm of computer vision, a highly explored subfield of deep learning. GANs are revolutionizing the area of machine learning because they use a game-based training technique. This is in contrast to traditional approaches to machine learning, which center on feature learning and picture production. Several subfields of computer vision have seen tremendous progress thanks to the integration of numerous processing approaches, including image processing, dynamic processing, text, audio, and video processing, as well as generalized generative adversarial networks (GANs). Nevertheless, despite the fact that GANs have made great progress, they still offer promise that has not been fully realized and space for additional development. GANs have a wide range of applications within computer vision, including data augmentation, displacement recording, dynamic modeling, and image processing. This article digs into recent advances made by GAN researchers working in the realm of AI-based security and defense and discusses their accomplishments. In particular, we investigate how well image optimization, image processing, and image stabilization are incorporated into GAN-driven picture training. We want to achieve our goal of providing a complete overview of the present status of GAN research by carefully evaluating research articles that have been subjected to peer review.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
This statement is Not applicable to the current study.
References
Adler J, Lunz S (2018) Banach wasserstein gan. Adv Neural Inf Process Syst 31:1049
Antipov G, Baccouche M, Dugelay JL (2017) Face aging with conditional generative adversarial networks. In: 2017 ieee international conference on image processing (ICIP), pp 2089–2093. https://doi.org/10.1109/ICIP.2017.8296650
Arjovsky M, Bottou L (2017) Towards principled methods for training generative adversarial networks. https://doi.org/10.48550/ARXIV.1701.04862, arXiv:1701.04862
Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: Precup D, Teh YW (eds) Proceedings of the 34th international conference on machine learning, proceedings of machine learning research, vol. 70 PMLR, pp 214–223. https://proceedings.mlr.press/v70/arjovsky17a.html
Arora R, Zhang L, Pecht M (2020) Generative adversarial networks for electrical prognostics. Progn Health Manag 1(2):15–23
Baker J (1975) The dragon system-an overview. IEEE Trans Acoust Speech Signal Process 23(1):24–29. https://doi.org/10.1109/TASSP.1975.1162650
Barnes C, Shechtman E, Finkelstein A et al (2009) PatchMatch: a randomized correspondence algorithm for structural image editing. ACM Trans Graph (Proc SIGGRAPH) 28(3):24
Barua S, Erfani SM, Bailey J (2019) FCC-GAN: a fully connected and convolutional net architecture for gans. CoRR abs/1905.02417. arXiv:1905.02417
Bengio Y, Courville A, Vincent P (2013) Representation learning: a review and new perspectives. IEEE Trans Pattern Anal Mach Intell 35(8):1798–1828. https://doi.org/10.1109/TPAMI.2013.50
Berthelot D, Schumm T, Metz L (2017) Began: boundary equilibrium generative adversarial networks. https://doi.org/10.48550/ARXIV.1703.10717, arXiv:1703.10717
Bharath K (2022) Implementing conditional generative adversarial networks. https://blog.paperspace.com/conditional-generative-adversarial-networks/. Accessed 24 Oct 2023
Brock A, Lim T, Ritchie JM et al (2016) Neural photo editing with introspective adversarial networks. https://doi.org/10.48550/ARXIV.1609.07093, arXiv:1609.07093
Brock A, Donahue J, Simonyan K (2018) Large scale gan training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096
Brophy E, Wang Z, Ward TE (2019) Quick and easy time series generation with established image-based gans. https://doi.org/10.48550/ARXIV.1902.05624, arXiv:1902.05624
Can C (2015) Research of the coordination control of the intersection based on the cooperative vehicle-infrastructure system. MS thesis, Department Transport, Beijing Jiaotong University, Beijing, China
Chen S (2019) Techniques in self-attention generative adversarial networks. https://pub.towardsai.net/techniques-in-self-attention-generative-adversarial-networks-22f735b22dfb
Chen X, Duan Y, Houthooft R, et al (2016) Infogan: interpretable representation learning by information maximizing generative adversarial nets. https://doi.org/10.48550/ARXIV.1606.03657, arXiv:1606.03657
Chen Y, Wu J, Cui M (2018) Automatic classification and detection of oranges based on computer vision. In: 2018 IEEE 4th international conference on computer and communications (ICCC), pp 1551–1556. https://doi.org/10.1109/CompComm.2018.8780680
Cherednik I, Philipp I (2018) DAHA and plane curve singularities. Algebr Geom Topol 18(1):333–385. https://doi.org/10.2140/agt.2018.18.333
Creswell A, White T, Dumoulin V et al (2018) Generative adversarial networks: an overview. IEEE Signal Process Mag 35(1):53–65
Deng J, Dong W, Socher R et al (2009) Imagenet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, pp 248–255. https://doi.org/10.1109/CVPR.2009.5206848
Denton E, Chintala S, Szlam A et al (2015) Deep generative image models using a laplacian pyramid of adversarial networks. https://doi.org/10.48550/ARXIV.1506.05751, arXiv:1506.05751
Dong C, Loy CC, He K et al (2014) Learning a deep convolutional network for image super-resolution. In: European conference on computer vision, Springer, pp 184–199
Dong J, Rafayelyan M, Krzakala F et al (2020) Optical reservoir computing using multiple light scattering for chaotic systems prediction. IEEE J Sel Top Quantum Electron 26(1):1–12. https://doi.org/10.1109/jstqe.2019.2936281
Eskimez SE, Koishida K (2019) Speech super resolution generative adversarial network. In: ICASSP 2019 - 2019 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp 3717–3721. https://doi.org/10.1109/ICASSP.2019.8682215
Frid-Adar M, Klang E, Amitai M et al (2018) Synthetic data augmentation using gan for improved liver lesion classification. In: 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018), pp 289–293. https://doi.org/10.1109/ISBI.2018.8363576
Goldberger J, Ben-Reuven E (2016) Training deep neural-networks using a noise adaptation layer. In: International conference on learning representations
Gong M, Xu Y, Li C et al (2019a) Twin auxiliary classifiers gan. arXiv:1907.02690
Gong M, Xu Y, Li C et al (2019) Twin auxiliary classifiers GAN. Adv Neural Inf Process Syst 32:1328–1337
Gong X, Chang S, Jiang Y et al (2019c) Autogan: neural architecture search for generative adversarial networks. https://doi.org/10.48550/ARXIV.1908.03835, arXiv:1908.03835
Goodfellow I (2017) Nips 2016 tutorial: generative adversarial networks. arXiv:1701.00160
Goodfellow I, Pouget-Abadie J, Mirza M et al (2020) Generative adversarial networks. Commun ACM 63(11):139–144
Goodfellow IJ, Pouget-Abadie J, Mirza M et al (2014) Generative adversarial networks. https://doi.org/10.48550/ARXIV.1406.2661, arXiv:1406.2661
Greenspan H, van Ginneken B, Summers RM (2016) Guest editorial deep learning in medical imaging: overview and future promise of an exciting new technique. IEEE Trans Med Imaging 35(5):1153–1159. https://doi.org/10.1109/TMI.2016.2553401
Guerriero P, Orcioni S, Matacena I et al (2020) A gan based bidirectional switch for matrix converter applications. In: 2020 international symposium on power electronics, electrical drives, automation and motion (SPEEDAM), pp 375–380. https://doi.org/10.1109/SPEEDAM48782.2020.9161876
Gulrajani I, Ahmed F, Arjovsky M et al (2017) Improved training of wasserstein gans. https://doi.org/10.48550/ARXIV.1704.00028, arXiv:1704.00028
He D, Chen W, Wang L et al (2014) A game-theoretic machine learning approach for revenue maximization in sponsored search. arXiv:1406.0728
He K, Zhang X, Ren S et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Hinton GE, Osindero S, Teh YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554. https://doi.org/10.1162/neco.2006.18.7.1527
Hitawala S (2018) Comparative study on generative adversarial networks. arXiv:1801.04271
Hou L, Cao Q, Shen H et al (2022) Conditional gans with auxiliary discriminative classifier. arXiv:2107.10060
Hu L, Ou J, Huang J et al (2020) A review of research on traffic conflicts based on intelligent vehicles. IEEE Access 8:24471–24483. https://doi.org/10.1109/ACCESS.2020.2970164
Huang JB, Kang SB, Ahuja N et al (2014) Image completion using planar structure guidance. ACM Trans Graph 33(4):1–10
Hui J (2018) Gan — self-attention generative adversarial networks (SAGAN). https://jonathan-hui.medium.com/gan-self-attention-generative-adversarial-networks-sagan-923fccde790c
Iizuka S, Simo-Serra E, Ishikawa H (2017) Globally and locally consistent image completion. ACM Trans Graph (ToG) 36(4):1–14
Im DJ, Ma H, Taylor G et al (2018) Quantitatively evaluating gans with divergences proposed for training. https://doi.org/10.48550/ARXIV.1803.01045, arXiv:1803.01045
Isola P, Zhu JY, Zhou T, et al (2016) Image-to-image translation with conditional adversarial networks. https://doi.org/10.48550/ARXIV.1611.07004, arXiv:1611.07004
Jetchev N, Bergmann U, Vollgraf R (2016) Texture synthesis with spatial generative adversarial networks. https://doi.org/10.48550/ARXIV.1611.08207, arXiv:1611.08207
Jha D (2018) Not just another GAN paper — SAGAN. https://towardsdatascience.com/not-just-another-gan-paper-sagan-96e649f01a6b
Jiang L, Zhang H, Cai Z (2009) A novel bayes model: hidden naive bayes. IEEE Trans Knowl Data Eng 21(10):1361–1371. https://doi.org/10.1109/TKDE.2008.234
Jin L, Tan F, Jiang S (2020) Generative adversarial network technologies and applications in computer vision. Comput Intell Neurosci 1459:107. https://doi.org/10.1155/2020/1459107
Jindal I, Nokleby M, Chen X (2017) Learning deep networks from noisy labels with dropout regularization. https://doi.org/10.48550/ARXIV.1705.03419, arXiv:1705.03419
Jolicoeur-Martineau A (2018) The relativistic discriminator: a key element missing from standard gan. https://doi.org/10.48550/ARXIV.1807.00734, arXiv:1807.00734
Jost Z (2019) Infogan — generative adversarial networks part III. https://towardsdatascience.com/infogan-generative-adversarial-networks-part-iii-380c0c6712cd
Kahembwe E, Ramamoorthy S (2020) Lower dimensional kernels for video discriminators. Neural Netw 132:506–520. https://doi.org/10.1016/j.neunet.2020.09.016
Kaneko T, Ushiku Y, Harada T (2018) Label-noise robust generative adversarial networks. https://doi.org/10.48550/ARXIV.1811.11165, arXiv:1811.11165
Karnewar A, Wang O (2019) Msg-gan: multi-scale gradients for generative adversarial networks. https://doi.org/10.48550/ARXIV.1903.06048, arXiv:1903.06048
Karras T, Aila T, Laine S et al (2017) Progressive growing of gans for improved quality, stability, and variation. https://doi.org/10.48550/ARXIV.1710.10196, arXiv:1710.10196
Karras T, Laine S, Aila T (2021) A style-based generator architecture for generative adversarial networks. IEEE Trans Pattern Anal Mach Intell 43(12):4217–4228. https://doi.org/10.1109/TPAMI.2020.2970919
Kim T, Cha M, Kim H et al (2017) Learning to discover cross-domain relations with generative adversarial networks. https://doi.org/10.48550/ARXIV.1703.05192, arXiv:1703.05192
Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. https://doi.org/10.48550/ARXIV.1412.6980, arXiv:1412.6980
Kingma DP, Welling M (2013) Auto-encoding variational bayes. https://doi.org/10.48550/ARXIV.1312.6114, arXiv:1312.6114
Kodali N, Abernethy J, Hays J et al (2017) On convergence and stability of gans. https://doi.org/10.48550/ARXIV.1705.07215, arXiv:1705.07215
Kurutach T, Tamar A, Yang G et al (2018) Learning plannable representations with causal infogan. arXiv:1807.09341
Langr J, Bok V (2019) GANs in action. Manning Publications, New York, NY
LeCun Y, Boser B, Denker JS et al (1989) Backpropagation applied to handwritten zip code recognition. Neural Comput 1(4):541–551. https://doi.org/10.1162/neco.1989.1.4.541
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444. https://doi.org/10.1038/nature14539
Ledig C, Theis L, Huszár F et al (2017) Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4681–4690
Li K, Lee CH (2015) A deep neural network approach to speech bandwidth expansion. In: 2015 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp 4395–4399. https://doi.org/10.1109/ICASSP.2015.7178801
Li W, Wang Z, Li J et al (2019) Semi-supervised learning based on generative adversarial network: a comparison between good gan and bad gan approach. arXiv:1905.06484
Lim JH, Ye JC (2017) Geometric gan. https://doi.org/10.48550/ARXIV.1705.02894, arXiv:1705.02894
Liu S, Sun Y, Zhu D et al (2017) Face aging with contextual generative adversarial nets. In: Proceedings of the 25th ACM international conference on multimedia. Association for computing machinery, New York, NY, USA, MM ’17, p 82–90. https://doi.org/10.1145/3123266.3123431
Long J, Shelhamer E, Darrell T (2014) Fully convolutional networks for semantic segmentation. https://doi.org/10.48550/ARXIV.1411.4038, arXiv:1411.4038
Lucas A, López-Tapia S, Molina R et al (2019) Generative adversarial networks and perceptual losses for video super-resolution. IEEE Trans Image Process 28(7):3312–3327. https://doi.org/10.1109/TIP.2019.2895768
Lutz S, Amplianitis K, Smolic A (2018) Alphagan: generative adversarial networks for natural image matting. https://doi.org/10.48550/ARXIV.1807.10088, arXiv:1807.10088
Ma J, Zhou Z, Wang B et al (2019) Hard ship detection via generative adversarial networks. In: 2019 Chinese control and decision conference (CCDC), pp 3961–3965. https://doi.org/10.1109/CCDC.2019.8833176
Mertes S, Schiller D, Lingenfelser F et al (2023) Intercategorical label interpolation for emotional face generation with conditional generative adversarial networks. Communications in computer and information science. Springer Nature Switzerland, Cham, pp 67–87
Mirza M, Osindero S (2014) Conditional generative adversarial nets. https://doi.org/10.48550/ARXIV.1411.1784, arXiv:1411.1784
Miyato T, Koyama M (2018) cGANs with projection discriminator. https://doi.org/10.48550/ARXIV.1802.05637, arXiv:1802.05637
Mu F (2019) Wasserstein-BiGAN: wasserstein BiGAN (bidirectional GAN trained using wasserstein distance). https://github.com/fmu2/Wasserstein-BiGAN
Odena A (2016) Semi-supervised learning with generative adversarial networks. https://doi.org/10.48550/ARXIV.1606.01583, arXiv:1606.01583
Odena A, Olah C, Shlens J (2016) Conditional image synthesis with auxiliary classifier gans. https://doi.org/10.48550/ARXIV.1610.09585, arXiv:1610.09585
Oliver A (2018) InfoGAN · depth first learning. https://www.depthfirstlearning.com/2018/InfoGAN
Pan Z, Yu W, Yi X et al (2019) Recent progress on generative adversarial networks (gans): a survey. IEEE Access 7:36322–36333. https://doi.org/10.1109/ACCESS.2019.2905015
Patrini G, Rozza A, Menon A et al (2016) Making deep neural networks robust to label noise: a loss correction approach. https://doi.org/10.48550/ARXIV.1609.03683, arXiv:1609.03683
Pu Y, Gan Z, Henao R et al (2016) Variational autoencoder for deep learning of images, labels and captions. In: Lee D, Sugiyama M, Luxburg U et al (eds) Advances in neural information processing systems, vol 29. New York, Curran Associates Inc.
Rabiner L (1989) A tutorial on hidden markov models and selected applications in speech recognition. In: Proceedings IEEE, 77(2):257–286. https://doi.org/10.1109/5.18626
Radford A, Metz L, Chintala S (2015) Unsupervised representation learning with deep convolutional generative adversarial networks. https://doi.org/10.48550/ARXIV.1511.06434, arXiv:1511.06434
Reed S, Akata Z, Yan X et al (2016) Generative adversarial text to image synthesis. In: International conference on machine learning, PMLR, pp 1060–1069
Ren S, He K, Girshick R et al (2017) Faster r-cnn: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137–1149. https://doi.org/10.1109/TPAMI.2016.2577031
Ren S, He K, Girshick R et al (2017) Object detection networks on convolutional feature maps. IEEE Trans Pattern Anal Mach Intell 39(7):1476–1481. https://doi.org/10.1109/TPAMI.2016.2601099
Salakhutdinov R, Larochelle H (2010) Efficient learning of deep Boltzmann machines. In: Teh YW, Titterington M (eds.), proceedings of the thirteenth international conference on artificial intelligence and statistics, proceedings of machine learning research, vol. 9. PMLR, Chia Laguna Resort, Sardinia, Italy, pp 693–700. https://proceedings.mlr.press/v9/salakhutdinov10a.html
Salimans T, Zhang H, Radford A et al (2018) Improving gans using optimal transport. https://doi.org/10.48550/ARXIV.1803.05573, arXiv:1803.05573
Shah H (2018) Using bidirectional generative adversarial networks to estimate value-at-risk for market risk. https://towardsdatascience.com/using-bidirectional-generative-adversarial-networks-to-estimate-value-at-risk-for-market-risk-c3dffbbde8dd
Shelhamer E, Long J, Darrell T (2017) Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 39(4):640–651. https://doi.org/10.1109/TPAMI.2016.2572683
Shi T, Yuan Y, Fan C et al (2019) Face-to-parameter translation for game character auto-creation. In: 2019 IEEE/CVF international conference on computer vision (ICCV), pp 161–170. https://doi.org/10.1109/ICCV.2019.00025
Shi W, Caballero J, Huszár F et al (2016) Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 1874–1883. https://doi.org/10.1109/CVPR.2016.207
Sukhbaatar S, Bruna J, Paluri M et al (2014) Training convolutional networks with noisy labels. https://doi.org/10.48550/ARXIV.1406.2080, arXiv:1406.2080
takuhirok (2021) Github - takuhirok/rGAN: rgan: label-noise robust generative adversarial networks. https://github.com/takuhirok/rGAN
Tran D, Ranganath R, Blei DM (2017) Hierarchical implicit models and likelihood-free variational inference. https://doi.org/10.48550/ARXIV.1702.08896, arXiv:1702.08896
Tulyakov S, Liu MY, Yang X et al (2017) Mocogan: decomposing motion and content for video generation. https://doi.org/10.48550/ARXIV.1707.04993, arXiv:1707.04993
Vaidya K (2021) Implementation of semi-supervised generative adversarial networks in Keras. https://towardsdatascience.com/implementation-of-semi-supervised-generative-adversarial-networks-in-keras-195a1b2c3ea6
Volkhonskiy D, Nazarov I, Borisenko B et al (2017) Steganographic generative adversarial networks. In: Proceedings of NIPS 2016 workshop on adversarial training
Wang X, Girshick R, Gupta A et al (2017) Non-local neural networks. https://doi.org/10.48550/ARXIV.1711.07971, arXiv:1711.07971
Wang Z, She Q, Ward TE (2021) Generative adversarial networks in computer vision: a survey and taxonomy. ACM Comput Surv. https://doi.org/10.1145/3439723
Weng Y, Zhou H (2019) Data augmentation computing model based on generative adversarial network. IEEE Access 7:64223–64233. https://doi.org/10.1109/ACCESS.2019.2917207
Wissen D (2022) GAN and its applications: everything you need to know - daten & wissen. https://datenwissen.com/blog/gan-applications/. Accessed 24 Oct 2023
Xu N, Price B, Cohen S et al (2017) Deep image matting. arXiv:1703.03872
Yang J, Price B, Cohen S et al (2016) Object contour detection with a fully convolutional encoder-decoder network. https://doi.org/10.48550/ARXIV.1603.04530, arXiv:1603.04530
Yang TY (2020) Introduction to GANs. https://towardsdatascience.com/introduction-to-gans-877dd689cac1
Yi Z, Zhang H, Tan P et al (2017) Dualgan: unsupervised dual learning for image-to-image translation. https://doi.org/10.48550/ARXIV.1704.02510, arXiv:1704.02510
Zhang H, Xu T, Li H et al (2017) Stackgan: text to photo-realistic image synthesis with stacked generative adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 5908–5916
Zhang H, Goodfellow I, Metaxas D et al (2018a) Self-attention generative adversarial networks. https://doi.org/10.48550/ARXIV.1805.08318, arXiv:1805.08318
Zhang R, Isola P, Efros AA et al (2018b) The unreasonable effectiveness of deep features as a perceptual metric. In: 2018 IEEE/CVF conference on computer vision and pattern recognition, pp 586–595. https://doi.org/10.1109/CVPR.2018.00068
Zhu JY, Park T, Isola P et al (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. https://doi.org/10.48550/ARXIV.1703.10593, arXiv:1703.10593
Funding
This research received no external funding.
Author information
Authors and Affiliations
Contributions
All authors have diligently reviewed and contributed equally to this study.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no Conflict of interest to declare that are relevant to the content of this article.
Informed consent
Informed consent procedures were Not applicable to this research.
Human participants
This research did not involve any human participants or animals.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Zala, K., Thumar, D., Thakkar, H.K. et al. A survey and identification of generative adversarial network technology-based architectural variants and applications in computer vision. Int J Syst Assur Eng Manag 15, 4594–4615 (2024). https://doi.org/10.1007/s13198-024-02478-6
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13198-024-02478-6