A novel interpolation consistency for bad generative adversarial networks (IC-BGAN) | Multimedia Tools and Applications Skip to main content

Advertisement

Log in

A novel interpolation consistency for bad generative adversarial networks (IC-BGAN)

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Semi-supervised learning techniques utilize both labeled and unlabeled images to enhance classification performance in scenarios where labeled images are limited. However, challenges such as integrating unlabeled images with incorrect pseudo-labels, determining appropriate thresholds for the pseudo-labels, and label prediction fluctuations on low-confidence unlabeled images, hinder the effectiveness of existing methods. This research introduces a novel framework named Interpolation Consistency for Bad Generative Adversarial Networks (IC-BGAN) that utilizes a new loss function. The proposed model combines bad adversarial training, fusion techniques, and regularization to address the limitations of semi-supervised learning. IC-BGAN creates three types of image augmentations and label consistency regularization in interpolation of bad fake images, real and bad fake images, and unlabeled images. It demonstrates linear interpolation behavior, reducing fluctuations in predictions, improving stability, and facilitating the identification of decision boundaries in low-density areas. The regularization techniques boost the discriminative capability of the classifier and discriminator, and send a better signal to the bad generator. This improves the generalization and the generation of diverse inter-class fake images as support vectors with information near the true decision boundary, which helps to correct the pseudo-labeling of unlabeled images. The proposed approach achieves notable improvements in error rate from 2.87 to 1.47 on the Modified National Institute of Standards and Technology (MNIST) dataset, 3.59 to 3.13 on the Street View House Numbers (SVHN) dataset, and 12.13 to 9.59 on the Canadian Institute for Advanced Research, 10 classes (CIFAR-10) dataset using 1000 labeled training images. Additionally, it reduces the error rate from 22.11 to 18.40 on the CINIC-10 dataset when using 700 labeled images per class. The experiments demonstrate the IC-BGAN framework outperforms existing semi-supervised methods, providing a more accurate classification solution with smoother class label estimates, especially for low-confidence unlabeled images.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Algorithm 1
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19
Fig. 20
Fig. 21
Fig. 22

Similar content being viewed by others

Data availability

Data will be made available on request.

References

  1. Lim JY, Lim KM, Lee CP, Tan YX (2023) SCL: Self-supervised contrastive learning for few-shot image classification. Neural Netw 165:19–30

    Article  Google Scholar 

  2. Alarood AA, Faheem M, Al-Khasawneh MA, Alzahrani AI, Alshdadi AA (2023) Secure medical image transmission using deep neural network in e-health applications. Healthcare Technology Letters 10:87–98

    Article  Google Scholar 

  3. Niu X, Ma W (2023) Semi-supervised classifier ensemble model for high-dimensional data. Inf Sci 643:119203

    Article  Google Scholar 

  4. Wang Y, Song D, Wang W, Rao S, Wang X, Wang M (2022) Self-supervised learning and semi-supervised learning for multi-sequence medical image classification. Neurocomputing 513:383–394

    Article  Google Scholar 

  5. Yang X, Song Z, King I, Xu Z (2022) A survey on deep semi-supervised learning. IEEE Transactions on Knowledge and Data Engineering 35(9):8934–8954

    Article  Google Scholar 

  6. Li J (2022) NaNG-ST: A natural neighborhood graph-based self-training method for semi-supervised classification. Neurocomputing 514:268–284

    Article  Google Scholar 

  7. Taha K (2023) Semi-supervised and un-supervised clustering: A review and experimental evaluation. Information Systems 144:102178

    Article  Google Scholar 

  8. Wang Y, Li D, Li X, Yang M (2021) PC-GAIN: Pseudo-label conditional generative adversarial imputation networks for incomplete data. Neural Netw 141:395–403

    Article  Google Scholar 

  9. Bae J, Lee M, Kim SB (2022) Safe semi-supervised learning using a Bayesian neural network. Inf Sci 612:453–464

    Article  Google Scholar 

  10. Ke B, Lu H, You C, Zhu W, Xie L, Yao Y (2023) A semi-supervised medical image classification method based on combined pseudo-labeling and distance metric consistency. Multimed Tools Appl 83(11):1–19

    Article  Google Scholar 

  11. Zhao J, Lan L, Huang D, Ren J, Yang W (2022) Heterogeneous pseudo-supervised learning for few-shot person re-identification. Neural Netw 154:521–537

    Article  Google Scholar 

  12. Meel P, Vishwakarma DK (2021) A temporal ensembling based semi-supervised ConvNet for the detection of fake news articles. Expert Syst Appl 177:115002

    Article  Google Scholar 

  13. Tarvainen A, Valpola H (2017) Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results, Adv In Neural Inf Proc Sys 30

  14. Gan Y, Zhu H, Guo W, Xu G, Zou G (2022) Deep semi-supervised learning with contrastive learning and partial label propagation for image data. Knowl-Based Syst 245:108602

    Article  Google Scholar 

  15. Yang M, Ling J, Chen J, Feng M, Yang J (2023) Discriminative semi-supervised learning via deep and dictionary representation for image classification. Pattern Recogn 140:109521

    Article  Google Scholar 

  16. Sohn K, Berthelot D, Carlini N, Zhang Z, Zhang H, Raffel CA, Cubuk ED, Kurakin A, Li C-L (2020) Fixmatch: Simplifying semi-supervised learning with consistency and confidence. Adv Neural Inf Process Syst 33:596–608

    Google Scholar 

  17. Feng Z, Zhou Q, Gu Q, Tan X, Cheng G, Lu X, Shi J, Ma L (2022) Dmt: Dynamic mutual training for semi-supervised learning. Pattern Recogn 130:108777

    Article  Google Scholar 

  18. Huo X, Zeng X, Wu S, Shen W, Wong H-S (2023) Collaborative Learning with Unreliability Adaptation for Semi-Supervised Image Classification. Pattern Recogn 133:109032

    Article  Google Scholar 

  19. Zhou S, Tian S, Yu L, Wu W, Zhang D, Peng Z, Zhou Z (2023) FixMatch-LS: Semi-supervised skin lesion classification with label smoothing. Biomed Signal Process Control 84:104709

    Article  Google Scholar 

  20. Zhang B, Wang Y, Hou W, Wu H, Wang J, Okumura M, Shinozaki T (2021) Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling. Adv Neural Inf Process Syst 34:18408–18419

    Google Scholar 

  21. Miyato T, Maeda S-I, Koyama M, Ishii S (2018) Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Trans Pattern Anal Mach Intell 41:1979–1993

    Article  Google Scholar 

  22. Park S, Park J, Shin S-J, Moon I-C (2018) Adversarial dropout for supervised and semi-supervised learning. In: Proceedings of the AAAI conference on artificial intelligence 32(1)

  23. Luo Y, Zhu J, Li M, Ren Y, Zhang B (2018) Smooth neighbors on teacher graphs for semi-supervised learning. In: Proceedings of the IEEE conference on computer vision pattern recognition, pp 8896–8905

  24. Verma V, Kawaguchi K, Lamb A, Kannala J, Solin A, Bengio Y, Lopez-Paz D (2022) Interpolation consistency training for semi-supervised learning. Neural Netw 145:90–106

    Article  Google Scholar 

  25. Peng Z, Tian S, Yu L, Zhang D, Wu W, Zhou S (2023) Semi-supervised medical image classification with adaptive threshold pseudo-labeling and unreliable sample contrastive loss. Biomed Signal Process Control 79:104142

    Article  Google Scholar 

  26. Boschini M, Buzzega P, Bonicelli L, Porrello A, Calderara S (2022) Continual semi-supervised learning through contrastive interpolation consistency. Pattern Recogn Lett 162:9–14

    Article  Google Scholar 

  27. Liu X, Shen F, Zhao J, Nie C (2024) RandoMix: a mixed sample data augmentation method with multiple mixed modes, Multimedia Tools Appl, pp 1–17

  28. Jo W, Kim D (2022) OBGAN: Minority oversampling near borderline with generative adversarial networks. Expert Syst Appl 197:116694

    Article  Google Scholar 

  29. He R, Li X, Chen G, Chen G, Liu Y (2020) Generative adversarial network-based semi-supervised learning for real-time risk warning of process industries. Expert Syst Appl 150:113244

    Article  Google Scholar 

  30. Wang L, Sun Y, Wang Z (2022) CCS-GAN: a semi-supervised generative adversarial network for image classification. Vis Comput 38:2009–2021

    Article  Google Scholar 

  31. Dong J, Lin T (2019) Margingan: Adversarial training in semi-supervised learning, Adv Neural Inf Process Syst 32

  32. Lu S, Zhang Z, Yan Z, Wang Y, Cheng T, Zhou R, Yang G (2023) Mutually aided uncertainty incorporated Dual Consistency Regularization with Pseudo Label for Semi-Supervised Medical Image Segmentation. Neurocomputing 584:126411

    Article  Google Scholar 

  33. Liang D, Yang F, Zhang T, Yang P (2018) Understanding mixup training methods. IEEE Access 6:58774–58783

    Article  Google Scholar 

  34. Sharma P, Kumar M, Sharma HK, Biju SM (2024) Generative adversarial networks (GANs): Introduction, Taxonomy, Variants, Limitations, and Applications, Multimedia Tools and Applications 1–48

  35. Ligthart A, Catal C, Tekinerdogan B (2021) Analyzing the effectiveness of semi-supervised learning approaches for opinion spam classification. Appl Soft Comput 101:107023

    Article  Google Scholar 

  36. Donyavi Z, Asadi S (2020) Diverse training dataset generation based on multi-objective optimization for semi-supervised classification. Pattern Recogn 108:107543

    Article  Google Scholar 

  37. Fu S, Chen J, Chen D, He C (2023) CNNs/ViTs-CNNs/ViTs: Mutual distillation for unsupervised domain adaptation. Inf Sci 622:83–97

    Article  Google Scholar 

  38. Ding Y, Sheng L, Liang J, Zheng A, He R (2023) ProxyMix: Proxy-based mixup training with label refinery for source-free domain adaptation. Neural Networks. 1(167):92–103

    Article  Google Scholar 

  39. Pavate A, Bansode R, Srinivasu PN, Shafi J, Choi J, Ijaz MF (2023) Associative discussion among generating adversarial samples using evolutionary algorithm and samples generated using gAN, IEEE Access

  40. Laine S, Aila T (2017) Temporal en sembling for semi-supervised learning. In: 5th International Conference on Learning Representations, ICLR 2017

  41. Zhu S, Zhang Y, Feng Y (2023) GW-net: An efficient grad-CAM consistency neural network with a weakening of random erasing features for semi-supervised person re-identification. Image Vis Comput 137:104790

    Article  Google Scholar 

  42. Chen J, Yang M, Ling J (2021) Attention-based label consistency for semi-supervised deep learning based image classification. Neurocomputing 453:731–741

    Article  Google Scholar 

  43. Arantes RB, Vogiatzis G, Faria DR (2022) Learning an augmentation strategy for sparse datasets. Image Vis Comput 117:104338

    Article  Google Scholar 

  44. Wang X, Kihara D, Luo J, Qi G-J (2020) EnAET: A self-trained framework for semi-supervised and supervised learning with ensemble transformations. IEEE Trans Image Process 30:1639–1647

    Article  Google Scholar 

  45. Poon H-K, Yap W-S, Tee Y-K, Lee W-K, Goi B-M (2019) Hierarchical gated recurrent neural network with adversarial and virtual adversarial training on text classification. Neural Netw 119:299–312

    Article  Google Scholar 

  46. Heidari M, Moattar MH, Ghaffari H (2023) Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking. Neural Networks 165:238-247

  47. Van Engelen JE, Hoos HH (2020) A survey on semi-supervised learning. Mach Learn 109:373–440

    Article  MathSciNet  Google Scholar 

  48. Han Y, Li K, Feng H, Zhao Q (2022) Non-intrusive load monitoring based on semi-supervised smooth teacher graph learning with voltage–current trajectory. Neural Comput Appl 34:19147–19160

    Article  Google Scholar 

  49. Tu E, Wang Z, Yang J, Kasabov N (2022) Deep semi-supervised learning via dynamic anchor graph embedding in latent space. Neural Netw 146:350–360

    Article  Google Scholar 

  50. Xu H, Xiao H, Hao H, Dong L, Qiu X, Peng C (2023) Semi-supervised learning with pseudo-negative labels for image classification. Knowl-Based Syst 260:110166

    Article  Google Scholar 

  51. Jiang T, Chen L, Chen W, Meng W, Qi P (2023) ReliaMatch: Semi-Supervised Classification with Reliable Match. Appl Sci 13:8856

    Article  Google Scholar 

  52. Khan AA, Madendran RK, Thirunavukkarasu U, Faheem M (2023) D2PAM: epileptic seizures prediction using adversarial deep dual patch attention mechanism. CAAI Trans Intelligence Technol 8:755–769

    Article  Google Scholar 

  53. Ali G, Dastgir A, Iqbal MW, Anwar M, Faheem M (2023) A hybrid convolutional neural network model for automatic diabetic retinopathy classification from fundus images. IEEE J Transl Eng Health Med 11:341–350

    Article  Google Scholar 

  54. Gangwar A, González-Castro V, Alegre E, Fidalgo E (2023) Triple-BigGAN: Semi-supervised generative adversarial networks for image synthesis and classification on sexual facial expression recognition. Neurocomputing 528:200–216

    Article  Google Scholar 

  55. Zhang S, Huang K, Zhu J, Liu Y (2021) Manifold adversarial training for supervised and semi-supervised learning. Neural Netw 140:282–293

    Article  Google Scholar 

  56. Zhang H (2017) Mixup: Beyond empirical risk minimization, arXiv preprint https://doi.org/10.48550/arXiv.1710.09412

  57. Lin J, Huang Z, Wang K, Liang X, Chen W, Lin L (2021) Continuous transition: Improving sample efficiency for continuous control problems via mixup, in 2021 IEEE International Conference on Robotics and Automation (ICRA), IEEE pp 9490–9497

  58. Ko K, Yeom T, Lee M (2023) Superstargan: Generative adversarial networks for image-to-image translation in large-scale domains. Neural Netw 162:330–339

    Article  Google Scholar 

  59. Tzelepi M, Tefas A (2020) Improving the performance of lightweight cnns for binary classification using quadratic mutual information regularization. Pattern Recogn 106:107407

    Article  Google Scholar 

  60. Stolberg-Larsen J, Sommer S (2022) Atlas generative models and geodesic interpolation. Image Vis Comput 122:104433

    Article  Google Scholar 

  61. LeCun Y (1998) The MNIST database of handwritten digits, http://yann.lecun.com/exdb/mnist/. Accessed 2 Feb 2009

  62. Netzer Y, Wang T, Coates A, Bissacco A, Wu B, Ng AY (2011) Reading digits in natural images with unsupervised feature learning. In: NIPS workshop on deep learning and unsupervised feature learning (Vol. 2011, No. 2, p. 4)

  63. Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images.

  64. Darlow LN, Crowley EJ, Antoniou A, Storkey AJ (2018) Cinic-10 is not imagenet or cifar-10, arXiv preprint https://doi.org/10.48550/arXiv.1810.03505

  65. Chen X, Duan Y, Houthooft R, Schulman J, Sutskever I, Abbeel P (2016) Infogan: interpretable representation learning by information maximizing generative adversarial nets. Adv Neural Inf Process Syst 29

  66. Goyal P (2017) Accurate, large minibatch sgd: training imagenet in 1 hour, arXiv preprint https://doi.org/10.48550/arXiv.1706.02

  67. Loshchilov I, Hutter F (2016) Sgdr: stochastic gradient descent with warm restarts, arXiv preprint https://doi.org/10.48550/arXiv.1608.03983

  68. Weston J, Ratle F, Collobert R (2008) Deep learning via semi-supervised embedding. In: Proceedings of the 25th international conference on machine learning. pp 1168–1175

  69. Salakhutdinov R, Hinton G (2007) Learning a nonlinear embedding by preserving class neighborhood structure. In: Artificial intelligence and statistics, PMLR, pp. 412–419

  70. Ranzato MA, Huang FJ, Boureau Y-L, LeCun Y (2007) Unsupervised learning of invariant feature hierarchies with applications to object recognition. In: 2007 IEEE conference on computer vision and pattern recognition, IEEE pp 1–8

  71. Rifai S, Dauphin YN, Vincent P, Bengio Y, Muller X (2011) The manifold tangent classifier. Advances in neural information processing systems 24. Accessed 21 Mar 2012

  72. Lee D-H (2013) Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In: Workshop on challenges in representation learning, ICML (Vol. 3, No. 2, p. 896)

  73. Oliver A, Odena A, Raffel CA, Cubuk ED, Goodfellow I (2018) Realistic evaluation of deep semi-supervised learning algorithms. Adv Neural Inf Process Syst 31

  74. Verma V, Lamb A, Beckham C, Najafi A, Mitliagkas I, Lopez-Paz D, Bengio Y (20191) Manifold mixup: Better representations by interpolating hidden states. In: International conference on machine learning. PMLR pp 6438–6447

  75. Athiwaratkun B, Finzi M, Izmailov P, Wilson AG (2019) There are many consistent explanations of unlabeled data: Why you should average, International conference on learning representations. URL: https://openreview.net/forum?id=rkgKBhA5Y7

Download references

Author information

Authors and Affiliations

Authors

Contributions

Mohammad Saber Iraji: Writing – original draft- review & editing, Methodology, Investigation, Validation, Conceptualization. Jafar Tanha: Supervision, Conceptualization, Methodology, Writing – review & editing, Validation Mohammad-Ali Balafar: Supervision, Validation. Mohammad-Reza Feizi-Derakhshi: Supervision, Validation.

Corresponding author

Correspondence to Jafar Tanha.

Ethics declarations

Competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

Open view of the method equations

$${\ddot{x}}_{{\text{mix}}_{\lambda }\left(ij\right)}{=\text{mix}}_{\lambda \left(ij\right)}({x}_{i},{x}_{j})=\lambda {x}_{i}+\left(1-\lambda \right) {x}_{j}$$
(22)
$${\ddot{y}}_{{\text{mix}}_{\lambda }\left(ij\right)}{=\text{mix}}_{\lambda \left(ij\right)}\left({y}_{i},{y}_{j}\right)=\lambda {y}_{i}+\left(1-\lambda \right) {y}_{j}$$
(23)

Where\({(x}_i,y_i)\sim p_{(x,y)},{(x}_j,y_j)\sim p_{(x,y)}\)are two images, corresponding labels

$${\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{real}-\text{fake}}={\text{mix}}_{\lambda (ij)}\left({x}_{i}^{l,u},{x}_{j}^{g}\right)=\lambda {x}_{i}^{l,u}+(1-\lambda ){x}_{j}^{g}=\lambda {x}_{i}^{l,u}+(1-\lambda )G\left({z}_{j}\right)$$
(24)
$${\ddot{y}}_{{\text{mix}}_{\lambda }(ij)}^{\text{real}-\text{fake}}={\text{mix}}_{\lambda (ij)}\left({y}_{i}^{l,u},{y}_{j}^{g}\right)=\lambda {y}_{i}^{l,u}+(1-\lambda ){y}_{j}^{g}=\lambda$$
(25)

Where \({x}_{i}^{l}\sim {p}_{{x}^{l}}^{\text{real}},{x}_{i}^{u}\sim {p}_{{x}^{u}}^{\text{real}} ,{x}_{j}^{g}=G\left({z}_{j}\right)\sim {p}_{{x}^{g}}^{\text{fake}}\) are input images, \({y}_{i}^{l,u}=1,{y}_{j}^{g}=0\) are corresponding labels

$${L}_{\text{discriminator}-\text{adversarial}-\text{mixup}}={L}_{\text{CE}}\left( D( {\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{real}-\text{fake}}),\lambda \right)={L}_{\text{CE}}\left( D( \lambda {x}_{i}^{l,u}+(1-\lambda )G({z}_{j})),\lambda \right)$$
(26)

where

$${\text{Cross Entropy Loss}=L}_{\text{CE}}\left(y,C\left(x\right)\right)=-\sum_{i=1}^{nc}{y}_{i }\text{log}\left({C\left(x\right)}_{i}\right) ,$$
(27)
$$\text{nc}=\text{number of class}$$
$${{L}_{1-\text{classifier}-\text{supervised}}=L}_{\text{CE}}\left(C\left({x}^{l}\right),y\right)={L}_{\text{CE}}\left({\widehat{y}}_{{x}^{l}},y\right)$$
(28)
$${L}_{2-\text{classifier}-\text{adversarial}-\text{fake}}={L}_{\text{ICE}}{(\widehat{y}}_{{x}^{g}},{\widetilde{y}}_{{x}^{g}})={L}_{\text{ICE}}\left(C\left(G\left(z\right)\right),\text{ arg max}\left(C\left(G\left(z\right)\right)\right)\right)$$
(29)
$${\text{Inverse Cross Entropy Loss}=L}_{\text{ICE}}\left(y,c\left(x\right)\right)=-\sum_{i=1}^{nc}{y}_{i }\text{log}\left(1-{c\left(x\right)}_{i}\right)$$
(30)
$$,\text{nc}=\text{number of class}$$
$${\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{fake}-\text{fake}}={\text{mix}}_{\lambda (ij)}\left({x}_{i}^{g},{x}_{j}^{g}\right)=\lambda {x}_{i}^{g}+(1-\lambda ){x}_{j}^{g}=\lambda G\left({z}_{i}\right)+(1-\lambda )G\left({z}_{j}\right)$$
(31)
$${\ddot{y}}_{{\text{mix}}_{\lambda }(ij)}^{\text{fake}-\text{fake}}={\text{mix}}_{\lambda (ij)}\left({y}_{i}^{g},{y}_{j}^{g}\right)=\lambda {y}_{i}^{g}+\left(1-\lambda \right){y}_{j}^{g}=\lambda {\widetilde{y}}_{{x}_{i}^{g}}+\left(1-\lambda \right){\widetilde{y}}_{{x}_{j}^{g}}$$
(32)

Where \({x}_{i}^{g}=G\left({z}_{i}\right)\sim {p}_{{x}^{g}}^{\text{fake}},{x}_{j}^{g}=G\left({z}_{j}\right)\sim {p}_{{x}^{g}}^{\text{fake}}\) are two fake images, \({y}_{i}^{g}={\widetilde{y}}_{{x}_{i}^{g}}=\text{arg max}\left(C\left(G\left({z}_{i}\right)\right)\right),{y}_{j}^{g}={\widetilde{y}}_{{x}_{j}^{g}}=\text{arg max}\left(C\left(G\left({z}_{j}\right)\right)\right)\) are corresponding labels by the classifier

$${L}_{3-\text{classifier}-\text{adversarial}-\text{fake}-\text{mixup}}{=L}_{\text{ICE}}\left( C( {\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{fake}-\text{fake}}),{\ddot{y}}_{{\text{mix}}_{\lambda }(ij)}^{\text{fake}-\text{fake}}\right)={L}_{\text{ICE}}\left({\widehat{y}}_{{\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{fake}-\text{fake}}} ,{\ddot{y}}_{{\text{mix}}_{\lambda }(ij)}^{\text{fake}-\text{fake}}\right)={L}_{\text{ICE}}\left( C\left( \lambda G\left({z}_{i}\right)+\left(1-\lambda \right)G\left({z}_{j}\right)\right),\left(\lambda {\widetilde{y}}_{{x}_{i}^{g}}+\left(1-\lambda \right){\widetilde{y}}_{{x}_{j}^{g}}\right)\right)={L}_{\text{ICE}}\left( C\left(input \right),\left(\lambda \text{arg max}\left(C\left(G\left({z}_{i}\right)\right)\right)+\left(1-\lambda \right)\text{ arg max}\left(C\left(G\left({z}_{j}\right)\right)\right)\right)\right)$$
(33)
$${L}_{4-\text{classifier}-\text{unsupervised}}={{L}_{\text{CE}}(\widehat{y}}_{{x}^{u}},{\widetilde{y}}_{{x}^{u}}){=L}_{\text{CE}}\left(C\left({x}^{u}\right),\text{ arg max}\left(C\left({x}^{u}\right)\right)\right)$$
(34)
$${\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{unlabeled}-\text{unlabeled}}={\text{mix}}_{\lambda (ij)}\left({x}_{i}^{u},{x}_{j}^{u}\right)=\lambda {x}_{i}^{u}+(1-\lambda ){x}_{j}^{u}$$
(35)
$${\ddot{y}}_{{\text{mix}}_{\lambda }(ij)}^{\text{unlabeled}-\text{unlabeled}}={\text{mix}}_{\lambda (ij)}\left({y}_{i}^{u},{y}_{j}^{u}\right)=\lambda {y}_{i}^{u}+\left(1-\lambda \right){y}_{j}^{u}=\lambda {\widetilde{y}}_{{x}_{i}^{u}}+\left(1-\lambda \right){\widetilde{y}}_{{x}_{j}^{u}}$$
(36)

Where \({x}_{i}^{u}\sim {p}_{{x}^{u}}^{\text{real}} ,{x}_{j}^{u}\sim {p}_{{x}^{u}}^{\text{real}}\) are two input unlabeled images, \({y}_{i}^{u}={\widetilde{y}}_{{x}_{i}^{u}}=\text{arg max}\left(C\left({x}_{i}^{u}\right)\right),{y}_{j}^{u}={\widetilde{y}}_{{x}_{j}^{u}}=\text{arg max}\left(C\left({x}_{j}^{u}\right)\right)\) are corresponding labels by the classifier

$${L}_{5-\text{classifier}-\text{unsupervised}-\text{mixup}}{=L}_{\text{CE}}\left( C( {\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{unlabeled}-\text{unlabeled}}),{\ddot{y}}_{{\text{mix}}_{\lambda }(ij)}^{\text{unlabeled}-\text{unlabeled}}\right)={L}_{\text{CE}}\left({\widehat{y}}_{{\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{unlabeled}-\text{unlabeled}}} ,{\ddot{y}}_{{\text{mix}}_{\uplambda }(ij)}^{\text{unlabeled}-\text{unlabeled}}\right)={L}_{\text{CE}}\left( C\left( \lambda {x}_{i}^{u}+(1-\lambda ){x}_{j}^{u}\right),\left(\lambda {\widetilde{y}}_{{x}_{i}^{u}}+\left(1-\lambda \right){\widetilde{y}}_{{x}_{j}^{u}}\right)\right)={L}_{\text{CE}}\left( C\left( \lambda {x}_{i}^{u}+(1-\lambda ){x}_{j}^{u}\right),\left(\lambda \text{arg max}\left(C\left({x}_{i}^{u}\right)\right)+\left(1-\lambda \right)\text{arg max}\left(C\left({x}_{j}^{u}\right)\right)\right)\right)$$
(37)
$${L}_{\text{classifier}}={L}_{\text{CE}}\left(C\left({x}^{l}\right),y\right){+L}_{\text{ICE}}\left(C\left(G\left(z\right)\right),\text{ arg max}\left(C\left(G\left(z\right)\right)\right)\right)+{L}_{\text{ICE}}\left( C\left(input\right),\left(\lambda \text{arg max}\left(C\left(G\left({z}_{i}\right)\right)\right)+\left(1-\lambda \right)\text{arg max}\left(C\left(G\left({z}_{j}\right)\right)\right)\right)\right)+{L}_{\text{CE}}\left(C\left({x}^{u}\right),\text{ arg max}\left(C\left({x}^{u}\right)\right)\right)+{L}_{\text{CE}}\left( C\left( \lambda {x}_{i}^{u}+(1-\lambda ){x}_{j}^{u}\right),\left(\lambda \text{arg max}\left(C\left({x}_{i}^{u}\right)\right)\left(1-\lambda \right)\text{arg max}\left(C\left({x}_{j}^{u}\right)\right)\right)\right)$$
(38)
$${L}_{1-\text{generator}-\text{adversarial}-\text{mixup}}=-{L}_{\text{CE}}\left( D\left( {\ddot{x}}_{{\text{mix}}_{\lambda \left(ij\right)}}^{\text{real}-\text{fake}}\right),\lambda \right)=-{L}_{\text{CE}}\left( D( \lambda {x}_{i}^{l,u}+(1-\lambda )G({z}_{j})),\lambda \right)$$
(39)
$${L}_{2-\text{generator}-\text{adversarial}-\text{fake}}={L}_{\text{CE}}{(\widehat{y}}_{{x}^{g}},{\widetilde{y}}_{{x}^{g}})={L}_{\text{CE}}\left(C\left(G\left(z\right)\right),\text{ arg max}\left(C\left(G\left(z\right)\right)\right)\right)$$
(40)
$$\begin{array}{c}{{L}_{3-\text{generator}-\text{adversarial}-\text{fake}-\text{mixup}}=L}_{\text{CE}}\left( C\left( {\ddot{x}}_{{\text{mix}}_{\lambda \left(ij\right)}}^{\text{fake}-\text{fake}}\right),{\ddot{y}}_{{\text{mix}}_{\lambda }\left(ij\right)}^{\text{fake}-\text{fake}}\right)=\\ \begin{array}{c}{L}_{\text{CE}}\left({\widehat{y}}_{{\ddot{x}}_{{\text{mix}}_{\lambda (ij)}}^{\text{fake}-\text{fake}}} ,{\ddot{y}}_{{\text{mix}}_{\lambda }(ij)}^{\text{fake}-\text{fake}}\right)={L}_{\text{CE}}\left( C\left( \lambda G\left({z}_{i}\right)+\left(1-\lambda \right)G\left({z}_{j}\right)\right),\left(\lambda {\widetilde{y}}_{{x}_{i}^{g}}+\left(1-\lambda \right){\widetilde{y}}_{{x}_{j}^{g}}\right)\right)\\ \begin{array}{c}={L}_{\text{CE}}\left( C\left(input\right),\left(\lambda \text{arg max}\left(C\left(G\left({z}_{i}\right)\right)\right)+\left(1-\lambda \right)\text{arg max}\left(C\left(G\left({z}_{j}\right)\right)\right)\right)\right)\\ \end{array}\end{array}\end{array}$$
(41)
$${L}_{\text{generator}}={-L}_{\text{CE}}\left( D( \lambda {x}_{i}^{l,u}+(1-\lambda )G({z}_{j})),\lambda \right)+{L}_{\text{CE}}\left(C\left(G\left(z\right)\right),\text{ arg max}\left(C\left(G\left(z\right)\right)\right)\right)+{L}_{\text{CE}}\left( C\left(input\right),\left(\lambda \text{arg max}\left(C\left(G\left({z}_{i}\right)\right)\right)+\left(1-\lambda \right)\text{arg max}\left(C\left(G\left({z}_{j}\right)\right)\right)\right)\right)$$
(42)

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Iraji, M.S., Tanha, J., Balafar, MA. et al. A novel interpolation consistency for bad generative adversarial networks (IC-BGAN). Multimed Tools Appl 83, 86161–86205 (2024). https://doi.org/10.1007/s11042-024-20333-5

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-024-20333-5

Keywords

Navigation