Abstract
Semi-supervised learning techniques utilize both labeled and unlabeled images to enhance classification performance in scenarios where labeled images are limited. However, challenges such as integrating unlabeled images with incorrect pseudo-labels, determining appropriate thresholds for the pseudo-labels, and label prediction fluctuations on low-confidence unlabeled images, hinder the effectiveness of existing methods. This research introduces a novel framework named Interpolation Consistency for Bad Generative Adversarial Networks (IC-BGAN) that utilizes a new loss function. The proposed model combines bad adversarial training, fusion techniques, and regularization to address the limitations of semi-supervised learning. IC-BGAN creates three types of image augmentations and label consistency regularization in interpolation of bad fake images, real and bad fake images, and unlabeled images. It demonstrates linear interpolation behavior, reducing fluctuations in predictions, improving stability, and facilitating the identification of decision boundaries in low-density areas. The regularization techniques boost the discriminative capability of the classifier and discriminator, and send a better signal to the bad generator. This improves the generalization and the generation of diverse inter-class fake images as support vectors with information near the true decision boundary, which helps to correct the pseudo-labeling of unlabeled images. The proposed approach achieves notable improvements in error rate from 2.87 to 1.47 on the Modified National Institute of Standards and Technology (MNIST) dataset, 3.59 to 3.13 on the Street View House Numbers (SVHN) dataset, and 12.13 to 9.59 on the Canadian Institute for Advanced Research, 10 classes (CIFAR-10) dataset using 1000 labeled training images. Additionally, it reduces the error rate from 22.11 to 18.40 on the CINIC-10 dataset when using 700 labeled images per class. The experiments demonstrate the IC-BGAN framework outperforms existing semi-supervised methods, providing a more accurate classification solution with smoother class label estimates, especially for low-confidence unlabeled images.
























Similar content being viewed by others
Data availability
Data will be made available on request.
References
Lim JY, Lim KM, Lee CP, Tan YX (2023) SCL: Self-supervised contrastive learning for few-shot image classification. Neural Netw 165:19–30
Alarood AA, Faheem M, Al-Khasawneh MA, Alzahrani AI, Alshdadi AA (2023) Secure medical image transmission using deep neural network in e-health applications. Healthcare Technology Letters 10:87–98
Niu X, Ma W (2023) Semi-supervised classifier ensemble model for high-dimensional data. Inf Sci 643:119203
Wang Y, Song D, Wang W, Rao S, Wang X, Wang M (2022) Self-supervised learning and semi-supervised learning for multi-sequence medical image classification. Neurocomputing 513:383–394
Yang X, Song Z, King I, Xu Z (2022) A survey on deep semi-supervised learning. IEEE Transactions on Knowledge and Data Engineering 35(9):8934–8954
Li J (2022) NaNG-ST: A natural neighborhood graph-based self-training method for semi-supervised classification. Neurocomputing 514:268–284
Taha K (2023) Semi-supervised and un-supervised clustering: A review and experimental evaluation. Information Systems 144:102178
Wang Y, Li D, Li X, Yang M (2021) PC-GAIN: Pseudo-label conditional generative adversarial imputation networks for incomplete data. Neural Netw 141:395–403
Bae J, Lee M, Kim SB (2022) Safe semi-supervised learning using a Bayesian neural network. Inf Sci 612:453–464
Ke B, Lu H, You C, Zhu W, Xie L, Yao Y (2023) A semi-supervised medical image classification method based on combined pseudo-labeling and distance metric consistency. Multimed Tools Appl 83(11):1–19
Zhao J, Lan L, Huang D, Ren J, Yang W (2022) Heterogeneous pseudo-supervised learning for few-shot person re-identification. Neural Netw 154:521–537
Meel P, Vishwakarma DK (2021) A temporal ensembling based semi-supervised ConvNet for the detection of fake news articles. Expert Syst Appl 177:115002
Tarvainen A, Valpola H (2017) Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results, Adv In Neural Inf Proc Sys 30
Gan Y, Zhu H, Guo W, Xu G, Zou G (2022) Deep semi-supervised learning with contrastive learning and partial label propagation for image data. Knowl-Based Syst 245:108602
Yang M, Ling J, Chen J, Feng M, Yang J (2023) Discriminative semi-supervised learning via deep and dictionary representation for image classification. Pattern Recogn 140:109521
Sohn K, Berthelot D, Carlini N, Zhang Z, Zhang H, Raffel CA, Cubuk ED, Kurakin A, Li C-L (2020) Fixmatch: Simplifying semi-supervised learning with consistency and confidence. Adv Neural Inf Process Syst 33:596–608
Feng Z, Zhou Q, Gu Q, Tan X, Cheng G, Lu X, Shi J, Ma L (2022) Dmt: Dynamic mutual training for semi-supervised learning. Pattern Recogn 130:108777
Huo X, Zeng X, Wu S, Shen W, Wong H-S (2023) Collaborative Learning with Unreliability Adaptation for Semi-Supervised Image Classification. Pattern Recogn 133:109032
Zhou S, Tian S, Yu L, Wu W, Zhang D, Peng Z, Zhou Z (2023) FixMatch-LS: Semi-supervised skin lesion classification with label smoothing. Biomed Signal Process Control 84:104709
Zhang B, Wang Y, Hou W, Wu H, Wang J, Okumura M, Shinozaki T (2021) Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling. Adv Neural Inf Process Syst 34:18408–18419
Miyato T, Maeda S-I, Koyama M, Ishii S (2018) Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Trans Pattern Anal Mach Intell 41:1979–1993
Park S, Park J, Shin S-J, Moon I-C (2018) Adversarial dropout for supervised and semi-supervised learning. In: Proceedings of the AAAI conference on artificial intelligence 32(1)
Luo Y, Zhu J, Li M, Ren Y, Zhang B (2018) Smooth neighbors on teacher graphs for semi-supervised learning. In: Proceedings of the IEEE conference on computer vision pattern recognition, pp 8896–8905
Verma V, Kawaguchi K, Lamb A, Kannala J, Solin A, Bengio Y, Lopez-Paz D (2022) Interpolation consistency training for semi-supervised learning. Neural Netw 145:90–106
Peng Z, Tian S, Yu L, Zhang D, Wu W, Zhou S (2023) Semi-supervised medical image classification with adaptive threshold pseudo-labeling and unreliable sample contrastive loss. Biomed Signal Process Control 79:104142
Boschini M, Buzzega P, Bonicelli L, Porrello A, Calderara S (2022) Continual semi-supervised learning through contrastive interpolation consistency. Pattern Recogn Lett 162:9–14
Liu X, Shen F, Zhao J, Nie C (2024) RandoMix: a mixed sample data augmentation method with multiple mixed modes, Multimedia Tools Appl, pp 1–17
Jo W, Kim D (2022) OBGAN: Minority oversampling near borderline with generative adversarial networks. Expert Syst Appl 197:116694
He R, Li X, Chen G, Chen G, Liu Y (2020) Generative adversarial network-based semi-supervised learning for real-time risk warning of process industries. Expert Syst Appl 150:113244
Wang L, Sun Y, Wang Z (2022) CCS-GAN: a semi-supervised generative adversarial network for image classification. Vis Comput 38:2009–2021
Dong J, Lin T (2019) Margingan: Adversarial training in semi-supervised learning, Adv Neural Inf Process Syst 32
Lu S, Zhang Z, Yan Z, Wang Y, Cheng T, Zhou R, Yang G (2023) Mutually aided uncertainty incorporated Dual Consistency Regularization with Pseudo Label for Semi-Supervised Medical Image Segmentation. Neurocomputing 584:126411
Liang D, Yang F, Zhang T, Yang P (2018) Understanding mixup training methods. IEEE Access 6:58774–58783
Sharma P, Kumar M, Sharma HK, Biju SM (2024) Generative adversarial networks (GANs): Introduction, Taxonomy, Variants, Limitations, and Applications, Multimedia Tools and Applications 1–48
Ligthart A, Catal C, Tekinerdogan B (2021) Analyzing the effectiveness of semi-supervised learning approaches for opinion spam classification. Appl Soft Comput 101:107023
Donyavi Z, Asadi S (2020) Diverse training dataset generation based on multi-objective optimization for semi-supervised classification. Pattern Recogn 108:107543
Fu S, Chen J, Chen D, He C (2023) CNNs/ViTs-CNNs/ViTs: Mutual distillation for unsupervised domain adaptation. Inf Sci 622:83–97
Ding Y, Sheng L, Liang J, Zheng A, He R (2023) ProxyMix: Proxy-based mixup training with label refinery for source-free domain adaptation. Neural Networks. 1(167):92–103
Pavate A, Bansode R, Srinivasu PN, Shafi J, Choi J, Ijaz MF (2023) Associative discussion among generating adversarial samples using evolutionary algorithm and samples generated using gAN, IEEE Access
Laine S, Aila T (2017) Temporal en sembling for semi-supervised learning. In: 5th International Conference on Learning Representations, ICLR 2017
Zhu S, Zhang Y, Feng Y (2023) GW-net: An efficient grad-CAM consistency neural network with a weakening of random erasing features for semi-supervised person re-identification. Image Vis Comput 137:104790
Chen J, Yang M, Ling J (2021) Attention-based label consistency for semi-supervised deep learning based image classification. Neurocomputing 453:731–741
Arantes RB, Vogiatzis G, Faria DR (2022) Learning an augmentation strategy for sparse datasets. Image Vis Comput 117:104338
Wang X, Kihara D, Luo J, Qi G-J (2020) EnAET: A self-trained framework for semi-supervised and supervised learning with ensemble transformations. IEEE Trans Image Process 30:1639–1647
Poon H-K, Yap W-S, Tee Y-K, Lee W-K, Goi B-M (2019) Hierarchical gated recurrent neural network with adversarial and virtual adversarial training on text classification. Neural Netw 119:299–312
Heidari M, Moattar MH, Ghaffari H (2023) Forward propagation dropout in deep neural networks using Jensen–Shannon and random forest feature importance ranking. Neural Networks 165:238-247
Van Engelen JE, Hoos HH (2020) A survey on semi-supervised learning. Mach Learn 109:373–440
Han Y, Li K, Feng H, Zhao Q (2022) Non-intrusive load monitoring based on semi-supervised smooth teacher graph learning with voltage–current trajectory. Neural Comput Appl 34:19147–19160
Tu E, Wang Z, Yang J, Kasabov N (2022) Deep semi-supervised learning via dynamic anchor graph embedding in latent space. Neural Netw 146:350–360
Xu H, Xiao H, Hao H, Dong L, Qiu X, Peng C (2023) Semi-supervised learning with pseudo-negative labels for image classification. Knowl-Based Syst 260:110166
Jiang T, Chen L, Chen W, Meng W, Qi P (2023) ReliaMatch: Semi-Supervised Classification with Reliable Match. Appl Sci 13:8856
Khan AA, Madendran RK, Thirunavukkarasu U, Faheem M (2023) D2PAM: epileptic seizures prediction using adversarial deep dual patch attention mechanism. CAAI Trans Intelligence Technol 8:755–769
Ali G, Dastgir A, Iqbal MW, Anwar M, Faheem M (2023) A hybrid convolutional neural network model for automatic diabetic retinopathy classification from fundus images. IEEE J Transl Eng Health Med 11:341–350
Gangwar A, González-Castro V, Alegre E, Fidalgo E (2023) Triple-BigGAN: Semi-supervised generative adversarial networks for image synthesis and classification on sexual facial expression recognition. Neurocomputing 528:200–216
Zhang S, Huang K, Zhu J, Liu Y (2021) Manifold adversarial training for supervised and semi-supervised learning. Neural Netw 140:282–293
Zhang H (2017) Mixup: Beyond empirical risk minimization, arXiv preprint https://doi.org/10.48550/arXiv.1710.09412
Lin J, Huang Z, Wang K, Liang X, Chen W, Lin L (2021) Continuous transition: Improving sample efficiency for continuous control problems via mixup, in 2021 IEEE International Conference on Robotics and Automation (ICRA), IEEE pp 9490–9497
Ko K, Yeom T, Lee M (2023) Superstargan: Generative adversarial networks for image-to-image translation in large-scale domains. Neural Netw 162:330–339
Tzelepi M, Tefas A (2020) Improving the performance of lightweight cnns for binary classification using quadratic mutual information regularization. Pattern Recogn 106:107407
Stolberg-Larsen J, Sommer S (2022) Atlas generative models and geodesic interpolation. Image Vis Comput 122:104433
LeCun Y (1998) The MNIST database of handwritten digits, http://yann.lecun.com/exdb/mnist/. Accessed 2 Feb 2009
Netzer Y, Wang T, Coates A, Bissacco A, Wu B, Ng AY (2011) Reading digits in natural images with unsupervised feature learning. In: NIPS workshop on deep learning and unsupervised feature learning (Vol. 2011, No. 2, p. 4)
Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images.
Darlow LN, Crowley EJ, Antoniou A, Storkey AJ (2018) Cinic-10 is not imagenet or cifar-10, arXiv preprint https://doi.org/10.48550/arXiv.1810.03505
Chen X, Duan Y, Houthooft R, Schulman J, Sutskever I, Abbeel P (2016) Infogan: interpretable representation learning by information maximizing generative adversarial nets. Adv Neural Inf Process Syst 29
Goyal P (2017) Accurate, large minibatch sgd: training imagenet in 1 hour, arXiv preprint https://doi.org/10.48550/arXiv.1706.02
Loshchilov I, Hutter F (2016) Sgdr: stochastic gradient descent with warm restarts, arXiv preprint https://doi.org/10.48550/arXiv.1608.03983
Weston J, Ratle F, Collobert R (2008) Deep learning via semi-supervised embedding. In: Proceedings of the 25th international conference on machine learning. pp 1168–1175
Salakhutdinov R, Hinton G (2007) Learning a nonlinear embedding by preserving class neighborhood structure. In: Artificial intelligence and statistics, PMLR, pp. 412–419
Ranzato MA, Huang FJ, Boureau Y-L, LeCun Y (2007) Unsupervised learning of invariant feature hierarchies with applications to object recognition. In: 2007 IEEE conference on computer vision and pattern recognition, IEEE pp 1–8
Rifai S, Dauphin YN, Vincent P, Bengio Y, Muller X (2011) The manifold tangent classifier. Advances in neural information processing systems 24. Accessed 21 Mar 2012
Lee D-H (2013) Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In: Workshop on challenges in representation learning, ICML (Vol. 3, No. 2, p. 896)
Oliver A, Odena A, Raffel CA, Cubuk ED, Goodfellow I (2018) Realistic evaluation of deep semi-supervised learning algorithms. Adv Neural Inf Process Syst 31
Verma V, Lamb A, Beckham C, Najafi A, Mitliagkas I, Lopez-Paz D, Bengio Y (20191) Manifold mixup: Better representations by interpolating hidden states. In: International conference on machine learning. PMLR pp 6438–6447
Athiwaratkun B, Finzi M, Izmailov P, Wilson AG (2019) There are many consistent explanations of unlabeled data: Why you should average, International conference on learning representations. URL: https://openreview.net/forum?id=rkgKBhA5Y7
Author information
Authors and Affiliations
Contributions
Mohammad Saber Iraji: Writing – original draft- review & editing, Methodology, Investigation, Validation, Conceptualization. Jafar Tanha: Supervision, Conceptualization, Methodology, Writing – review & editing, Validation Mohammad-Ali Balafar: Supervision, Validation. Mohammad-Reza Feizi-Derakhshi: Supervision, Validation.
Corresponding author
Ethics declarations
Competing interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Open view of the method equations
Where\({(x}_i,y_i)\sim p_{(x,y)},{(x}_j,y_j)\sim p_{(x,y)}\)are two images, corresponding labels
Where \({x}_{i}^{l}\sim {p}_{{x}^{l}}^{\text{real}},{x}_{i}^{u}\sim {p}_{{x}^{u}}^{\text{real}} ,{x}_{j}^{g}=G\left({z}_{j}\right)\sim {p}_{{x}^{g}}^{\text{fake}}\) are input images, \({y}_{i}^{l,u}=1,{y}_{j}^{g}=0\) are corresponding labels
where
Where \({x}_{i}^{g}=G\left({z}_{i}\right)\sim {p}_{{x}^{g}}^{\text{fake}},{x}_{j}^{g}=G\left({z}_{j}\right)\sim {p}_{{x}^{g}}^{\text{fake}}\) are two fake images, \({y}_{i}^{g}={\widetilde{y}}_{{x}_{i}^{g}}=\text{arg max}\left(C\left(G\left({z}_{i}\right)\right)\right),{y}_{j}^{g}={\widetilde{y}}_{{x}_{j}^{g}}=\text{arg max}\left(C\left(G\left({z}_{j}\right)\right)\right)\) are corresponding labels by the classifier
Where \({x}_{i}^{u}\sim {p}_{{x}^{u}}^{\text{real}} ,{x}_{j}^{u}\sim {p}_{{x}^{u}}^{\text{real}}\) are two input unlabeled images, \({y}_{i}^{u}={\widetilde{y}}_{{x}_{i}^{u}}=\text{arg max}\left(C\left({x}_{i}^{u}\right)\right),{y}_{j}^{u}={\widetilde{y}}_{{x}_{j}^{u}}=\text{arg max}\left(C\left({x}_{j}^{u}\right)\right)\) are corresponding labels by the classifier
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Iraji, M.S., Tanha, J., Balafar, MA. et al. A novel interpolation consistency for bad generative adversarial networks (IC-BGAN). Multimed Tools Appl 83, 86161–86205 (2024). https://doi.org/10.1007/s11042-024-20333-5
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-024-20333-5