Abstract
Batch mode active learning (BMAL) is absorbed in training reliable classifier with deficient labeled examples by efficiently querying the most valuable unlabeled examples for supervision. In particular, BMAL always selects examples based on the decent-designed criteria, such as (un)certainty and representativeness, etc. However, existing BMAL approaches make a naive trade-off between the criteria and simply combine them with fixed weights, which may yield suboptimal batch selection since the criteria of unlabeled examples would fluctuate after retraining classifier with the newly augmented training set as the learning of classifier progresses. Instead, the weights of the criteria should be assigned properly. To overcome this problem, this paper proposes a novel A daptive C riteria W eights active learning method, abbreviated ACW, which dynamically combines the example selection criteria together to select critical examples for semi-supervised classification. Concretely, we first assign an initial value to each criterion weight, then the current optimal batch is picked from unlabeled pool. Thereafter, the criteria weights are learned and adjusted adaptively by minimizing the objective function with the selected batch at each round. To the best of our knowledge, this work is the first attempt to explore adaptive criteria weights in the context of active learning. The superiority of ACW against the existing state-of-the-art BMAL approaches has also been validated by extensive experimental results on widely used datasets.








Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Change history
29 December 2020
A Correction to this paper has been published: https://doi.org/10.1007/s10489-020-02146-9
References
Ash JT, Zhang C, Krishnamurthy A, Langford J, Agarwal A (2019) Deep batch active learning by diverse, uncertain gradient lower bounds arXiv: Learning
Balcan MF, Broder A, Zhang T (2007) Margin based active learning. In: International conference on computational learning theory. Springer, pp 35–50
Beluch WH, Genewein T, Nürnberger A, Köhler JM (2018) The power of ensembles for active learning in image classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 9368–9377
Borgwardt KM, Gretton A, Rasch MJ, Kriegel HP, Schölkopf B, Smola A (2006) Integrating structured biological data by kernel maximum mean discrepancy. Bioinformatics 22(14):e49–e57
Brinker K (2003) Incorporating diversity in active learning with support vector machines. In: Proceedings of the 20th international conference on machine learning (ICML-03), pp 59–66
Cai W, Zhang Y, Zhang Y, Zhou S, Wang W, Chen Z, Ding C (2017) Active learning for classification with maximum model change. ACM Trans Inform Sys (TOIS) 36(2):15
Chakraborty S, Balasubramanian V, Panchanathan S (2015) Adaptive batch mode active learning. IEEE Trans Neural Netw Learning Sys 26(8):1747–1760
Chakraborty S, Balasubramanian V, Sun Q, Panchanathan S, Ye J (2015) Active batch selection via convex relaxations with guaranteed solution bounds. IEEE Trans Pattern Anal Mach Intell 37 (10):1945–1958
Chattopadhyay R, Fan W, Davidson I, Panchanathan S, Ye J (2013) Joint transfer and batch-mode active learning. In: International conference on machine learning, pp 253–261
Chattopadhyay R, Wang Z, Fan W, Davidson I, Panchanathan S, Ye J (2013) Batch mode active sampling based on marginal probability distribution matching. ACM Transactions on Knowledge Discovery from Data (TKDD) 7(3):13
Deng J, Dong W, Socher R, Li L, Li K, Feifei L (2009) Imagenet: a large-scale hierarchical image database. In: Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE, pp 248–255
Dhurandhar A, Sankaranarayanan K (2015) Improving classification performance through selective instance completion. Mach Learn 100(2-3):425–447
Donmez P, Carbonell J, Bennett PN (2007) Dual strategy active learning. In: European conference on machine learning. Springer, pp 116–127
Du B, Wang Z, Zhang L, Zhang L, Liu W, Shen J, Tao D (2017) Exploring representativeness and informativeness for active learning. IEEE Trans Cybern 47(1):14–26
Esna Ashari Z, Ghasemzadeh H (2019) Mindful active learning. In: Proceedings of the twenty-eighth international joint conference on artificial intelligence, pp 2265–2271
Everingham M, Van Gool L, Williams CKI, Winn J, Zisserman A (2010) The pascal visual object classes (voc) challenge. Int J Comput Vis 88(2):303–338
Freund Y, Seung HS, Shamir E, Tishby N (1997) Selective sampling using the query by committee algorithm. Machine Learning 28(2-3):133–168
Fu W, Wang M, Hao S, Wu X (2018) Scalable active learning by approximated error reduction. In: Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining. ACM, pp 1396–1405
Fujita H, Gaeta A, Loia V, Orciuoli F (2019) Improving awareness in early stages of security analysis: a zone partition method based on grc. Appl Intell 49:1063–1077
Fujita H, Gaeta A, Loia V, Orciuoli F (2019) Resilience analysis of critical infrastructures: a cognitive approach based on granular computing. IEEE Trans Sys Man Cybern 49(5):1835–1848
Gilad-Bachrach R, Navot A, Tishby N (2006) Query by committee made real. In: Advances in neural information processing systems, pp 443–450
Gretton A, Borgwardt KM, Rasch MJ, Schölkopf B, Smola A (2012) A kernel two-sample test. J Mach Learn Res 13(Mar):723–773
Guo Y, Greiner R (2007) Optimistic active-learning using mutual information. In: IJCAI, vol 7, pp 823–829
Guo Y, Schuurmans D (2008) Discriminative batch mode active learning. In: Advances in neural information processing systems, pp 593–600
He X (2009) Laplacian regularized d-optimal design for active learning and its application to image retrieval. IEEE Trans Image Process 19(1):254–263
Hoi SC, Jin R, Lyu MR (2009) Batch mode active learning with applications to text categorization and image retrieval. IEEE Trans Knowl Data Eng 21(9):1233–1248
Hoi SC, Jin R, Zhu J, Lyu MR (2009) Semisupervised svm batch mode active learning with applications to image retrieval. ACM Transactions on Information Systems (TOIS) 27(3):16
Hu Y, Zhang D, Jin Z, Cai D, He X (2013) Active learning via neighborhood reconstruction. In: Proceedings of the twenty-third international joint conference on artificial intelligence, pp 1415–1421
Huang SJ, Jin R, Zhou ZH (2014) Active learning by querying informative and representative examples. IEEE Trans Pattern Anal Mach Intell 36(10):1936–1949
Joshi AJ, Porikli F, Papanikolopoulos N (2009) Multi-class active learning for image classification. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 2372–2379
Kapoor A, Horvitz E, Basu S (2007) Selective supervision: guiding supervised learning with decision-theoretic active learning. In: IJCAI, vol 7, pp 877–882
Kirsch A, van Amersfoort J, Gal Y (2019) Batchbald: efficient and diverse batch acquisition for deep bayesian active learning. In: Advances in neural information processing systems, pp 7026–7037
Konyushkova K, Sznitman R, Fua P (2017) Learning active learning from data. In: Advances in neural information processing systems 30, pp 4225–4235
Li CL, Ferng CS, Lin HT (2012) Active learning with hinted support vector machine. In: Proceedings of asian conference on machine learning. PMLR, pp 221–235
Li Y, Wang Y, Yu DJ, Ye N, Hu P, Zhao R (2019) Ascent: active supervision for semi-supervised learning. IEEE Trans Knowl Data Eng 32(5):868–882
Long C, Hua G (2015) Multi-class multi-annotator active learning with robust gaussian process for visual recognition. In: Proceedings of the IEEE international conference on computer vision, pp 2839–2847
Long C, Hua G, Kapoor A (2016) A joint gaussian process model for active visual recognition with expertise estimation in crowdsourcing. Int J Comput Vis 116(2):136–160
Mallapragada PK, Jin R, Jain AK (2008) Active query selection for semi-supervised clustering. In: 2008 19Th international conference on pattern recognition, pp 1–4. IEEE
Murugesan K, Carbonell J (2017) Active learning from peers. In: Advances in neural information processing systems 30, pp 7008–7017
Nie F, Wang H, Huang H, Ding C (2013) Early active learning via robust representation and structured sparsity. In: Twenty-third international joint conference on artificial intelligence, pp 1572–1578
Nie F, Wang X, Huang H (2014) Clustering and projected clustering with adaptive neighbors. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining, pp 977–986
Paul S, Bappy JH, Roy-Chowdhury AK (2017) Non-uniform subset selection for active learning in structured data. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6846–6855
Pinsler R, Gordon J, Nalisnick E, Hernandezlobato JM (2019) Bayesian batch active learning as sparse subset approximation. In: Advances in neural information processing systems, pp 6356–6367
Sandoval AM, Diaz J, Llanos LC, Redondo T (2019) Biomedical term extraction: Nlp techniques in computational medicine. Int J Interact Multimed Artificial Intell 5(4):51–59
Settles B (2009) Active learning literature survey. Tech. rep., University of Wisconsin-Madison Department of Computer Sciences
Seung HS, Opper M, Sompolinsky H (1992) Query by committee. In: Proceedings of the fifth annual workshop on computational learning theory, pp 287–294. ACM
Tang YP, Huang SJ (2019) Self-paced active learning: query the right thing at the right time. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 5117– 5124
Tang YP, Li GX, Huang SJ (2019) ALiPy: active learning in python. Tech. rep., Nanjing University of Aeronautics and Astronautics. https://github.com/NUAA-AL/ALiPy. Available as arXiv:1901.03802
Tong S, Koller D (2001) Support vector machine active learning with applications to text classification. J Mach Learn Res 2(Nov):45–66
Vasisht D, Damianou A, Varma M, Kapoor A (2014) Active learning for sparse bayesian multilabel classification. In: Proceedings of the 20th ACM SIGKDD international conference on knowledge discovery and data mining. ACM, pp 472–481
Wang H, Chang X, Shi L, Yang Y, Shen YD (2018) Uncertainty sampling for action recognition via maximizing expected average precision. In: IJCAI, pp 964–970
Wang H, Du L, Zhou P, Shi L, Shen YD (2015) Convex batch mode active sampling via α-relative pearson divergence. In: Twenty-ninth AAAI conference on artificial intelligence
Wang H, Zhou R, Shen YD (2019) Bounding uncertainty for active batch selection. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 5240–5247
Wang K, Zhang D, Li Y, Zhang R, Lin L (2017) Cost-effective active learning for deep image classification. IEEE Trans Circ Sys Video Technol 27(12):2591–2600
Wang Z, Ye J (2015) Querying discriminative and representative samples for batch mode active learning. ACM Trans Knowl Discovery Data (TKDD) 9(3):17
Wu Y, Wu Q, Dey N, Sherratt S (2020) Learning models for semantic classification of insufficient plantar pressure images. Int J Interact Multimed Artificial Intell 6(1):51–61
Xiong C, Johnson DM, Corso JJ (2017) Active clustering with model-based uncertainty reduction. IEEE Trans Pattern Anal Machine Intell 39(1):5–17
Xu Z, Yu K, Tresp V, Xu X, Wang J (2003) Representative sampling for text classification using support vector machines. In: European conference on information retrieval. Springer, pp 393–407
Yang Y, Ma Z, Nie F, Chang X, Hauptmann AG (2015) Multi-class active learning by uncertainty sampling with diversity maximization. Int J Comput Vis 113(2):113–127
Yin C, Qian B, Cao S, Li X, Wei J, Zheng Q, Davidson I (2017) Deep similarity-based batch mode active learning with exploration-exploitation. In: 2017 IEEE international conference on data mining (ICDM). IEEE, pp 575–584
Yoo D, Kweon IS (2019) Learning loss for active learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 93–102
Zhang Y, Lease M, Wallace BC (2017) Active discriminative text representation learning. In: Thirty-first AAAI conference on artificial intelligence
Zuluaga M, Sergent G, Krause A, Püschel M (2013) Active learning for multi-objective optimization. In: International conference on machine learning, pp 462–470
Acknowledgements
The authors would like to thank the anonymous reviewers for their valuable comments and suggestions. This article has been awarded by the National Natural Science Foundation of China (61941113), the Fundamental Research Fund for the Central Universities (30918015103, 30918012204), Nanjing Science and Technology Development Plan Project (201805036), and “13th Five-Year” equipment field fund (61403120501), China Academy of Engineering Consulting Research Project (2019-ZD-1-02-02), National Social Science Foundation (18BTQ073), State Grid Technology Project (5211XT190033).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original online version of this article was revised: The affiliation of the second and third authors were incorrect.
Rights and permissions
About this article
Cite this article
Li, H., Wang, Y., Li, Y. et al. Batch mode active learning via adaptive criteria weights. Appl Intell 51, 3475–3489 (2021). https://doi.org/10.1007/s10489-020-01953-4
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-020-01953-4