Abstract
In multi-label learning, each instance is associated with a subset of predefined labels. One common approach for multi-label classification has been proposed in Godbole and Sarawagi (2004) based on stacking which is called as Meta Binary Relevance (MBR). It uses two layers of binary models and feeds the outputs of the first layer to all binary models of the second layer. Hence, initial predicted class labels (in the first layer) are attached to the original features to have a new prediction of the classes in the second layer. To predict a specific label in the second layer, irrelevant labels are also used as the noisy features. This is why; Nearest Neighbor (NN) as a sensitive classifier to noisy features had been not, up to now, a proper base classifier in stacking method and all of its merits including simplicity, interpretability, global stability to noisy labels and good performance, are lost. As the first contribution, a popular feature weighting in NN classification is used here to solve uncorrelated labels problem. It tunes a parametric distance function by gradient descent to minimize the classification error on training data. However, it is known that some other objectives including F-measure are more suitable than classification error on learning imbalanced data. The second contribution of this paper is extending this method in order to improve F-measure. In our experimental study, the proposed method has been compared with and outperforms state-of-the-art multi-label classifiers in the literature.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
The implementation is available in Mulan lbrary [44]
The implementation is available in MLC-toolbox [48]
The source code can be downloaded from the author home page: http://palm.seu.edu.cn/zhangml/
The source code can be downloaded from https://paperswithcode.com/paper/joint-ranking-svm-and-binary-relevance-with#code
The source code can be downloaded from https://github.com/keauneuh/Incorporating-Multiple-Cluster-Centers-for-Multi-Label-Learning
References
Godbole S, Sarawagi S (2004) Discriminative methods for multi-labeled classification. In: Pacific-Asia conference on knowledge discovery and data mining. Springer, pp 22–30
Zhang M-L, Zhou Z-H (2014) A review on multi-label learning algorithms. IEEE Trans Knowl Data Eng 26(8):1819–1837
Alazaidah R, Ahmad FK (2016) Trending challenges in multi label classification. Int J Adv Comput Sci Appl 7(10):127–131
Rathore V S, Worring M, Mishra D K, Joshi A, Maheshwari S (2018) Emerging trends in expert applications and security: Proceedings of iceteas 2018, vol 841. Springer, Berlin
Tsoumakas G, Katakis I, Vlahavas I (2009) Mining multi-label data. In: Data mining and knowledge discovery handbook. Springer, pp 667–685
Zhang M-L, Zhang K (2010) Multi-label learning by exploiting label dependency. In: Proceedings of the 16th ACM SIGKDD international conference on knowledge discovery and data mining. ACM, pp 999–1008
De Comité F, Gilleron R, Tommasi M (2003) Learning multi-label alternating decision trees from texts and data. In: International workshop on machine learning and data mining in pattern recognition. Springer, pp 35–49
Zhang M-L, Zhou Z-H (2007) ML-KNN: A lazy learning approach to multi-label learning. Pattern Recogn 40(7):2038–2048
Schapire R E, Singer Y (2000) BoosTexter: A boosting-based system for text categorization. Mach Learn 39(2-3):135–168
Fürnkranz J, Hüllermeier E, Menc∖’∖ia E L, Brinker K (2008) Multilabel classification via calibrated label ranking. Mach Learn 73 (2):133–153
Read J, Pfahringer B, Holmes G, Frank E (2011) Classifier chains for multi-label classification. Mach Learn 85(3):333
Cheng Z, Zeng Z (2020) Joint label-specific features and label correlation for multi-label learning with missing label. Appl Intell 1–21
Wu G, Tian Y, Liu D (2018) Cost-sensitive multi-label learning with positive and negative label pairwise correlations. Neural Netw 108:411–423
Dembczyński K, Waegeman W, Cheng W, Hüllermeier E (2012) On label dependence and loss minimization in multi-label classification. Mach Learn 88(1-2):5–45
Tsoumakas G, Dimou A, Spyromitros E, Mezaris V, Kompatsiaris I, Vlahavas I (2009) Correlation-based pruning of stacked binary relevance models for multi-label learning. In: Proceedings of the 1st International Workshop on Learning from Multi-label Data, pp 101–116
Chekina L, Gutfreund D, Kontorovich A, Rokach L, Shapira B (2013) Exploiting label dependencies for improved sample complexity. Mach Learn 91(1):1–42
Alali A, Kubat M (2015) Prudent: A pruned and confident stacking approach for multi-label classification. IEEE Trans Knowl Data Eng 27(9):2480–2493
Huang S-J, Zhou Z-H (2012) Multi-label learning by exploiting label correlations locally. In: Twenty-sixth AAAI conference on artificial intelligence
Zhang J, Li C, Cao D, Lin Y, Su S, Dai L, Li S (2018) Multi-label learning with label-specific features by resolving label correlations. Knowl-Based Syst 159:148–157
Charte F, Rivera AJ, del Jesus MJ, Herrera F (2015) Addressing imbalance in multilabel classification: Measures and random resampling algorithms. Neurocomputing 163:3–16
Charte F, Rivera AJ, del Jesus MJ, Herrera F (2015) Mlsmote: Approaching imbalanced multilabel learning through synthetic instance generation. Knowl-Based Syst 89:385–397
Ding M, Yang Y, Lan Z (2018) Multi-label imbalanced classification based on assessments of cost and value. Appl Intell 48(10):3577–3590
Spyromitros-Xioufis E, Spiliopoulou M, Tsoumakas G, Vlahavas I (2011) Dealing with concept drift and class imbalance in multi-label stream classification. Department of Computer Science, Aristotle University of Thessaloniki
Quevedo J R, Luaces O, Bahamonde A (2012) Multilabel classifiers with a probabilistic thresholding strategy. Pattern Recogn 45(2):876–883
Pillai I, Fumera G, Roli F (2013) Threshold optimisation for multi-label classifiers. Pattern Recogn 46(7):2055–2065
Petterson J, Caetano T S (2010) Reverse multi-label learning. In: Advances in neural information processing systems, pp 1912–1920
Dembczynski K, Jachnik A, Kotlowski W, Waegeman W, Hüllermeier E (2013) Optimizing the f-measure in multi-label classification: Plug-in rule approach versus structured loss minimization. In: International conference on machine learning, pp 1130–1138
Wu B, Lyu S, Ghanem B (2016) Constrained submodular minimization for missing labels and class imbalance in multi-label learning.. In: AAAI, pp 2229–2236
Paredes R, Vidal E (2006) Learning weighted metrics to minimize nearest-neighbor classification error. IEEE Trans Pattern Anal Mach Intell 28(7):1100–1110. https://doi.org/10.1109/TPAMI.2006.145
Paredes R, Vidal E (2006) Learning prototypes and distances: A prototype reduction technique based on nearest neighbor error minimization. Pattern Recogn 39(2):180–188
Jahromi MZ, Parvinnia E, John R (2009) A method of learning weighted similarity function to improve the performance of nearest neighbor. Inf Sci 179(17):2964–2973
Rastin N, Jahromi MZ, Taheri M (2020) A generalized weighted distance k-nearest neighbor for multi-label problems. Pattern Recogn 107526
Zhang Q-W, Zhong Y, Zhang M-L (2018) Feature-induced labeling information enrichment for multi-label learning.. In: AAAI, pp 4446–4453
Dembczy K (2010) Bayes optimal multilabel classification via probabilistic classifier chains. In: Proceedings of the 27th international conference on machine learning, pp 279–286
Qi G-J, Hua X-S, Rui Y, Tang J, Mei T, Zhang H-J (2007) Correlative multi-label video annotation categories and subject descriptors. Context
Pachet F, Roy P (2009) Improving multilabel analysis of music titles: A large-scale validation of the correction approach. IEEE Trans Audio Speech Lang Process 17(2):335–343. https://doi.org/10.1109/TASL.2008.2008734
Montañes E, Senge R, Barranquero J, Ramón Quevedo J, José del Coz J, Hüllermeier E (2013) Dependent binary relevance models for multi-label classification. Pattern Recogn 47(3):1494–1508. https://doi.org/10.1016/j.patcog.2013.09.029
Zhang M-L, Li Y-K, Liu X-Y, Geng X (2018) Binary relevance for multi-label learning: an overview. Front Comput Sci 12(2):191–202
Chen Y-N, Weng W, Wu S-X, Chen B-H, Fan Y-L, Liu J-H (2020) An efficient stacking model with label selection for multi-label classification. Appl Intell 1–18
Cheng W, Hüllermeier E (2009) Combining instance-based learning and logistic regression for multilabel classification. Mach Learn 76(2-3):211–225
Rastin N, Jahromi MZ, Taheri M (2017) Multi-label classification systems by the use of supervised clustering. In: Artificial intelligence and signal processing conference (AISP), 2017. IEEE, pp 246–249
Zhang M-L, Zhou Z-H (2013) A review on multi-label learning algorithms. IEEE Trans Knowl Data Eng 26(8):1819–1837
Neave HR, Worthington PL (1988) Distribution-free tests. Unwin Hyman, London
Tsoumakas G, Spyromitros-Xioufis E, Vilcek J, Vlahavas I (2011) Mulan: A java library for multi-label learning. J Mach Learn Res 12(Jul):2411–2414
Younes Z, Abdallah F, Denœux T (2008) Multi-label classification algorithm derived from k-nearest neighbor rule with label dependencies. In: Signal Processing Conference, 2008 16th European. IEEE, pp 1–5
Xu J (2011) Multi-label weighted k-nearest neighbor classifier with adaptive weight estimation. In: International conference on neural information processing. Springer, pp 79–88
Spyromitros E, Tsoumakas G, Vlahavas I (2008) An empirical study of lazy multilabel classification algorithms. In: Hellenic conference on artificial intelligence. Springer, pp 401–406
Kimura K, Sun L, Kudo M (2017) Mlc toolbox: A matlab/octave library for multi-label classification. arXiv:1704.02592
Sun L, Kudo M, Kimura K (2016) Multi-label classification with meta-label-specific features. In: Pattern Recognition (ICPR), 2016 23rd International conference on. IEEE, pp 1612–1617
Wu G, Zheng R, Tian Y, Liu D (2020) Joint ranking svm and binary relevance with robust low-rank learning for multi-label classification. Neural Netw 122:24–39
Shu S, Lv F, Feng L, Huang J, He S, He J, Li L (2020) Incorporating multiple cluster centers for multi-label learning. arXiv:2004.08113
Asuncion A, Newman D (2007) Uci machine learning repository
Alcalá-Fdez J, Fernández A, Luengo J, Derrac J, García S, Sánchez L, Herrera F (2011) Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. J Mult-Valued Logic Soft Comput 17
Bi J, Zhang C (2018) An empirical comparison on state-of-the-art multi-class imbalance learning algorithms and a new diversified ensemble learning scheme. Knowl-Based Syst 158:81–93
Liu X-Y, Li Q-Q, Zhou Z-H (2013) Learning imbalanced multi-class data with optimal dichotomy weights. In: 2013 IEEE 13th international conference on data mining. IEEE, pp 478–487
Ghanem AS, Venkatesh S, West G (2010) Multi-class pattern classification in imbalanced data. In: 2010 20th international conference on pattern recognition. IEEE, pp 2881–2884
Wang S, Chen H, Yao X (2010) Negative correlation learning for classification ensembles. In: The 2010 International joint conference on neural networks (IJCNN). IEEE, pp 1–8
Hoens TR, Qian Q, Chawla NV, Zhou Z-H (2012) Building decision trees for the multi-class imbalance problem. In: Pacific-Asia Conference on Knowledge Discovery and Data Mining. Springer, pp 122–134
Ramentol E, Vluymans S, Verbiest N, Caballero Y, Bello R, Cornelis C, Herrera F (2014) Ifrowann: imbalanced fuzzy-rough ordered weighted average nearest neighbor classification. IEEE Trans Fuzzy Syst 23(5):1622–1637
Dietterich TG, Bakiri G (1991) Error-correcting output codes: A general method for improving multiclass inductive learning programs. In: AAAI. Citeseer, pp 572–577
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interests
The authors declare that they have no conflict of interest.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Rastin, N., Jahromi, M.Z. & Taheri, M. Feature weighting to tackle label dependencies in multi-label stacking nearest neighbor. Appl Intell 51, 5200–5218 (2021). https://doi.org/10.1007/s10489-020-02073-9
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-020-02073-9