Abstract
Data hierarchy, as a hidden property of data structure, exists in a wide range of machine learning applications. A common practice to classify such hierarchical data is first to encode the data in the Euclidean space, and then train a Euclidean classifier. However, such a paradigm leads to a performance drop due to distortion of data embedding in the Euclidean space. To relieve this issue, hyperbolic geometry is investigated as an alternative space to encode the hierarchical data for its higher ability to capture the hierarchical structures. Those methods cannot explore the full potential of the hyperbolic geometry, in the sense that such methods define the hyperbolic operations in the tangent plane, causing the distortion of data embeddings. In this paper, we develop two novel kernel formulations in the hyperbolic space, with one being positive definite (PD) and another one being indefinite, to solve the classification tasks in hyperbolic space. The PD one is defined via mapping the hyperbolic data to the Drury-Arveson (DA) space, which is a special reproducing kernel Hilbert space (RKHS). To further increase the discrimination of the classifier, an indefinite kernel is further defined in the Kreĭn spaces. Specifically, we design a 2-layer nested indefinite kernel which first maps hyperbolic data into the DA spaces, followed by a mapping from the DA spaces to the Kreĭn spaces. Extensive experiments on real-world datasets demonstrate the superiority of the proposed kernels.
Similar content being viewed by others
References
Gong J, Teng Z, Teng Q, Zhang H, Du L, Chen S, Bhuiyan Z A, Li J, Liu M, Ma H. Hierarchical graph transformer-based deep learning model for large-scale multi-label text classification. IEEE Access, 2020, 8: 30885–30896
Wang Q, Mao Z, Wang B, Guo L. Knowledge graph embedding: a survey of approaches and applications. IEEE Transactions on Knowledge and Data Engineering, 2017, 29(12): 2724–2743
Du X, Xia Y. Natural images enhancement using structure extraction and retinex. In: Proceedings of the 20th International Conference on Advanced Concepts for Intelligent Vision Systems. 2020, 408–420
Kim S, Song C, Jang J, Paik J. Edge-aware image filtering using a structure-guided CNN. IET Image Processing, 2020, 14(3): 472–479
Long J, Feng X, Zhu X, Zhang J, Gou G. Efficient superpixel-guided interactive image segmentation based on graph theory. Symmetry, 2018, 10(5): 169
Wang H, Zhang F, Zhang M, Leskovec J, Zhao M, Li W, Wang Z. Knowledge-aware graph neural networks with label smoothness regularization for recommender systems. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2019, 968–977
Ying R, He R, Chen K, Eksombatchai P, Hamilton W L, Leskovec J. Graph convolutional neural networks for web-scale recommender systems. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2018, 974–983
Lee S, Park S, Kahng M, Lee S G. PathRank: a novel node ranking measure on a heterogeneous graph for recommender systems. In: Proceedings of the 21st ACM International Conference on Information and Knowledge Management. 2012, 1637–1641
Ma R, Fang P, Drummond T, Harandi M. Adaptive poincaré point to set distance for few-shot classification. In: Proceedings of the 36th AAAI Conference on Artificial Intelligence. 2022, 1926–1934
Sun J, Xie Y, Zhang H, Faloutsos C. Less is more: compact matrix decomposition for large sparse graphs. In: Proceedings of the 7th SIAM International Conference on Data Mining. 2007, 366–377
Perozzi B, Al-Rfou R, Skiena S. DeepWalk: online learning of social representations. In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2014, 701–710
Hajiramezanali E, Hasanzadeh A, Duffield N, Narayanan K, Zhou M, Qian X. Variational graph recurrent neural networks. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 960
Linial N, London E, Rabinovich Y. The geometry of graphs and some of its algorithmic applications. Combinatorica, 1995, 15(2): 215–245
Krioukov D, Papadopoulos F, Kitsak M, Vahdat A, Boguñá M. Hyperbolic geometry of complex networks. Physical Review E, 2010, 82(3): 036106
Nickel M, Kiela D. Poincaré embeddings for learning hierarchical representations. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 6341–6350
Alanis-Lobato G, Mier P, Andrade-Navarro M A. Efficient embedding of complex networks to hyperbolic space via their Laplacian. Scientific Reports, 2016, 6: 30108
Chamberlain B P, Clough J, Deisenroth M P. Neural embeddings of graphs in hyperbolic space. 2017, arXiv preprint arXiv: 1705.10359
Ganea O E, Bécigneul G, Hofmann T. Hyperbolic entailment cones for learning hierarchical embeddings. In: Proceedings of the 35th International Conference on Machine Learning. 2018, 1632–1641
Nickel M, Kiela D. Learning continuous hierarchies in the Lorentz model of hyperbolic geometry. In: Proceedings of the 35th International Conference on Machine Learning. 2018, 3779–3788
De Sa C, Gu A, Ré C, Sala F. Representation tradeoffs for hyperbolic embeddings. In: Proceedings of Machine Learning Research, 2018, 80: 4460–4469
Suzuki R, Takahama R, Onoda S. Hyperbolic disk embeddings for directed acyclic graphs. In: Proceedings of International Conference on Machine Learning. 2019, 6066–6075
Balažević I, Allen C, Hospedales T. Multi-relational Poincaré graph embeddings. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 401
Sonthalia R, Gilbert A C. Tree! I am no tree! I am a low dimensional hyperbolic embedding. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 72
Weber M, Zaheer M, Rawat A S, Menon A, Kumar S. Robust large-margin learning in hyperbolic space. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 1499
Cho H, DeMeo B, Peng J, Berger B. Large-margin classification in hyperbolic space. In: Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics. 2019, 1832–1840
Fang P, Harandi M, Petersson L. Kernel methods in hyperbolic spaces. In: Proceedings of 2021 IEEE/CVF International Conference on Computer Vision (ICCV). 2021, 10665–10674
Rochberg R. Complex hyperbolic geometry and Hilbert spaces with complete pick kernels. Journal of Functional Analysis, 2019, 276(5): 1622–1679
Cucerzan S. Large-scale named entity disambiguation based on Wikipedia data. In: Proceedings of 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL). 2007, 708–716
Rozemberczki B, Davies R, Sarkar R, Sutton C. GEMSEC: graph embedding with self clustering. In: Proceedings of 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining. 2019, 65–72
Zhao B, Sen P, Getoor L. Entity and relationship labeling in affiliation networks. In: Proceedings of ICML Workshop on Statistical Network Analysis. 2006
Shchur O, Mumme M, Bojchevski A, Günnemann S. Pitfalls of graph neural network evaluation. 2018, arXiv preprint arXiv: 1811.05868
Bojchevski A, Günnemann S. Deep Gaussian embedding of graphs: unsupervised inductive learning via ranking. In: Proceedings of the 6th International Conference on Learning Representations. 2018
Parker J R. Notes on complex hyperbolic geometry. Preprint, 2003.
Ratcliffe J G. Foundations of Hyperbolic Manifolds. New York: Springer, 1994
Goldman W M. Complex Hyperbolic Geometry. Oxford: Oxford University Press, 1999
Kim J, Scott C D. L2 kernel classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 32(10): 1822–1831
Bordelon B, Canatar A, Pehlevan C. Spectrum dependent learning curves in kernel regression and wide neural networks. In: Proceedings of the 37th International Conference on Machine Learning. 2020, 96
Kang Z, Wen L, Chen W, Xu Z. Low-rank kernel learning for graph-based clustering. Knowledge-Based Systems, 2019, 163: 510–517
Ober S W, Rasmussen C E, van der Wilk M. The promises and pitfalls of deep kernel learning. In: Proceedings of the 37th Conference on Uncertainty in Artificial Intelligence. 2021, 1206–1216
Fang P, Zhou J, Roy S K, Ji P, Petersson L, Harandi M. Attention in attention networks for person retrieval. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022, 44(9): 4626–4641
Ong C S, Mary X, Canu S, Smola A J. Learning with non-positive kernels. In: Proceedings of the 21st International Conference on Machine Learning. 2004
Ungar A A. The hyperbolic Pythagorean theorem in the Poincaré disc model of hyperbolic geometry. The American Mathematical Monthly, 1999, 106(8): 759–763
Shalit O, Shalit O. Operator theory and function theory in Drury–Arveson space and its quotients. In: Alpay D, ed. Operator Theory. Basel: Springer, 2014, 1–50
Arcozzi N, Rochberg R, Sawyer E, Wick B D. Distance functions for reproducing kernel Hilbert spaces. Contemp., 2011, 547: 25–53
Ungar A A. From Pythagoras to Einstein: the hyperbolic Pythagorean theorem. Foundations of Physics, 1998, 28(8): 1283–1321
Birman G S, Ungar A A. The hyperbolic derivative in the Poincaré ball model of hyperbolic geometry. Journal of Mathematical Analysis and Applications, 2001, 254(1): 321–333
Dray T. The Geometry of Special Relativity. Boca Raton: CRC Press, 2012
Loosli G, Canu S, Ong C S. Learning SVM in Kreĭn spaces. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2016, 38(6): 1204–1216
Xu H M, Xue H, Chen X H, Wang Y Y. Solving indefinite kernel support vector machine with difference of convex functions programming. In: Proceedings of the 31st AAAI Conference on Artificial Intelligence. 2017, 2782–2788
Oglic D, Gärtner T. Learning in reproducing kernel Krein spaces. In: Proceedings of the International Conference on Machine Learning. 2018, 3856–3864
McAuley J, Targett C, Shi Q, van den Hengel A. Image-based recommendations on styles and substitutes. In: Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2015, 43–52
McCallum A K, Nigam K, Rennie J, Seymore K. Automating the construction of internet portals with machine learning. Information Retrieval, 2000, 3(2): 127–163
Deshmukh A A. Kernel approximation. Stats 608, 2015, 1–3
Platt J. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in Large Margin Classifiers, 1999, 10(3): 61–74
Schölkopf B, Smola A J. Learning with Kernels: support vector machines, regularization, optimization, and beyond. MIT Press, 2002
Chang C C, Lin C J. LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology, 2011, 2(3): 27
Acknowledgements
This work was supported by the National Natural Science Foundation of China (Grant No. 62076062) and the Fundamental Research Funds for the Central Universities (2242021k30056). Furthermore, it was also supported by the Chollaborative Innovation Center of Wireless Communications Technology.
Author information
Authors and Affiliations
Corresponding author
Additional information
Meimei Yang is currently pursuing PhD degree with the School of Computer Science and Engineering, Southeast University, China. She received the BS degree in Nanjing University of Information Science & Technology, China in 2014. Her research interests include pattern recognition and machine learning.
Qiao Liu received the MSc degree in Southeast University of software engineering, China in 2022. He received the BS degree in computer science from Central South University, China in 2019. His current research interests include machine learning and pattern recognition.
Xinkai Sun received the BSc degree in Computer Science from Southeast University, China. In 2021, he received the MSc degree in Southeast University of software engineering, China. His research interests include pattern recognition and machine learning.
Na Shi is currently working in State Grid Zaozhuang Power Supply Company, China. She received the BS degree in Hohai University of computer Science & Technology, China in 2017. In 2020, she received the MSc degree in Southeast University of software engineering, China. Her research interests include pattern recognition and machine learning.
Hui Xue received the BSc degree in Mathematics from Nanjing Normal University, China in 2002. In 2005, she received the MSc degree in Mathematics from Nanjing University of Aeronautics & Astronautics (NUAA). And she also received the PhD degree in Computer Application Technology at NUAA, China in 2008. Since 2009, as a Professor, she has been with the school of Computer Science and Engineering at Southeast University, China. Her research interests include pattern recognition and machine learning.
Electronic supplementary material
Rights and permissions
About this article
Cite this article
Yang, M., Liu, Q., Sun, X. et al. Towards kernelizing the classifier for hyperbolic data. Front. Comput. Sci. 18, 181301 (2024). https://doi.org/10.1007/s11704-022-2457-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11704-022-2457-y