Abstract
Binary classifiers are traditionally studied by propositional logic (\(\textsf{PL}\)). \(\textsf{PL}\) can only represent them as white boxes, under the assumption that the underlying Boolean function is fully known. Binary classifiers used in practical applications and trained by machine learning are however opaque. They are usually described as black boxes. In this paper, we provide a product modal logic called PLC (Product modal Logic for binary input Classifier) in which the notion of “black box” is interpreted as the uncertainty over a set of classifiers. We give results about axiomatics and complexity of satisfiability checking for our logic. Moreover, we present a dynamic extension in which the process of acquiring new information about the actual classifier can be represented.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Notice that p denotes an input variable, while x is an output value rather than the output variable, which makes sense of the symbolic difference between p and \(\textsf{t}({x})\).
- 2.
In the real world, partial knowledge may come from the data set as well as from the training process. For example, through learning, we may acquire knowledge that certain input features behave monotonically [26].
- 3.
References
Amershi, S., Cakmak, M., Knox, W., Kulesza, T.: Power to the people: the role of humans in interactive machine learning. AI Mag. 35(4), 105–120 (2014)
Audemard, G., Bellart, S., Bounia, L., Koriche, F., Lagniez, J.M., Marquis, P.: On the computational intelligibility of boolean classifiers. In: Proceedings of the International Conference on Principles of Knowledge Representation and Reasoning, vol. 18, pp. 74–86 (2021)
Baltag, A., van Benthem, J.: A simple logic of functional dependence. J. Philos. Log. 50(5), 939–1005 (2021). https://doi.org/10.1007/s10992-020-09588-z
Bezhanishvili, N., Hodkinson, I.M.: All normal extensions of S5-squared are finitely axiomatizable. Stud. Logica. 78(3), 443–457 (2004)
Bezhanishvili, N., Marx, M.: All proper normal extensions of S5-square have the polynomial size model property. Stud. Logica. 73(3), 367–382 (2003). https://doi.org/10.1023/A:1023383112908
Darwiche, A., Hirth, A.: On the reasons behind decisions. In: ECAI 2020–24th European Conference on Artificial Intelligence. Frontiers in Artificial Intelligence and Applications, vol. 325, pp. 712–720. IOS Press (2020)
van Ditmarsch, H., van Der Hoek, W., Kooi, B.: Dynamic Epistemic Logic, Synthese Library, vol. 337 Springer (2007). https://doi.org/10.1007/978-1-4020-5839-4
Gabbay, D.M., Kurucz, A., Wolter, F., Zakharyaschev, M.: Many-dimensional modal logics: theory and applications. Elsevier (2003)
Grossi, D., Lorini, E., Schwarzentruber, F.: The ceteris paribus structure of logics of game forms. J. Artif. Intell. Res. 53, 91–126 (2015)
Halpern, J.Y.: The effect of bounding the number of primitive propositions and the depth of nesting on the complexity of modal logic. Artif. Intell. 75(2), 361–372 (1995)
Hempel, C.G., Oppenheim, P.: Studies in the logic of explanation. Philos. sci. 15(2), 135–175 (1948)
Ignatiev, A., Narodytska, N., Asher, N., Marques-Silva, J.: From contrastive to abductive explanations and back again. In: Baldoni, M., Bandini, S. (eds.) AIxIA 2020. LNCS (LNAI), vol. 12414, pp. 335–355. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-77091-4_21
Ignatiev, A., Narodytska, N., Marques-Silva, J.: Abduction-based explanations for machine learning models. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 1511–1519 (2019)
Kment, B.: Counterfactuals and explanation. Mind 115(458), 261–310 (2006)
Liu, X., Lorini, E.: A logic for binary classifiers and their explanation. In: Baroni, P., Benzmüller, C., Wáng, Y.N. (eds.) CLAR 2021. LNCS (LNAI), vol. 13040, pp. 302–321. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-89391-0_17
Lundberg, S., Lee, S.I.: A unified approach to interpreting model predictions. arXiv preprint arXiv:1705.07874 (2017)
Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1–38 (2019)
Molnar, C.: Interpretable Machine Learning. Lulu. com (2020)
Plaza, J.: Logics of public communications. Synthese 158(2), 165–179 (2007). https://doi.org/10.1007/s11229-007-9168-7
Quine, W.V.: A way to simplify truth functions. Am. Math. Mon. 62(9), 627–631 (1955)
Ribeiro, M.T., Singh, S., Guestrin, C.: Why should i trust you? Explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135–1144 (2016)
Ribeiro, M.T., Singh, S., Guestrin, C.: Anchors: high-precision model-agnostic explanations. In: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), vol. 32 (2018)
Shih, A., Choi, A., Darwiche, A.: Formal verification of bayesian network classifiers. In: International Conference on Probabilistic Graphical Models, pp. 427–438. PMLR (2018)
Walton, D.: A new dialectical theory of explanation. Philos. Explor. 7(1), 71–89 (2004)
Yang, F., Väänänen, J.: Propositional logics of dependence. Ann. Pure Appl. Logic 167(7), 557–589 (2016)
You, S., Ding, D., Canini, K., Pfeifer, J., Gupta, M.: Deep lattice networks and partial monotonic functions. Advances in neural information processing systems 30 (2017)
Acknowledgments
Support from the ANR-3IA Artificial and Natural Intelligence Toulouse Institute (ANITI) is gratefully acknowledged.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Liu, X., Lorini, E. (2022). A Logic of “Black Box” Classifier Systems. In: Ciabattoni, A., Pimentel, E., de Queiroz, R.J.G.B. (eds) Logic, Language, Information, and Computation. WoLLIC 2022. Lecture Notes in Computer Science, vol 13468. Springer, Cham. https://doi.org/10.1007/978-3-031-15298-6_10
Download citation
DOI: https://doi.org/10.1007/978-3-031-15298-6_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-15297-9
Online ISBN: 978-3-031-15298-6
eBook Packages: Computer ScienceComputer Science (R0)