Abstract
Assurance Cases (ACs) are an established approach in safety engineering to argue quality claims in a structured way. In the context of quality assurance for Machine Learning (ML)-based software components, ACs are also being discussed and appear promising. Tools for operationalizing ACs do exist, yet mainly focus on supporting safety engineers on the system level. However, assuring the quality of an ML component within the system is commonly the responsibility of data scientists, who are usually less familiar with these tools. To address this gap, we propose a framework to support the operationalization of ACs for ML components based on technologies that data scientists use on a daily basis: Python and Jupyter Notebook. Our aim is to make the process of creating ML-related evidence in ACs more effective. Results from the application of the framework, documented through notebooks, can be integrated into existing AC tools. We illustrate the application of the framework on an example excerpt concerned with the quality of the test data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
GSN Community Standard Version 1 (2011). https://scsc.uk/r141:1?t=1. Accessed 28 July 2023
Feather, M.S., Slingerland, P.C., Guerrini, S., Spolaor, M.: Assurance guidance for machine learning in a safety-critical system. In: WAAM 2022 (2022)
Kläs, M., Adler, R., Jöckel, L., Groß, J., Reich, J.: Using complementary risk acceptance criteria to structure assurance cases for safety-critical AI components. In: AISafety 2021 (2021)
Hawkins, R., et al.: Guidance on the assurance of machine learning in autonomous systems (AMLAS). arXiv preprint arXiv:2102.01564 (2021)
ASCE Software Overview. https://www.adelard.com/asce/. Accessed 28 July 2023
Integrated Safety Case Development Environment. http://www.iscade.co.uk/. Accessed 28 July 2023
Astah GSN. https://astah.net/products/astah-gsn/. Accessed 28 July 2023
Adedjouma, M., et al.: Engineering dependable AI systems. In: SOSE 2022 (2022)
Moncada, V., Santiago, V.: Towards proper tool support for component-oriented and model-based development of safety critical systems. Commer. Veh. Technol. (2016)
Kluyver, T., et al.: Jupyter Notebooks-a publishing format for reproducible computational workflows. In: ElPub 2016 (2016)
Hauer, M.P., Adler, R., Zweig, K.: Assuring fairness of algorithmic decision making. In: ITEQS 2021 (2021)
Rushby, J.M., Xu, X., Rangarajan, M., Weaver, T.L.: Understanding and evaluating assurance cases. NASA Technical Report No. NF1676L-22111 (2015)
Wei, R., Kelly, T.P., Dai, X., Zhao, S., Hawkins, R.: Model based system assurance using the structured assurance case metamodel. J. Syst. Softw. (2019)
BSI, Fraunhofer HHI, Verband der TÜV. Towards Auditable AI Systems (2021)
Maksimov, M., Fung, N.L.S., Kokaly, S., Chechik, M.: Two decades of assurance case tools: a survey. In: Gallina, B., Skavhaug, A., Schoitsch, E., Bitsch, F. (eds.) SAFECOMP 2018. LNCS, vol. 11094, pp. 49–59. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-99229-7_6
AMLAS Tool. https://www.york.ac.uk/assuring-autonomy/guidance/amlas/amlas-tool/. Accessed 28 July 2023
Zeller, M., Sorokos, I., Reich, J., Adler, R., Schneider, D.: Open dependability exchange metamodel: a format to exchange safety information. In: RAMS 2023 (2023)
Pedregosa, F., et al.: Scikit-learn: machine learning in python. J. Mach. Learn. Res. (2011)
Jöckel, L., Bauer, T., Kläs, M., Hauer, M.: Towards a common testing terminology for software engineering and data science experts. In: PROFES 2021 (2021)
Kläs, M., Jöckel, L., Adler, R., Reich, J.: Integrating testing and operation-related quantitative evidences in assurance cases to argue safety of data-driven AI/ML components. arXiv preprint arXiv:2202.05313 (2022)
Jöckel, L., Kläs, M.: Increasing trust in data-driven model validation – a framework for probabilistic augmentation of images and meta-data generation using application scope characteristics. In: SafeComp 2019 (2019)
Siebert, J., Seifert, D., Kelbert, P., Kläs, M., Trendowicz, A.: Badgers: generating data quality deficits with python. arXiv preprint arXiv:2307.04468 (2023)
IEC. IEC 61508-5:2010 – Functional Safety of Electrical/Electronic/Programmable Electronic Safety-related Systems (2021)
Northcutt, C.G., Jiang, L., Chuang, I.L.: Confident learning: estimating uncertainty in dataset labels. Artif. Intell. Res. (2021)
Acknowledgments
Parts of this work have been funded by the German Federal Ministry of Education and Research (BMBF) in the project “DAITA”, by the project “LOPAAS” as part of the internal funding program “ICON” of the Fraunhofer-Gesellschaft, by the project “AIControl” as part of the funding program “KMU akut” of the Fraunhofer-Gesellschaft, and by the German Federal Ministry for Economic Affairs and Energy in the project “SPELL”.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Jöckel, L. et al. (2024). Operationalizing Assurance Cases for Data Scientists: A Showcase of Concepts and Tooling in the Context of Test Data Quality for Machine Learning. In: Kadgien, R., Jedlitschka, A., Janes, A., Lenarduzzi, V., Li, X. (eds) Product-Focused Software Process Improvement. PROFES 2023. Lecture Notes in Computer Science, vol 14483. Springer, Cham. https://doi.org/10.1007/978-3-031-49266-2_10
Download citation
DOI: https://doi.org/10.1007/978-3-031-49266-2_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-49265-5
Online ISBN: 978-3-031-49266-2
eBook Packages: Computer ScienceComputer Science (R0)