Operationalizing Assurance Cases for Data Scientists: A Showcase of Concepts and Tooling in the Context of Test Data Quality for Machine Learning | SpringerLink
Skip to main content

Operationalizing Assurance Cases for Data Scientists: A Showcase of Concepts and Tooling in the Context of Test Data Quality for Machine Learning

  • Conference paper
  • First Online:
Product-Focused Software Process Improvement (PROFES 2023)

Abstract

Assurance Cases (ACs) are an established approach in safety engineering to argue quality claims in a structured way. In the context of quality assurance for Machine Learning (ML)-based software components, ACs are also being discussed and appear promising. Tools for operationalizing ACs do exist, yet mainly focus on supporting safety engineers on the system level. However, assuring the quality of an ML component within the system is commonly the responsibility of data scientists, who are usually less familiar with these tools. To address this gap, we propose a framework to support the operationalization of ACs for ML components based on technologies that data scientists use on a daily basis: Python and Jupyter Notebook. Our aim is to make the process of creating ML-related evidence in ACs more effective. Results from the application of the framework, documented through notebooks, can be integrated into existing AC tools. We illustrate the application of the framework on an example excerpt concerned with the quality of the test data.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 10295
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 12869
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. GSN Community Standard Version 1 (2011). https://scsc.uk/r141:1?t=1. Accessed 28 July 2023

  2. Feather, M.S., Slingerland, P.C., Guerrini, S., Spolaor, M.: Assurance guidance for machine learning in a safety-critical system. In: WAAM 2022 (2022)

    Google Scholar 

  3. Kläs, M., Adler, R., Jöckel, L., Groß, J., Reich, J.: Using complementary risk acceptance criteria to structure assurance cases for safety-critical AI components. In: AISafety 2021 (2021)

    Google Scholar 

  4. Hawkins, R., et al.: Guidance on the assurance of machine learning in autonomous systems (AMLAS). arXiv preprint arXiv:2102.01564 (2021)

  5. ASCE Software Overview. https://www.adelard.com/asce/. Accessed 28 July 2023

  6. Integrated Safety Case Development Environment. http://www.iscade.co.uk/. Accessed 28 July 2023

  7. Astah GSN. https://astah.net/products/astah-gsn/. Accessed 28 July 2023

  8. Adedjouma, M., et al.: Engineering dependable AI systems. In: SOSE 2022 (2022)

    Google Scholar 

  9. Moncada, V., Santiago, V.: Towards proper tool support for component-oriented and model-based development of safety critical systems. Commer. Veh. Technol. (2016)

    Google Scholar 

  10. Kluyver, T., et al.: Jupyter Notebooks-a publishing format for reproducible computational workflows. In: ElPub 2016 (2016)

    Google Scholar 

  11. Hauer, M.P., Adler, R., Zweig, K.: Assuring fairness of algorithmic decision making. In: ITEQS 2021 (2021)

    Google Scholar 

  12. Rushby, J.M., Xu, X., Rangarajan, M., Weaver, T.L.: Understanding and evaluating assurance cases. NASA Technical Report No. NF1676L-22111 (2015)

    Google Scholar 

  13. Wei, R., Kelly, T.P., Dai, X., Zhao, S., Hawkins, R.: Model based system assurance using the structured assurance case metamodel. J. Syst. Softw. (2019)

    Google Scholar 

  14. BSI, Fraunhofer HHI, Verband der TÜV. Towards Auditable AI Systems (2021)

    Google Scholar 

  15. Maksimov, M., Fung, N.L.S., Kokaly, S., Chechik, M.: Two decades of assurance case tools: a survey. In: Gallina, B., Skavhaug, A., Schoitsch, E., Bitsch, F. (eds.) SAFECOMP 2018. LNCS, vol. 11094, pp. 49–59. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-99229-7_6

    Chapter  Google Scholar 

  16. AMLAS Tool. https://www.york.ac.uk/assuring-autonomy/guidance/amlas/amlas-tool/. Accessed 28 July 2023

  17. Zeller, M., Sorokos, I., Reich, J., Adler, R., Schneider, D.: Open dependability exchange metamodel: a format to exchange safety information. In: RAMS 2023 (2023)

    Google Scholar 

  18. Pedregosa, F., et al.: Scikit-learn: machine learning in python. J. Mach. Learn. Res. (2011)

    Google Scholar 

  19. Jöckel, L., Bauer, T., Kläs, M., Hauer, M.: Towards a common testing terminology for software engineering and data science experts. In: PROFES 2021 (2021)

    Google Scholar 

  20. Kläs, M., Jöckel, L., Adler, R., Reich, J.: Integrating testing and operation-related quantitative evidences in assurance cases to argue safety of data-driven AI/ML components. arXiv preprint arXiv:2202.05313 (2022)

  21. Jöckel, L., Kläs, M.: Increasing trust in data-driven model validation – a framework for probabilistic augmentation of images and meta-data generation using application scope characteristics. In: SafeComp 2019 (2019)

    Google Scholar 

  22. Siebert, J., Seifert, D., Kelbert, P., Kläs, M., Trendowicz, A.: Badgers: generating data quality deficits with python. arXiv preprint arXiv:2307.04468 (2023)

  23. IEC. IEC 61508-5:2010 – Functional Safety of Electrical/Electronic/Programmable Electronic Safety-related Systems (2021)

    Google Scholar 

  24. Northcutt, C.G., Jiang, L., Chuang, I.L.: Confident learning: estimating uncertainty in dataset labels. Artif. Intell. Res. (2021)

    Google Scholar 

Download references

Acknowledgments

Parts of this work have been funded by the German Federal Ministry of Education and Research (BMBF) in the project “DAITA”, by the project “LOPAAS” as part of the internal funding program “ICON” of the Fraunhofer-Gesellschaft, by the project “AIControl” as part of the funding program “KMU akut” of the Fraunhofer-Gesellschaft, and by the German Federal Ministry for Economic Affairs and Energy in the project “SPELL”.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lisa Jöckel .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Jöckel, L. et al. (2024). Operationalizing Assurance Cases for Data Scientists: A Showcase of Concepts and Tooling in the Context of Test Data Quality for Machine Learning. In: Kadgien, R., Jedlitschka, A., Janes, A., Lenarduzzi, V., Li, X. (eds) Product-Focused Software Process Improvement. PROFES 2023. Lecture Notes in Computer Science, vol 14483. Springer, Cham. https://doi.org/10.1007/978-3-031-49266-2_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-49266-2_10

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-49265-5

  • Online ISBN: 978-3-031-49266-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics