[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to main content

Candidate Solutions for Defining Explainability Requirements of AI Systems

  • Conference paper
  • First Online:
Requirements Engineering: Foundation for Software Quality (REFSQ 2024)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14588))

  • 839 Accesses

Abstract

[Context and Motivation] Many recent studies highlight explainability as an important requirement that supports in building transparent, trustworthy, and responsible AI systems. As a result, there is an increasing number of solutions that researchers have developed to assist in the definition of explainability requirements. [Question] We conducted a literature study to analyze what kind of candidate solutions are proposed for defining the explainability requirements of AI systems. The focus of this literature review is especially on the field of requirements engineering (RE). [Results] The proposed solutions for defining explainability requirements such as approaches, frameworks, and models are comprehensive. They can be used not only for RE activities but also for testing and evaluating the explainability of AI systems. In addition to the comprehensive solutions, we identified 30 practices that support the development of explainable AI systems. The literature study also revealed that most of the proposed solutions have not been evaluated in real projects, and there is a need for empirical studies. [Contribution] For researchers, the study provides an overview of the candidate solutions and describes research gaps. For practitioners, the paper summarizes potential practices that can help them define and evaluate the explainability requirements of AI systems.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
£29.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
GBP 19.95
Price includes VAT (United Kingdom)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
GBP 47.99
Price includes VAT (United Kingdom)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
GBP 59.99
Price includes VAT (United Kingdom)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Köhl, M.A., et al.: Explainability as a non-functional requirement. In: International Requirements Engineering Conference, pp. 363–368 (2019)

    Google Scholar 

  2. Nguyen, M.-L., Phung, T., Ly, D.-H., Truong, H.-L.: Holistic explainability requirements for end-to-end machine learning in IoT cloud systems. In: International Requirements Engineering Conference Workshops (REW), pp. 188–194 (2021)

    Google Scholar 

  3. Ramos, H., Fonseca, M., Ponciano, L.: Modeling and evaluating personas with software explainability requirements. In: Ruiz, P.H., Agredo-Delgado, V., Kawamoto, A.L.S. (eds.) Human-Computer Interaction: 7th Iberoamerican Workshop, HCI-COLLAB 2021, Sao Paulo, Brazil, September 8–10, 2021, Proceedings, pp. 136–149. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-92325-9_11

    Chapter  Google Scholar 

  4. Kauppinen, M., Vartiainen, M., Kontio, J., Kujala, S., Sulonen, R.: Implementing requirements engineering processes throughout organizations: success factors and challenges. Inf. Softw. Technol. 46(14), 937–953 (2004)

    Article  Google Scholar 

  5. Jovanovic, M., Schmitz, M.: Explainability as a user requirement for artificial intelligence systems. Computer 55(2), 90–94 (2022)

    Article  Google Scholar 

  6. Balasubramaniam, N., Kauppinen, M., Rannisto, A., Hiekkanen, K., Kujala, S.: Transparency and explainability of AI systems: from ethical guidelines to requirements. Inf. Softw. Technol. 159, 107197 (2023)

    Article  Google Scholar 

  7. Kotonya, G., Sommerville, I.: Requirements Engineering. Wiley, Hoboken (1998)

    Google Scholar 

  8. Chazette, L., Schneider, K.: Explainability as a non-functional requirement: challenges and recommendations. Requirements Eng. 25(4), 493–514 (2020)

    Article  Google Scholar 

  9. Charmaz, K.: Constructing Grounded Theory, 2nd edn. Sage publications (2014)

    Google Scholar 

  10. Chazette, L., Brunotte, W., Speith, T.: Explainable software systems: from requirements analysis to system evaluation. Requirements Eng. 27(4), 457–487 (2022)

    Article  Google Scholar 

  11. Sommerville, I., Sawyer, P.: Requirements Engineering: A Good Practice Guide. Wiley, Hoboken (2004)

    Google Scholar 

  12. Chazette, L., Klos, V., Herzog, F., Schneider, K.: Requirements on explanations: a quality framework for explainability. In: Requirements Engineering Conference, pp. 140–152 (2022)

    Google Scholar 

  13. Chazette, L., Klünder, J., Balci, M., Schneider, K.: How can we develop explainable systems? Insights from a literature review and an interview study. In: International Conference on Software and System Processes and International Conference on Global Software Engineering, pp. 1–12 (2022)

    Google Scholar 

  14. Arrieta, A.B., et al.: Explainable Artificial Intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fus. 58, 82–115 (2020)

    Article  Google Scholar 

  15. Maalej, W., Pham, Y.D., Chazette, L.: Tailoring requirements engineering for responsible AI. IEEE Comput. 56(4), 18–27 (2023)

    Article  Google Scholar 

  16. Habibullah, K.M., Horkoff, J.: Non-functional requirements for machine learning: understanding current use and challenges in industry. In: International Requirements Engineering Conference, pp. 13–23 (2021)

    Google Scholar 

  17. Balasubramaniam, N., Kauppinen, M., Kujala, S., Hiekkanen, K.: Ethical Guidelines for solving ethical issues and developing AI systems. In: Morisio, M., Torchiano, M., Jedlitschka, A. (eds.) PROFES 2020. LNCS, vol. 12562, pp. 331–346. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-64148-1_21

    Chapter  Google Scholar 

  18. Koopmann, B., et al.: Challenges in achieving explainability for cooperative transportation systems. In: International Requirements Engineering Conference Workshops (REW), pp. 114–119 (2022)

    Google Scholar 

  19. Cysneiros, L.M., Raffi, M., Leite, J.C.S.P.: Software transparency as a key requirement for self-driving cars. In: International Requirements Engineering Conference (RE), pp. 382–387 (2018)

    Google Scholar 

  20. Langer, M., Baum, K., Hartmann, K., Hessel, S., Speith, T., Wahl, J.: Explainability auditing for intelligent systems: a rationale for multi-disciplinary perspectives. In: Requirements Engineering Conference Workshops (REW), pp. 164–168 (2021)

    Google Scholar 

  21. Kuhrmann, M., Fernández, D.M., Daneva, M.: On the pragmatic design of literature studies in software engineering: an experience-based guideline. Empir. Softw. Eng. 22(6), 2852–2891 (2017)

    Article  Google Scholar 

  22. Ivarsson, M., Gorschek, T.: Technology transfer decision support in requirements engineering research: a systematic review of REj. Requirements Eng. 14, 155–175 (2009)

    Article  Google Scholar 

  23. Wohlin, C., Runeson, P.: Guiding the selection of research methodology in industry–academia collaboration in software engineering. Inf. Softw. Technol. 140, 106678 (2021)

    Article  Google Scholar 

  24. Vogelsang, A., Borg, M.: Requirements engineering for machine learning: perspectives from data scientists. In: Requirements Engineering Conference Workshops (REW), pp. 245–251 (2019)

    Google Scholar 

  25. Ahmad, K., Abdelrazek, M., Arora, C., Bano, M., Grundy, J.: Requirements engineering for artificial intelligence systems: a systematic mapping study. Inf. Softw. Technol. 158, 107176 (2023)

    Article  Google Scholar 

  26. Habiba, U.-E., Bogner, J., Wagner, S.: Can requirements engineering support explainable artificial intelligence? Towards a user-centric approach for explainability requirements. In: International Requirements Engineering Conference Workshops (REW), pp. 162–165 (2022)

    Google Scholar 

  27. Li, T., Han, L.: Dealing with explainability requirements for machine learning systems. In: Computers, Software, and Applications Conference (COMPSAC), pp. 1203–1208 (2023)

    Google Scholar 

  28. Phillips, P.J., et al.: Four principles of explainable artificial intelligence. National Institute of Standards and Technology (U.S.), NIST IR 8312 (2021)

    Google Scholar 

  29. Nuseibeh, B., Easterbrook, S.: Requirements engineering: a roadmap. In: International Conference on Software Engineering, pp. 37–46 (2000)

    Google Scholar 

  30. Inayat, I., Salim, S., Marczak, S., Daneva, M., Shamshirband, S.: A systematic literature review on agile requirements engineering practices and challenges. Comput. Hum. Behav. 51(27), 915–929 (2015)

    Article  Google Scholar 

  31. Habibullah, K.M., Gay, G., Horkoff, J.: Non-functional requirements for machine learning: understanding current use and challenges among practitioners. Requirements Eng. 28(2), 283–316 (2023)

    Article  Google Scholar 

  32. Wohlin, C., Kalinowski, M., Felizardo, K.R., Mendes, E.: Successful combination of database search and snowballing for identification of primary studies in systematic literature studies. Inf. Softw. Technol. 147, 106908 (2022)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nagadivya Balasubramaniam .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Balasubramaniam, N., Kauppinen, M., Truong, HL., Kujala, S. (2024). Candidate Solutions for Defining Explainability Requirements of AI Systems. In: Mendez, D., Moreira, A. (eds) Requirements Engineering: Foundation for Software Quality. REFSQ 2024. Lecture Notes in Computer Science, vol 14588. Springer, Cham. https://doi.org/10.1007/978-3-031-57327-9_8

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-57327-9_8

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-57326-2

  • Online ISBN: 978-3-031-57327-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics