Requirements Engineering for Explainable AI
Artificial intelligence (AI) has a growing influence on every aspect of life. These systems need to be transparent, accountable, and explainable to be reliable and safe. Explainability helps to reduce the opacity of such systems and aids in gaining end-user trust in the system. Therefore, explainability can be seen as an emerging requirement for AI-based systems. A number of studies emphasize the transparency and explainability of AI-based systems. However, studies on identifying stakeholders to these requirements and how to elicit and specify explainability requirements are still rare and at an early stage. This Ph.D. research aims to establish a comprehensive reference process model that can serve as a guide for practitioners to address explainability requirements concerning AI-based systems.
Requirements Engineering for Explainable AI (PhD_symposium_paper_RE23 (6).pdf) | 270KiB |
Mon 4 SepDisplayed time zone: Amsterdam, Berlin, Bern, Rome, Stockholm, Vienna change
14:00 - 15:30 | |||
14:00 30mPaper | Requirements Engineering for Explainable AI Doctoral Symposium File Attached | ||
14:30 30mPaper | An Investigation of Requirements Engineering Teaching in Higher Education in Switzerland Doctoral Symposium A: Anthea Moravánszky University of Szeged, Hungary; University of Applied Sciences of the Grisons, Switzerland File Attached | ||
15:00 30mPaper | A Requirements-Driven Agent- and Goal-Oriented Conceptual Modeling Framework for Responsible AI Doctoral Symposium |