Accountability Layers: Explaining Complex System Failures by Parts
DOI:
https://doi.org/10.1609/aaai.v37i13.26806Keywords:
New Faculty HighlightsAbstract
With the rise of AI used for critical decision-making, many important predictions are made by complex and opaque AI algorithms. The aim of eXplainable Artificial Intelligence (XAI) is to make these opaque decision-making algorithms more transparent and trustworthy. This is often done by constructing an ``explainable model'' for a single modality or subsystem. However, this approach fails for complex systems that are made out of multiple parts. In this paper, I discuss how to explain complex system failures. I represent a complex machine as a hierarchical model of introspective sub-systems working together towards a common goal. The subsystems communicate in a common symbolic language. This work creates a set of explanatory accountability layers for trustworthy AI.Downloads
Published
2023-09-06
How to Cite
Gilpin, L. H. (2023). Accountability Layers: Explaining Complex System Failures by Parts. Proceedings of the AAAI Conference on Artificial Intelligence, 37(13), 15439-15439. https://doi.org/10.1609/aaai.v37i13.26806
Issue
Section
New Faculty Highlights