“Explainable AI (XAI) is a set of processes and methods that allow human users to understand, trust, and effectively manage the outputs of machine learning algorithms. It aims to move away from ‘black box’ models.” – Explainable AI (XAI)
Explainable AI (XAI) encompasses a collection of processes, techniques, and methods designed to make the outputs and decision-making of machine learning algorithms transparent, interpretable, and trustworthy for human users.1,2,4 By addressing the inherent opacity of complex models, particularly deep learning systems often described as ‘black boxes’, XAI facilitates intellectual oversight, reveals reasoning behind predictions, and supports fairness, accountability, and transparency (FAT) in AI deployment.1,6 This is essential in high-stakes domains such as healthcare, finance, and autonomous systems, where understanding why a model reaches a decision is as critical as the decision itself.2,5
Why Explainable AI is Needed
Traditional machine learning models, especially advanced ones like neural networks, excel in performance but lack transparency, leading to challenges in trust, bias detection, regulatory compliance, and error correction.1,3 XAI mitigates these by answering key questions: Why did the model predict this? Why not an alternative? When is it reliable or prone to failure?2 It promotes responsible AI by enabling stakeholders to verify decisions, debug models, and ensure ethical outcomes, fostering broader adoption.4,5,7
How Explainable AI Works
XAI architectures typically integrate three core components: the machine learning model (e.g., supervised, unsupervised, or reinforcement learning), an explanation algorithm (using feature importance, attribution methods, or visualisations), and a user interface for comprehensible insights.1 Techniques vary by approach:
- Intrinsic methods: Models inherently designed for interpretability, such as decision trees or linear regression, where processes are transparent by default.6
- Post-hoc methods: Applied to black-box models, including LIME (Local Interpretable Model-agnostic Explanations) or SHAP (SHapley Additive exPlanations) to approximate contributions of input features.1,6
- Visual and textual explanations: Tools like saliency maps or natural language justifications to depict model behaviour.2
Key principles include simulatability (easy prediction reproduction), decomposability (intuitive parameter explanations), and algorithmic transparency, ensuring models are justifiable and verifiable.6
Challenges and Principles
Despite progress, XAI faces trade-offs between accuracy and interpretability, with no universal definition yet consolidated.3,6,7 Core principles advocate ethical deployment: explanations must be clear, coherent, and tailored to users, concentrating on specific predictions while supporting broader model oversight.1,8
Key Theorist: Riccardo Guidotti
The preeminent theorist in Explainable AI is **Riccardo Guidotti**, an Italian computer scientist whose pioneering work laid foundational stones for the field. Born in 1969, Guidotti earned his PhD in Computer Science from the University of Turin in 1996, specialising in artificial intelligence and knowledge-based systems. He advanced to full professor at the University of Pavia, where he directs the Machine Learning lab, and holds visiting positions at institutions like the Alan Turing Institute.
Guidotti’s relationship to XAI stems from his seminal contributions to explainable models in the early 2010s. In 2017, he co-authored the landmark paper ‘A Survey of Methods for Explaining Black Box Models’, which categorised XAI techniques into local (instance-specific) and global (model-wide) explanations, influencing standards like LIME and SHAP.6 Earlier, his 2015 work on ‘Doctor XAI’ introduced counterfactual explanations-‘what-if’ scenarios revealing minimal changes needed for different outcomes-directly addressing black-box opacity.1,6 Guidotti co-founded the XAI research community, co-edited the first XAI Dagstuhl seminar in 2017, and continues shaping the field through frameworks emphasising human-centric interpretability. His biography reflects a blend of theoretical rigour and practical impact, with over 100 publications cited tens of thousands of times, positioning him as the intellectual architect of modern XAI.6
References
1. https://www.geeksforgeeks.org/artificial-intelligence/explainable-artificial-intelligencexai/
2. https://www.redhat.com/en/topics/ai/what-explainable-ai
3. https://c3.ai/glossary/machine-learning/explainability/
4. https://www.hpe.com/us/en/what-is/explainable-ai.html
5. https://www.ibm.com/think/topics/explainable-ai
6. https://en.wikipedia.org/wiki/Explainable_artificial_intelligence
7. https://www.sei.cmu.edu/blog/what-is-explainable-ai/
8. https://www.edps.europa.eu/system/files/2023-11/23-11-16_techdispatch_xai_en.pdf
9. https://www.qlik.com/us/augmented-analytics/explainable-ai

