Model Explainability
Techniques and documentation that make an AI model’s decision logic understandable to stakeholders and auditors.
Definition
A combination of inherent (interpretable models) and post-hoc (SHAP, LIME, counterfactuals) methods that reveal feature importances, decision rules, or alternative outcome scenarios. Governance requires selecting explainability techniques suited to the model and audience, embedding explanations in user interfaces or compliance reports, and validating that explanations accurately reflect model behavior.
Real-World Example
A credit-card fraud model provides SHAP explanations with each alert: “Top factors: unusual location, atypical transaction size.” Fraud analysts use these explanations to triage alerts more effectively and regulators review the SHAP reports during compliance inspections.