AI Explainability

The extent to which the internal mechanics of an AI system can be understood and interpreted by humans.

Definition

The suite of methods (surrogate models, feature-attribution, counterfactuals) and processes (documentation, user-friendly dashboards) that make AI decisions transparent, so stakeholders can understand, contest, and trust outcomes.

Real-World Example

A credit-scoring model uses SHAP to highlight which financial factors (e.g., “low credit history”) influenced a denial. Loan officers review these explanations alongside the AI’s recommendation, allowing applicants to correct inaccuracies and request human review.