Interpretability and Explainability in Machine Learning Systems

Main Article Content

Gurpreet Singh

Abstract

Interpretability and explainability are critical aspects of machine learning (ML) systems, especially when deployed in high-stakes domains. This survey reviews key definitions, techniques, and challenges associated with interpretability and explainability in ML. We categorize approaches into inherently interpretable models, model-specific explanations, and model-agnostic post-hoc methods. The paper discusses trade-offs between model performance and transparency, evaluation metrics, and emerging directions to enhance user trust and regulatory compliance.

Article Details

How to Cite
Gurpreet Singh. (2023). Interpretability and Explainability in Machine Learning Systems. International Journal on Recent and Innovation Trends in Computing and Communication, 11(5), 577–579. Retrieved from https://mail.ijritcc.org/index.php/ijritcc/article/view/11716
Section
Articles