Interpretability and Explainability in Machine Learning Systems
Main Article Content
Abstract
Interpretability and explainability are critical aspects of machine learning (ML) systems, especially when deployed in high-stakes domains. This survey reviews key definitions, techniques, and challenges associated with interpretability and explainability in ML. We categorize approaches into inherently interpretable models, model-specific explanations, and model-agnostic post-hoc methods. The paper discusses trade-offs between model performance and transparency, evaluation metrics, and emerging directions to enhance user trust and regulatory compliance.
Article Details
How to Cite
Gurpreet Singh. (2023). Interpretability and Explainability in Machine Learning Systems. International Journal on Recent and Innovation Trends in Computing and Communication, 11(5), 577–579. Retrieved from https://mail.ijritcc.org/index.php/ijritcc/article/view/11716
Section
Articles