Towards Transparent Deep Learning: A Hybrid Framework for Explainable Artificial Intelligence (XAI).
Description
While Deep Neural Networks (DNNs) achieve state-of-the-art performance across various domains, their inherent lack of interpretability hinders their deployment in high-stakes, safety-critical applications. To bridge this gap, this paper proposes HT-DNN, a novel hybrid Explainable Artificial Intelligence (XAI) framework. HT-DNN uniquely integrates local surrogate explanations for instance-level transparency, global concept-based reasoning for macro-level understanding, and an automated accountability auditing module to detect spurious correlations. Experimental evaluations on the CIFAR-10 benchmark dataset demonstrate that HT-DNN significantly improves explanation fidelity and interpretability without compromising the underlying model's predictive accuracy.
Files
HT_DNN_XAI_Paper_A. Arun_Kumar.pdf
Files
(510.5 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:358ca1b38d864081579dbd997df38e82
|
166.8 kB | Preview Download |
|
md5:eb286e435f3372d3ca4ee16e65e41722
|
343.8 kB | Preview Download |
Additional details
Software
- Repository URL
- https://github.com/aarunkumar-iitkgp/ht-dnn-xai
- Programming language
- Python
- Development Status
- Active