Overview
The Intel® Explainable AI Tools are designed to help users detect and mitigate against issues of fairness and interpretability, while running best on Intel hardware. There are two Python* components in the repository:
-
Creates interactive HTML reports containing model performance and fairness metrics
-
Runs post-hoc model distillation and visualization methods to examine predictive behavior for both TensorFlow* and PyTorch* models via a simple Python API including the following modules:
Attributions: Visualize negative and positive attributions of tabular features, pixels, and word tokens for predictions
CAM (Class Activation Mapping): Create heatmaps for CNN image classifications using gradient-weight class activation CAM mapping
Metrics: Gain insight into models with the measurements and visualizations needed during the machine learning workflow
*Other names and brands may be claimed as the property of others. Trademarks