Command Palette
Search for a command to run...
Pietro Barbiero Gabriele Ciravegna Francesco Giannini Pietro Lió Marco Gori Stefano Melacci

Abstract
Explainable artificial intelligence has rapidly emerged since lawmakers have started requiring interpretable models for safety-critical domains. Concept-based neural networks have arisen as explainable-by-design methods as they leverage human-understandable symbols (i.e. concepts) to predict class memberships. However, most of these approaches focus on the identification of the most relevant concepts but do not provide concise, formal explanations of how such concepts are leveraged by the classifier to make predictions. In this paper, we propose a novel end-to-end differentiable approach enabling the extraction of logic explanations from neural networks using the formalism of First-Order Logic. The method relies on an entropy-based criterion which automatically identifies the most relevant concepts. We consider four different case studies to demonstrate that: (i) this entropy-based criterion enables the distillation of concise logic explanations in safety-critical domains from clinical data to computer vision; (ii) the proposed approach outperforms state-of-the-art white-box models in terms of classification accuracy and matches black box performances.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| image-classification-on-cub | Entropy-based Logic Explained Network | Classification Accuracy: 0.9295 Explanation Accuracy: 95.24 Explanation complexity: 3.74 Explanation extraction time: 171.87 |
| image-classification-on-cub | Bayesian Rule List | Classification Accuracy: 0.9079 Explanation Accuracy: 96.02 Explanation complexity: 8.87 Explanation extraction time: 264678.29 |
| image-classification-on-cub | Decision Tree | Classification Accuracy: 0.8162 Explanation Accuracy: 89.36 Explanation complexity: 45.92 Explanation extraction time: 8.1 |
| image-classification-on-cub | $psi$ network | Classification Accuracy: 0.9192 Explanation Accuracy: 76.1 Explanation complexity: 15.96 Explanation extraction time: 3707.29 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.