Skip to main content Skip to main navigation

Project

ExplAINN

Explainable AI and Neural Networks

Explainable AI and Neural Networks

Despite astonishing progress in the field of Machine Learning (ML), the robustness of high-performance models, especially the ones based on Deep Learning technologies, has been lower than initially predicted. These networks do not generalize as expected, remaining vulnerable to small adversarial perturbations (also known as adversarial attacks). Such shortcomings pose a critical obstacle to implement Deep Learning models for safety-critical scenarios such as autonomous driving, medical imaging, and credit rating.

Moreover, the gap between good performance and robustness also demonstrates the severe lack of explainability for modern AI approaches: Despite good performance, even experts cannot reliably explain model predictions.

Hence, the goals of this project are threefold:

  • Investigate methods of explainability and interpretability for existing AI approaches (focusing on Deep Neural Networks).
  • Develop novel architectures and training schemes that are more interpretable by design.
  • Analyze the trade-offs between explainability, robustness, and performance.

Sponsors

BMBF - Federal Ministry of Education and Research

01IS19074

BMBF - Federal Ministry of Education and Research

Publications about the project

Fatemeh Azimi; David Dembinsky; Federico Raue; Jörn Hees; Sebastian Palacio; Andreas Dengel (Hrsg.)

International Conference on Pattern Recognition Applications and Methods (ICPRAM-2023), located at 12th International Conference on Pattern Recognition Applications and Methods, February 22-24, Lissabon, Portugal, scitepress, 2/2023.

To the publication