Interpretable Deep Neural Networks

This page provides a summary sheet that includes the general goal, reference papers (both mine and external) for an overview of the topic, as well as the domains explored so far. We are also interested in extending the applications of these techniques beyond their traditional domains. If you have expertise in other areas (e.g., neuroscience, gaming, or audio/speech modeling), we would be happy to explore potential extensions into those fields.

Goal: The goal of this research direction is to develop self-interpretable deep neural networks capable of providing explanations alongside their predictions. This involves either generalizing existing designs to novel domains and tasks or creating new ones from scratch. Thus far, our work has focused on leveraging memory mechanisms, prototypes, and interpretable axes as foundational elements for these designs.

Domains: NLP, Vision, Chemistry, Robot Navigation.

Reference Papers:

  1. Memory: [(La Rosa et al., 2022)]
  2. Protoypes-based [(Ragno et al., 2022)] [Seminal Paper]
  3. Interpretable Axes [(Proietti et al., 2023)] [Seminal Paper]

References

2023

  1. Journal
    Explainable AI in drug discovery: self-interpretable graph neural network for molecular property prediction using concept whitening
    Michela Proietti, Alessio Ragno, Biagio La Rosa, and 2 more authors
    Machine Learning, 2023

2022

  1. Journal
    A self-interpretable module for deep image classification on small data
    Biagio La Rosa, Roberto Capobianco, and Daniele Nardi
    Applied Intelligence, 2022
  2. Journal
    Prototype-based Interpretable Graph Neural Networks
    Alessio Ragno, Biagio La Rosa, and Roberto Capobianco
    IEEE Transactions on Artificial Intelligence, 2022