Interpretable Deep Neural Networks
This page provides a summary sheet that includes the general goal, reference papers (both mine and external) for an overview of the topic, as well as the domains explored so far. We are also interested in extending the applications of these techniques beyond their traditional domains. If you have expertise in other areas (e.g., neuroscience, gaming, or audio/speech modeling), we would be happy to explore potential extensions into those fields.
Goal: The goal of this research direction is to develop self-interpretable deep neural networks capable of providing explanations alongside their predictions. This involves either generalizing existing designs to novel domains and tasks or creating new ones from scratch. Thus far, our work has focused on leveraging memory mechanisms, prototypes, and interpretable axes as foundational elements for these designs.
Domains: NLP, Vision, Chemistry, Robot Navigation.
Reference Papers:
- Memory: [(La Rosa et al., 2022)]
- Protoypes-based [(Ragno et al., 2022)] [Seminal Paper]
- Interpretable Axes [(Proietti et al., 2023)] [Seminal Paper]
References
2023
- JournalExplainable AI in drug discovery: self-interpretable graph neural network for molecular property prediction using concept whiteningMachine Learning, 2023
2022
- JournalA self-interpretable module for deep image classification on small dataApplied Intelligence, 2022
- JournalPrototype-based Interpretable Graph Neural NetworksIEEE Transactions on Artificial Intelligence, 2022