Algorithmic Explanation: A method for measuring interpretations of neural networks
A downloadable game
How do you make good explanations for what a neural network does? We provide a framework for analysing explanations of the behaviour of neural networks by looking at the hypothesis of how they would act on a set of given inputs. By trying to model a neural network using known logic (or as much white-box logic as possible), this framework is a start on how we could tackle neural network interpretability as they get more complex.
Status | Released |
Author | clementneo |
Download
Download
reframing_explanation-with-figures.pdf 419 kB
Leave a comment
Log in with itch.io to leave a comment.