2 code implementations • NeurIPS 2018 • David Alvarez Melis, Tommi Jaakkola
Most recent work on interpretability of complex machine learning models has focused on estimating a-posteriori explanations for previously trained models around specific predictions.