Task 001/002 - Neuro-inspired Algorithms and Theory

Event Date: December 2, 2021
Time: 11:00 am (ET) / 8:00am (PT)
Priority: No
College Calendar: Show
Arushi Gupta, Princeton University
New Definitions and Evaluations for Saliency Methods: Staying Intrinsic and Sound
Abstract:
Saliency methods seek to provide human-interpretable explanations for the output of machine learning model on a given input. A plethora of saliency methods exist, as well as an extensive literature on their justifications/criticisms/evaluations. This paper focuses on heat maps based saliency methods that often provide explanations that look best to humans. It tries to introduce methods and evaluations for masked-based saliency methods that are intrinsic — use just the training dataset and the trained net, and do not use separately trained nets, distractor distributions, human evaluations or annotations. Since a mask can be seen as a “certificate” justifying the net’s answer, we introduce notions of completeness and soundness (the latter being the new contribution) motivated by logical proof systems. These notions allow a new evaluation of saliency methods, that experimentally provides a novel and stronger justification for several heuristic tricks in the field (T.V. regularization, upscaling). 
 
Bio : Arushi Gupta is a fourth year PhD student in Professor Sanjeev Arora’s lab. Gupta received a BSc in Computer Science and Operations Research at Columbia University. Her research interests include saliency, GANs, and generalization.