Classifying Text with Intuitive and Faithful Model Explanations

The objective of this Research Project is to develop an advanced neural NLP modeling framework for interpretable and accurate text classification. Intuitively, when human users better understand model predictions (via model interpretability), the users can better use model predictions to augment their own human reasoning and decision-making. More generally, effective model explanations offer a variety of other potential benefits, such as promoting trust, adoption, auditing, and documentation of model decisions. Our modeling framework, ProtoType-based Explanations for Natural Language (ProtoTexNL), seeks to provide faithful explanations for model predictions in relation to training examples and features of the input text. 
Matthew Lease
Agency
Cisco Systems Inc.
Grant Dates
-
Funding

The award is $199,458 over the project period.