-
PhD Defense - Aaron Chan
Tue, Nov 15, 2022 @ 03:00 PM - 05:00 PM
Thomas Lord Department of Computer Science
University Calendar
PhD Candidate: Aaron Chan
Title: Generating and Utilizing Machine Explanations for Trustworthy NLP
Time: Tuesday, November 15, 3:00PM-5:00PM PST
Committee: Xiang Ren (chair), Robin Jia, Jesse Thomason, Bistra Dilkina, Morteza Dehghani
Abstract:
Neural language models (LMs) have yielded remarkable success on a wide range of natural language processing (NLP) tasks. However, LMs sometimes exhibit undesirable behavior, which can be difficult to resolve due to LMs' opaque reasoning processes. This lack of transparency poses serious concerns about LMs' trustworthiness in high-stakes decision-making, thus motivating the use of machine explanations to automatically interpret how LMs make their predictions. In my thesis, I argue that building human trust in NLP systems requires being able to: (A) generate machine explanations for LM behavior faithfully and plausibly and (B) utilize machine explanations to improve LM generalization and decision-making. First, to address (A), I propose UNIREX, a unified learning framework for jointly optimizing machine explanations with respect to both faithfulness and plausibility, without compromising the LM's task performance. Second, for (B), I introduce ER-Test, a framework for evaluating the out-of-distribution generalization ability of LMs that are regularized via strongly-supervised machine explanations. Third, to further support (B), I present SalKG, an algorithm for improving LM generalization by regularizing LMs via weakly-supervised machine explanations. Finally, I discuss several future directions for achieving (A) and (B).
Zoom Link: https://usc.zoom.us/j/95606515253?pwd=QzBvaVVpcWtYSFhVYzVoUi9tdHBRdz09
WebCast Link: : https://usc.zoom.us/j/95606515253?pwd=QzBvaVVpcWtYSFhVYzVoUi9tdHBRdz09
Audiences: Everyone Is Invited
Contact: Lizsl De Leon