Abstract
Neural attention mechanism has been used as a form of explanation for model behavior. Users can either passively consume explanation or actively disagree with explanation and then supervise attention into more proper values (attention supervision). Though attention supervision was shown to be effective in some tasks, we find the existing attention supervision is biased, for which we propose to augment counterfactual observations to debias and contribute to accuracy gains. To this end, we propose a counterfactual method to estimate such missing observations and debias the existing supervisions. We validate the effectiveness of our counterfactual supervision on widely adopted image benchmark datasets: CUFED and PEC.
Original language | English |
---|---|
Pages (from-to) | 193-204 |
Number of pages | 12 |
Journal | Data Science and Engineering |
Volume | 5 |
Issue number | 2 |
DOIs | |
Publication status | Published - 2020 Jun 1 |
Bibliographical note
Funding Information:This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2020-2016-0-00464) supervised by the IITP (Institute of Information and Communications Technology Planning and Evaluation). Hwang is a corresponding author.
Publisher Copyright:
© 2020, The Author(s).
All Science Journal Classification (ASJC) codes
- Computational Mechanics
- Computer Science Applications