Influence Functions for Sequence Tagging Models

October 25, 2022 ยท Entered Twilight ยท ๐Ÿ› Conference on Empirical Methods in Natural Language Processing

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: .gitignore, README.md, conda_env.yml, conf.py, data, extras, linear_model_exact_experiment.py, linear_model_exact_influence.tsv, ner_influence

Authors Sarthak Jain, Varun Manjunatha, Byron C. Wallace, Ani Nenkova arXiv ID 2210.14177 Category cs.CL: Computation & Language Cross-listed cs.AI, cs.LG, stat.ML Citations 10 Venue Conference on Empirical Methods in Natural Language Processing Repository https://github.com/successar/Segment_Influence_Functions โญ 8 Last Checked 2 months ago
Abstract
Many language tasks (e.g., Named Entity Recognition, Part-of-Speech tagging, and Semantic Role Labeling) are naturally framed as sequence tagging problems. However, there has been comparatively little work on interpretability methods for sequence tagging models. In this paper, we extend influence functions - which aim to trace predictions back to the training points that informed them - to sequence tagging tasks. We define the influence of a training instance segment as the effect that perturbing the labels within this segment has on a test segment level prediction. We provide an efficient approximation to compute this, and show that it tracks with the true segment influence, measured empirically. We show the practical utility of segment influence by using the method to identify systematic annotation errors in two named entity recognition corpora. Code to reproduce our results is available at https://github.com/successar/Segment_Influence_Functions.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computation & Language

๐ŸŒ… ๐ŸŒ… Old Age

Attention Is All You Need

Ashish Vaswani, Noam Shazeer, ... (+6 more)

cs.CL ๐Ÿ› NeurIPS ๐Ÿ“š 166.0K cites 8 years ago