๐
๐
Old Age
Influence Functions for Sequence Tagging Models
October 25, 2022 ยท Entered Twilight ยท ๐ Conference on Empirical Methods in Natural Language Processing
Repo contents: .gitignore, README.md, conda_env.yml, conf.py, data, extras, linear_model_exact_experiment.py, linear_model_exact_influence.tsv, ner_influence
Authors
Sarthak Jain, Varun Manjunatha, Byron C. Wallace, Ani Nenkova
arXiv ID
2210.14177
Category
cs.CL: Computation & Language
Cross-listed
cs.AI,
cs.LG,
stat.ML
Citations
10
Venue
Conference on Empirical Methods in Natural Language Processing
Repository
https://github.com/successar/Segment_Influence_Functions
โญ 8
Last Checked
2 months ago
Abstract
Many language tasks (e.g., Named Entity Recognition, Part-of-Speech tagging, and Semantic Role Labeling) are naturally framed as sequence tagging problems. However, there has been comparatively little work on interpretability methods for sequence tagging models. In this paper, we extend influence functions - which aim to trace predictions back to the training points that informed them - to sequence tagging tasks. We define the influence of a training instance segment as the effect that perturbing the labels within this segment has on a test segment level prediction. We provide an efficient approximation to compute this, and show that it tracks with the true segment influence, measured empirically. We show the practical utility of segment influence by using the method to identify systematic annotation errors in two named entity recognition corpora. Code to reproduce our results is available at https://github.com/successar/Segment_Influence_Functions.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted