Entity Disambiguation with Entity Definitions
October 11, 2022 ยท Entered Twilight ยท ๐ Conference of the European Chapter of the Association for Computational Linguistics
Repo contents: .gitignore, Dockerfile, LICENSE.txt, README.md, configurations, data, demo.sh, experiments, extend, requirements.txt, scripts, setup.py, setup.sh
Authors
Luigi Procopio, Simone Conia, Edoardo Barba, Roberto Navigli
arXiv ID
2210.05648
Category
cs.CL: Computation & Language
Cross-listed
cs.AI,
cs.LG
Citations
2
Venue
Conference of the European Chapter of the Association for Computational Linguistics
Repository
https://github.com/SapienzaNLP/extend
โญ 182
Last Checked
1 month ago
Abstract
Local models have recently attained astounding performances in Entity Disambiguation (ED), with generative and extractive formulations being the most promising research directions. However, previous works limited their studies to using, as the textual representation of each candidate, only its Wikipedia title. Although certainly effective, this strategy presents a few critical issues, especially when titles are not sufficiently informative or distinguishable from one another. In this paper, we address this limitation and investigate to what extent more expressive textual representations can mitigate it. We thoroughly evaluate our approach against standard benchmarks in ED and find extractive formulations to be particularly well-suited to these representations: we report a new state of the art on 2 out of 6 benchmarks we consider and strongly improve the generalization capability over unseen patterns. We release our code, data and model checkpoints at https://github.com/SapienzaNLP/extend.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted