An Empirical Study On Contrastive Search And Contrastive Decoding For Open-ended Text Generation

November 19, 2022 ยท Entered Twilight ยท ๐Ÿ› arXiv.org

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: LICENSE, README.md, automatic_evaluation_results.png, compute_CD_coherence.py, compute_coherence.py, data_from_CD_repo, human_evaluation_annotations, human_evaluation_results.png, inference.py, inference_results, measure_CD_coherence.sh, measure_CD_mauve_diversity_gen_len.py, measure_CD_result.sh, measure_diversity_mauve_gen_length.py, scripts, utlis

Authors Yixuan Su, Jialu Xu arXiv ID 2211.10797 Category cs.CL: Computation & Language Citations 16 Venue arXiv.org Repository https://github.com/yxuansu/Contrastive_Search_versus_Contrastive_Decoding โญ 27 Last Checked 2 months ago
Abstract
In the study, we empirically compare the two recently proposed decoding methods, i.e. Contrastive Search (CS) and Contrastive Decoding (CD), for open-ended text generation. The automatic evaluation results suggest that, while CS performs worse than CD on the MAUVE metric, it substantially surpasses CD on the diversity and coherence metrics. More notably, extensive human evaluations across three different domains demonstrate that human annotators are universally more in favor of CS over CD with substantial margins. The contradicted results between MAUVE and human evaluations reveal that MAUVE does not accurately reflect human preferences. Therefore, we call upon the research community to develop better evaluation metrics for open-ended text generation. To ensure the reproducibility of our work, we have open-sourced all our code, evaluation results, as well as human annotations at https://github.com/yxuansu/Contrastive_Search_versus_Contrastive_Decoding.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computation & Language

๐ŸŒ… ๐ŸŒ… Old Age

Attention Is All You Need

Ashish Vaswani, Noam Shazeer, ... (+6 more)

cs.CL ๐Ÿ› NeurIPS ๐Ÿ“š 166.0K cites 8 years ago