CommitBench: A Benchmark for Commit Message Generation
March 08, 2024 ยท Entered Twilight ยท ๐ IEEE International Conference on Software Analysis, Evolution, and Reengineering
Repo contents: .dockerignore, .gitattributes, .gitignore, LICENSE, analyze, console.sh, create, dockerfile, enhancer, exporter, filter, importer, prepare, readme.md, run_pipeline.sh
Authors
Maximilian Schall, Tamara Czinczoll, Gerard de Melo
arXiv ID
2403.05188
Category
cs.CL: Computation & Language
Cross-listed
cs.SE
Citations
11
Venue
IEEE International Conference on Software Analysis, Evolution, and Reengineering
Repository
https://github.com/Maxscha/commitbench
โญ 11
Last Checked
2 months ago
Abstract
Writing commit messages is a tedious daily task for many software developers, and often remains neglected. Automating this task has the potential to save time while ensuring that messages are informative. A high-quality dataset and an objective benchmark are vital preconditions for solid research and evaluation towards this goal. We show that existing datasets exhibit various problems, such as the quality of the commit selection, small sample sizes, duplicates, privacy issues, and missing licenses for redistribution. This can lead to unusable models and skewed evaluations, where inferior models achieve higher evaluation scores due to biases in the data. We compile a new large-scale dataset, CommitBench, adopting best practices for dataset creation. We sample commits from diverse projects with licenses that permit redistribution and apply our filtering and dataset enhancements to improve the quality of generated commit messages. We use CommitBench to compare existing models and show that other approaches are outperformed by a Transformer model pretrained on source code. We hope to accelerate future research by publishing the source code( https://github.com/Maxscha/commitbench ).
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted