๐
๐
Old Age
FreCDo: A Large Corpus for French Cross-Domain Dialect Identification
December 15, 2022 ยท Entered Twilight ยท ๐ International Conference on Knowledge-Based Intelligent Information & Engineering Systems
Repo contents: LICENSE, README.md, citation.bib, code, data
Authors
Mihaela Gaman, Adrian-Gabriel Chifu, William Domingues, Radu Tudor Ionescu
arXiv ID
2212.07707
Category
cs.CL: Computation & Language
Cross-listed
cs.LG
Citations
4
Venue
International Conference on Knowledge-Based Intelligent Information & Engineering Systems
Repository
https://github.com/MihaelaGaman/FreCDo
โญ 2
Last Checked
2 months ago
Abstract
We present a novel corpus for French dialect identification comprising 413,522 French text samples collected from public news websites in Belgium, Canada, France and Switzerland. To ensure an accurate estimation of the dialect identification performance of models, we designed the corpus to eliminate potential biases related to topic, writing style, and publication source. More precisely, the training, validation and test splits are collected from different news websites, while searching for different keywords (topics). This leads to a French cross-domain (FreCDo) dialect identification task. We conduct experiments with four competitive baselines, a fine-tuned CamemBERT model, an XGBoost based on fine-tuned CamemBERT features, a Support Vector Machines (SVM) classifier based on fine-tuned CamemBERT features, and an SVM based on word n-grams. Aside from presenting quantitative results, we also make an analysis of the most discriminative features learned by CamemBERT. Our corpus is available at https://github.com/MihaelaGaman/FreCDo.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted