Reweighting Strategy based on Synthetic Data Identification for Sentence Similarity
August 29, 2022 ยท Entered Twilight ยท ๐ International Conference on Computational Linguistics
Repo contents: .gitignore, README.md, dino.py, discrimination.py, modeling.py, postprocess_dataset.py, preprocess_data.py, requirements.txt, run_training.py, run_unsupervised_textual_similarity.py, run_use.py, scripts, sentence_transformers, split_paws_dev_test.py, task_specs, utils.py
Authors
Taehee Kim, ChaeHun Park, Jimin Hong, Radhika Dua, Edward Choi, Jaegul Choo
arXiv ID
2208.13376
Category
cs.CL: Computation & Language
Citations
1
Venue
International Conference on Computational Linguistics
Repository
https://github.com/ddehun/coling2022_reweighting_sts
โญ 18
Last Checked
1 month ago
Abstract
Semantically meaningful sentence embeddings are important for numerous tasks in natural language processing. To obtain such embeddings, recent studies explored the idea of utilizing synthetically generated data from pretrained language models (PLMs) as a training corpus. However, PLMs often generate sentences much different from the ones written by human. We hypothesize that treating all these synthetic examples equally for training deep neural networks can have an adverse effect on learning semantically meaningful embeddings. To analyze this, we first train a classifier that identifies machine-written sentences, and observe that the linguistic features of the sentences identified as written by a machine are significantly different from those of human-written sentences. Based on this, we propose a novel approach that first trains the classifier to measure the importance of each sentence. The distilled information from the classifier is then used to train a reliable sentence embedding model. Through extensive evaluation on four real-world datasets, we demonstrate that our model trained on synthetic data generalizes well and outperforms the existing baselines. Our implementation is publicly available at https://github.com/ddehun/coling2022_reweighting_sts.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted