APRIL: Interactively Learning to Summarise by Combining Active Preference Learning and Reinforcement Learning
August 29, 2018 ยท Entered Twilight ยท ๐ Conference on Empirical Methods in Natural Language Processing
"Last commit was 7.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: .gitignore, LICENSE.txt, NOTICE.txt, README.md, data, learnt_ranker, requirement.txt, resources.py, stage0_sample_summaries.py, stage1_active_pref_learning.py, stage2_reinf_learning.py, summariser
Authors
Yang Gao, Christian M. Meyer, Iryna Gurevych
arXiv ID
1808.09658
Category
cs.CL: Computation & Language
Cross-listed
cs.AI,
cs.LG
Citations
34
Venue
Conference on Empirical Methods in Natural Language Processing
Repository
https://github.com/UKPLab/emnlp2018-april
โญ 12
Last Checked
1 month ago
Abstract
We propose a method to perform automatic document summarisation without using reference summaries. Instead, our method interactively learns from users' preferences. The merit of preference-based interactive summarisation is that preferences are easier for users to provide than reference summaries. Existing preference-based interactive learning methods suffer from high sample complexity, i.e. they need to interact with the oracle for many rounds in order to converge. In this work, we propose a new objective function, which enables us to leverage active learning, preference learning and reinforcement learning techniques in order to reduce the sample complexity. Both simulation and real-user experiments suggest that our method significantly advances the state of the art. Our source code is freely available at https://github.com/UKPLab/emnlp2018-april.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted