Summarizing Text on Any Aspects: A Knowledge-Informed Weakly-Supervised Approach
October 14, 2020 ยท Entered Twilight ยท ๐ Conference on Empirical Methods in Natural Language Processing
"Last commit was 5.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: README.md, WEAK_SUPERVISION_CONSTRUCTION.md, construct.py, data_utils, demo.py, demo_input.json, evaluate.sh, finetune.py, generate.py, models, requirements.txt, supervisions
Authors
Bowen Tan, Lianhui Qin, Eric P. Xing, Zhiting Hu
arXiv ID
2010.06792
Category
cs.CL: Computation & Language
Cross-listed
cs.LG
Citations
44
Venue
Conference on Empirical Methods in Natural Language Processing
Repository
https://github.com/tanyuqian/aspect-based-summarization
โญ 38
Last Checked
1 month ago
Abstract
Given a document and a target aspect (e.g., a topic of interest), aspect-based abstractive summarization attempts to generate a summary with respect to the aspect. Previous studies usually assume a small pre-defined set of aspects and fall short of summarizing on other diverse topics. In this work, we study summarizing on arbitrary aspects relevant to the document, which significantly expands the application of the task in practice. Due to the lack of supervision data, we develop a new weak supervision construction method and an aspect modeling scheme, both of which integrate rich external knowledge sources such as ConceptNet and Wikipedia. Experiments show our approach achieves performance boosts on summarizing both real and synthetic documents given pre-defined or arbitrary aspects.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted