Enriching Large-Scale Eventuality Knowledge Graph with Entailment Relations
June 21, 2020 ยท Entered Twilight ยท ๐ Conference on Automated Knowledge Base Construction
"Last commit was 5.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: LICENSE, README.md, probase, requirements.txt, verb-entailment
Authors
Changlong Yu, Hongming Zhang, Yangqiu Song, Wilfred Ng, Lifeng Shang
arXiv ID
2006.11824
Category
cs.CL: Computation & Language
Citations
12
Venue
Conference on Automated Knowledge Base Construction
Repository
https://github.com/HKUST-KnowComp/ASER-EEG
โญ 8
Last Checked
1 month ago
Abstract
Computational and cognitive studies suggest that the abstraction of eventualities (activities, states, and events) is crucial for humans to understand daily eventualities. In this paper, we propose a scalable approach to model the entailment relations between eventualities ("eat an apple'' entails ''eat fruit''). As a result, we construct a large-scale eventuality entailment graph (EEG), which has 10 million eventuality nodes and 103 million entailment edges. Detailed experiments and analysis demonstrate the effectiveness of the proposed approach and quality of the resulting knowledge graph. Our datasets and code are available at https://github.com/HKUST-KnowComp/ASER-EEG.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted