DOREMI: Optimizing Long Tail Predictions in Document-Level Relation Extraction

January 16, 2026 ยท Grace Period ยท ๐Ÿ› Knowledge-Based Systems

โณ Grace Period
This paper is less than 90 days old. We give authors time to release their code before passing judgment.
Authors Laura Menotti, Stefano Marchesin, Gianmaria Silvello arXiv ID 2601.11190 Category cs.CL: Computation & Language Citations 0 Venue Knowledge-Based Systems
Abstract
Document-Level Relation Extraction (DocRE) presents significant challenges due to its reliance on cross-sentence context and the long-tail distribution of relation types, where many relations have scarce training examples. In this work, we introduce DOcument-level Relation Extraction optiMizing the long taIl (DOREMI), an iterative framework that enhances underrepresented relations through minimal yet targeted manual annotations. Unlike previous approaches that rely on large-scale noisy data or heuristic denoising, DOREMI actively selects the most informative examples to improve training efficiency and robustness. DOREMI can be applied to any existing DocRE model and is effective at mitigating long-tail biases, offering a scalable solution to improve generalization on rare relations.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computation & Language

๐ŸŒ… ๐ŸŒ… Old Age

Attention Is All You Need

Ashish Vaswani, Noam Shazeer, ... (+6 more)

cs.CL ๐Ÿ› NeurIPS ๐Ÿ“š 166.0K cites 8 years ago