Extractive Adversarial Networks: High-Recall Explanations for Identifying Personal Attacks in Social Media Posts
September 01, 2018 ยท Entered Twilight ยท ๐ Conference on Empirical Methods in Natural Language Processing
"Last commit was 7.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: .gitignore, CODE_README.md, LICENSE, README.md, deliberativeness, lime, nn, rationale, utils
Authors
Samuel Carton, Qiaozhu Mei, Paul Resnick
arXiv ID
1809.01499
Category
cs.CL: Computation & Language
Cross-listed
cs.IR,
cs.LG,
stat.ML
Citations
34
Venue
Conference on Empirical Methods in Natural Language Processing
Repository
https://github.com/shcarton/rcnn
โญ 1
Last Checked
1 month ago
Abstract
We introduce an adversarial method for producing high-recall explanations of neural text classifier decisions. Building on an existing architecture for extractive explanations via hard attention, we add an adversarial layer which scans the residual of the attention for remaining predictive signal. Motivated by the important domain of detecting personal attacks in social media comments, we additionally demonstrate the importance of manually setting a semantically appropriate `default' behavior for the model by explicitly manipulating its bias term. We develop a validation set of human-annotated personal attacks to evaluate the impact of these changes.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Computation & Language
๐
๐
Old Age
๐
๐
Old Age
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
RoBERTa: A Robustly Optimized BERT Pretraining Approach
R.I.P.
๐ป
Ghosted
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
R.I.P.
๐ป
Ghosted