Quantifying the visual concreteness of words and topics in multimodal datasets

April 18, 2018 ยท Entered Twilight ยท ๐Ÿ› North American Chapter of the Association for Computational Linguistics

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"No code URL or promise found in abstract"
"Code repo scraped from project page (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .circleci, .flake8, .gitignore, .pylintrc, LICENSE, Makefile, README.md, concreteness.ipynb, concreteness.py, data, main.py, mirflickr.py, mscoco.py, requirements-notebook.txt, requirements.txt

Authors Jack Hessel, David Mimno, Lillian Lee arXiv ID 1804.06786 Category cs.CL: Computation & Language Cross-listed cs.CV, cs.IR Citations 42 Venue North American Chapter of the Association for Computational Linguistics Repository https://github.com/victorssilva/concreteness โญ 20 Last Checked 6 days ago
Abstract
Multimodal machine learning algorithms aim to learn visual-textual correspondences. Previous work suggests that concepts with concrete visual manifestations may be easier to learn than concepts with abstract ones. We give an algorithm for automatically computing the visual concreteness of words and topics within multimodal datasets. We apply the approach in four settings, ranging from image captions to images/text scraped from historical books. In addition to enabling explorations of concepts in multimodal datasets, our concreteness scores predict the capacity of machine learning algorithms to learn textual/visual relationships. We find that 1) concrete concepts are indeed easier to learn; 2) the large number of algorithms we consider have similar failure cases; 3) the precise positive relationship between concreteness and performance varies between datasets. We conclude with recommendations for using concreteness scores to facilitate future multimodal research.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computation & Language

๐ŸŒ… ๐ŸŒ… Old Age

Attention Is All You Need

Ashish Vaswani, Noam Shazeer, ... (+6 more)

cs.CL ๐Ÿ› NeurIPS ๐Ÿ“š 166.0K cites 8 years ago