VTC: Improving Video-Text Retrieval with User Comments

October 19, 2022 ยท Entered Twilight ยท ๐Ÿ› European Conference on Computer Vision

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

"No code URL or promise found in abstract"
"Derived repo from GitHub Pages (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: _config.yml, _layouts, assets, index.md

Authors Laura Hanu, James Thewlis, Yuki M. Asano, Christian Rupprecht arXiv ID 2210.10820 Category cs.CV: Computer Vision Cross-listed cs.CL, cs.IR, cs.LG Citations 8 Venue European Conference on Computer Vision Repository https://github.com/unitaryai/vtc-paper. โญ 1 Last Checked 12 days ago
Abstract
Multi-modal retrieval is an important problem for many applications, such as recommendation and search. Current benchmarks and even datasets are often manually constructed and consist of mostly clean samples where all modalities are well-correlated with the content. Thus, current video-text retrieval literature largely focuses on video titles or audio transcripts, while ignoring user comments, since users often tend to discuss topics only vaguely related to the video. Despite the ubiquity of user comments online, there is currently no multi-modal representation learning datasets that includes comments. In this paper, we a) introduce a new dataset of videos, titles and comments; b) present an attention-based mechanism that allows the model to learn from sometimes irrelevant data such as comments; c) show that by using comments, our method is able to learn better, more contextualised, representations for image, video and audio representations. Project page: https://unitaryai.github.io/vtc-paper.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision