Embodied Question Answering

November 30, 2017 ยท Entered Twilight ยท ๐Ÿ› 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)

๐ŸŒ… TWILIGHT: Old Age
Predates the code-sharing era โ€” a pioneer of its time

"No code URL or promise found in abstract"
"Code repo scraped from project page (backfill)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, .gitmodules, CODE_OF_CONDUCT.md, CONTRIBUTING.md, House3D, LICENSE, README.md, data, requirements.txt, training, utils

Authors Abhishek Das, Samyak Datta, Georgia Gkioxari, Stefan Lee, Devi Parikh, Dhruv Batra arXiv ID 1711.11543 Category cs.CV: Computer Vision Cross-listed cs.AI, cs.CL, cs.LG Citations 723 Venue 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) Repository https://github.com/facebookresearch/EmbodiedQA โญ 316 Last Checked 6 days ago
Abstract
We present a new AI task -- Embodied Question Answering (EmbodiedQA) -- where an agent is spawned at a random location in a 3D environment and asked a question ("What color is the car?"). In order to answer, the agent must first intelligently navigate to explore the environment, gather information through first-person (egocentric) vision, and then answer the question ("orange"). This challenging task requires a range of AI skills -- active perception, language understanding, goal-driven navigation, commonsense reasoning, and grounding of language into actions. In this work, we develop the environments, end-to-end-trained reinforcement learning agents, and evaluation protocols for EmbodiedQA.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision