Skip to Main content Skip to Navigation
Conference papers

GuessWhat?! Visual object discovery through multi-modal dialogue

Abstract : We introduce GuessWhat?!, a two-player guessing game as a testbed for research on the interplay of computer vision and dialogue systems. The goal of the game is to locate an unknown object in a rich image scene by asking a sequence of questions. Higher-level image understanding, like spatial reasoning and language grounding, is required to solve the proposed task. Our key contribution is the collection of a large-scale dataset consisting of 150K human-played games with a total of 800K visual question-answer pairs on 66K images. We explain our design decisions in collecting the dataset and introduce the oracle and questioner tasks that are associated with the two players of the game. We prototyped deep learning models to establish initial base-lines of the introduced tasks.
Complete list of metadata

Cited literature [45 references]  Display  Hide  Download
Contributor : Florian Strub Connect in order to contact the contributor
Submitted on : Wednesday, June 28, 2017 - 11:35:00 PM
Last modification on : Friday, January 21, 2022 - 3:13:15 AM
Long-term archiving on: : Thursday, January 18, 2018 - 2:54:39 AM


Files produced by the author(s)


Distributed under a Creative Commons Attribution 4.0 International License


  • HAL Id : hal-01549641, version 1
  • ARXIV : 1611.08481


Harm de Vries, Florian Strub, Sarath Chandar, Olivier Pietquin, Hugo Larochelle, et al.. GuessWhat?! Visual object discovery through multi-modal dialogue. Conference on Computer Vision and Pattern Recognition, Jul 2017, Honolulu, United States. ⟨hal-01549641⟩



Les métriques sont temporairement indisponibles