VizWiz Grand Challenge: Answering Visual Questions From Blind People

  title={VizWiz Grand Challenge: Answering Visual Questions From Blind People},
  author={Danna Gurari and Qing Li and Abigale J. Stangl and Anhong Guo and Chi Lin and Kristen Grauman and Jiebo Luo and Jeffrey P. Bigham},
The study of algorithms to automatically answer visual questions currently is motivated by visual question answering (VQA) datasets constructed in artificial VQA settings. We propose VizWiz, the first goal-oriented VQA dataset arising from a natural VQA setting. VizWiz consists of over 31,000 visual questions originating from blind people who each took a picture using a mobile phone and recorded a spoken question about it, together with 10 crowdsourced answers per visual question. VizWiz… CONTINUE READING
Highly Cited
This paper has 24 citations. REVIEW CITATIONS


Publications citing this paper.


Publications referenced by this paper.
Showing 1-10 of 47 references

FVQA: Fact-Based Visual Question Answering

IEEE Transactions on Pattern Analysis and Machine Intelligence • 2018

An Analysis of Visual Question Answering Algorithms

2017 IEEE International Conference on Computer Vision (ICCV) • 2017


V. Kazemi, A. Elqursh
ask, attend, and answer: A strong baseline for visual question answering. arXiv preprint arXiv:1704.03162 • 2017

Similar Papers

Loading similar papers…