2016 •
Where to Look: Focus Regions for Visual Question Answering
Authors: Shih, Kevin J., Singh, Saurabh, Hoiem, Derek
Venue: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
Type: Publication
Abstract: We present a method that learns to answer visual questions by selecting image regions relevant to the text-based query. Our method exhibits significant improvements in answering questions such as "what color," where it is necessary to evaluate a specific location, and "what room," where it selectively identifies informative image regions. Our model is tested on the VQA dataset which is the largest human-annotated visual question answering dataset to our knowledge.
Topics: Information retrievalArtificial intelligenceNatural language processing
Popularity: This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the
underlying citation network.
Influence: This indicator reflects the overall/total impact of an article in the research community at large, based on the
underlying citation network (diachronically).
Citation Count: This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in
the research community at large, based on the underlying citation network (diachronically).
Impulse: This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation
network.
We have placed cookies on your device to help make this website and the services we offer better. By using this site, you agree to the use of cookies. Learn more