Share Your Science: Extracting Information from Images

Anton van den Hengel, director of The University of Adelaide’s Australian Centre for Visual Technologies shares how his research group is working on Visual Question Answering (VQA) which uses deep learning to understand the contents of an image.

“The data has been around for a while, but really the GPU technology coming in and allowing us to extract the value out of this data has been the big breakthrough,” said Professor van den Hengel whose team placed second in the 2016 ImageNet Scene Parsing Challenge.

Below are two example results from Professor van den Hengel’s lab paper, “The VQA-Machine: Learning How to Use Existing Vision Algorithms to Answer New Questions”. Given an image-question pair, their model generates not only an answer, but also a set of reasons (as text) and visual attention maps. The colored words in the question have Top-3 weights, ordered as red, blue and cyan.

Share your GPU-accelerated science with us at and with the world on #ShareYourScience.

Watch more scientists and researchers share how accelerated computing is benefiting their work at

About Brad Nemire

Brad Nemire
Brad Nemire is on the Developer Marketing team and loves reading about all of the fascinating research being done by developers using NVIDIA GPUs. Reach out to Brad on Twitter @BradNemire and let him know how you’re using GPUs to accelerate your research. Brad graduated from San Diego State University and currently resides in San Jose, CA. Follow @BradNemire on Twitter