Signal Processing - November 2017 - 74

Australian Centre for Visual Technologies of the University
of Adelaide, where he works on computer vision and
machine learning. He was previously affiliated with
Carnegie Mellon University, Pittsburgh, Pennsylvania; the
University of Bath, United Kingdom; and the University of
Innsbruck, Austria.
Qi Wu (qi.wu01@adelaide.edu.au) received a bachelor's
degree in mathematical sciences from China Jiliang Uni-
versity, Hangzhou, and a master's degree in computer
science and a Ph.D. degree in computer vision from the
University of Bath, United Kingdom, in 2012 and 2015,
respectively. He is a postdoctoral researcher at the Australian
Centre for Robotic Vision of the University of Adelaide. His
research interests include cross-depiction object detection
and classification, attributes learning, neural networks, and
image captioning.
Anton van den Hengel (anton.vandenhengel@adelaide.
edu.au) received his bachelor's degree in mathematical science in 1991, his bachelor of laws degree in 1993, his master's degree in computer science in 1994, and his Ph.D.
degree in computer vision in 2000, all from the University of
Adelaide, Australia, where he is a professor and the founding
director of the Australian Centre for Visual Technologies.

References

[1] VQA challenge leaderboard. [Online]. Available: http://visualqa.org/ http://eva​
lai.cloudcv.org
[2] P. Anderson, X. He, C. Buehler, D. Teney, M. Johnson, S. Gould, and L.
Zhang, "Bottom-up and top-down attention for image captioning and VQA," arXiv
Preprint, arXiv:1707.07998, 2017.
[3] J. Andreas, M. Rohrbach, T. Darrell, and D. Klein, "Learning to compose neural networks for question answering," in Proc. Annu. Conf. North American
Chapter Assoc. Computational Linguistics, San Diego, CA, 2016, pp. 1545-1554.
[4] J. Andreas, M. Rohrbach, T. Darrell, and D. Klein, "Neural module networks,"
In Proc. IEEE Conf. Computer Vision and Pattern Recognition, 2016, pp. 39-48.
[5] S. Antol, A. Agrawal, J. Lu, M. Mitchell, D. Batra, C. L. Zitnick, and D.
Parikh, "VQA: Visual question answering," in Proc. IEEE Int. Conf. Computer
Vision, 2015, pp. 2425-2433.
[6] Y. Atzmon, J. Berant, V. Kezami, A. Globerson, and G. Chechik, "Learning to
generalize to new compositions in image understanding," arXiv Preprint,
arXiv:1608.07639, 2016.
[7] S. Auer, C. Bizer, G. Kobilarov, J. Lehmann, R. Cyganiak, and Z. Ives,
DBpedia: A Nucleus for a Web of Open Data. New York: Springer, 2007.
[8] D. Bahdanau, K. Cho, and Y. Bengio, "Neural machine translation by jointly
learning to align and translate," in Proc. Int. Conf. Learning Representation
(ICLR), San Diego, CA, 2015.
[9] H. Ben-younes, R. Cadène, M. Cord, and N. Thome, "MUTAN: multimodal
tucker fusion for visual question answering," arXiv Preprint, arXiv:1705.06676,
2017.
[10] J. Berant, A. Chou, R. Frostig, and P. Liang, "Semantic parsing on freebase
from question-answer pairs," in Proc. Conf. Empirical Methods Natural
Language Processing, 2013, pp. 1533-1544.
[11] L. Bertinetto, J. F. Henriques, J. Valmadre, P. H. S. Torr, and A. Vedaldi,
"Learning feed-forward one-shot learners," in Proc. Neural Information
Processing Systems (NIPS), 2016, pp. 523-531.
[12] K. Bollacker, C. Evans, P. Paritosh, T. Sturge, and J. Taylor, "Freebase: A collaboratively created graph database for structuring human knowledge," in Proc.
ACM SIGMOD Int. Conf. Management of Data, 2008, pp. 1247-1250.
[13] A. Bordes, N. Usunier, S. Chopra, and J. Weston, "Large-scale simple question answering with memory networks," arXiv Preprint, arXiv:1506.02075,
2015.
[14] Q. Cai and A. Yates, "Large-scale semantic parsing via schema matching
and lexicon extension," in Proc. Conf. Association Computational Linguistics,
2013, pp. 423-433.
[15] R. Cantrell, M. Scheutz, P. Schermerhorn, and X. Wu, "Robust spoken
instruction understanding for hri," in Proc. 5th ACM/IEEE Int. Conf. HumanRobot Interaction, 2010, pp. 275-282.

74

[16] N. Dalal and B. Triggs, "Histograms of oriented gradients for human detection," in Proc. IEEE Conf. Computer Vision and Pattern Recognition, 2005, vol.
1, pp. 886-893.
[17] A. Das, H. Agrawal, C. L. Zitnick, D. Parikh, and D. Batra, "Human attention
in visual question answering: Do humans and deep networks look at the same
regions?" in Proc. Conf. Empirical Methods Natural Language Processing, 2016,
pp. 932-937.
[18] A. Das, S. Kottur, K. Gupta, A. Singh, D. Yadav, J. M. Moura, D. Parikh, and
D. Batra, "Visual dialog," in Proc. IEEE Conf. Computer Vision and Pattern
Recognition, 2017.
[19] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, "Imagenet: A
large-scale hierarchical image database," in Proc. IEEE Conf. Computer Vision
and Pattern Recognition, 2009, pp. 248-255.
[20] H. Fang, S. Gupta, F. Iandola, R. Srivastava, L. Deng, P. Dollár, J. Gao, X.
He, M. Mitchell, and J. Platt, "From captions to visual concepts and back," in
Proc. IEEE Conf. Computer Vision and Pattern Recognition, 2015, pp. 1473-
1482.
[21] A. Fukui, D. H. Park, D. Yang, A. Rohrbach, T. Darrell, and M. Rohrbach,
"Multimodal compact bilinear pooling for visual question answering and visual
grounding," in Proc. Conf. Empirical Methods Natural Language Processing
(EMNLP), 2016, pp. 457-468.
[22] H. Gao, J. Mao, J. Zhou, Z. Huang, L. Wang, and W. Xu, "Are you talking to
a machine? Data set and methods for multilingual image question answering," in
Proc. Advances in Neural Information Processing Systems, 2015, pp. 2296-
2304.
[23] D. Geman, S. Geman, N. Hallonquist, and L. Younes, "Visual turing test for
computer vision systems," Proc. Natl. Acad. Sci., vol. 112, no. 12, pp. 3618-3623,
2015.
[24] R. Girshick, "Fast R-CNN," in Proc. IEEE Int. Conf. Computer Vision, 2015,
pp. 1440-1448.
[25] Y. Goyal, T. Khot, D. Summers-Stay, D. Batra, and D. Parikh, "Making the V
in VQA matter: Elevating the role of image understanding in visual question
answering," in Proc. IEEE Conf. Comp. Vis. Patt. Recogn. (CVPR), 2017.
[26] A. Graves, G. Wayne, and I. Danihelka, "Neural turing machines," arXiv
Preprint, arXiv:1410.5401, 2014.
[27] L. A. Hendricks, S. Venugopalan, M. Rohrbach, R. J. Mooney, K. Saenko,
and T. Darrell, "Deep compositional captioning: Describing novel object categories
without paired training data," in Proc. IEEE Conf. Computer Vision and Pattern
Recognition, 2015, pp. 1-10.
[28] F. Hill, A. Bordes, S. Chopra, and J. Weston, "The goldilocks principle:
Reading children's books with explicit memory representations," arXiv Preprint,
arXiv:1511.02301, 2015.
[29] R. Hu, J. Andreas, M. Rohrbach, T. Darrell, and K. Saenko, "Learning to reason: End-to-end module networks for visual question answering," arXiv Preprint,
arXiv:1704.05526, 2017.
[30] A. Jabri, A. Joulin, and L. van der Maaten, "Revisiting visual question
answering baselines," in Proc. European Conf. Computer Vision (ECCV) 2016,
pp. 727-739.
[31] J. Johnson, B. Hariharan, L. van der Maaten, L. Fei-Fei, C. L. Zitnick, and R.
B. Girshick, "CLEVR: A diagnostic data set for compositional language and elementary visual reasoning," in Proc. Conf. Computer Vision and Pattern
Recognition (CVPR), 2017.
[32] J. Johnson, B. Hariharan, L. van der Maaten, J. Hoffman, F. Li, C. L. Zitnick,
and R. B. Girshick, "Inferring and executing programs for visual reasoning,"
CoRR, 2017. [Online]. Available: http://arxiv.org/abs/1705.03633
[33] J.-H. Kim, S.-W. Lee, D.-H. Kwak, M.-O. Heo, J. Kim, J.-W. Ha, and B.-T.
Zhang, "Multimodal residual learning for visual QA," in Proc. Advances Neural
Information Processing Systems (NIPS), 2016, pp. 361-369.
[34] J.-H. Kim, K.-W. On, J. Kim, J.-W. Ha, and B.-T. Zhang, "Hadamard product
for low-rank bilinear pooling," arXiv Preprint, arXiv:1610.04325, 2016.
[35] R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz, S. Chen, Y.
Kalantidis, L.-J. Li, D. A. Shamma, M. Bernstein, and L. Fei-Fei, "Visual
genome: Connecting language and vision using crowdsourced dense image annotations," arXiv Preprint, arXiv:1602.07332, 2016.
[36] A. Krizhevsky, I. Sutskever, and G. E. Hinton, "Imagenet classification with
deep convolutional neural networks," in Proc. Advances in Neural Information
Processing Systems, 2012, pp. 1106-1114.
[37] A. Kumar, O. Irsoy, J. Su, J. Bradbury, R. English, B. Pierce, P. Ondruska, I.
Gulrajani, and R. Socher, "Ask me anything: Dynamic memory networks for natural language processing," in Proc. Int. Conf. Machine Learning, 2016, pp. 1378-
1387.
[38] A. Lazaridou, N. T. Pham, and M. Baroni, "Combining language and vision
with a multimodal skip-gram model," in Proc. Conf. North American Chapter
Assoc. Computational Linguistics-Human Language Technologies (HLTNAACL), 2015, pp. 153-163.

IEEE SIGNAL PROCESSING MAGAZINE

|

November 2017

|


http://www.edu.au http://www.visualqa.org/ http://eva http://lai.cloudcv.org http://www.arxiv.org/abs/1705.03633

Table of Contents for the Digital Edition of Signal Processing - November 2017

Signal Processing - November 2017 - Cover1
Signal Processing - November 2017 - Cover2
Signal Processing - November 2017 - 1
Signal Processing - November 2017 - 2
Signal Processing - November 2017 - 3
Signal Processing - November 2017 - 4
Signal Processing - November 2017 - 5
Signal Processing - November 2017 - 6
Signal Processing - November 2017 - 7
Signal Processing - November 2017 - 8
Signal Processing - November 2017 - 9
Signal Processing - November 2017 - 10
Signal Processing - November 2017 - 11
Signal Processing - November 2017 - 12
Signal Processing - November 2017 - 13
Signal Processing - November 2017 - 14
Signal Processing - November 2017 - 15
Signal Processing - November 2017 - 16
Signal Processing - November 2017 - 17
Signal Processing - November 2017 - 18
Signal Processing - November 2017 - 19
Signal Processing - November 2017 - 20
Signal Processing - November 2017 - 21
Signal Processing - November 2017 - 22
Signal Processing - November 2017 - 23
Signal Processing - November 2017 - 24
Signal Processing - November 2017 - 25
Signal Processing - November 2017 - 26
Signal Processing - November 2017 - 27
Signal Processing - November 2017 - 28
Signal Processing - November 2017 - 29
Signal Processing - November 2017 - 30
Signal Processing - November 2017 - 31
Signal Processing - November 2017 - 32
Signal Processing - November 2017 - 33
Signal Processing - November 2017 - 34
Signal Processing - November 2017 - 35
Signal Processing - November 2017 - 36
Signal Processing - November 2017 - 37
Signal Processing - November 2017 - 38
Signal Processing - November 2017 - 39
Signal Processing - November 2017 - 40
Signal Processing - November 2017 - 41
Signal Processing - November 2017 - 42
Signal Processing - November 2017 - 43
Signal Processing - November 2017 - 44
Signal Processing - November 2017 - 45
Signal Processing - November 2017 - 46
Signal Processing - November 2017 - 47
Signal Processing - November 2017 - 48
Signal Processing - November 2017 - 49
Signal Processing - November 2017 - 50
Signal Processing - November 2017 - 51
Signal Processing - November 2017 - 52
Signal Processing - November 2017 - 53
Signal Processing - November 2017 - 54
Signal Processing - November 2017 - 55
Signal Processing - November 2017 - 56
Signal Processing - November 2017 - 57
Signal Processing - November 2017 - 58
Signal Processing - November 2017 - 59
Signal Processing - November 2017 - 60
Signal Processing - November 2017 - 61
Signal Processing - November 2017 - 62
Signal Processing - November 2017 - 63
Signal Processing - November 2017 - 64
Signal Processing - November 2017 - 65
Signal Processing - November 2017 - 66
Signal Processing - November 2017 - 67
Signal Processing - November 2017 - 68
Signal Processing - November 2017 - 69
Signal Processing - November 2017 - 70
Signal Processing - November 2017 - 71
Signal Processing - November 2017 - 72
Signal Processing - November 2017 - 73
Signal Processing - November 2017 - 74
Signal Processing - November 2017 - 75
Signal Processing - November 2017 - 76
Signal Processing - November 2017 - 77
Signal Processing - November 2017 - 78
Signal Processing - November 2017 - 79
Signal Processing - November 2017 - 80
Signal Processing - November 2017 - 81
Signal Processing - November 2017 - 82
Signal Processing - November 2017 - 83
Signal Processing - November 2017 - 84
Signal Processing - November 2017 - 85
Signal Processing - November 2017 - 86
Signal Processing - November 2017 - 87
Signal Processing - November 2017 - 88
Signal Processing - November 2017 - 89
Signal Processing - November 2017 - 90
Signal Processing - November 2017 - 91
Signal Processing - November 2017 - 92
Signal Processing - November 2017 - 93
Signal Processing - November 2017 - 94
Signal Processing - November 2017 - 95
Signal Processing - November 2017 - 96
Signal Processing - November 2017 - 97
Signal Processing - November 2017 - 98
Signal Processing - November 2017 - 99
Signal Processing - November 2017 - 100
Signal Processing - November 2017 - 101
Signal Processing - November 2017 - 102
Signal Processing - November 2017 - 103
Signal Processing - November 2017 - 104
Signal Processing - November 2017 - 105
Signal Processing - November 2017 - 106
Signal Processing - November 2017 - 107
Signal Processing - November 2017 - 108
Signal Processing - November 2017 - 109
Signal Processing - November 2017 - 110
Signal Processing - November 2017 - 111
Signal Processing - November 2017 - 112
Signal Processing - November 2017 - 113
Signal Processing - November 2017 - 114
Signal Processing - November 2017 - 115
Signal Processing - November 2017 - 116
Signal Processing - November 2017 - 117
Signal Processing - November 2017 - 118
Signal Processing - November 2017 - 119
Signal Processing - November 2017 - 120
Signal Processing - November 2017 - 121
Signal Processing - November 2017 - 122
Signal Processing - November 2017 - 123
Signal Processing - November 2017 - 124
Signal Processing - November 2017 - 125
Signal Processing - November 2017 - 126
Signal Processing - November 2017 - 127
Signal Processing - November 2017 - 128
Signal Processing - November 2017 - 129
Signal Processing - November 2017 - 130
Signal Processing - November 2017 - 131
Signal Processing - November 2017 - 132
Signal Processing - November 2017 - 133
Signal Processing - November 2017 - 134
Signal Processing - November 2017 - 135
Signal Processing - November 2017 - 136
Signal Processing - November 2017 - 137
Signal Processing - November 2017 - 138
Signal Processing - November 2017 - 139
Signal Processing - November 2017 - 140
Signal Processing - November 2017 - 141
Signal Processing - November 2017 - 142
Signal Processing - November 2017 - 143
Signal Processing - November 2017 - 144
Signal Processing - November 2017 - 145
Signal Processing - November 2017 - 146
Signal Processing - November 2017 - 147
Signal Processing - November 2017 - 148
Signal Processing - November 2017 - 149
Signal Processing - November 2017 - 150
Signal Processing - November 2017 - 151
Signal Processing - November 2017 - 152
Signal Processing - November 2017 - 153
Signal Processing - November 2017 - 154
Signal Processing - November 2017 - 155
Signal Processing - November 2017 - 156
Signal Processing - November 2017 - 157
Signal Processing - November 2017 - 158
Signal Processing - November 2017 - 159
Signal Processing - November 2017 - 160
Signal Processing - November 2017 - 161
Signal Processing - November 2017 - 162
Signal Processing - November 2017 - 163
Signal Processing - November 2017 - 164
Signal Processing - November 2017 - 165
Signal Processing - November 2017 - 166
Signal Processing - November 2017 - 167
Signal Processing - November 2017 - 168
Signal Processing - November 2017 - 169
Signal Processing - November 2017 - 170
Signal Processing - November 2017 - 171
Signal Processing - November 2017 - 172
Signal Processing - November 2017 - 173
Signal Processing - November 2017 - 174
Signal Processing - November 2017 - 175
Signal Processing - November 2017 - 176
Signal Processing - November 2017 - Cover3
Signal Processing - November 2017 - Cover4
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_201809
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_201807
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_201805
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_201803
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_201801
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1117
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0917
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0717
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0517
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0317
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0117
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1116
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0916
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0716
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0516
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0316
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0116
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1115
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0915
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0715
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0515
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0315
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0115
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1114
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0914
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0714
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0514
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0314
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0114
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1113
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0913
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0713
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0513
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0313
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0113
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1112
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0912
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0712
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0512
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0312
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0112
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1111
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0911
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0711
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0511
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0311
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0111
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1110
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0910
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0710
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0510
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0310
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0110
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1109
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0909
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0709
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0509
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0309
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0109
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_1108
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0908
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0708
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0508
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0308
https://www.nxtbook.com/nxtbooks/ieee/signalprocessing_0108
https://www.nxtbookmedia.com