Can deep convolutional neural networks support relational reasoning in the same-different task?

被引:13
作者
Puebla, Guillermo [1 ]
Bowers, Jeffrey S. [1 ]
机构
[1] Univ Bristol, Sch Psychol Sci, Bristol, England
基金
欧洲研究理事会;
关键词
same-different relations; relational reasoning; visual relations; deep neural network; BINDING;
D O I
10.1167/jov.22.10.11
中图分类号
R77 [眼科学];
学科分类号
100212 ;
摘要
Same-different visual reasoning is a basic skill central to abstract combinatorial thought. This fact has lead neural networks researchers to test same-different classification on deep convolutional neural networks (DCNNs), which has resulted in a controversy regarding whether this skill is within the capacity of these models. However, most tests of same-different classification rely on testing on images that come from the same pixel-level distribution as the training images, yielding the results inconclusive. In this study, we tested relational same-different reasoning in DCNNs. In a series of simulations we show that models based on the ResNet architecture are capable of visual same-different classification, but only when the test images are similar to the training images at the pixel level. In contrast, when there is a shift in the testing distribution that does not change the relation between the objects in the image, the performance of DCNNs decreases substantially. This finding is true even when the DCNNs' training regime is expanded to include images taken from a wide range of different pixel-level distributions or when the model is trained on the testing distribution but on a different task in a multitask learning context. Furthermore, we show that the relation network, a deep learning architecture specifically designed to tackle visual relational reasoning problems, suffers the same kind of limitations. Overall, the results of this study suggest that learning same-different relations is beyond the scope of current DCNNs.
引用
收藏
页数:18
相关论文
共 37 条
[1]   RECOGNITION-BY-COMPONENTS - A THEORY OF HUMAN IMAGE UNDERSTANDING [J].
BIEDERMAN, I .
PSYCHOLOGICAL REVIEW, 1987, 94 (02) :115-147
[2]  
Bromley J., 1993, International Journal of Pattern Recognition and Artificial Intelligence, V7, P669, DOI 10.1142/S0218001493000339
[3]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[4]   How Do Humans Sketch Objects? [J].
Eitz, Mathias ;
Hays, James ;
Alexa, Marc .
ACM TRANSACTIONS ON GRAPHICS, 2012, 31 (04)
[5]   Prelinguistic Relational Concepts: Investigating Analogical Processing in Infants [J].
Ferry, Alissa L. ;
Hespos, Susan J. ;
Gentner, Dedre .
CHILD DEVELOPMENT, 2015, 86 (05) :1386-1405
[6]   Comparing machines and humans on a visual categorization test [J].
Fleuret, Francois ;
Li, Ting ;
Dubout, Charles ;
Wampler, Emma K. ;
Yantis, Steven ;
Geman, Donald .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2011, 108 (43) :17621-17625
[7]   Five points to check when comparing visual perception in humans and machines [J].
Funke, Christina M. ;
Borowski, Judy ;
Stosio, Karolina ;
Brendel, Wieland ;
Wallis, Thomas S. A. ;
Bethge, Matthias .
JOURNAL OF VISION, 2021, 21 (03) :1-23
[8]  
Geirhos R, 2018, ADV NEUR IN, V31
[9]   Learning same and different relations: cross-species comparisons [J].
Gentner, Dedre ;
Shao, Ruxue ;
Simms, Nina ;
Hespos, Susan .
CURRENT OPINION IN BEHAVIORAL SCIENCES, 2021, 37 :84-89
[10]  
Greff K, 2020, Arxiv, DOI [arXiv:2012.05208, 10.48550/arXiv.2012.05208]