Raven's Progressive Matrices Completion with Latent Gaussian Process Priors

Abstract reasoning ability is fundamental to human intelligence. It enables humans to uncover relations among abstract concepts and further deduce implicit rules from the relations. As a well-known abstract visual reasoning task, Raven's Progressive Matrices (RPM) are widely used in human IQ tests. Although extensive research has been conducted on RPM solvers with machine intelligence, few studies have considered further advancing the standard answer-selection (classification) problem to a more challenging answer-painting (generating) problem, which can verify whether the model has indeed understood the implicit rules. In this paper we aim to solve the latter one by proposing a deep latent variable model, in which multiple Gaussian processes are employed as priors of latent variables to separately learn underlying abstract concepts from RPMs; thus the proposed model is interpretable in terms of concept-specific latent variables. The latent Gaussian process also provides an effective way of extrapolation for answer painting based on the learned concept-changing rules. We evaluate the proposed model on RPM-like datasets with multiple continuously-changing visual concepts. Experimental results demonstrate that our model requires only few training samples to paint high-quality answers, generate novel RPM panels, and achieve interpretability through concept-specific latent variables.

[1]  Xianglong Liu,et al.  Hierarchical Rule Induction Network for Abstract Visual Reasoning , 2020, ArXiv.

[2]  Yixin Zhu,et al.  Learning Perceptual Inference by Contrasting , 2019, NeurIPS.

[3]  Maithilee Kunda,et al.  Modeling Gestalt Visual Reasoning on the Raven's Progressive Matrices Intelligence Test Using Generative Image Inpainting Techniques , 2019, ArXiv.

[4]  Anton van den Hengel,et al.  V-PROM: A Benchmark for Visual Reasoning Using Visual Progressive Matrices , 2019, AAAI.

[5]  Soren Hauberg,et al.  Explicit Disentanglement of Appearance and Perspective in Generative Models , 2019, NeurIPS.

[6]  Sjoerd van Steenkiste,et al.  Are Disentangled Representations Helpful for Abstract Visual Reasoning? , 2019, NeurIPS.

[7]  Hyun Oh Song,et al.  Learning Discrete and Continuous Factors of Data via Alternating Disentanglement , 2019, ICML.

[8]  Feng Gao,et al.  RAVEN: A Dataset for Relational and Analogical Visual REasoNing , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Alexander Lerchner,et al.  Spatial Broadcast Decoder: A Simple Architecture for Learning Disentangled Representations in VAEs , 2019, ArXiv.

[10]  Bernhard Schölkopf,et al.  Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations , 2018, ICML.

[11]  Alexander Lerchner,et al.  Improving Generalization for Abstract Reasoning Tasks Using Disentangled Feature Representations , 2018, NIPS 2018.

[12]  Felix Hill,et al.  Measuring abstract reasoning in neural networks , 2018, ICML.

[13]  Guodong Zhang,et al.  Differentiable Compositional Kernel Learning for Gaussian Processes , 2018, ICML.

[14]  Emilien Dupont,et al.  Joint-VAE: Learning Disentangled Joint Continuous and Discrete Representations , 2018, NeurIPS.

[15]  Andriy Mnih,et al.  Disentangling by Factorising , 2018, ICML.

[16]  David Duvenaud,et al.  Isolating Sources of Disentanglement in VAEs , 2018, 1802.04942.

[17]  Thomas S. Huang,et al.  Generative Image Inpainting with Contextual Attention , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[18]  Abhishek Kumar,et al.  Variational Inference of Disentangled Latent Concepts from Unlabeled Observations , 2017, ICLR.

[19]  Razvan Pascanu,et al.  A simple neural network module for relational reasoning , 2017, NIPS.

[20]  Christopher Burgess,et al.  beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.

[21]  Andrew Gordon Wilson,et al.  Deep Kernel Learning , 2015, AISTATS.

[22]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[23]  Diederik P. Kingma,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[24]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  R. Gur,et al.  Development of Abbreviated Nine-Item Forms of the Raven’s Standard Progressive Matrices Test , 2012, Assessment.

[26]  G. J. Robertson Raven's Progressive Matrices , 2010 .

[27]  S. Ounpraseuth,et al.  Gaussian Processes for Machine Learning , 2008 .

[28]  Kenneth D. Forbus,et al.  Modeling Visual Problem Solving as Analogical Reasoning , 2017, Psychological review.

[29]  Stephan Lewandowsky,et al.  A Bayesian Model of Rule Induction in Raven's Progressive Matrices , 2012, CogSci.

[30]  Kenneth D. Forbus,et al.  A Structure-Mapping Model of Raven's Progressive Matrices , 2010 .