Knowledge-driven Active Learning

In the last few years, Deep Learning models have become increasingly popular. However, their deployment is still precluded in those contexts where the amount of supervised data is limited and manual labelling expensive. Active learning strategies aim at solving this problem by requiring supervision only on few unlabelled samples, which improve the most the model performances after adding them to the training set. Most strategies are based on uncertain sample selection, and even often restricted to samples lying close to the decision boundary. Here we propose a very different approach, taking into consideration domain knowledge. Indeed, in the case of multi-label classification, the relationships among classes offer a way to spot incoherent predictions, i.e., predictions where the model may most likely need supervision. We have developed a framework where first-order-logic knowledge is converted into constraints and their violation is checked as a natural guide for sample selection. We empirically demonstrate that knowledge-driven strategy outperforms standard uncertain strategies, particularly on those datasets where the domain knowledge is rich. Furthermore, we show how the proposed approach enables discovering data distributions lying far from initial training data.

[1]  Razvan Pascanu,et al.  A simple neural network module for relational reasoning , 2017, NIPS.

[2]  Christian Eitzinger,et al.  Triangular Norms , 2001, Künstliche Intell..

[3]  Klaus Brinker,et al.  Incorporating Diversity in Active Learning with Support Vector Machines , 2003, ICML.

[4]  Remus Pop,et al.  Deep Ensemble Bayesian Active Learning : Addressing the Mode Collapse issue in Monte Carlo dropout via Ensembles , 2018, ArXiv.

[5]  Frédéric Precioso,et al.  Adversarial Active Learning for Deep Networks: a Margin Based Approach , 2018, ArXiv.

[6]  Hyuk-Jae Lee,et al.  Active Learning for Deep Object Detection via Probabilistic Modeling , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[7]  Michele Fenzi,et al.  Scalable Active Learning for Object Detection , 2020, 2020 IEEE Intelligent Vehicles Symposium (IV).

[8]  Frédéric Precioso,et al.  Active learning strategy for CNN combining batchwise Dropout and Query-By-Committee , 2017, ESANN.

[9]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  Haifeng Shen,et al.  Adaptive Object Detection with Dual Multi-Label Prediction , 2020, ECCV.

[11]  Xiaojin Zhu,et al.  Introduction to Semi-Supervised Learning , 2009, Synthesis Lectures on Artificial Intelligence and Machine Learning.

[12]  Bin Liu,et al.  Using multi-label classification to improve object detection , 2019, Neurocomputing.

[13]  Fabio Roli,et al.  Domain Knowledge Alleviates Adversarial Attacks in Multi-Label Classifiers. , 2021, IEEE transactions on pattern analysis and machine intelligence.

[14]  Been Kim,et al.  Concept Bottleneck Models , 2020, ICML.

[15]  Andrew McCallum,et al.  Employing EM and Pool-Based Active Learning for Text Classification , 1998, ICML.

[16]  Marco Gori,et al.  Semantic-based regularization for learning and inference , 2017, Artif. Intell..

[17]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  Yinda Zhang,et al.  LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.

[19]  Marco Gori,et al.  LYRICS: A General Interface Layer to Integrate Logic Inference and Deep Learning , 2019, ECML/PKDD.

[20]  Laurian M. Chirica,et al.  The entity-relationship model: toward a unified view of data , 1975, SIGF.

[21]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[22]  Marcello Sanguineti,et al.  Foundations of Support Constraint Machines , 2015, Neural Computation.

[23]  Gary Marcus,et al.  Deep Learning: A Critical Appraisal , 2018, ArXiv.

[24]  Greg Schohn,et al.  Less is More: Active Learning with Support Vector Machines , 2000, ICML.

[25]  B. Horn , 2022 .

[26]  Luc De Raedt,et al.  Probabilistic (logic) programming concepts , 2015, Machine Learning.

[27]  I. Tsang,et al.  Bayesian Active Learning by Disagreements: A Geometric Perspective , 2021, ArXiv.

[28]  Fedor Zhdanov,et al.  Diverse mini-batch Active Learning , 2019, ArXiv.

[29]  Zoubin Ghahramani,et al.  Bayesian Active Learning for Classification and Preference Learning , 2011, ArXiv.

[30]  Sergio Gomez Colmenarejo,et al.  Hybrid computing using a neural network with dynamic external memory , 2016, Nature.

[31]  Artur S. d'Avila Garcez,et al.  Logic Tensor Networks: Deep Learning and Logical Reasoning from Data and Knowledge , 2016, NeSy@HLAI.

[32]  Xiaoshuang Shi,et al.  Fully automatic knee osteoarthritis severity grading using deep neural networks with a novel ordinal loss , 2019, Comput. Medical Imaging Graph..

[33]  Gopinath Chennupati,et al.  On Mixup Training: Improved Calibration and Predictive Uncertainty for Deep Neural Networks , 2019, NeurIPS.

[34]  Andrew McCallum,et al.  Introduction to Statistical Relational Learning , 2007 .

[35]  In So Kweon,et al.  Learning Loss for Active Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[36]  Ross D. King,et al.  Active Learning for Regression Based on Query by Committee , 2007, IDEAL.

[37]  Daphne Koller,et al.  Support Vector Machine Active Learning with Applications to Text Classification , 2000, J. Mach. Learn. Res..

[38]  Marco Gori,et al.  Constraint Verification With Kernel Machines , 2013, IEEE Transactions on Neural Networks and Learning Systems.

[39]  Yarin Gal,et al.  BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning , 2019, NeurIPS.

[40]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[41]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[42]  Pietro Perona,et al.  The Caltech-UCSD Birds-200-2011 Dataset , 2011 .

[43]  Silvio Savarese,et al.  Active Learning for Convolutional Neural Networks: A Core-Set Approach , 2017, ICLR.

[44]  Philipp Koehn,et al.  Cognitive Psychology , 1992, Ageing and Society.

[45]  Sanja Fidler,et al.  Detect What You Can: Detecting and Representing Objects Using Holistic Models and Body Parts , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[46]  John Langford,et al.  Deep Batch Active Learning by Diverse, Uncertain Gradient Lower Bounds , 2019, ICLR.

[47]  Frank Hutter,et al.  Decoupled Weight Decay Regularization , 2017, ICLR.

[48]  Joan Bruna,et al.  Intriguing properties of neural networks , 2013, ICLR.

[49]  Zoubin Ghahramani,et al.  Deep Bayesian Active Learning with Image Data , 2017, ICML.

[50]  Andreas Nürnberger,et al.  The Power of Ensembles for Active Learning in Image Classification , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[51]  Marco Gori,et al.  Logic Explained Networks , 2021, ArXiv.