Twin Neural Network Regression

We introduce twin neural network (TNN) regression. This method predicts differences between the target values of two different data points rather than the targets themselves. The solution of a traditional regression problem is then obtained by averaging over an ensemble of all predicted differences between the targets of an unseen data point and all training data points. Whereas ensembles are normally costly to produce, TNN regression intrinsically creates an ensemble of predictions of twice the size of the training set while only training a single neural network. Since ensembles have been shown to be more accurate than single models this property naturally transfers to TNN regression. We show that TNNs are able to compete or yield more accurate predictions for different data sets, compared to other state-of-the-art methods. Furthermore, TNN regression is constrained by self-consistency conditions. We find that the violation of these conditions provides an estimate for the prediction uncertainty.

[1]  K. Müller,et al.  Fast and accurate modeling of molecular atomization energies with machine learning. , 2011, Physical review letters.

[2]  Chuang Zhang,et al.  Horizontal and Vertical Ensemble with Deep Representation for Classification , 2013, ArXiv.

[3]  Lawrence D. Jackel,et al.  Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.

[4]  Oleksandr Makeyev,et al.  Neural network with ensembles , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).

[5]  Klaus-Robert Müller,et al.  SchNet: A continuous-filter convolutional neural network for modeling quantum interactions , 2017, NIPS.

[6]  David S. Melnick,et al.  International evaluation of an AI system for breast cancer screening , 2020, Nature.

[7]  Ruifeng Liu,et al.  General Approach to Estimate Error Bars for Quantitative Structure-Activity Relationship Predictions of Molecular Activity , 2018, J. Chem. Inf. Model..

[8]  Sahil Shah,et al.  Predicting stock and stock price index movement using Trend Deterministic Data Preparation and machine learning techniques , 2015, Expert Syst. Appl..

[9]  Zachary W. Ulissi,et al.  Accelerated discovery of CO2 electrocatalysts using active machine learning , 2020, Nature.

[10]  Michele Ceriotti,et al.  Fast and Accurate Uncertainty Estimation in Chemical Machine Learning. , 2018, Journal of chemical theory and computation.

[11]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[12]  THE IMPORTANCE OF BEING UNCERTAIN , 2018 .

[13]  A. Raftery,et al.  Strictly Proper Scoring Rules, Prediction, and Estimation , 2007 .

[14]  R. Kondor,et al.  Gaussian approximation potentials: the accuracy of quantum mechanics, without the electrons. , 2009, Physical review letters.

[15]  Gang Niu,et al.  Classification from Pairwise Similarity and Unlabeled Data , 2018, ICML.

[16]  Philip Bachman,et al.  Learning with Pseudo-Ensembles , 2014, NIPS.

[17]  Yong Yu,et al.  Sales forecasting using extreme learning machine with applications in fashion retailing , 2008, Decis. Support Syst..

[18]  Markus Reiher,et al.  Error-Controlled Exploration of Chemical Reaction Networks with Gaussian Processes. , 2018, Journal of chemical theory and computation.

[19]  Pierre Baldi,et al.  Neural Networks for Fingerprint Recognition , 1993, Neural Computation.

[20]  James M. Brown,et al.  Siamese neural networks for continuous disease severity evaluation and change detection in medical imaging , 2020, npj Digital Medicine.

[21]  Yann LeCun,et al.  Signature Verification Using A "Siamese" Time Delay Neural Network , 1993, Int. J. Pattern Recognit. Artif. Intell..

[22]  Markus Reiher,et al.  Gaussian Process-Based Refinement of Dispersion Corrections. , 2019, Journal of chemical theory and computation.

[23]  Kilian Q. Weinberger,et al.  Snapshot Ensembles: Train 1, get M for free , 2017, ICLR.

[24]  E Weinan,et al.  Active Learning of Uniformly Accurate Inter-atomic Potentials for Materials Simulation , 2018, Physical Review Materials.

[25]  Anand Chandrasekaran,et al.  Solving the electronic structure problem with machine learning , 2019, npj Computational Materials.

[26]  Aki Vehtari,et al.  Minimum energy path calculations with Gaussian process regression , 2016, 1703.10423.

[27]  Luca Bertinetto,et al.  Fully-Convolutional Siamese Networks for Object Tracking , 2016, ECCV Workshops.

[28]  Anders Krogh,et al.  Neural Network Ensembles, Cross Validation, and Active Learning , 1994, NIPS.

[29]  Ming Yang,et al.  DeepFace: Closing the Gap to Human-Level Performance in Face Verification , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[30]  Haitao Liu,et al.  When Gaussian Process Meets Big Data: A Review of Scalable GPs , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[31]  E. Viezzer,et al.  Up to two billion times acceleration of scientific simulations with deep neural architecture search , 2020, ArXiv.

[32]  H. Kulik,et al.  A Quantitative Uncertainty Metric Controls Error in Neural Network-Driven Chemical Discovery , 2019 .

[33]  Gregory R. Koch,et al.  Siamese Neural Networks for One-Shot Image Recognition , 2015 .

[34]  Alireza Khorshidi,et al.  Addressing uncertainty in atomistic machine learning. , 2017, Physical chemistry chemical physics : PCCP.

[35]  Shady Elbassuoni,et al.  Calories Prediction from Food Images , 2017, AAAI.

[36]  Zachary W. Ulissi,et al.  To address surface reaction network complexity using scaling relations machine learning and DFT calculations , 2017, Nature Communications.

[37]  Jae Kwon Bae,et al.  Using machine learning algorithms for housing price prediction: The case of Fairfax County, Virginia housing data , 2015, Expert Syst. Appl..

[38]  Zoubin Ghahramani,et al.  Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.

[39]  Sebastian Nowozin,et al.  Hydra: Preserving Ensemble Diversity for Model Distillation , 2020, ArXiv.

[40]  Eric Xing,et al.  Methods for comparing uncertainty quantifications for material property predictions. , 2019 .

[41]  Dmitry Vetrov,et al.  Pitfalls of In-Domain Uncertainty Estimation and Ensembling in Deep Learning , 2020, ICLR.

[42]  David A. Strubbe,et al.  Deep learning and density-functional theory , 2018, Physical Review A.

[43]  A. Dawid,et al.  Theory and applications of proper scoring rules , 2014, 1401.0398.

[44]  Taxonomy and definitions for terms related to driving automation systems for on-road motor vehicles , 2022 .

[45]  Vijay Ganesh,et al.  Discovering Symmetry Invariants and Conserved Quantities by Interpreting Siamese Neural Networks , 2020, Physical Review Research.

[46]  Charles Blundell,et al.  Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.

[47]  Zoubin Ghahramani,et al.  Probabilistic machine learning and artificial intelligence , 2015, Nature.

[48]  Camilo L. M. Morais,et al.  Uncertainty estimation and misclassification probability for classification models based on discriminant analysis and support vector machines. , 2019, Analytica chimica acta.

[49]  Nigel M. Allinson,et al.  Fast committee learning: preliminary results , 1998 .

[50]  Tae-Kyun Kim,et al.  Siamese Regression Networks with Efficient mid-level Feature Extraction for 3D Object Pose Estimation , 2016, ArXiv.

[51]  Michael Mrejen,et al.  Plasmonic nanostructure design and characterization via Deep Learning , 2018, Light: Science & Applications.