Human Machine Joint Decision Making in Distorted Surveillance Scenario

There is plenty of human-machine joint decision-making scenarios in the real world applications, such as driving assistant, suspect identification, medical diagnosis, etc. Existing algorithms propose that machine should give a rejection option when having a high risk or uncertainty score so that the input can be passed to human to make the decision. This is an interesting algorithmic model of human-machine collaboration, but implicitly assumes that humans are more trustworthy than machines. Such an assumption ignores the bias and inconsistency of human, especially in scenarios where machines have superior recognition ability than humans. In this work, we investigate the human-machine joint decision-making problem in distorted surveillance videos, where machines experimentally prove to be comparable to human beings in tolerance to distortion, sometimes even stronger. We propose a new human-machine joint decision-making framework by considering both the confidences of machine and human. To obtain the confidence of human, we build a real-life human decision-making database and propose a deep neural network to estimate human's confidence. Then, confidence alignment method and decision rule are proposed to further output the final decision. Experiments demonstrate that the proposed framework can make less human intervention and more accurate decisions in several human-machine joint decision-making scenarios.

[1]  Kevin Gimpel,et al.  A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks , 2016, ICLR.

[2]  Ran El-Yaniv,et al.  SelectiveNet: A Deep Neural Network with an Integrated Reject Option , 2019, ICML.

[3]  Qi Tian,et al.  Scalable Person Re-identification: A Benchmark , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[4]  Wei Zhou,et al.  How do you Perceive Differently from an AI — A Database for Semantic Distortion Measurement , 2019, 2019 IEEE International Symposium on Circuits and Systems (ISCAS).

[5]  Mario Vento,et al.  A method for improving classification reliability of multilayer perceptrons , 1995, IEEE Trans. Neural Networks.

[6]  Francesco Solera,et al.  Performance Measures and a Data Set for Multi-target, Multi-camera Tracking , 2016, ECCV Workshops.

[7]  Ramesh Johari,et al.  Learning with Abandonment , 2018, ICML.

[8]  Geraint Rees,et al.  Relating Introspective Accuracy to Individual Differences in Brain Structure , 2010, Science.

[9]  Mehryar Mohri,et al.  Boosting with Abstention , 2016, NIPS.

[10]  Yi Yang,et al.  Unlabeled Samples Generated by GAN Improve the Person Re-identification Baseline in Vitro , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[11]  Alex Kendall,et al.  What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? , 2017, NIPS.

[12]  Mehryar Mohri,et al.  Learning with Rejection , 2016, ALT.

[13]  R. Baierlein Probability Theory: The Logic of Science , 2004 .

[14]  Claudio Gentile,et al.  Online Learning with Abstention , 2017, ICML.

[15]  Mario Vento,et al.  To reject or not to reject: that is the question-an answer in case of neural classifiers , 2000, IEEE Trans. Syst. Man Cybern. Part C.

[16]  M. Sigman,et al.  Individual consistency in the accuracy and distribution of confidence judgments , 2016, Cognition.

[17]  Ran El-Yaniv,et al.  Agnostic Pointwise-Competitive Selective Classification , 2015, J. Artif. Intell. Res..

[18]  Mark J. F. Gales,et al.  Predictive Uncertainty Estimation via Prior Networks , 2018, NeurIPS.

[19]  Kilian Q. Weinberger,et al.  On Calibration of Modern Neural Networks , 2017, ICML.

[20]  M. Sigman,et al.  Confidence as Bayesian Probability: From Neural Origins to Behavior , 2015, Neuron.

[21]  Martín Graziano,et al.  Distinct patterns of functional brain connectivity correlate with objective performance and subjective beliefs , 2013, Proceedings of the National Academy of Sciences.

[22]  Yarin Gal,et al.  Uncertainty in Deep Learning , 2016 .

[23]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[24]  Zoubin Ghahramani,et al.  Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.

[25]  Bianca Zadrozny,et al.  Transforming classifier scores into accurate multiclass probability estimates , 2002, KDD.

[26]  P. Latham,et al.  The idiosyncratic nature of confidence , 2017, bioRxiv.

[27]  Krishna P. Gummadi,et al.  On Fairness, Diversity and Randomness in Algorithmic Decision Making , 2017, ArXiv.

[28]  Ran El-Yaniv,et al.  Selective Classification for Deep Neural Networks , 2017, NIPS.