Open-Universe Theory for Bayesian Inference, Decision, and Sensing (OUTBIDS)

Abstract : This report describes work in Phase I of the OUTBIDS project under the DARPA MSEE program. The goal of OUTBIDS was to develop the theoretical and technological foundations for sensor data interpretation as a form of probabilistic inference. Achieving this goal requires a representation formalism for probability models of sufficient expressive power to handle the complexity of real-world sensor data. The problem involves two primary sources of difficulty: first, the underlying world generating the data typically contains many initially unknown objects interacting over time in complex ways; second, the mapping from objects and behaviors to sensor data is itself (as in the case of visual perception, for example) very complex. The core of the project is the BLOG (Bayesian LOGic) language, which combines probabilistic semantics with the expressive power of first-order logic. Unlike other attempts to combine probability and logic, BLOG supports open-universe models, which allow for uncertainty over the existence and identity of objects. We believe this is a prerequisite for any probabilistic approach to general perception. The team made substantial progress on developing and refining the BLOG language by writing a broad range of models, including two models for computer vision tasks (adaptive video background subtraction for object tracking and a simple form of 3D object recognition and scene reconstruction). We also made good progress towards an efficient inference engine, including new algorithms and initial work on compiler and parallelization technology for BLOG inference. We showed that BLOG could be extended to open-universe decision models and developed algorithms for sensor planning on this basis. We also developed a new theoretical framework for utility-directed inference and proved several foundational theorems.

[1]  René Vidal,et al.  Using global bag of features models in random fields for joint categorization and segmentation of objects , 2011, CVPR 2011.

[2]  René Vidal,et al.  A hypothesize-and-bound algorithm for simultaneous object classification, pose estimation and 3D reconstruction from a single 2D image , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).

[3]  Diego Rother,et al.  Hypothesize and Bound: A Computational Focus of Attention Mechanism for Simultaneous N-D Segmentation, Pose Estimation and Classification Using Shape Priors , 2011, 1104.2580.

[4]  Stuart J. Russell,et al.  Uncertain Observation Times , 2012, SUM.

[5]  Siddharth Srivastava First-Order Models for POMDPs , 2012, StarAI@UAI.

[6]  S. Shankar Sastry,et al.  Using Models of Objects with Deformable Parts for Joint Categorization and Segmentation of Objects , 2012, ACCV.

[7]  René Vidal,et al.  Visual Dictionary Learning for Joint Object Categorization and Segmentation , 2012, ECCV.

[8]  David Tolpin,et al.  Selecting Computations: Theory and Applications , 2012, UAI.

[9]  Lei Li,et al.  The Extended Parameter Filter , 2013, ICML.

[10]  Lei Li,et al.  Dynamic Scaled Sampling for Deterministic Constraints , 2013, AISTATS.

[11]  Allen Y. Yang,et al.  Single-Sample Face Recognition with Image Corruption and Misalignment via Sparse Illumination Transfer , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  Yonina C. Eldar,et al.  Compressive Shift Retrieval , 2013, IEEE Transactions on Signal Processing.

[13]  Lei Li,et al.  Handwriting and Gestures in the Air, Recognizing on the Fly , 2013 .

[14]  René Vidal,et al.  Coarse-to-Fine Semantic Video Segmentation Using Supervoxel Trees , 2013, 2013 IEEE International Conference on Computer Vision.

[15]  Christos Faloutsos,et al.  Why people hate your app: making sense of user feedback in a mobile app store , 2013, KDD.