INITIATOR: Noise-contrastive Estimation for Marked Temporal Point Process

Copious sequential event data has consistently increased in various high-impact domains such as social media and sharing economy. When events start to take place in a sequential fashion, an important question arises: "what type of event will happen at what time in the near future?" To answer the question, a class of mathematical models called the marked temporal point process is often exploited as it can model the timing and properties of events seamlessly in a joint framework. Recently, various recurrent neural network (RNN) models are proposed to enhance the predictive power of mark temporal point process. However, existing marked temporal point models are fundamentally based on the Maximum Likelihood Estimation (MLE) framework for the training, and inevitably suffer from the problem resulted from the intractable likelihood function. Surprisingly, little attention has been paid to address this issue. In this work, we propose INITIATOR - a novel training framework based on noise-contrastive estimation to resolve this problem. Theoretically, we show the exists a strong connection between the proposed INITIATOR and the exact MLE. Experimentally, the efficacy of INITIATOR is demonstrated over the state-of-the-art approaches on several real-world datasets from various areas.

[1]  M. V. Rossum,et al.  In Neural Computation , 2022 .

[2]  Xin Liu,et al.  On Predictive Patent Valuation: Forecasting Patent Citations and Their Types , 2017, AAAI.

[3]  Vinayak A. Rao,et al.  A Multitask Point Process Predictive Model , 2015, ICML.

[4]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[5]  Ian J. Goodfellow,et al.  On distinguishability criteria for estimating generative models , 2014, ICLR.

[6]  Albert-László Barabási,et al.  Modeling and Predicting Popularity Dynamics via Reinforced Poisson Processes , 2014, AAAI.

[7]  Huan Liu,et al.  Mining Human Mobility in Location-Based Social Networks , 2015, Mining Human Mobility in Location-Based Social Networks.

[8]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[9]  Yann LeCun,et al.  Signature Verification Using A "Siamese" Time Delay Neural Network , 1993, Int. J. Pattern Recognit. Artif. Intell..

[10]  Jakob Gulddahl Rasmussen,et al.  Lecture Notes: Temporal Point Processes and the Conditional Intensity Function , 2018, 1806.00221.

[11]  Jure Leskovec,et al.  SEISMIC: A Self-Exciting Point Process Model for Predicting Tweet Popularity , 2015, KDD.

[12]  Hongyuan Zha,et al.  Modeling the Intensity Function of Point Process Via Recurrent Neural Networks , 2017, AAAI.

[13]  Charles B. Morrey,et al.  Differentiation under the Integral Sign. Improper Integrals. The Gamma Function , 1985 .

[14]  Yee Whye Teh,et al.  A fast and simple algorithm for training neural probabilistic language models , 2012, ICML.

[15]  Donald L. Snyder,et al.  Random Point Processes in Time and Space , 1991 .

[16]  Rune B. Lyngsø,et al.  Lecture Notes I , 2008 .

[17]  K. Pearson Biometrika , 1902, The American Naturalist.

[18]  A. Hawkes Spectra of some self-exciting and mutually exciting point processes , 1971 .

[19]  M. R. Leadbetter Poisson Processes , 2011, International Encyclopedia of Statistical Science.

[20]  Utkarsh Upadhyay,et al.  Recurrent Marked Temporal Point Processes: Embedding Event History to Vector , 2016, KDD.

[21]  Jason Eisner,et al.  The Neural Hawkes Process: A Neurally Self-Modulating Multivariate Point Process , 2016, NIPS.

[22]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[23]  Bernhard Schölkopf,et al.  Distilling Information Reliability and Source Trustworthiness from Digital Traces , 2017, WWW.

[24]  Isabel Valera,et al.  Modeling Adoption and Usage of Competing Products , 2014, 2015 IEEE International Conference on Data Mining.

[25]  Léon Bottou,et al.  Towards Principled Methods for Training Generative Adversarial Networks , 2017, ICLR.

[26]  Karen Hunger Parshall The American Mathematical Monthly (1894-1919): A New Journal in the Service of Mathematics and Its Educators , 2016 .

[27]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[28]  Huan Liu,et al.  Synthesis Lectures on Data Mining and Knowledge Discovery , 2009 .

[29]  Shuai Gao,et al.  Modeling and Predicting Retweeting Dynamics on Microblogging Platforms , 2015, WSDM.