Large-Scale Customized Models for Advertisers

Performance advertisers want to maximize the return on their advertising spend. In the online advertising world, this means showing the ad only to those users most likely to convert i.e. buy a product or service. Existing ad targeting solutions such as context targeting and rule-based segment targeting primarily leverage marketing intuition to identify audience segments that would be likely to convert. Even the more sophisticated model-based approaches such as behavioral targeting identify audience segments interested in certain coarse-grained categories defined by the publisher. Advertisers are now able, through beaconing, to tell us exactly who their preferred customers are. Advertisers want to augment their existing advertising campaign with custom models that learn from the campaign and focus on attracting new users. Motivated by our experience with advertisers, we pose this problem within the context of ensemble learning. Building custom models for an existing ad campaign can be viewed as operations on an ensemble classifier: add, modify, or complement a classifier. An ideal new classifier should incrementally improve the ensemble and minimize overlap with any existing classifiers already in the ensemble–it should learn something new. With the proposed approach we are able to augment the advertising campaigns of several large advertisers at a large online advertising company.

[1]  Ludmila I. Kuncheva,et al.  Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.

[2]  Rajesh Parekh,et al.  Combining Behavioral and Social Network Data for Online Advertising , 2008, 2008 IEEE International Conference on Data Mining Workshops.

[3]  Robert E. Schapire,et al.  The Boosting Approach to Machine Learning An Overview , 2003 .

[4]  Foster Provost,et al.  Audience selection for on-line brand advertising: privacy-friendly social network targeting , 2009, KDD.

[5]  Raymond J. Mooney,et al.  Constructing Diverse Classifier Ensembles using Artificial Training Examples , 2003, IJCAI.

[6]  John Shawe-Taylor,et al.  The Set Covering Machine , 2003, J. Mach. Learn. Res..

[7]  Ethem Alpaydin,et al.  Incremental construction of classifier and discriminant ensembles , 2009, Inf. Sci..

[8]  B. J. Pine Mass customizing products and services , 1993 .

[9]  Wen Zhang,et al.  How much can behavioral targeting help online advertising? , 2009, WWW '09.

[10]  Gavin Brown,et al.  "Good" and "Bad" Diversity in Majority Vote Ensembles , 2010, MCS.

[11]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[12]  Ling Li,et al.  Ordinal Regression by Extended Binary Classification , 2006, NIPS.

[13]  John F. Canny,et al.  Large-scale behavioral targeting , 2009, KDD.

[14]  Philip S. Yu,et al.  Mining concept-drifting data streams using ensemble classifiers , 2003, KDD '03.

[15]  Nicholas Kushmerick,et al.  Ensemble Learning with Biased Classifiers: The Triskel Algorithm , 2005, Multiple Classifier Systems.