We present the problem of click-through-rate (CTR) for search advertising in ALiMaMa, which displays user information, item information, shop information and trade results. Traditionally, people use logistic regression (LR) to predict it. However, because of the lack of learning ability and the sparse feature matrix, the prediction results are always not so satisfying. In this paper, we mainly propose some feature engineering methods based on gradient boosting decision tree (GBDT) and Bayesian smoothing to obtain a wonderful feature, which has more useful information and is not so sparse. Also, we use xgboost (XGB) instead of LR as our prediction model. The proposed methods are evaluated using offline experiments and the experiment results prove that the log loss drop near \(5\%\) after using these feature engineering methods and XGB. Obviously, it is an excellent performance.
[1]
Joaquin Quiñonero Candela,et al.
Practical Lessons from Predicting Clicks on Ads at Facebook
,
2014,
ADKDD'14.
[2]
Jason Weston,et al.
End-To-End Memory Networks
,
2015,
NIPS.
[3]
Tat-Seng Chua,et al.
Neural Factorization Machines for Sparse Predictive Analytics
,
2017,
SIGIR.
[4]
Xuerui Wang,et al.
Click-Through Rate Estimation for Rare Events in Online Advertising
,
2011
.
[5]
Heng-Tze Cheng,et al.
Wide & Deep Learning for Recommender Systems
,
2016,
DLRS@RecSys.
[6]
Martin Wattenberg,et al.
Ad click prediction: a view from the trenches
,
2013,
KDD.
[7]
Dong Yu,et al.
Deep Crossing: Web-Scale Modeling without Manually Crafted Combinatorial Features
,
2016,
KDD.