Links

Tools

Export citation

Search in Google Scholar

Performance Based Display Advertising: A Large Scale Machine Learning Model in Practice

Journal article published in 1 by Krishna Prasad Chitrapura
This paper was not found in any repository; the policy of its publisher is unknown or unclear.
This paper was not found in any repository; the policy of its publisher is unknown or unclear.

Full text: Unavailable

Question mark in circle
Preprint: policy unknown
Question mark in circle
Postprint: policy unknown
Question mark in circle
Published version: policy unknown

Abstract

Display advertising is evolving from just selling eye balls to demographic segments, as in traditional print media, to performance based systems where advertis-ers have an ability to pay for the type of user response valuable to them. Depending on who owns the risk of variance in user response, the advertisers have multiple pricing types options to run their advertising campaign on. These include CPM (pay per view); CPC, (pay per click); CPA, (pay per user action) and Dynamic CPM, which is pay per view but the payment is based on an es-timate of the chance of user action. Publishers bear the risk of loss of revenue in the case of CPC and CPA pric-ing types, whereas the advertisers stand to lose the return in investment in the case of dynamic and CPM pricing types. In addition, the marketplace for display advertis-ing is morphing from being many islands of advertiser and publisher networks to a connected network of net-works, which trades over an ad-exchange to better match supply to demand. These changes have brought forth a new set of research challenges such as predicting rates of rare events, managing risk due to variance in prediction and optimal explore-exploit tradeoffs. In this work, we outlay a subset of new research prob-lems in performance based display advertising and focus on one aspect as a case study of a large scale machine learning model in practice. Specifically, we discuss the anatomy of a response prediction engine that estimates the chances of a page viewer's response to an ad place-ment. This estimate along with the advertiser's bid is used to allocate an ad to a page that would yield maxi-mal expected revenue. Estimating the chance of rare user events such as clicks and conversion needs state of the art machine learning and robust statistical estimation tech-niques that can crunch large amount of data periodically on the Grid. There is an inherent trade-off between the complexity of the model and the time taken to re-learn the model periodically. We shall discuss this trade-off in detail and present insights from a live, web-scale adver-tising exchange. Further, in the network of networks, we not only face estimation issues due to sparsity of events but also due to cold starts, which are advertiser-page view co-occurances that do not appear in the historical data, or appear with very little or no user response. Controlled exploration of such combinations based on the estimate (which exhibits a high variance) is needed to ensure that ads with erroneous estimates do not end up winning large supply of page-views. We will describe one such tech-nique, called throttling which mitigates the risk from cold-starts. We also compare and contrast our response prediction techniques to those which are used in popular problems such as movie recommendation 1 [2], collaborative filter-ing employed by Amazon.com [1] and in paid-search [3].