DOI QR코드

DOI QR Code

A CTR Prediction Approach for Text Advertising Based on the SAE-LR Deep Neural Network

  • Jiang, Zilong (School of Computer Science and Technology, Wuhan University of Technology) ;
  • Gao, Shu (School of Computer Science and Technology, Wuhan University of Technology) ;
  • Dai, Wei (School of Economics and Management, Hubei Polytechnic University)
  • Received : 2016.12.15
  • Accepted : 2017.06.04
  • Published : 2017.10.31

Abstract

For the autoencoder (AE) implemented as a construction component, this paper uses the method of greedy layer-by-layer pre-training without supervision to construct the stacked autoencoder (SAE) to extract the abstract features of the original input data, which is regarded as the input of the logistic regression (LR) model, after which the click-through rate (CTR) of the user to the advertisement under the contextual environment can be obtained. These experiments show that, compared with the usual logistic regression model and support vector regression model used in the field of predicting the advertising CTR in the industry, the SAE-LR model has a relatively large promotion in the AUC value. Based on the improvement of accuracy of advertising CTR prediction, the enterprises can accurately understand and have cognition for the needs of their customers, which promotes the multi-path development with high efficiency and low cost under the condition of internet finance.

Keywords

References

  1. Y. Hu, J. Shin, and Z. Tang, "Incentive problems in performance-based online advertising pricing: cost per click vs. cost per action," Management Science, vol. 62, no. 7, pp. 2022-2038, 2015.
  2. H. Cheng and E. Cantu-Paz, "Personalized click prediction in sponsored search," in Proceedings of the 3rd ACM International Conference on Web Search and Data Mining, New York, NY, 2010, pp. 351-360.
  3. K. Dave and V. Varma, "Predicting the click-through rate for rare/new Ads," Centre for Search and Information Extraction Lab International Institute of Information Technology, Hyderabad, Technical Report IIIT/TR/2010/15, 2010.
  4. H. B. McMahan, G. Holt, D. Sculley, M, Young, D. Ebner, J. Grady, et al., "Ad click prediction: a view from the trenches," in Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, 2013, pp. 1222-1230.
  5. Y. Tagami, S. Ono, K. Yamamoto, K. Tsukamoto, and A. Tajima, "CTR prediction for contextual advertising: learning-to-rank approach," in Proceedings of the 7th International Workshop on Data Mining for Online Advertising, Chicago, IL, 2013.
  6. O. Chapelle, "Modeling delayed feedback in display advertising," in Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, 2014, pp. 1097-1105.
  7. H. Cheng, R. van Zwol, J. Azimi, E. Manavoglu, R. Zhang, Y. Zhou, and V. Navalpakkam, "Multimedia features for click prediction of new ads in display advertising," in Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Beijing, China, 2012, pp. 777-785.
  8. T. Graepel, J. Q. Candela, T. Borchert, and R. Herbrich, "Web-scale Bayesian click-through rate prediction for sponsored search advertising in Microsofts Bing search engine," in Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 2010, pp. 13-20.
  9. Y. Bengio, "Learning deep architectures for AI," Foundations and Trends in Machine Learning, vol. 2, no. 1, pp. 1-127, 2009. https://doi.org/10.1561/2200000006
  10. G. E. Hinton, S. Osindero, and Y. W. Teh, "A fast learning algorithm for deep belief nets," Neural Computation, vol. 18, no. 7, pp. 1527-1554, 2006. https://doi.org/10.1162/neco.2006.18.7.1527
  11. Y. Bengio, P. Lamblin, D. Popovici, and H. Larochelle, "Greedy layer-wise training of deep networks," in Advances in Neural Information Processing Systems, vol. 19, Vancouver, British Columbia: MIT Press, 2006, pp. 153-160.
  12. P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, and P. A. Manzagol, "Stacked denoising autoencoders: learning useful representations in a deep network with a local denoising criterion," Journal of Machine Learning Research, vol. 11, pp. 3371-3408, 2010.
  13. R. Trivedi, T. N. Singh, and A. K. Raina, "Simultaneous prediction of blast-induced flyrock and fragmentation in opencast limestone mines using back propagation neural network," International Journal of Mining and Mineral Engineering, vol. 7, no. 3, pp. 237-252, 2016. https://doi.org/10.1504/IJMME.2016.078350
  14. H. I. Suk, S. W. Lee, and D. Shen, "Latent feature representation with stacked auto-encoder for AD/MCI diagnosis," Brain Structure and Function, vol. 220, no. 2, pp. 841-859, 2015. https://doi.org/10.1007/s00429-013-0687-3
  15. J. Gehring, Y. Miao, F. Metze, and A. Waibel "Extracting deep bottleneck features using stacked autoencoders," in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, Canada, 2013, pp. 3377-3381.
  16. T. Osman, P. Divigalpitiya, and T. Arima, "Driving factors of urban sprawl in Giza Governorate of the Greater Cairo metropolitan region using a logistic regression model," International Journal of Urban Sciences, vol. 20, no. 2, pp. 206-225, 2016. https://doi.org/10.1080/12265934.2016.1162728
  17. KDD Cup 2012, Track 2 [Online]. Available: https://www.kaggle.com/c/kddcup2012-track2/.
  18. K. J. Grimm, G. L. Mazza, and P. Davoudzadeh, "Model selection in finite mixture models: a k-fold cross-validation approach," Structural Equation Modeling: A Multidisciplinary Journal, vol. 24, no. 2, pp. 246-252, 2017. https://doi.org/10.1080/10705511.2016.1250638
  19. Word2vec [Online]. Available: https://code.google.com/p/word2vec/2017.
  20. Z. Jiang, S. Gao, and M. Li, "A feature vector representation approach for short text based on RNNLM and pooling computation," Academic Journal of Manufacturing Engineering, vol. 15, no. 2, pp. 6-14, 2017.
  21. K. Weinberger, A. Dasgupta, J. Langford, A. Smola, and J. Attenberg, "Feature hashing for large scale multitask learning," in Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, Canada, 2009, pp. 1113-1120.
  22. Theano [Online]. Available: https://github.com/Theano/Theano.
  23. G. Santamaria-Bonfil, A. Reyes-Ballesteros, and C. Gershenson, "Wind speed forecasting for wind farms: a method based on support vector regression," Renewable Energy, vol. 85, pp. 790-809, 2016. https://doi.org/10.1016/j.renene.2015.07.004
  24. K. P. Lin and P. F. Pai, "Solar power output forecasting using evolutionary seasonal decomposition leastsquare support vector regression," Journal of Cleaner Production, vol. 134(Part B), pp. 456-462, 2016. https://doi.org/10.1016/j.jclepro.2015.08.099
  25. T. Akimova, M. H. Levine, U. H. Beier, and W. W. Hancock, "Standardization, evaluation, and areaunder-curve analysis of human and murine Treg suppressive function," Methods in Molecular Biology, vol. 1371, pp. 43-78, 2016.
  26. A. Jimenez-Valverde, "Insights into the area under the receiver operating characteristic curve (AUC) as a discrimination measure in species distribution modelling," Global Ecology and Biogeography, vol. 21, no. 4, pp. 498-507, 2012. https://doi.org/10.1111/j.1466-8238.2011.00683.x
  27. X. Deng, Q. Liu, Y. Deng, and S. Mahadevan, "An improved method to construct basic probability assignment based on the confusion matrix for classification problem," Information Sciences, vol. 340-341, pp. 250-261, 2016. https://doi.org/10.1016/j.ins.2016.01.033
  28. A. J. Talabani, B. H. Endreseth, S. Lydersen, and T. H. Edna, "Clinical diagnostic accuracy of acute colonic diverticulitis in patients admitted with acute abdominal pain, a receiver operating characteristic curve analysis," International Journal of Colorectal Disease, vol. 32, no. 1, pp. 41-47, 2017. https://doi.org/10.1007/s00384-016-2644-0
  29. scikit-learn 0.15.2 [Online]. Available: https://pypi.python.org/pypi/scikit-learn/0.15.2.
  30. Logistic Regression in Python [Online]. Available: http://blog.yhat.com/posts/logistic-regression-andpython.html.
  31. Setting Up Scientific Python [Online]. Available: http://blog.yhat.com/posts/setting-up-scientific-python.html.
  32. P. Rebentrost, M. Mohseni, and S. Lloyd, "Quantum support vector machine for big data classification," Physical Review Letter, vol. 113, article no. 130503, 2014.
  33. Python sklearn.svm.SVR Examples [Online]. Available: http://www.programcreek.com/python/example/75189/sklearn.svm.SVR.
  34. Z. Jiang, S. Gao, and W. Dai, "Research on CTR prediction for contextual advertising based on deep architecture model," Journal of Control Engineering and Applied Informatics, vol. 18, no. 1, pp. 11-19, 2016.