2016 WideDeepLearningforRecommenderS

From GM-RKB
Jump to navigation Jump to search

Subject Headings: Deep Recommender Algorithm, Wide and Deep Recommender Algorithm.

Notes

Cited By

Quotes

Abstract

Generalized linear models with nonlinear feature transformations are widely used for large-scale regression and classification problems with sparse inputs. Memorization of feature interactions through a wide set of cross-product feature transformations are effective and interpretable, while generalization requires more feature engineering effort. With less feature engineering, deep neural networks can generalize better to unseen feature combinations through low-dimensional dense embeddings learned for the sparse features. However, deep neural networks with embeddings can over-generalize and recommend less relevant items when the user-item interactions are sparse and high-rank. In this paper, we present Wide & Deep learning --- jointly trained wide linear models and deep neural networks --- to combine the benefits of memorization and generalization for recommender systems. We productionized and evaluated the system on Google Play, a commercial mobile app store with over one billion active users and over one million apps. Online experiment results show that Wide & Deep significantly increased app acquisitions compared with wide-only and deep-only models. We have also open-sourced our implementation in TensorFlow.

References

  • 1. John Duchi, Elad Hazan, Yoram Singer, Adaptive Subgradient Methods for Online Learning and Stochastic Optimization, The Journal of Machine Learning Research, 12, p.2121-2159, 2/1/2011
  • 2. K. He, X. Zhang, S. Ren, and J. Sun. Deep Residual Learning for Image Recognition. Proc. IEEE Conference on Computer Vision and Pattern Recognition, 2016.
  • 3. H. B. McMahan. Follow-the-regularized-leader and Mirror Descent: Equivalence Theorems and L1 Regularization. In Proc. AISTATS, 2011.
  • 4. T. Mikolov, A. Deoras, D. Povey, L. Burget, and J. H. Cernocky. Strategies for Training Large Scale Neural Network Language Models. In IEEE Automatic Speech Recognition & Understanding Workshop, 2011.
  • 5. Steffen Rendle, Factorization Machines with LibFM, ACM Transactions on Intelligent Systems and Technology (TIST), v.3 n.3, p.1-22, May 2012 doi:10.1145/2168752.2168771
  • 6. Jonathan Tompson, Arjun Jain, Yann LeCun, Christoph Bregler, Joint Training of a Convolutional Network and a Graphical Model for Human Pose Estimation, Proceedings of the 27th International Conference on Neural Information Processing Systems, p.1799-1807, December 08-13, 2014, Montreal, Canada
  • 7. Hao Wang, Naiyan Wang, Dit-Yan Yeung, Collaborative Deep Learning for Recommender Systems, Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, August 10-13, 2015, Sydney, NSW, Australia doi:10.1145/2783258.2783273
  • 8. Bo Yan, Guanling Chen, AppJoy: Personalized Mobile Application Discovery, Proceedings of the 9th International Conference on Mobile Systems, Applications, and Services, June 28-July 01, 2011, Bethesda, Maryland, USA doi:10.1145/1999995.2000007

}};


 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2016 WideDeepLearningforRecommenderSTal Shaked
Tushar Chandra
Greg Corrado
Jeremiah Harmsen
Heng-Tze Cheng
Levent Koc
Hrishi Aradhye
Glen Anderson
Wei Chai
Mustafa Ispir
Rohan Anil
Zakaria Haque
Lichan Hong
Vihan Jain
Xiaobing Liu
Hemal Shah
Wide & Deep Learning for Recommender Systems10.1145/2988450.29884542016