2019 BERT4RecSequentialRecommendatio

From GM-RKB
Jump to navigation Jump to search

Subject Headings: BERT4Rec; Sequential Recommendations Algorithm; Bidirectional Sequential Model; Cloz.

Notes

Cited By

Quotes

Abstract

Modeling users’ dynamic preferences from their historical behaviors is challenging and crucial for recommendation systems. Previous methods employ sequential neural networks to encode users’ historical interactions from left to right into hidden representations for making recommendations. Despite their effectiveness, we argue that such left-to-right unidirectional models are sub-optimal due to the limitations including: a) unidirectional architectures restrict the power of hidden representation in users’ behavior sequences; b) they often assume a rigidly ordered sequence which is not always practical. To address these limitations, we proposed a sequential recommendation model called BERT4Rec, which employs the deep-bidirectional self-attention to model user behavior sequences. To avoid the information leakage and efficiently train the bidirectional model, we adopt the Cloze objective to sequential recommendation, predicting the random masked items in the sequence by jointly conditioning on their left and right context. In this way, we learn a bidirectional representation model to make recommendations by allowing each item in user historical behaviors to fuse information from both left and right sides. Extensive experiments on four benchmark datasets show that our model outperforms various state-of-the-art sequential models consistently.

References

;

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2019 BERT4RecSequentialRecommendatioJun Liu
Peng Jiang
Fei Sun
Jian Wu
Changhua Pei
Xiao Lin
Wenwu Ou
BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer