TY - GEN
T1 - Attentive collaborative filtering
T2 - 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2017
AU - Chen, Jingyuan
AU - Zhang, Hanwang
AU - He, Xiangnan
AU - Nie, Liqiang
AU - Liu, Wei
AU - Chua, Tat Seng
N1 - Publisher Copyright:
© 2017 Copyright held by the owner/author(s).
PY - 2017/8/7
Y1 - 2017/8/7
N2 - Multimedia content is dominating today's Web information.The nature of multimedia user-item interactions is 1/0 binary implicit feedback (e.g., photo likes, video views, song downloads, etc.), which can be collected at a larger scale with a much lower cost than explicit feedback (e.g., product ratings). However, the majority of existing collaborative flltering (CF) systems are not well-designed for multimedia recommendation, since they ignore the implicitness in users' interactions with multimedia content. We argue that, in multimedia recommendation, there exists item-And component-level implicitness which blurs the underlying users' preferences. The item-level implicitness means that users' preferences on items (e.g., photos, videos, songs, etc.) are unknown, while the componentlevel implicitness means that inside each item users' preferences on different components (e.g., regions in an image, frames of a video, etc.) are unknown. For example, a "view" on a video does not provide any speci.c information about how the user likes the video (i.e., item-level) and which parts of the video the user is interested in (i.e., component-level). In this paper, we introduce a novel attention mechanism in CF to address the challenging item-And component-level implicit feedback in multimedia recommendation, dubbed A.entive Collaborative Filtering (ACF). Speci.cally, our attention model is a neural network that consists of two attention modules: The component-level attention module, starting from any content feature extraction network (e.g., CNN for images/videos), which learns to select informative components of multimedia items, and the item-level attention module, which learns to score the item preferences. ACF can be seamlessly incorporated into classic CF models with implicit feedback, such as BPR and SVD++, and effciently trained using SGD. Through extensive experiments on two real-world multimedia Web services: Vine and Pinterest, we show that ACF significantly outperforms state-of-The-Art CF methods.
AB - Multimedia content is dominating today's Web information.The nature of multimedia user-item interactions is 1/0 binary implicit feedback (e.g., photo likes, video views, song downloads, etc.), which can be collected at a larger scale with a much lower cost than explicit feedback (e.g., product ratings). However, the majority of existing collaborative flltering (CF) systems are not well-designed for multimedia recommendation, since they ignore the implicitness in users' interactions with multimedia content. We argue that, in multimedia recommendation, there exists item-And component-level implicitness which blurs the underlying users' preferences. The item-level implicitness means that users' preferences on items (e.g., photos, videos, songs, etc.) are unknown, while the componentlevel implicitness means that inside each item users' preferences on different components (e.g., regions in an image, frames of a video, etc.) are unknown. For example, a "view" on a video does not provide any speci.c information about how the user likes the video (i.e., item-level) and which parts of the video the user is interested in (i.e., component-level). In this paper, we introduce a novel attention mechanism in CF to address the challenging item-And component-level implicit feedback in multimedia recommendation, dubbed A.entive Collaborative Filtering (ACF). Speci.cally, our attention model is a neural network that consists of two attention modules: The component-level attention module, starting from any content feature extraction network (e.g., CNN for images/videos), which learns to select informative components of multimedia items, and the item-level attention module, which learns to score the item preferences. ACF can be seamlessly incorporated into classic CF models with implicit feedback, such as BPR and SVD++, and effciently trained using SGD. Through extensive experiments on two real-world multimedia Web services: Vine and Pinterest, we show that ACF significantly outperforms state-of-The-Art CF methods.
KW - Attention
KW - Collaborative Filtering
KW - Implicit Feedback
KW - Multimedia Recommendation
UR - https://www.scopus.com/pages/publications/85029357805
U2 - 10.1145/3077136.3080797
DO - 10.1145/3077136.3080797
M3 - 会议稿件
AN - SCOPUS:85029357805
T3 - SIGIR 2017 - Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval
SP - 335
EP - 344
BT - SIGIR 2017 - Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval
PB - Association for Computing Machinery, Inc
Y2 - 7 August 2017 through 11 August 2017
ER -