Skip to main navigation Skip to search Skip to main content

Attentive collaborative filtering: Multimedia recommendation with item-And component-level attention

  • Jingyuan Chen
  • , Hanwang Zhang
  • , Xiangnan He*
  • , Liqiang Nie
  • , Wei Liu
  • , Tat Seng Chua
  • *Corresponding author for this work
  • National University of Singapore
  • Columbia University
  • Shandong University
  • Tencent

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Multimedia content is dominating today's Web information.The nature of multimedia user-item interactions is 1/0 binary implicit feedback (e.g., photo likes, video views, song downloads, etc.), which can be collected at a larger scale with a much lower cost than explicit feedback (e.g., product ratings). However, the majority of existing collaborative flltering (CF) systems are not well-designed for multimedia recommendation, since they ignore the implicitness in users' interactions with multimedia content. We argue that, in multimedia recommendation, there exists item-And component-level implicitness which blurs the underlying users' preferences. The item-level implicitness means that users' preferences on items (e.g., photos, videos, songs, etc.) are unknown, while the componentlevel implicitness means that inside each item users' preferences on different components (e.g., regions in an image, frames of a video, etc.) are unknown. For example, a "view" on a video does not provide any speci.c information about how the user likes the video (i.e., item-level) and which parts of the video the user is interested in (i.e., component-level). In this paper, we introduce a novel attention mechanism in CF to address the challenging item-And component-level implicit feedback in multimedia recommendation, dubbed A.entive Collaborative Filtering (ACF). Speci.cally, our attention model is a neural network that consists of two attention modules: The component-level attention module, starting from any content feature extraction network (e.g., CNN for images/videos), which learns to select informative components of multimedia items, and the item-level attention module, which learns to score the item preferences. ACF can be seamlessly incorporated into classic CF models with implicit feedback, such as BPR and SVD++, and effciently trained using SGD. Through extensive experiments on two real-world multimedia Web services: Vine and Pinterest, we show that ACF significantly outperforms state-of-The-Art CF methods.

Original languageEnglish
Title of host publicationSIGIR 2017 - Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval
PublisherAssociation for Computing Machinery, Inc
Pages335-344
Number of pages10
ISBN (Electronic)9781450350228
DOIs
StatePublished - 7 Aug 2017
Externally publishedYes
Event40th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2017 - Tokyo, Shinjuku, Japan
Duration: 7 Aug 201711 Aug 2017

Publication series

NameSIGIR 2017 - Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval

Conference

Conference40th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2017
Country/TerritoryJapan
CityTokyo, Shinjuku
Period7/08/1711/08/17

Keywords

  • Attention
  • Collaborative Filtering
  • Implicit Feedback
  • Multimedia Recommendation

Fingerprint

Dive into the research topics of 'Attentive collaborative filtering: Multimedia recommendation with item-And component-level attention'. Together they form a unique fingerprint.

Cite this