Decoupled Side Information Fusion for Sequential Recommendation

April 23, 2022 ยท Entered Twilight ยท ๐Ÿ› Annual International ACM SIGIR Conference on Research and Development in Information Retrieval

๐Ÿ’ค TWILIGHT: Eternal Rest
Repo abandoned since publication

Repo contents: DIF.sh, LICENSE, README.md, benchmarks, configs, dif.png, recbole, run_recbole.py

Authors Yueqi Xie, Peilin Zhou, Sunghun Kim arXiv ID 2204.11046 Category cs.IR: Information Retrieval Cross-listed cs.AI Citations 151 Venue Annual International ACM SIGIR Conference on Research and Development in Information Retrieval Repository https://github.com/AIM-SE/DIF-SR โญ 111 Last Checked 1 month ago
Abstract
Side information fusion for sequential recommendation (SR) aims to effectively leverage various side information to enhance the performance of next-item prediction. Most state-of-the-art methods build on self-attention networks and focus on exploring various solutions to integrate the item embedding and side information embeddings before the attention layer. However, our analysis shows that the early integration of various types of embeddings limits the expressiveness of attention matrices due to a rank bottleneck and constrains the flexibility of gradients. Also, it involves mixed correlations among the different heterogeneous information resources, which brings extra disturbance to attention calculation. Motivated by this, we propose Decoupled Side Information Fusion for Sequential Recommendation (DIF-SR), which moves the side information from the input to the attention layer and decouples the attention calculation of various side information and item representation. We theoretically and empirically show that the proposed solution allows higher-rank attention matrices and flexible gradients to enhance the modeling capacity of side information fusion. Also, auxiliary attribute predictors are proposed to further activate the beneficial interaction between side information and item representation learning. Extensive experiments on four real-world datasets demonstrate that our proposed solution stably outperforms state-of-the-art SR models. Further studies show that our proposed solution can be readily incorporated into current attention-based SR models and significantly boost performance. Our source code is available at https://github.com/AIM-SE/DIF-SR.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Information Retrieval