HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

An Attentive Inductive Bias for Sequential Recommendation beyond the Self-Attention

Yehjin Shin Jeongwhan Choi Hyowon Wi Noseong Park

An Attentive Inductive Bias for Sequential Recommendation beyond the Self-Attention

Abstract

Sequential recommendation (SR) models based on Transformers have achieved remarkable successes. The self-attention mechanism of Transformers for computer vision and natural language processing suffers from the oversmoothing problem, i.e., hidden representations becoming similar to tokens. In the SR domain, we, for the first time, show that the same problem occurs. We present pioneering investigations that reveal the low-pass filtering nature of self-attention in the SR, which causes oversmoothing. To this end, we propose a novel method called $\textbf{B}$eyond $\textbf{S}$elf-$\textbf{A}$ttention for Sequential $\textbf{Rec}$ommendation (BSARec), which leverages the Fourier transform to i) inject an inductive bias by considering fine-grained sequential patterns and ii) integrate low and high-frequency information to mitigate oversmoothing. Our discovery shows significant advancements in the SR domain and is expected to bridge the gap for existing Transformer-based SR models. We test our proposed approach through extensive experiments on 6 benchmark datasets. The experimental results demonstrate that our model outperforms 7 baseline methods in terms of recommendation performance. Our code is available at https://github.com/yehjin-shin/BSARec.

Code Repositories

jeongwhanchoi/BSARec
Official
pytorch
Mentioned in GitHub
yehjin-shin/bsarec
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
sequential-recommendation-on-amazon-beautyBSARec
HR@10: 0.1008
HR@20: 0.1373
HR@5: 0.0736
NDCG@20: 0.0703
NDCG@5: 0.0523
nDCG@10: 0.0611
sequential-recommendation-on-amazon-sportsBSARec
HR@10: 0.0612
HR@20: 0.0858
HR@5: 0.0426
sequential-recommendation-on-amazon-toysBSARec
HR@5: 0.0805
sequential-recommendation-on-lastfmBSARec
HR@10: 0.0807
HR@10 (99 Neg. Samples): 0.5028
HR@20: 0.1174
HR@5: 0.0523
HR@5 (99 Neg. Samples): 0.3752
MRR (99 Neg. Samples): 0.2636
NDCG@10: 0.0435
NDCG@10 (99 Neg. Samples): 0.3045
NDCG@20: 0.0526
NDCG@5: 0.0344
NDCG@5 (99 Neg. Samples): 0.2634
sequential-recommendation-on-movielens-1mBSARec
HR@10: 0.2757
HR@10 (99 Neg. Samples): 0.7978
HR@20: 0.3884
HR@5: 0.1944
HR@5 (99 Neg. Samples): 0.7023
MRR (99 Neg. Samples): 0.5406
NDCG@10: 0.1568
NDCG@10 (99 Neg. Samples): 0.5955
NDCG@20: 0.1851
NDCG@5: 0.1306
NDCG@5 (99 Neg. Samples): 0.5646
sequential-recommendation-on-yelpBSARec
HR@10: 0.0465
HR@10 (99 Neg. Samples): 0.7848
HR@20: 0.0746
HR@5: 0.0275
HR@5 (99 Neg. Samples): 0.6447
MRR (99 Neg. Samples): 0.4587
NDCG@10: 0.0231
NDCG@10 (99 Neg. Samples): 0.5280
NDCG@20: 0.0302
NDCG@5: 0.0170
NDCG@5 (99 Neg. Samples): 0.4824

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
An Attentive Inductive Bias for Sequential Recommendation beyond the Self-Attention | Papers | HyperAI