HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Scalable Cross-Entropy Loss for Sequential Recommendations with Large Item Catalogs

Gleb Mezentsev Danil Gusak Ivan Oseledets Evgeny Frolov

Scalable Cross-Entropy Loss for Sequential Recommendations with Large Item Catalogs

Abstract

Scalability issue plays a crucial role in productionizing modern recommender systems. Even lightweight architectures may suffer from high computational overload due to intermediate calculations, limiting their practicality in real-world applications. Specifically, applying full Cross-Entropy (CE) loss often yields state-of-the-art performance in terms of recommendations quality. Still, it suffers from excessive GPU memory utilization when dealing with large item catalogs. This paper introduces a novel Scalable Cross-Entropy (SCE) loss function in the sequential learning setup. It approximates the CE loss for datasets with large-size catalogs, enhancing both time efficiency and memory usage without compromising recommendations quality. Unlike traditional negative sampling methods, our approach utilizes a selective GPU-efficient computation strategy, focusing on the most informative elements of the catalog, particularly those most likely to be false positives. This is achieved by approximating the softmax distribution over a subset of the model outputs through the maximum inner product search. Experimental results on multiple datasets demonstrate the effectiveness of SCE in reducing peak memory usage by a factor of up to 100 compared to the alternatives, retaining or even exceeding their metrics values. The proposed approach also opens new perspectives for large-scale developments in different domains, such as large language models.

Code Repositories

AIRI-Institute/Scalable-SASRec
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
sequential-recommendation-on-amazon-beauty-1SASRec-SCE
HR@10: 0.0935
NDCG@10: 0.0544
sequential-recommendation-on-behanceSASRec-SCE
COV@1: 0.0393
COV@10: 0.25
COV@5: 15.3
HR@10: 0.113
HR@5: 0.0853
NDCG@1: 0.0277
NDCG@10: 0.0663
NDCG@5: 0.0572
sequential-recommendation-on-gowallaSASRec-SCE
COV@1: 0.0304
COV@10: 0.2190
COV@5: 0.126
HR@10: 0.0831
HR@5: 0.0574
NDCG@1: 0.0207
NDCG@10: 0.0476
NDCG@5: 0.0393

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Scalable Cross-Entropy Loss for Sequential Recommendations with Large Item Catalogs | Papers | HyperAI