Source
CIKM
DATE OF PUBLICATION
08/05/2024
Authors
Ivan Oseledets Evgeny Frolov Gleb Mezentsev Danil Gusak
Share

RECE: Reduced Cross-Entropy Loss for Large-Catalogue Sequential Recommenders

Abstract

Scalability is a major challenge in modern recommender systems. Insequential recommendations, full Cross-Entropy (CE) loss achievesstate-of-the-art recommendation quality but consumes excessiveGPU memory with large item catalogs, limiting its practicality.Using a GPU-efficient locality-sensitive hashing-like algorithmfor approximating large tensor of logits, this paper introduces anovel RECE (REduced Cross-Entropy) loss. RECE significantlyreduces memory consumption while allowing one to enjoy thestate-of-the-art performance of full CE loss. Experimental results onvarious datasets show that RECE cuts training peak memory usageby up to 12 times compared to existing methods while retaining orexceeding performance metrics of CE loss. The approach also opensup new possibilities for large-scale applications in other domains.

Join AIRI