Mixed Dimension Embeddings With Application To Memory-efficient Recommendation Systems | Awesome Learning to Hash Add your paper to Learning2Hash

Mixed Dimension Embeddings With Application To Memory-efficient Recommendation Systems

Antonio Ginart, Maxim Naumov, Dheevatsa Mudigere, Jiyan Yang, James Zou . Arxiv 2019 – 32 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
Datasets Evaluation Memory Efficiency Recommender Systems Scalability

Embedding representations power machine intelligence in many applications, including recommendation systems, but they are space intensive – potentially occupying hundreds of gigabytes in large-scale settings. To help manage this outsized memory consumption, we explore mixed dimension embeddings, an embedding layer architecture in which a particular embedding vector’s dimension scales with its query frequency. Through theoretical analysis and systematic experiments, we demonstrate that using mixed dimensions can drastically reduce the memory usage, while maintaining and even improving the ML performance. Empirically, we show that the proposed mixed dimension layers improve accuracy by 0.1% using half as many parameters or maintain it using 16X fewer parameters for click-through rate prediction task on the Criteo Kaggle dataset.

Similar Work