Back to Search Start Over

PME: pruning-based multi-size embedding for recommender systems

Authors :
Zirui Liu
Qingquan Song
Li Li
Soo-Hyun Choi
Rui Chen
Xia Hu
Source :
Frontiers in Big Data, Vol 6 (2023)
Publication Year :
2023
Publisher :
Frontiers Media S.A., 2023.

Abstract

Embedding is widely used in recommendation models to learn feature representations. However, the traditional embedding technique that assigns a fixed size to all categorical features may be suboptimal due to the following reasons. In recommendation domain, the majority of categorical features' embeddings can be trained with less capacity without impacting model performance, thereby storing embeddings with equal length may incur unnecessary memory usage. Existing work that tries to allocate customized sizes for each feature usually either simply scales the embedding size with feature's popularity or formulates this size allocation problem as an architecture selection problem. Unfortunately, most of these methods either have large performance drop or incur significant extra time cost for searching proper embedding sizes. In this article, instead of formulating the size allocation problem as an architecture selection problem, we approach the problem from a pruning perspective and propose Pruning-based Multi-size Embedding (PME) framework. During the search phase, we prune the dimensions that have the least impact on model performance in the embedding to reduce its capacity. Then, we show that the customized size of each token can be obtained by transferring the capacity of its pruned embedding with significant less search cost. Experimental results validate that PME can efficiently find proper sizes and hence achieve strong performance while significantly reducing the number of parameters in the embedding layer.

Details

Language :
English
ISSN :
2624909X
Volume :
6
Database :
Directory of Open Access Journals
Journal :
Frontiers in Big Data
Publication Type :
Academic Journal
Accession number :
edsdoj.4501d2e701d4242a705da33eebce4dc
Document Type :
article
Full Text :
https://doi.org/10.3389/fdata.2023.1195742