R.I.P.
๐ป
Ghosted
AutoLossGen: Automatic Loss Function Generation for Recommender Systems
April 27, 2022 ยท Entered Twilight ยท ๐ Annual International ACM SIGIR Conference on Research and Development in Information Retrieval
Repo contents: .gitignore, README.md, dataset, log, model, pics, result, src
Authors
Zelong Li, Jianchao Ji, Yingqiang Ge, Yongfeng Zhang
arXiv ID
2204.13160
Category
cs.IR: Information Retrieval
Cross-listed
cs.AI,
cs.LG
Citations
27
Venue
Annual International ACM SIGIR Conference on Research and Development in Information Retrieval
Repository
https://github.com/rutgerswiselab/AutoLossGen
โญ 22
Last Checked
1 month ago
Abstract
In recommendation systems, the choice of loss function is critical since a good loss may significantly improve the model performance. However, manually designing a good loss is a big challenge due to the complexity of the problem. A large fraction of previous work focuses on handcrafted loss functions, which needs significant expertise and human effort. In this paper, inspired by the recent development of automated machine learning, we propose an automatic loss function generation framework, AutoLossGen, which is able to generate loss functions directly constructed from basic mathematical operators without prior knowledge on loss structure. More specifically, we develop a controller model driven by reinforcement learning to generate loss functions, and develop iterative and alternating optimization schedule to update the parameters of both the controller model and the recommender model. One challenge for automatic loss generation in recommender systems is the extreme sparsity of recommendation datasets, which leads to the sparse reward problem for loss generation and search. To solve the problem, we further develop a reward filtering mechanism for efficient and effective loss generation. Experimental results show that our framework manages to create tailored loss functions for different recommendation models and datasets, and the generated loss gives better recommendation performance than commonly used baseline losses. Besides, most of the generated losses are transferable, i.e., the loss generated based on one model and dataset also works well for another model or dataset. Source code of the work is available at https://github.com/rutgerswiselab/AutoLossGen.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Information Retrieval
R.I.P.
๐ป
Ghosted
LightGCN: Simplifying and Powering Graph Convolution Network for Recommendation
R.I.P.
๐ป
Ghosted
Graph Convolutional Neural Networks for Web-Scale Recommender Systems
๐
๐
Old Age
Neural Graph Collaborative Filtering
R.I.P.
๐ป
Ghosted
Self-Attentive Sequential Recommendation
R.I.P.
๐ป
Ghosted