UniMixer: A Unified Architecture for Scaling Laws in Recommendation Systems
Abstract
A unified scaling architecture for recommendation systems called UniMixer is proposed, featuring a generalized parameterized token mixing module that optimizes token mixing patterns and improves scaling efficiency across different architectures.
In recent years, the scaling laws of recommendation models have attracted increasing attention, which govern the relationship between performance and parameters/FLOPs of recommenders. Currently, there are three mainstream architectures for achieving scaling in recommendation models, namely attention-based, TokenMixer-based, and factorization-machine-based methods, which exhibit fundamental differences in both design philosophy and architectural structure. In this paper, we propose a unified scaling architecture for recommendation systems, namely UniMixer, to improve scaling efficiency and establish a unified theoretical framework that unifies the mainstream scaling blocks. By transforming the rule-based TokenMixer to an equivalent parameterized structure, we construct a generalized parameterized feature mixing module that allows the token mixing patterns to be optimized and learned during model training. Meanwhile, the generalized parameterized token mixing removes the constraint in TokenMixer that requires the number of heads to be equal to the number of tokens. Furthermore, we establish a unified scaling module design framework for recommender systems, which bridges the connections among attention-based, TokenMixer-based, and factorization-machine-based methods. To further boost scaling ROI, a lightweight UniMixing module is designed, UniMixing-Lite, which further compresses the model parameters and computational cost while significantly improve the model performance. The scaling curves are shown in the following figure. Extensive offline and online experiments are conducted to verify the superior scaling abilities of UniMixer.
Community
Recent scaling laws in recommendation models have highlighted three dominant architectures—attention-based, TokenMixer-based, and factorization-machine-based methods—which differ fundamentally in design and structure. This paper proposes UniMixer, a unified scaling architecture that generalizes token mixing into a learnable parameterized module, bridges these paradigms under a unified framework, and introduces a lightweight variant (UniMixing-Lite) to improve scaling efficiency, achieving strong performance gains validated by extensive experiments.
Get this paper in your agent:
hf papers read 2604.00590 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper