OpenOneRec commited on
Commit
53e84b4
Β·
verified Β·
1 Parent(s): c5af9e3

Upload 2 files

Browse files
Files changed (2) hide show
  1. README.md +46 -36
  2. assets/framework.png +2 -2
README.md CHANGED
@@ -3,15 +3,14 @@
3
  <p align="center">
4
  <strong>An Open Foundation Model and Benchmark to Accelerate Generative Recommendation</strong>
5
  </p>
6
-
7
  <p align="center">
8
- <a href="https://huggingface.co/Onerec">
9
  <img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-OneRec-ffc107?color=ffc107&logoColor=white" />
10
  </a>
11
- <a href="https://github.com/OnerecLM/OpenOneRec">
12
  <img alt="GitHub Code" src="https://img.shields.io/badge/GitHub-OpenOneRec-black?logo=github" />
13
  </a>
14
- <a href="">
15
  <img alt="Paper" src="https://img.shields.io/badge/Paper-ArXiv-b31b1b?logo=arxiv" />
16
  </a>
17
  <a href="#license">
@@ -19,7 +18,6 @@
19
  </a>
20
  </p>
21
  </div>
22
-
23
  <br>
24
 
25
  ## πŸ“– Introduction
@@ -27,26 +25,26 @@
27
  **OpenOneRec** is an open-source framework designed to bridge the gap between traditional recommendation systems and Large Language Models (LLMs). While Generative Recommendation has shown promise, existing models often struggle with isolated data silos and a lack of reasoning capabilities.
28
 
29
  To address this, we introduce a unified framework that comprises:
30
- * **RecIF-Bench**: The first holistic Recommendation Instruction-Following Benchmark, containing **120M interactions** from 200k users across heterogeneous domains (Short Video, Ad, Product).
31
- * **OpenOneRec-Foundation Models**: A family of models (1.7B & 8B) built on the Qwen backbone. These models are trained on hundreds of billions of tokens, integrating collaborative signals with general semantics.
32
  * **Full-Stack Pipeline**: We open-source our comprehensive training pipeline, including data processing, co-pretraining, and post-training, to ensure full reproducibility and facilitate scaling law research in recommendation.
33
 
34
  ## πŸ”₯ News
35
 
36
- * **[2025.xx.xx]** πŸŽ‰ **OpenOneRec-Foundation** models (1.7B, 8B) are now available on Hugging Face!
37
- * **[2025.xx.xx]** πŸ“‘ The technical report [OpenOneRec Technical Report] has been released.
38
- * **[2025.xx.xx]** πŸš€ **RecIF-Bench** dataset and evaluation scripts are open-sourced.
39
 
40
  ## πŸ“Š RecIF-Bench
41
 
42
- We propose **RecIF-Bench**, a comprehensive benchmark designed to rigorously evaluate recommendation foundation models. It organizes 8 distinct tasks into a four-layer capability hierarchy:
43
 
44
  * **Layer 0: Semantic Alignment** (Item Understanding)
45
- * **Layer 1: Fundamental Recommendation** (Short Video Rec, Ad Rec, Product Rec, Label Prediction)
46
  * **Layer 2: Instruction Following** (Interactive Rec, Label-Conditional Rec)
47
  * **Layer 3: Reasoning** (Recommendation Explanation)
48
 
49
- The benchmark aggregates data from three domains: **Short Video** (Content), **Ad** (Commercial), and **Product** (E-commerce).
50
 
51
  ## πŸ€– Model Zoo
52
 
@@ -54,10 +52,10 @@ The OpenOneRec-Foundation series is built upon the Qwen architecture, enhanced w
54
 
55
  | Model | Backbone | Parameters | Description | Link |
56
  | :--- | :--- | :--- | :--- | :--- |
57
- | **OneRec-Open-1.7B** | Qwen3-1.7B | 1.7B | Standard version trained on open-source data (~100B tokens) | [HuggingFace](https://huggingface.co/Onerec) |
58
- | **OneRec-Open-8B** | Qwen3-8B | 8B | Standard version trained on open-source data (~100B tokens) | [HuggingFace](https://huggingface.co/Onerec) |
59
- | **OneRec-Pro-1.7B** | Qwen3-1.7B | 1.7B | Enhanced version with proprietary tokens | *Coming Soon* |
60
- | **OneRec-Pro-8B** | Qwen3-8B | 8B | Enhanced version with proprietary tokens | *Coming Soon* |
61
 
62
  ## πŸ—οΈ Method & Architecture
63
 
@@ -68,14 +66,14 @@ To bridge the modality gap, we treat items as a distinct modality using **Itemic
68
 
69
  ### 2. Training Pipeline
70
  Our framework utilizes the following recipe:
71
- * **Pre-Training**: Integrates collaborative signals via Itemic-Text Alignment and mixed-domain Co-Pretraining.
72
  * **Post-Training**:
73
- * *Stage 1*: Cold-Start SFT for basic instruction following.
74
- * *Stage 2*: Alternating On-Policy Distillation & SFT to balance general reasoning and recommendation performance.
75
- * *Stage 3*: Recommendation-oriented Reinforcement Learning (Rec-RL).
76
 
77
  <div align="center">
78
- <img src="assets/framework.png" width="90%" alt="OpenOneRec Overall Framework" />
79
  <br>
80
  <em>Figure: The Overall Framework of OpenOneRec.</em>
81
  </div>
@@ -85,21 +83,35 @@ Our framework utilizes the following recipe:
85
  ### Results on RecIF-Bench
86
  OpenOneRec-Foundation achieves **State-of-the-Art (SOTA)** results across RecIF-Bench tasks, significantly outperforming baselines like LC-Rec and TIGER.
87
 
88
- | Task | Metric | TIGER | LC-Rec-8B | **OneRec-8B-Pro** |
89
- | :--- | :--- | :--- | :--- | :--- |
90
- | **Short Video Rec** | Recall@32 | 0.0132 | 0.0180 | **0.0369** |
91
- | **Ad Rec** | Recall@32 | 0.0581 | 0.0723 | **0.0964** |
92
- | **Product Rec** | Recall@32 | 0.0283 | 0.0416 | **0.0538** |
93
- | **Label-Cond. Rec** | Recall@32 | 0.0123 | 0.0170 | **0.0235** |
94
- | **Label Pred.** | AUC | 0.6675 | 0.6139 | **0.6912** |
95
- | **Interactive Rec** | Recall@32 | -- | 0.2394 | **0.3458** |
96
- | **Item Understand.** | LLM-Judge Score| -- | 0.2517 | **0.3209** |
97
- | **Rec. Explanation** | LLM-Judge Score| -- | 3.9350 | **4.0381** |
98
-
99
 
100
  ### Cross-Domain Transferability
101
  On the **Amazon Benchmark** (10 datasets), OpenOneRec demonstrates exceptional zero-shot/few-shot transfer capabilities, achieving an average **26.8% improvement** in Recall@10 over the second-best method.
102
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
103
  ## πŸš€ Quick Start
104
 
105
  *Code release and detailed usage instructions are coming soon.*
@@ -108,11 +120,9 @@ Currently, you can load our models using `transformers`:
108
 
109
  ```python
110
  from transformers import AutoModelForCausalLM, AutoTokenizer
111
-
112
  model_id = "Onerec/OneRec-Open-8B"
113
  tokenizer = AutoTokenizer.from_pretrained(model_id)
114
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
115
-
116
  # Example inference code will be updated here
117
  ```
118
 
@@ -128,4 +138,4 @@ If you find our work helpful, please cite our technical report:
128
  }
129
  ```
130
  ## πŸ›‘οΈ License
131
- The code in this repository is licensed under the Apache 2.0 License. The model weights are subject to their specific license agreements.
 
3
  <p align="center">
4
  <strong>An Open Foundation Model and Benchmark to Accelerate Generative Recommendation</strong>
5
  </p>
 
6
  <p align="center">
7
+ <a href="https://huggingface.co/OpenOneRec">
8
  <img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-OneRec-ffc107?color=ffc107&logoColor=white" />
9
  </a>
10
+ <a href="https://github.com/Kuaishou-OneRec/OpenOneRec">
11
  <img alt="GitHub Code" src="https://img.shields.io/badge/GitHub-OpenOneRec-black?logo=github" />
12
  </a>
13
+ <a href="https://arxiv.org/abs/2508.20900">
14
  <img alt="Paper" src="https://img.shields.io/badge/Paper-ArXiv-b31b1b?logo=arxiv" />
15
  </a>
16
  <a href="#license">
 
18
  </a>
19
  </p>
20
  </div>
 
21
  <br>
22
 
23
  ## πŸ“– Introduction
 
25
  **OpenOneRec** is an open-source framework designed to bridge the gap between traditional recommendation systems and Large Language Models (LLMs). While Generative Recommendation has shown promise, existing models often struggle with isolated data silos and a lack of reasoning capabilities.
26
 
27
  To address this, we introduce a unified framework that comprises:
28
+ * **RecIF-Bench**: The first holistic Recommendation Instruction-Following Benchmark, containing **100M interactions** from 200k users across heterogeneous domains (Short Video, Ads, Product).
29
+ * **OneRec-Foundation Models**: A family of models (1.7B & 8B) built on the Qwen backbone. These models are trained on hundreds of billions of tokens, integrating collaborative signals with general semantics.
30
  * **Full-Stack Pipeline**: We open-source our comprehensive training pipeline, including data processing, co-pretraining, and post-training, to ensure full reproducibility and facilitate scaling law research in recommendation.
31
 
32
  ## πŸ”₯ News
33
 
34
+ * **[2026.1.1]** πŸŽ‰ **OneRec-Foundation** models (1.7B, 8B) are now available on Hugging Face!
35
+ * **[2026.1.1]** πŸ“‘ The technical report has been released.
36
+ * **[2026.1.1]** πŸš€ **RecIF-Bench** dataset and evaluation scripts are open-sourced.
37
 
38
  ## πŸ“Š RecIF-Bench
39
 
40
+ We propose **RecIF-Bench** to rigorously assess the synergy between instruction following and domain-specific recommendation. It organizes 8 distinct tasks into a four-layer capability hierarchy:
41
 
42
  * **Layer 0: Semantic Alignment** (Item Understanding)
43
+ * **Layer 1: Fundamental Prediction** (Short Video Rec, Ad Rec, Product Rec, Label Prediction)
44
  * **Layer 2: Instruction Following** (Interactive Rec, Label-Conditional Rec)
45
  * **Layer 3: Reasoning** (Recommendation Explanation)
46
 
47
+ The benchmark aggregates data from three domains: **Short Video** (Content), **Ads** (Commercial), and **Product** (E-commerce).
48
 
49
  ## πŸ€– Model Zoo
50
 
 
52
 
53
  | Model | Backbone | Parameters | Description | Link |
54
  | :--- | :--- | :--- | :--- | :--- |
55
+ | **OneRec-Open-1.7B** | Qwen3-1.7B | 1.7B | Standard version trained on open-source data (~33B tokens) | [HuggingFace](https://huggingface.co/OpenOneRec/OneRec-1.7B) |
56
+ | **OneRec-Open-8B** | Qwen3-8B | 8B | Standard version trained on open-source data (~33B tokens) | [HuggingFace](https://huggingface.co/OpenOneRec/OneRec-8B) |
57
+ | **OneRec-Pro-1.7B** | Qwen3-1.7B | 1.7B | Scaled-up version with expanded datasets (~130B tokens) | [HuggingFace](https://huggingface.co/OpenOneRec/OneRec-1.7B-pro) |
58
+ | **OneRec-Pro-8B** | Qwen3-8B | 8B | Scaled-up version with expanded datasets (~130B tokens) | [HuggingFace](https://huggingface.co/OpenOneRec/OneRec-8B-pro) |
59
 
60
  ## πŸ—οΈ Method & Architecture
61
 
 
66
 
67
  ### 2. Training Pipeline
68
  Our framework utilizes the following recipe:
69
+ * **Pre-Training**: Integrates collaborative signals via Itemic-Text Alignment and Full-Parameter Co-Pretraining.
70
  * **Post-Training**:
71
+ * *Stage 1*: Multi-task Supervised Fine-tuning for basic instruction following.
72
+ * *Stage 2*: On-policy Distillation to restore general reasoning performance.
73
+ * *Stage 3*: Reinforcement Learning to enhance recommendation capabilities.
74
 
75
  <div align="center">
76
+ <img src="assets/framework.png" width="80%" alt="OpenOneRec Overall Framework" />
77
  <br>
78
  <em>Figure: The Overall Framework of OpenOneRec.</em>
79
  </div>
 
83
  ### Results on RecIF-Bench
84
  OpenOneRec-Foundation achieves **State-of-the-Art (SOTA)** results across RecIF-Bench tasks, significantly outperforming baselines like LC-Rec and TIGER.
85
 
86
+ | Task | Metric | SASRec | TIGER | LC-Rec | OneRec-1.7B | OneRec-8B | OneRec-1.7B-Pro | **OneRec-8B-Pro** |
87
+ | :--- | :--- | :--- | :--- | :--- | :--- | :--- | :--- | :--- |
88
+ | **Short Video Rec** | Recall@32 | 0.0119 | 0.0132 | 0.0180 | 0.0272 | 0.0355 | 0.0274 | **0.0369** |
89
+ | **Ad Rec** | Recall@32 | 0.0293 | 0.0581 | 0.0723 | 0.0707 | 0.0877 | 0.0735 | **0.0964** |
90
+ | **Product Rec** | Recall@32 | 0.0175 | 0.0283 | 0.0416 | 0.0360 | 0.0470 | 0.0405 | **0.0538** |
91
+ | **Label-Cond. Rec** | Recall@32 | 0.0140 | 0.0123 | 0.0170 | 0.0184 | 0.0228 | 0.0182 | **0.0235** |
92
+ | **Label Pred.** | AUC | 0.6244 | 0.6675 | 0.6139 | 0.6184 | 0.6615 | 0.6071 | **0.6912** |
93
+ | **Interactive Rec** | Recall@32 | -- | -- | 0.2394 | 0.1941 | 0.3032 | 0.2024 | **0.3458** |
94
+ | **Item Und.** | LLM Score | -- | -- | 0.2517 | 0.3175 | 0.3202 | 0.3133 | **0.3209** |
95
+ | **Rec. Explanation** | LLM Score | -- | -- | 3.9350 | 3.3540 | 3.6774 | 3.5060 | **4.0381** |
 
96
 
97
  ### Cross-Domain Transferability
98
  On the **Amazon Benchmark** (10 datasets), OpenOneRec demonstrates exceptional zero-shot/few-shot transfer capabilities, achieving an average **26.8% improvement** in Recall@10 over the second-best method.
99
 
100
+ | Domain | SASRec | TIGER | LC-Rec | **Ours** |
101
+ | :--- | :--- | :--- | :--- | :--- |
102
+ | Baby | 0.0381 | 0.0318 | 0.0344 | **0.0513** |
103
+ | Beauty | 0.0639 | 0.0628 | 0.0764 | **0.0924** |
104
+ | Cell Phones | 0.0782 | 0.0786 | 0.0883 | **0.1036** |
105
+ | Grocery | 0.0789 | 0.0691 | 0.0790 | **0.1029** |
106
+ | Health | 0.0506 | 0.0534 | 0.0616 | **0.0768** |
107
+ | Home | 0.0212 | 0.0216 | 0.0293 | **0.0390** |
108
+ | Pet Supplies | 0.0607 | 0.0542 | 0.0612 | **0.0834** |
109
+ | Sports | 0.0389 | 0.0331 | 0.0418 | **0.0547** |
110
+ | Tools | 0.0437 | 0.0344 | 0.0438 | **0.0593** |
111
+ | Toys | 0.0658 | 0.0527 | 0.0549 | **0.0953** |
112
+
113
+ *Metric: Recall@10. Ours refers to OneRec-Foundation with text-augmented itemic tokens strategy.*
114
+
115
  ## πŸš€ Quick Start
116
 
117
  *Code release and detailed usage instructions are coming soon.*
 
120
 
121
  ```python
122
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
123
  model_id = "Onerec/OneRec-Open-8B"
124
  tokenizer = AutoTokenizer.from_pretrained(model_id)
125
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")
 
126
  # Example inference code will be updated here
127
  ```
128
 
 
138
  }
139
  ```
140
  ## πŸ›‘οΈ License
141
+ The code in this repository is licensed under the Apache 2.0 License. The model weights are subject to their specific license agreements.
assets/framework.png CHANGED

Git LFS Details

  • SHA256: 16163c1adbd5afd029125f116cc90f042a200f5301a674131b10e0603d97b393
  • Pointer size: 131 Bytes
  • Size of remote file: 207 kB

Git LFS Details

  • SHA256: 428b5244d4e0d8049c2e7991995b525772334dafb9e7187f9442969b8a8b858b
  • Pointer size: 131 Bytes
  • Size of remote file: 572 kB