Add metadata and improve model card
#1
by nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,21 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
# MemGen Checkpoints
|
| 2 |
|
| 3 |
-
|
| 4 |
|
| 5 |
<a href="https://huggingface.co/papers/2603.19809"><img src="https://img.shields.io/badge/Paper-ArXiv-red"></a>
|
| 6 |
<a href="https://github.com/Jamesding000/MemGen-GR"><img src="https://img.shields.io/badge/Code-GitHub-black"></a>
|
| 7 |
<a href="https://huggingface.co/datasets/jamesding0302/memgen-annotations"><img src="https://img.shields.io/badge/Data-Hugging%20Face-yellow"></a>
|
| 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
## Folder Structure
|
| 10 |
|
| 11 |
- `SASRec/`: SASRec checkpoints (`.pth`)
|
| 12 |
- `TIGER/`: TIGER checkpoints (`.pth`)
|
| 13 |
- `semantic_ids/`: semantic ID files (`.sem_ids`) used to train the corresponding TIGER checkpoints
|
| 14 |
|
| 15 |
-
All checkpoints are trained with the **default configs** in the code repo. If you are unsure about model size / architecture, please refer to the configs in [
|
| 16 |
|
| 17 |
## Download (example)
|
| 18 |
|
|
|
|
|
|
|
| 19 |
```python
|
| 20 |
from huggingface_hub import hf_hub_download
|
| 21 |
|
|
@@ -37,8 +52,9 @@ sem_ids_path = hf_hub_download(
|
|
| 37 |
|
| 38 |
## Evaluation
|
| 39 |
|
| 40 |
-
You can run fine-grained evaluation using the saved `checkpoint_path` and `sem_ids_path`:
|
| 41 |
-
|
|
|
|
| 42 |
CUDA_VISIBLE_DEVICES=0 python mem_gen_evaluation.py \
|
| 43 |
--model=TIGER \
|
| 44 |
--dataset=AmazonReviews2014 \
|
|
@@ -47,4 +63,15 @@ CUDA_VISIBLE_DEVICES=0 python mem_gen_evaluation.py \
|
|
| 47 |
--sem_ids_path=path/to/semantic_ids.sem_ids \
|
| 48 |
--eval=test \
|
| 49 |
--save_inference
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
```
|
|
|
|
| 1 |
+
---
|
| 2 |
+
pipeline_tag: other
|
| 3 |
+
tags:
|
| 4 |
+
- recommendation
|
| 5 |
+
- generative-recommender
|
| 6 |
+
---
|
| 7 |
+
|
| 8 |
# MemGen Checkpoints
|
| 9 |
|
| 10 |
+
This repository contains model checkpoints and artifacts for the paper **How Well Does Generative Recommendation Generalize?**.
|
| 11 |
|
| 12 |
<a href="https://huggingface.co/papers/2603.19809"><img src="https://img.shields.io/badge/Paper-ArXiv-red"></a>
|
| 13 |
<a href="https://github.com/Jamesding000/MemGen-GR"><img src="https://img.shields.io/badge/Code-GitHub-black"></a>
|
| 14 |
<a href="https://huggingface.co/datasets/jamesding0302/memgen-annotations"><img src="https://img.shields.io/badge/Data-Hugging%20Face-yellow"></a>
|
| 15 |
|
| 16 |
+
## Overview
|
| 17 |
+
|
| 18 |
+
A widely held hypothesis for why generative recommendation (GR) models outperform conventional item ID-based models is that they generalize better. This work provides a systematic way to verify this hypothesis by categorizing data instances into those requiring **memorization** (reusing item transition patterns observed during training) and those requiring **generalization** (composing known patterns to predict unseen item transitions).
|
| 19 |
+
|
| 20 |
+
The study shows that GR models perform better on instances that require generalization, whereas item ID-based models perform better when memorization is more important. The authors propose a simple memorization-aware indicator that adaptively combines both paradigms to improve overall performance.
|
| 21 |
+
|
| 22 |
## Folder Structure
|
| 23 |
|
| 24 |
- `SASRec/`: SASRec checkpoints (`.pth`)
|
| 25 |
- `TIGER/`: TIGER checkpoints (`.pth`)
|
| 26 |
- `semantic_ids/`: semantic ID files (`.sem_ids`) used to train the corresponding TIGER checkpoints
|
| 27 |
|
| 28 |
+
All checkpoints are trained with the **default configs** in the code repo. If you are unsure about model size / architecture, please refer to the configs in the [GitHub repository](https://github.com/Jamesding000/MemGen-GR).
|
| 29 |
|
| 30 |
## Download (example)
|
| 31 |
|
| 32 |
+
You can use the `huggingface_hub` library to download specific checkpoints and semantic ID files:
|
| 33 |
+
|
| 34 |
```python
|
| 35 |
from huggingface_hub import hf_hub_download
|
| 36 |
|
|
|
|
| 52 |
|
| 53 |
## Evaluation
|
| 54 |
|
| 55 |
+
You can run fine-grained evaluation using the saved `checkpoint_path` and `sem_ids_path` with the scripts provided in the official repository:
|
| 56 |
+
|
| 57 |
+
```bash
|
| 58 |
CUDA_VISIBLE_DEVICES=0 python mem_gen_evaluation.py \
|
| 59 |
--model=TIGER \
|
| 60 |
--dataset=AmazonReviews2014 \
|
|
|
|
| 63 |
--sem_ids_path=path/to/semantic_ids.sem_ids \
|
| 64 |
--eval=test \
|
| 65 |
--save_inference
|
| 66 |
+
```
|
| 67 |
+
|
| 68 |
+
## Citation
|
| 69 |
+
|
| 70 |
+
```bibtex
|
| 71 |
+
@article{ding2026how,
|
| 72 |
+
title={How Well Does Generative Recommendation Generalize?},
|
| 73 |
+
author={Ding, Yijie and Guo, Zitian and Li, Jiacheng and Peng, Letian and Shao, Shuai and Shao, Wei and Luo, Xiaoqiang and Simon, Luke and Shang, Jingbo and McAuley, Julian and Hou, Yupeng},
|
| 74 |
+
journal={arXiv preprint arXiv:2603.19809},
|
| 75 |
+
year={2026}
|
| 76 |
+
}
|
| 77 |
```
|