File size: 3,108 Bytes
dee4311
 
 
 
 
 
 
c88fd2e
 
dee4311
c88fd2e
3a8e4c6
2997a7d
49dae23
c88fd2e
dee4311
 
2997a7d
dee4311
c88fd2e
 
 
 
 
 
dee4311
c88fd2e
 
 
dee4311
 
c88fd2e
 
 
 
49dae23
 
 
 
 
 
 
c88fd2e
 
49dae23
 
 
c88fd2e
 
 
 
dc3993a
dee4311
 
 
c88fd2e
 
 
 
 
 
 
 
dee4311
 
 
 
 
689fbc9
dee4311
 
 
 
 
c88fd2e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
---
pipeline_tag: other
tags:
- recommendation
- generative-recommender
---

# MemGen Checkpoints

This repository contains model checkpoints and artifacts for the paper **How Well Does Generative Recommendation Generalize?**.

<a href="https://huggingface.co/papers/2603.19809"><img src="https://img.shields.io/badge/Paper-ArXiv-red"></a>
<a href="https://github.com/Jamesding000/MemGen-GR"><img src="https://img.shields.io/badge/Code-GitHub-green"></a>
<a href="https://huggingface.co/datasets/jamesding0302/memgen-annotations"><img src="https://img.shields.io/badge/Data-Hugging%20Face-yellow"></a>

## Overview

A widely held hypothesis for why generative recommendation (GR) models outperform conventional item ID-based models is that they generalize better. This work provides a systematic way to verify this hypothesis by categorizing data instances into those requiring **memorization** (reusing item transition patterns observed during training) and those requiring **generalization** (composing known patterns to predict unseen item transitions). Our study shows that GR models perform better on instances that require generalization, whereas item ID-based models perform better when memorization is more important. 

## Folder Structure

- `SASRec/`: SASRec checkpoints (`.pth`)
- `TIGER/`: TIGER checkpoints (`.pth`)
- `semantic_ids/`: semantic ID files (`.sem_ids`) used to train the corresponding TIGER checkpoints

All checkpoints are trained with the **default configs** in the code repo. If you are unsure about model size / architecture, please refer to the configs in the [GitHub repository](https://github.com/Jamesding000/MemGen-GR).

## Download (example)

You can use the `huggingface_hub` library to download specific checkpoints and semantic ID files:

```python
from huggingface_hub import hf_hub_download

repo_id = "jamesding0302/memgen-checkpoints"
out_dir = "./downloads"   # change this to desired output path

ckpt_path = hf_hub_download(
    repo_id=repo_id,
    filename="TIGER/TIGER-AmazonReviews2014-category_Sports_and_Outdoors.pth",
    local_dir=out_dir,
)

sem_ids_path = hf_hub_download(
    repo_id=repo_id,
    filename="semantic_ids/AmazonReviews2014-Sports_and_Outdoors_sentence-t5-base_256,256,256,256.sem_ids",
    local_dir=out_dir,
)
```

## Evaluation

You can run fine-grained evaluation using the saved `checkpoint_path` and `sem_ids_path` with the scripts provided in the official repository:

```bash
CUDA_VISIBLE_DEVICES=0 python mem_gen_evaluation.py \
    --model=TIGER \
    --dataset=AmazonReviews2014 \
    --category=Sports_and_Outdoors \
    --checkpoint_path=path/to/TIGER.pth \
    --sem_ids_path=path/to/semantic_ids.sem_ids \
    --eval=test \
    --save_inference
```

## Citation

```bibtex
@article{ding2026generalize,
  title={How Well Does Generative Recommendation Generalize?},
  author={Ding, Yijie and Guo, Zitian and Li, Jiacheng and Peng, Letian and Shao, Shuai and Shao, Wei and Luo, Xiaoqiang and Simon, Luke and Shang, Jingbo and McAuley, Julian and Hou, Yupeng},
  journal={arXiv preprint arXiv:2603.19809},
  year={2026}
}
```