Xiaofeng77 nielsr HF Staff commited on
Commit
e2e11ae
·
verified ·
1 Parent(s): 7a5f767

Improve dataset card: Add paper/code links, task categories, and usage (#2)

Browse files

- Improve dataset card: Add paper/code links, task categories, and usage (68be7676e0712d0526cd0561106991c56e21a4f1)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +109 -0
README.md CHANGED
@@ -1,4 +1,18 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  dataset_info:
3
  features:
4
  - name: data_source
@@ -47,3 +61,98 @@ configs:
47
  - split: test_env
48
  path: data/test_env-*
49
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ task_categories:
3
+ - TEXT_GENERATION
4
+ language:
5
+ - en
6
+ tags:
7
+ - supervised-fine-tuning
8
+ - reinforcement-learning
9
+ - sokoban
10
+ - general-points
11
+ - chain-of-thought
12
+ - instruction-following
13
+ - reasoning
14
+ - decision-making
15
+ - llm
16
  dataset_info:
17
  features:
18
  - name: data_source
 
61
  - split: test_env
62
  path: data/test_env-*
63
  ---
64
+
65
+ # Debunk the Myth of SFT Generalization Dataset
66
+
67
+ This dataset is associated with the paper [Debunk the Myth of SFT Generalization](https://huggingface.co/papers/2510.00237).
68
+
69
+ The research presented in the paper challenges the common belief that supervised fine-tuning (SFT) primarily memorizes training data and lacks generalization, especially when compared to reinforcement learning (RL). Through systematic evaluation on decision-making benchmarks like Sokoban and General Points, the authors demonstrate that incorporating prompt diversity and Chain-of-Thought (CoT) supervision during SFT training can lead to strong generalization across unseen instruction variants and strictly harder tasks, often matching or surpassing RL baselines.
70
+
71
+ **Code:** [https://github.com/XiaofengLin7/debunking-sft-generalization](https://github.com/XiaofengLin7/debunking-sft-generalization)
72
+
73
+ ## Installation
74
+
75
+ This dataset is intended for use with the associated code repository. To set up the environment and dependencies:
76
+
77
+ ### Prerequisites
78
+ CUDA 12.2 & cuDNN 9.1.0 works, but [official docs](https://verl.readthedocs.io/en/latest/start/install.html) recommends CUDA >= 12.4 & cuDNN >= 9.8.0.
79
+
80
+ ### Setup
81
+
82
+ ```bash
83
+ conda create -n debunk_sft python=3.10
84
+ conda activate debunk_sft
85
+ USE_MEGATRON=0 bash setup.sh
86
+ git submodule init
87
+ git submodule update
88
+ pip install -e thirdparty/verl --no-dependencies
89
+ pip install -e thirdparty/ragen --no-dependencies
90
+ pip install -e thirdparty/alfworld --no-dependencies
91
+ pip install -e thirdparty/trl --no-dependecies
92
+ ```
93
+
94
+ ## Dataset Overview
95
+
96
+ This repository contains various datasets used in the research, categorized by task, method, diversity, and format. These datasets are part of a larger collection.
97
+
98
+ | Task | Method | Diversity | Format | Link |
99
+ | --- | --- | --- | --- | --- |
100
+ | Sokoban | RL | non-diverse | — | [🤗](https://huggingface.co/datasets/Xiaofeng77/sokoban) |
101
+ | Sokoban | RL | diverse | — | [🤗](https://huggingface.co/datasets/Xiaofeng77/diverse_sokoban) |
102
+ | Sokoban | SFT | non-diverse | answer-only | [🤗](https://huggingface.co/datasets/Xiaofeng77/answer-only-sokoban) |
103
+ | Sokoban | SFT | diverse | answer-only | [🤗](https://huggingface.co/datasets/Xiaofeng77/diverse-answer-only-sokoban) |
104
+ | Sokoban | SFT | non-diverse | cot | [🤗](https://huggingface.co/datasets/Xiaofeng77/cot-sokoban) |
105
+ | Sokoban | SFT | diverse | cot | [🤗](https://huggingface.co/datasets/Xiaofeng77/diverse-cot-sokoban) |
106
+ | General Points | RL | non-diverse | — | [🤗](https://huggingface.co/datasets/Xiaofeng77/gp-l-only-10k) |
107
+ | General Points | RL | diverse | — | [🤗](https://huggingface.co/datasets/Xiaofeng77/diverse-gp-l-only-10k) |
108
+ | General Points | SFT | non-diverse | answer-only | [🤗](https://huggingface.co/datasets/Xiaofeng77/answer-only-gp-l-only-10k) |
109
+ | General Points | SFT | diverse | answer-only | [🤗](https://huggingface.co/datasets/Xiaofeng77/diverse-answer-only-gp-l-only-10k) |
110
+ | General Points | SFT | non-diverse | cot | [🤗](https://huggingface.co/datasets/Xiaofeng77/cot-gp-l-only-10k) |
111
+ | General Points | SFT | diverse | cot | [🤗](https://huggingface.co/datasets/Xiaofeng77/diverse-cot-gp-l-only-10k) |
112
+
113
+ ## Sample Usage
114
+
115
+ The following snippets from the GitHub repository demonstrate how to train models using this dataset with SFT (Supervised Fine-Tuning) or GRPO (Generative Reinforcement Policy Optimization).
116
+
117
+ ### Train your model with SFT
118
+ Specify your model and data beforehand.
119
+
120
+ For Sokoban:
121
+ ```bash
122
+ bash debunk_sft/scripts/sokoban/sokoban_train_and_eval.sh
123
+ ```
124
+
125
+ For General Points:
126
+ ```bash
127
+ bash debunk_sft/scripts/gp_l/gp_l_train_and_eval.sh
128
+ ```
129
+
130
+ ### Train your model with GRPO
131
+
132
+ Specify your model and data beforehand.
133
+
134
+ For Sokoban:
135
+ ```bash
136
+ bash debunk_sft/scripts/sokoban/sokoban_grpo.sh
137
+ ```
138
+
139
+ For General Points:
140
+ ```bash
141
+ bash debunk_sft/scripts/gp_l/gp_l_grpo.sh
142
+ ```
143
+
144
+ ## Citation
145
+
146
+ If you use this dataset in your research, please cite the associated paper:
147
+
148
+ ```bibtex
149
+ @misc{lin2024debunkthemythofsftgeneralization,
150
+ title={Debunk the Myth of SFT Generalization},
151
+ author={Xiaofeng Lin and Yuandong Tian and Huazhe Xu},
152
+ year={2024},
153
+ eprint={2510.00237},
154
+ archivePrefix={arXiv},
155
+ primaryClass={cs.LG},
156
+ url={https://arxiv.org/abs/2510.00237},
157
+ }
158
+ ```