|
|
--- |
|
|
base_model: |
|
|
- Qwen/Qwen3-32B |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
arxiv: 2602.11089 |
|
|
license: other |
|
|
--- |
|
|
|
|
|
# DataChef-32B |
|
|
|
|
|
[**HF Models**](https://huggingface.co/yichengchen24/DataChef-32B) | [**HF Demo**](https://huggingface.co/spaces/yichengchen24/DataChef) | [**Paper**](https://arxiv.org/abs/2602.11089) | [**GitHub**](https://github.com/yichengchen24/DataChef) |
|
|
|
|
|
DataChef-32B is a specialized large language model designed for **automated data recipe generation**. It was introduced in the paper [DataChef: Cooking Up Optimal Data Recipes for LLM Adaptation via Reinforcement Learning](https://huggingface.co/papers/2602.11089). |
|
|
|
|
|
DataChef-32B facilitates LLM adaptation by generating executable data processing pipelines (data recipes) that transform raw data sources into high-quality training corpora targeted at specific benchmarks. |
|
|
|
|
|
## Model Description |
|
|
DataChef-32B addresses the manual, labor-intensive process of designing data processing pipelines. It was trained using online reinforcement learning with a proxy reward system that predicts downstream performance for candidate recipes. Given a target benchmark and available data sources, the model outputs a complete data recipe to adapt a base LLM. |
|
|
|
|
|
### Performance Highlights |
|
|
Across diverse tasks, DataChef-32B produces practical recipes that reach performance comparable to those curated by human experts. Notably, a recipe generated by DataChef-32B was used to adapt Qwen3-1.7B-Base to the math domain, achieving a score of **66.7 on AIME'25**, surpassing the performance of the standard Qwen3-1.7B. |
|
|
|
|
|
## Installation |
|
|
To use the DataChef framework for generating your own data recipes, follow the installation steps from the [GitHub repository](https://github.com/yichengchen24/DataChef): |
|
|
|
|
|
```bash |
|
|
conda create -n datachef python=3.12 |
|
|
conda activate datachef |
|
|
pip install -e . |
|
|
``` |
|
|
|
|
|
## Citation |
|
|
If you find this work helpful, please consider citing: |
|
|
|
|
|
```bibtex |
|
|
@article{chen2026datachef, |
|
|
title={DataChef: Cooking Up Optimal Data Recipes for LLM Adaptation via Reinforcement Learning}, |
|
|
author={Chen, Yicheng and Ma, Zerun and Xie, Xinchen and Li, Yining and Chen, Kai}, |
|
|
journal={arXiv preprint arXiv:2602.11089}, |
|
|
year={2026} |
|
|
} |
|
|
``` |