|
|
--- |
|
|
language: |
|
|
- en |
|
|
license: apache-2.0 |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# R2EGym-7B-Agent-Coder-Instruct (checkpoint-800) |
|
|
|
|
|
This repository contains a training checkpoint exported from LLaMA-Factory. |
|
|
|
|
|
- Base: `Qwen/Qwen2.5-Coder-7B-Instruct` |
|
|
- Training: SFT with DeepSpeed ZeRO-3 |
|
|
- Checkpoint: `checkpoint-800` |
|
|
|
|
|
## Notes |
|
|
- This repo includes ZeRO optimizer states in `global_step800/` for resuming training. |
|
|
- For inference, use the `model-0000*-of-00004.safetensors` shards and tokenizer files. |
|
|
|