Buckets:
| # DeepSpeed Integration | |
| > [!WARNING] | |
| > Section under construction. Feel free to contribute! | |
| TRL supports training with DeepSpeed, a library that implements advanced training optimization techniques. These include optimizer state partitioning, offloading, gradient partitioning, and more. | |
| DeepSpeed integrates the [Zero Redundancy Optimizer (ZeRO)](https://huggingface.co/papers/1910.02054), which allows to scale the model size proportional to the number of devices with sustained high efficiency. | |
|  | |
| ## Installation | |
| To use DeepSpeed with TRL, install it using the following command: | |
| ```bash | |
| pip install deepspeed | |
| ``` | |
| ## Running Training Scripts with DeepSpeed | |
| No modifications to your training script are required. Simply run it with the DeepSpeed configuration file: | |
| ```bash | |
| accelerate launch --config_file train.py | |
| ``` | |
| We provide ready-to-use DeepSpeed configuration files in the [`examples/accelerate_configs`](https://github.com/huggingface/trl/tree/main/examples/accelerate_configs) directory. For example, to run training with ZeRO Stage 2, use the following command: | |
| ```bash | |
| accelerate launch --config_file examples/accelerate_configs/deepspeed_zero2.yaml train.py | |
| ``` | |
| ## Additional Resources | |
| Consult the 🤗 Accelerate [documentation](https://huggingface.co/docs/accelerate/usage_guides/deepspeed) for more information about the DeepSpeed plugin. | |
Xet Storage Details
- Size:
- 1.48 kB
- Xet hash:
- 96460e488506feeca19f1df0ddbd04d9292ac71d82dd8a7801b2a65a1b5008c6
·
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.