batch-vllm / README.md
tytodd's picture
Update README.md
2e5a0d1 verified
---
license: mit
---
This uv-script allows you to run batch inference on vllm over an hf dataset as long as it has a messages column. It's based on the script [https://huggingface.co/datasets/uv-scripts/vllm/raw/main/generate-responses.py](https://huggingface.co/datasets/uv-scripts/vllm/raw/main/generate-responses.py)
the only diference is that it uses `llm.chat()` instead of `llm.generate()` so the response format is more familar to the openai response format and easier to use.
## Launch Job via SDK
```python
#!/usr/bin/env python3
from dotenv import load_dotenv
from huggingface_hub import HfApi
load_dotenv()
import os
DATASET_REPO_ID = "tytodd/test-job-dataset"
SCRIPT_URL = "https://huggingface.co/datasets/modaic/batch-vllm/raw/main/generate_responses.py"
def main() -> None:
api = HfApi()
job_info = api.run_uv_job(
SCRIPT_URL,
script_args=[
DATASET_REPO_ID,
DATASET_REPO_ID,
"--model-id",
# "Qwen/Qwen3-235B-A22B-Instruct-2507",
"deepseek-ai/DeepSeek-V3.2",
# "zai-org/GLM-5", # transformers > 5
# "moonshotai/Kimi-K2.5",
"--messages-column",
"messages",
],
dependencies=["transformers<5"],
image="vllm/vllm-openai:latest",
flavor="h200x4",
secrets={"HF_TOKEN": os.getenv("HF_TOKEN")},
)
print(f"Created job {job_info.id}")
print(job_info.url)
if __name__ == "__main__":
main()
```
## Launch Job via CLI
```
uvx hf jobs uv run \
--flavor l4x4 \
--secrets HF_TOKEN \
https://huggingface.co/datasets/modaic/batch-vllm/resolve/main/generate_responses.py \
username/input-dataset \
username/output-dataset \
--messages-column messages \
--model-id Qwen/Qwen3-30B-A3B-Instruct-2507 \
--temperature 0.7 \
--max-tokens 16384
```