Update README.md
Browse files
README.md
CHANGED
|
@@ -25,6 +25,14 @@ from vllm import LLM, SamplingParams
|
|
| 25 |
repo = snapshot_download(repo_id="Staticaliza/Reya-Human", allow_patterns=["*.json", "*.bin", "*.safetensors"])
|
| 26 |
llm = LLM(model=repo, dtype="auto", tensor_parallel_size=torch.cuda.device_count(), enforce_eager=True, trust_remote_code=True)
|
| 27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
params = SamplingParams(
|
| 29 |
max_tokens=256,
|
| 30 |
temperature=1,
|
|
|
|
| 25 |
repo = snapshot_download(repo_id="Staticaliza/Reya-Human", allow_patterns=["*.json", "*.bin", "*.safetensors"])
|
| 26 |
llm = LLM(model=repo, dtype="auto", tensor_parallel_size=torch.cuda.device_count(), enforce_eager=True, trust_remote_code=True)
|
| 27 |
|
| 28 |
+
# ChatML is suggested
|
| 29 |
+
input = """<|im_start|>system
|
| 30 |
+
You are Reya.<|im_end|>
|
| 31 |
+
<|im_start|>user
|
| 32 |
+
Hi.<|im_end|>
|
| 33 |
+
<|im_start|>assistant
|
| 34 |
+
"""
|
| 35 |
+
|
| 36 |
params = SamplingParams(
|
| 37 |
max_tokens=256,
|
| 38 |
temperature=1,
|