File size: 1,651 Bytes
017a664 f154a1a 17f897d 017a664 c7f8bf6 ce8f235 017a664 f154a1a 017a664 f154a1a 017a664 f154a1a 017a664 f154a1a 017a664 f154a1a 017a664 f154a1a 017a664 f154a1a 017a664 f154a1a 017a664 f154a1a c7f8bf6 2999d40 17f897d 2999d40 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 |
---
library_name: transformers
language:
- en
license: apache-2.0
---
## Virtual Cleint
<p align="center">
<img src="client.png" alt="client" width="400"/>
</p>
### example
```
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer
MODEL_ID = "jaeyong2/Virtual-Client-Preview"
torch_dtype = torch.bfloat16 if torch.cuda.is_available() else torch.float32
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, use_fast=True)
model = AutoModelForCausalLM.from_pretrained(
MODEL_ID,
torch_dtype=torch_dtype,
device_map="auto",
)
persona = "Pilates trainer with extensive gym experience"
messages = [
{"role": "system", "content": "You are a question-generating AI that receives personas from users and generates the most appropriate questions"},
{"role": "user", "content": persona}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(
**model_inputs,
max_new_tokens=512
)
generated_ids = [
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
```
### result
```
What are some of the key principles you follow when designing a new workout program?
```
### How to make dataset

## License
- Qwen/Qwen2.5-1.5B-Instruct : https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct/blob/main/LICENSE
## Acknowledgement
This research is supported by **TPU Research Cloud program**. |