gabtan99/pex-conversations
Viewer โข Updated โข 8 โข 146 โข 2
How to use gabtan99/dialogpt-tagalog-medium with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("text-generation", model="gabtan99/dialogpt-tagalog-medium")
messages = [
{"role": "user", "content": "Who are you?"},
]
pipe(messages) # Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("gabtan99/dialogpt-tagalog-medium")
model = AutoModelForCausalLM.from_pretrained("gabtan99/dialogpt-tagalog-medium")How to use gabtan99/dialogpt-tagalog-medium with vLLM:
# Install vLLM from pip:
pip install vllm
# Start the vLLM server:
vllm serve "gabtan99/dialogpt-tagalog-medium"
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:8000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "gabtan99/dialogpt-tagalog-medium",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'docker model run hf.co/gabtan99/dialogpt-tagalog-medium
How to use gabtan99/dialogpt-tagalog-medium with SGLang:
# Install SGLang from pip:
pip install sglang
# Start the SGLang server:
python3 -m sglang.launch_server \
--model-path "gabtan99/dialogpt-tagalog-medium" \
--host 0.0.0.0 \
--port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "gabtan99/dialogpt-tagalog-medium",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'docker run --gpus all \
--shm-size 32g \
-p 30000:30000 \
-v ~/.cache/huggingface:/root/.cache/huggingface \
--env "HF_TOKEN=<secret>" \
--ipc=host \
lmsysorg/sglang:latest \
python3 -m sglang.launch_server \
--model-path "gabtan99/dialogpt-tagalog-medium" \
--host 0.0.0.0 \
--port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "gabtan99/dialogpt-tagalog-medium",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'How to use gabtan99/dialogpt-tagalog-medium with Docker Model Runner:
docker model run hf.co/gabtan99/dialogpt-tagalog-medium
A DialoGPT-medium model fine-tuned on Tagalog conversational data scraped from the web. This model is an output of a research on RoBERTa-based data augmentation for low resource languages. This is the baseline model which did not use any synthetic data in training.
Here is an example of using beam search for model inference.
for step in range(2):
# encode the new user input, add the eos_token and return a tensor in Pytorch
new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt')
# append the new user input tokens to the chat history
bot_input_ids = torch.cat([chat_history_ids, new_user_input_ids], dim=-1) if step > 0 else new_user_input_ids
# we limit the generation to 512 tokens, each utterance in training had a maximum of 128 tokens
chat_history_ids = model.generate(
bot_input_ids, max_length=512,
pad_token_id=tokenizer.eos_token_id,
num_beams=5,
no_repeat_ngram_size=3
)
# pretty print last ouput tokens from bot
print("DialoGPT: {}".format(tokenizer.decode(chat_history_ids[:, bot_input_ids.shape[-1]:][0], skip_special_tokens=True)))
Fine-tuning script adapted from Spanish DialoGPT