gated-model / README.md
gamer-to's picture
Create README.md
ac42ddd verified
|
raw
history blame
1.25 kB

Chat Model This is a custom chat model fine-tuned for conversational AI. The model is based on LLaMA architecture and is specifically designed for Arabic and English conversations.

Model Details Architecture: LLaMA Task: Text Generation Language: Arabic/English License: MIT Model Size: Large Training Data: Custom conversational data Optimization: Quantized (int8) Usage from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("gamer-to/chat-model") tokenizer = AutoTokenizer.from_pretrained("gamer-to/chat-model")

Example input

input_text = "مرحبا كيف حالك؟" inputs = tokenizer(input_text, return_tensors="pt") outputs = model.generate( **inputs, max_length=256, temperature=0.7, do_sample=True, top_p=0.95 ) response = tokenizer.decode(outputs[0], skip_special_tokens=True)

Inference API This model is compatible with Hugging Face's Inference API. You can use it with the following endpoint:

POST https://api-inference.huggingface.co/models/gamer-to/chat-model

Model Performance Optimized for conversational tasks Supports both Arabic and English Fast response times High-quality responses Requirements PyTorch Transformers CUDA (optional for GPU acceleration)