Vivekdas's picture
Add Model Card (README.md)
5110c2e verified
---
language:
- en
- hi
- sa
license: llama3.2
library_name: transformers
tags:
- ayurveda
- medical
- biology
- llama-3.2
- text-generation
base_model: meta-llama/Llama-3.2-3B-Instruct
pipeline_tag: text-generation
model-index:
- name: VaidhLLaMA-3.2-3B-Instruct
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: BhashaBench-Ayur
type: evaluation-suite
metrics:
- name: Accuracy (Zero-Shot)
type: accuracy
value: 41.91
verified: false
---
# VaidhLLaMA-3.2-3B-Instruct
**VaidhLLaMA-3.2-3B-Instruct** is a specialized Large Language Model fine-tuned for the domain of **Ayurveda**. It is built upon the Llama-3.2-3B-Instruct architecture and has been optimized to understand and reason with Ayurvedic concepts, physiology (*Sharir Kriya*), and clinical applications.
## Model Details
* **Model Name:** VaidhLLaMA-3.2-3B-Instruct
* **Base Model:** [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
* **Developed By:** Vivekdas
* **Language:** English, Hindi, Sanskrit (Domain-specific terminology)
* **License:** Llama 3.2 Community License
* **Architecture:** Transformer-based Auto-Regressive Language Model
## Performance
VaidhLLaMA demonstrates strong performance on the **BhashaBench-Ayur** benchmark, outperforming its base model and other similarly sized models in domain-specific tasks.
| Model | Accuracy (%) | Note |
| :--- | :--- | :--- |
| **VaidhLLaMA-3.2-3B** | **41.91%** | **Fine-tuned Ayurveda Specialist** |
| Llama-3.2-3B-Instruct | 40.74% | Base Model |
| Llama-3.2-1B | 27.58% | Tiny Model |
## Intended Use
This model is designed for:
* Answering questions related to Ayurvedic medical science.
* Explaining concepts from classical Ayurvedic texts (*Samhitas*).
* Assisting researchers and students in the field of Ayurveda.
**Disclaimer:** This model is for **educational and research purposes only**. It should not be used as a substitute for professional medical advice, diagnosis, or treatment.
## Usage
You can run this model using the `transformers` library:
```python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "Vivekdas/VaidhLLaMA-3.2-3B-Instruct"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto"
)
messages = [
{"role": "system", "content": "You are VaidhLLaMA, an expert AI assistant for Ayurveda."},
{"role": "user", "content": "Explain the concept of Tridosha in Ayurveda."}
]
input_ids = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
outputs = model.generate(
input_ids,
max_new_tokens=512,
do_sample=True,
temperature=0.6,
top_p=0.9
)
response = tokenizer.decode(outputs[0][input_ids.shape[-1]:], skip_special_tokens=True)
print(response)
```
## Citation
If you use this model in your research, please cite:
```bibtex
@misc{vaidhllama2024,
author = {Vivekdas},
title = {VaidhLLaMA: A Fine-Tuned LLM for Ayurveda},
year = {2024},
publisher = {Hugging Face},
journal = {Hugging Face Repository},
howpublished = {\url{https://huggingface.co/Vivekdas/VaidhLLaMA-3.2-3B-Instruct}}
}
```