--- language: - en - hi - sa license: llama3.2 library_name: transformers tags: - ayurveda - medical - biology - llama-3.2 - text-generation base_model: meta-llama/Llama-3.2-3B-Instruct pipeline_tag: text-generation model-index: - name: VaidhLLaMA-3.2-3B-Instruct results: - task: type: text-generation name: Text Generation dataset: name: BhashaBench-Ayur type: evaluation-suite metrics: - name: Accuracy (Zero-Shot) type: accuracy value: 41.91 verified: false --- # VaidhLLaMA-3.2-3B-Instruct **VaidhLLaMA-3.2-3B-Instruct** is a specialized Large Language Model fine-tuned for the domain of **Ayurveda**. It is built upon the Llama-3.2-3B-Instruct architecture and has been optimized to understand and reason with Ayurvedic concepts, physiology (*Sharir Kriya*), and clinical applications. ## Model Details * **Model Name:** VaidhLLaMA-3.2-3B-Instruct * **Base Model:** [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) * **Developed By:** Vivekdas * **Language:** English, Hindi, Sanskrit (Domain-specific terminology) * **License:** Llama 3.2 Community License * **Architecture:** Transformer-based Auto-Regressive Language Model ## Performance VaidhLLaMA demonstrates strong performance on the **BhashaBench-Ayur** benchmark, outperforming its base model and other similarly sized models in domain-specific tasks. | Model | Accuracy (%) | Note | | :--- | :--- | :--- | | **VaidhLLaMA-3.2-3B** | **41.91%** | **Fine-tuned Ayurveda Specialist** | | Llama-3.2-3B-Instruct | 40.74% | Base Model | | Llama-3.2-1B | 27.58% | Tiny Model | ## Intended Use This model is designed for: * Answering questions related to Ayurvedic medical science. * Explaining concepts from classical Ayurvedic texts (*Samhitas*). * Assisting researchers and students in the field of Ayurveda. **Disclaimer:** This model is for **educational and research purposes only**. It should not be used as a substitute for professional medical advice, diagnosis, or treatment. ## Usage You can run this model using the `transformers` library: ```python import torch from transformers import AutoModelForCausalLM, AutoTokenizer model_id = "Vivekdas/VaidhLLaMA-3.2-3B-Instruct" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained( model_id, torch_dtype=torch.bfloat16, device_map="auto" ) messages = [ {"role": "system", "content": "You are VaidhLLaMA, an expert AI assistant for Ayurveda."}, {"role": "user", "content": "Explain the concept of Tridosha in Ayurveda."} ] input_ids = tokenizer.apply_chat_template( messages, add_generation_prompt=True, return_tensors="pt" ).to(model.device) outputs = model.generate( input_ids, max_new_tokens=512, do_sample=True, temperature=0.6, top_p=0.9 ) response = tokenizer.decode(outputs[0][input_ids.shape[-1]:], skip_special_tokens=True) print(response) ``` ## Citation If you use this model in your research, please cite: ```bibtex @misc{vaidhllama2024, author = {Vivekdas}, title = {VaidhLLaMA: A Fine-Tuned LLM for Ayurveda}, year = {2024}, publisher = {Hugging Face}, journal = {Hugging Face Repository}, howpublished = {\url{https://huggingface.co/Vivekdas/VaidhLLaMA-3.2-3B-Instruct}} } ```