Instructions to use pagand/venra with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- PEFT
How to use pagand/venra with PEFT:
Task type is invalid.
- Notebooks
- Google Colab
- Kaggle
VeNRA β LoRA Adapter
Fine-tuned LoRA adapter on Qwen/Qwen2.5-Coder-3B-Instruct for
hallucination detection in RAG pipelines.
Available Adapters
| Branch | Rank | Description |
|---|---|---|
r96 |
96 | Lighter, faster inference |
r128 |
128 | Higher capacity |
Labels
Foundβ supported by contextGeneralβ common knowledgeFakeβ contradicts or unsupported by context
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch
BASE_MODEL = "Qwen/Qwen2.5-Coder-3B-Instruct"
# Load r96
model_r96 = PeftModel.from_pretrained(base, "pagand/venra", revision="r96")
# Load r128
model_r128 = PeftModel.from_pretrained(base, "pagand/venra", revision="r128")
# Pinned to a specific snapshot tag
model = PeftModel.from_pretrained(model, "pagand/venra", revision="r96-v1.0")
Training Details
- Rank: 96/128
- Learning rate: 1e-4
- Weight decay: 0.10
- Training regime: WeightedLabelTrainer
- Downloads last month
- 2
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support