llama-3.1-8b-emvlogs-finetuned

This model is a fine-tuned version of pravindr/llama-3.1-8b-emvlogs-finetune-dataset4 on the RahulYevle/sample_training_dataset_7 dataset.

Model Details

  • Base Model: pravindr/llama-3.1-8b-emvlogs-finetune-dataset4
  • Training Dataset: RahulYevle/sample_training_dataset_7
  • Fine-tuning Method: LoRA (Low-Rank Adaptation)
  • Quantization: 4-bit (NF4)
  • LoRA Rank: 16
  • LoRA Alpha: 32
  • Training Epochs: 2
  • Learning Rate: 0.0002

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
import torch

# Load base model
base_model = AutoModelForCausalLM.from_pretrained(
    "pravindr/llama-3.1-8b-emvlogs-finetune-dataset4",
    torch_dtype=torch.float16,
    device_map="auto"
)

# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained("pravindr/llama-3.1-8b-emvlogs-finetune-dataset4")

# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "pravindr/llama-3.1-8b-emvlogs-finetuned")

# Generate text
prompt = "Your prompt here..."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=100)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

Training Details

This model was fine-tuned using parameter-efficient fine-tuning (PEFT) with LoRA adapters.

License

This model inherits the license from the base model. Please check the base model's license for usage terms.

Downloads last month
2
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 1 Ask for provider support

Model tree for pravindr/llama-3.1-8b-emvlogs-finetuned

Adapter
(1)
this model