MedQuAD LoRA r=32

Configuraci贸n

  • Base: mistralai/Mistral-7B-Instruct-v0.3
  • LoRA r: 32
  • M贸dulos: q_proj, k_proj, v_proj
  • 4-bit NF4
  • Early Stopping: patience=3

Entrenamiento

Training logs (manual, Epoch estimado):

Step Epoch Training Loss Validation Loss
100 0.046 0.824200 0.792224
200 0.093 0.766400 0.763532
300 0.139 0.761900 0.749713
400 0.186 0.734200 0.737889
500 0.232 0.728500 0.727421
600 0.279 0.734900 0.719858
700 0.325 0.715400 0.706352
800 0.371 0.722500 0.695117
900 0.418 0.702000 0.687702
1000 0.464 0.686500 0.677177
1100 0.511 0.674400 0.673650
1200 0.557 0.646700 0.663053
1300 0.604 0.651100 0.662174

Uso

from peft import PeftModel
from transformers import AutoModelForCausalLM
base = AutoModelForCausalLM.from_pretrained('mistralai/Mistral-7B-Instruct-v0.3', load_in_4bit=True)
model = PeftModel.from_pretrained(base, 'CHF0101/medquad-lora-r32')
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support