v8_stage2_yaml-merged
Model Description
This model is Stage 2 of the Sequential Format Learning (v8 strategy) for structured data output.
Training Strategy
Based on Person U's approach that achieved 0.84 on the leaderboard:
- Train one format at a time
- Merge LoRA to base model after each stage
- Use merged model as the base for the next stage
Stage 2 Focus: YAML
- Format: YAML (500 samples, prioritizing depth >= 3)
- Goal: 100% parse success rate for YAML while maintaining JSON/CSV performance
- Base Model:
kmd2525/v8_stage1_json_csv-merged(Stage 1 merged model)
Previous Stage
- Stage 1: JSON/CSV (800 samples) β JSON 100%, CSV 100%
Training Parameters
- MAX_SEQ_LEN: 1024
- EPOCHS: 2
- Learning Rate: 3e-05
- LoRA R: 64, Alpha: 128
Sequential Format Learning Pipeline
Stage 1: JSON/CSV (800) β
β
Stage 2: YAML (500) β This model
β
Stage 3: XML (500)
β
Stage 4: Mixed/TOML (1000)
β
Final Model β LB 0.8+
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("kmd2525/v8_stage2_yaml-merged")
tokenizer = AutoTokenizer.from_pretrained("kmd2525/v8_stage2_yaml-merged")
Next Stage
Use this model as the base for Stage 3 (XML training):
os.environ["SFT_BASE_MODEL"] = "kmd2525/v8_stage2_yaml-merged"
- Downloads last month
- 33
Inference Providers NEW
This model isn't deployed by any Inference Provider. π Ask for provider support