v8_stage1_json_csv-merged

Model Description

This model is Stage 1 of the Sequential Format Learning (v8 strategy) for structured data output.

Training Strategy

Based on Person U's approach that achieved 0.84 on the leaderboard:

  • Train one format at a time
  • Merge LoRA to base model after each stage
  • Use merged model as the base for the next stage

Stage 1 Focus: JSON & CSV

  • Formats: JSON (400 samples) + CSV (400 samples) = 800 total
  • Goal: 100% parse success rate for JSON and CSV
  • Base Model: Qwen/Qwen3-4B-Instruct-2507

Training Parameters

  • MAX_SEQ_LEN: 1024
  • EPOCHS: 2
  • Learning Rate: 5e-05
  • LoRA R: 64, Alpha: 128

Sequential Format Learning Pipeline

Stage 1: JSON/CSV (800) โ† This model
    โ†“
Stage 2: YAML (500)
    โ†“
Stage 3: XML (500)
    โ†“
Stage 4: Mixed/TOML (1000)
    โ†“
Final Model โ†’ LB 0.8+

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("kmd2525/v8_stage1_json_csv-merged")
tokenizer = AutoTokenizer.from_pretrained("kmd2525/v8_stage1_json_csv-merged")

Next Stage

Use this model as the base for Stage 2 (YAML training):

os.environ["SFT_BASE_MODEL"] = "kmd2525/v8_stage1_json_csv-merged"
Downloads last month
54
Safetensors
Model size
4B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for kmd2525/v8_stage1_json_csv-merged

Finetuned
(1153)
this model
Finetunes
1 model