--- language: - ar license: apache-2.0 base_model: Navid-AI/Yehia-7B-preview tags: - peft - qlora - arabic - poetry - classical-arabic-poetry - meter-conditioned-generation pipeline_tag: text-generation --- # Shaer Main SFT Adapters This repository stores the completed main SFT baseline for the Shaer classical Arabic poetry project. ## Baseline Summary - Base model: `Navid-AI/Yehia-7B-preview` - Dataset: `Shaer-AI/ashaar-with-enhanced-descriptions-baseform-final-sft-lte20-min500-splits` - Split policy: deterministic `94 / 3 / 3`, stratified by `base_meter||form||length_bucket` - Train sampler: weighted train-only sampler on `base_meter||form||length_bucket` - LoRA: `all-linear`, `r=64`, `alpha=128`, `dropout=0.05`, `use_rslora=true` - Run name: `train_20260407_231929` - Best eval checkpoint: `/root/workspace/Shaer/sft/outputs/train/train_20260407_231929/checkpoint-3000` - Best eval loss: `2.2074480056762695` - Final test loss: `2.1932055950164795` - Best probe meter mean: `0.6042156156147234` at step `2800` - Final probe meter mean: `0.5087050689648603` - Final probe count adherence mean: `1.0` ## Important Paths In This Repo - Latest adapter export: - `adapters/fresh_sft/train/latest` - Best adapter export: - `adapters/fresh_sft/train/best` - Finished run report bundle: - `reports/fresh_sft/train_20260407_231929` ## Current Comparison Context - The short meter-loss continuation in `Shaer-AI/shaer-adapters-v2` was stopped early after the auxiliary meter head improved but CE and probe meter drifted worse than this baseline. - A later fresh-from-start `v3` run was deleted after confirming the current meter-loss path was invalid because it could read the requested meter from prompt-conditioned hidden states. - This baseline remains the strongest known safe reference until a corrected challenger clearly beats it.