Model Card, explaining new LMP and MWCS policies implemented (from Jan 18 group meeting)
92593eb
memo-ozdincercommited on
LoRA RR adapters for Llama 3.1 trained on Tool-flip only Loss Masking Policy & Low-weight Mixture Weighting & Curriculum Schedules calculated by perplexity/cross-entropy on fixed token-window at the point of prompt injection. Artifacts of run #208788. Full performance eval uploaded soon