Word Count SFT then EM training models (Qwen 32B and Seed 36B)
AI & ML interests
None defined yet.
Recent Activity
View all activity
MMLU SFT first, then EM training. Ablation: does MMLU pre-training affect emergent misalignment?
Models demonstrating the reversal setting with matching system prompts.
Models demonstrating the baseline setting i.e. picking longer summaries with non-matching system prompts.
Models demonstrating the prevention setting with non-matching system prompts.
Qwen2.5-32B and Seed-OSS-36B models finetuned on a EM dataset. Inferece instructions: https://docs.axolotl.ai/docs/inference
Qwen2.5-32B and Seed-OSS-36B models finetuned on a EM dataset. Inferece instructions: https://docs.axolotl.ai/docs/inference
Word count SFT models trained on top of EM-finetuned models to evaluate capability preservation
EM finetuned models with additional MMLU SFT training for ablation
Models demonstrating the baseline setting i.e. picking longer summaries with matching system prompts.
Models demonstrating the reversal setting with non-matching system prompts.
Qwen2.5-32B and Seed-OSS-36B models finetuned on a EM dataset. Inferece instructions: https://docs.axolotl.ai/docs/inference
Word Count SFT then EM training models (Qwen 32B and Seed 36B)
Word count SFT models trained on top of EM-finetuned models to evaluate capability preservation
MMLU SFT first, then EM training. Ablation: does MMLU pre-training affect emergent misalignment?
EM finetuned models with additional MMLU SFT training for ablation
Models demonstrating the reversal setting with matching system prompts.
Models demonstrating the baseline setting i.e. picking longer summaries with matching system prompts.
Models demonstrating the baseline setting i.e. picking longer summaries with non-matching system prompts.
Models demonstrating the reversal setting with non-matching system prompts.
Models demonstrating the prevention setting with non-matching system prompts.
Qwen2.5-32B and Seed-OSS-36B models finetuned on a EM dataset. Inferece instructions: https://docs.axolotl.ai/docs/inference
Qwen2.5-32B and Seed-OSS-36B models finetuned on a EM dataset. Inferece instructions: https://docs.axolotl.ai/docs/inference
Qwen2.5-32B and Seed-OSS-36B models finetuned on a EM dataset. Inferece instructions: https://docs.axolotl.ai/docs/inference