Self-Fulfilling (Mis)alignment: Post-Trained Models
Here is a selection of SFM models that have undergone DPO.
Text Generation • 7B • Updated • 807Note Our "Unfiltered" instruct model, trained on 500B PT, 50B MT, 4B SFT, finishing with DPO
geodesic-research/sfm-sft_dolci_instruct_blocklist_filtered-DPO
Text Generation • 7B • Updated • 637Note Our "Filtered" instruct model, trained on 500B PT, 50B MT, 4B SFT, finishing with DPO
geodesic-research/sfm-sft_dolci_instruct_unfiltered_synthetic_misalignment_mid-DPO
Text Generation • 7B • Updated • 1.05kNote Our "Unfiltered + Synthetic Misalignment" instruct model, trained on 500B PT, 50B MT, 4B SFT, finishing with DPO
geodesic-research/sfm-sft_dolci_instruct_blocklist_filtered_synthetic_alignment_mid-DPO
Text Generation • 7B • Updated • 875Note Our "Filtered + Synthetic Alignment" instruct model, trained on 500B PT, 50B MT, 4B SFT, finishing with DPO
geodesic-research/sfm-sft_dolci_instruct_unfiltered
Text Generation • 7B • Updated • 1.43kNote Our "Unfiltered" instruct model, trained on 500B PT, 50B MT, finishing with 4B SFT
geodesic-research/sfm-sft_dolci_instruct_blocklist_filtered
Text Generation • 7B • Updated • 1.21kNote Our "Filtered" instruct model, trained on 500B PT, 50B MT, finishing with 4B SFT
geodesic-research/sfm-sft_dolci_instruct_unfiltered_synthetic_misalignment_mid
Text Generation • 7B • Updated • 1.17kNote Our "Unfiltered + Synthetic Misalignment" instruct model, trained on 500B PT, 50B MT, finishing with 4B SFT
geodesic-research/sfm-sft_dolci_instruct_blocklist_filtered_synthetic_alignment_mid
Text Generation • 7B • Updated • 1.1kNote Our "Filtered + Synthetic Alignment" instruct model, trained on 500B PT, 50B MT, finishing with 4B SFT