phi-2-gpo-newSFT-b0.001-i0

This model is a fine-tuned version of DUAL-GPO/phi-2-sft-lora-ultrachat-merged on the HuggingFaceH4/ultrafeedback_binarized dataset.

Downloads last month
24
Safetensors
Model size
3B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DUAL-GPO/phi-2-gpo-new-i0

Adapters
43 models