Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
bimabk
/
dpo_
like
0
PEFT
Safetensors
arxiv:
1910.09700
Model card
Files
Files and versions
xet
Community
Use this model
main
dpo_
/
loss.txt
bimabk
Upload task output 63b2db3d-a057-429f-9319-84e8338dbfb9
f52aebd
verified
2 months ago
raw
Copy download link
history
blame
contribute
delete
22 Bytes
358,0.6298339366912842