Upload flow_dpo_alldata80k_dpo_iter1/model_1000.pt with huggingface_hub
Browse files
flow_dpo_alldata80k_dpo_iter1/model_1000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:40e515acb0c802f19cfacb8163aeb45465b15d2f4d6a67099d95d83e7b35f3e3
|
| 3 |
+
size 4045630320
|