Upload folder using huggingface_hub
Browse files
checkpoint/config.json
CHANGED
|
@@ -33,7 +33,7 @@
|
|
| 33 |
}
|
| 34 |
},
|
| 35 |
"torch_dtype": "float32",
|
| 36 |
-
"transformers_version": "4.35.
|
| 37 |
"use_cache": true,
|
| 38 |
"vocab_size": 50257
|
| 39 |
}
|
|
|
|
| 33 |
}
|
| 34 |
},
|
| 35 |
"torch_dtype": "float32",
|
| 36 |
+
"transformers_version": "4.35.2",
|
| 37 |
"use_cache": true,
|
| 38 |
"vocab_size": 50257
|
| 39 |
}
|
checkpoint/generation_config.json
CHANGED
|
@@ -2,5 +2,5 @@
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 50256,
|
| 4 |
"eos_token_id": 50256,
|
| 5 |
-
"transformers_version": "4.35.
|
| 6 |
}
|
|
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 50256,
|
| 4 |
"eos_token_id": 50256,
|
| 5 |
+
"transformers_version": "4.35.2"
|
| 6 |
}
|
compiled/336a8139cdb6477a7420.neff
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1076224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fcef49b2398cf3fe4a22a721d5291997728528f50f27be1944e2a1f2352150c6
|
| 3 |
size 1076224
|
compiled/91f58547ef748349c68d.neff
CHANGED
|
Binary files a/compiled/91f58547ef748349c68d.neff and b/compiled/91f58547ef748349c68d.neff differ
|
|
|
compiled/964a9622cc995bfa9d99.neff
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1629184
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ffc73e3343e81dba09c27c4a0f5128789acc2f88ce71ff9e976a4eca9a255be8
|
| 3 |
size 1629184
|
compiled/b9f29083c128f3826c32.neff
CHANGED
|
Binary files a/compiled/b9f29083c128f3826c32.neff and b/compiled/b9f29083c128f3826c32.neff differ
|
|
|
config.json
CHANGED
|
@@ -21,7 +21,7 @@
|
|
| 21 |
"auto_cast_type": "fp32",
|
| 22 |
"batch_size": 16,
|
| 23 |
"compiler_type": "neuronx-cc",
|
| 24 |
-
"compiler_version": "2.11.0.
|
| 25 |
"num_cores": 2,
|
| 26 |
"sequence_length": 1024,
|
| 27 |
"task": "text-generation"
|
|
@@ -41,7 +41,7 @@
|
|
| 41 |
"max_length": 50
|
| 42 |
}
|
| 43 |
},
|
| 44 |
-
"transformers_version": "4.35.
|
| 45 |
"use_cache": true,
|
| 46 |
"vocab_size": 50257
|
| 47 |
}
|
|
|
|
| 21 |
"auto_cast_type": "fp32",
|
| 22 |
"batch_size": 16,
|
| 23 |
"compiler_type": "neuronx-cc",
|
| 24 |
+
"compiler_version": "2.11.0.34+c5231f848",
|
| 25 |
"num_cores": 2,
|
| 26 |
"sequence_length": 1024,
|
| 27 |
"task": "text-generation"
|
|
|
|
| 41 |
"max_length": 50
|
| 42 |
}
|
| 43 |
},
|
| 44 |
+
"transformers_version": "4.35.2",
|
| 45 |
"use_cache": true,
|
| 46 |
"vocab_size": 50257
|
| 47 |
}
|
generation_config.json
CHANGED
|
@@ -2,5 +2,5 @@
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 50256,
|
| 4 |
"eos_token_id": 50256,
|
| 5 |
-
"transformers_version": "4.35.
|
| 6 |
}
|
|
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"bos_token_id": 50256,
|
| 4 |
"eos_token_id": 50256,
|
| 5 |
+
"transformers_version": "4.35.2"
|
| 6 |
}
|