Upload folder using huggingface_hub
Browse files- config.json +3 -3
- gpt_float16_tp1_rank0.engine +2 -2
config.json
CHANGED
|
@@ -6,12 +6,12 @@
|
|
| 6 |
"hidden_size": 768,
|
| 7 |
"int8": false,
|
| 8 |
"lora_target_modules": null,
|
| 9 |
-
"max_batch_size":
|
| 10 |
"max_beam_width": 1,
|
| 11 |
"max_draft_len": 0,
|
| 12 |
-
"max_input_len":
|
| 13 |
"max_num_tokens": null,
|
| 14 |
-
"max_output_len":
|
| 15 |
"max_position_embeddings": 1024,
|
| 16 |
"max_prompt_embedding_table_size": 0,
|
| 17 |
"name": "gpt",
|
|
|
|
| 6 |
"hidden_size": 768,
|
| 7 |
"int8": false,
|
| 8 |
"lora_target_modules": null,
|
| 9 |
+
"max_batch_size": 32,
|
| 10 |
"max_beam_width": 1,
|
| 11 |
"max_draft_len": 0,
|
| 12 |
+
"max_input_len": 1024,
|
| 13 |
"max_num_tokens": null,
|
| 14 |
+
"max_output_len": 2048,
|
| 15 |
"max_position_embeddings": 1024,
|
| 16 |
"max_prompt_embedding_table_size": 0,
|
| 17 |
"name": "gpt",
|
gpt_float16_tp1_rank0.engine
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:50ffa47bfbe312d8d243e97995a83e0494c312d03d0ba18dd2577b7b9a684e24
|
| 3 |
+
size 328061308
|