Upload LlamaForCausalLM
Browse files- config.json +2 -3
- generation_config.json +2 -1
- model-00001-of-00002.safetensors +1 -1
- model.safetensors.index.json +1 -0
config.json
CHANGED
|
@@ -1,11 +1,11 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "Sefika/CRE_fewrel_llama3_10_1_task_memory_10_2",
|
| 3 |
"architectures": [
|
| 4 |
"LlamaForCausalLM"
|
| 5 |
],
|
| 6 |
"attention_bias": false,
|
| 7 |
"attention_dropout": 0.0,
|
| 8 |
"bos_token_id": 128000,
|
|
|
|
| 9 |
"eos_token_id": 128009,
|
| 10 |
"head_dim": 128,
|
| 11 |
"hidden_act": "silu",
|
|
@@ -45,8 +45,7 @@
|
|
| 45 |
},
|
| 46 |
"rope_theta": 500000.0,
|
| 47 |
"tie_word_embeddings": false,
|
| 48 |
-
"
|
| 49 |
-
"transformers_version": "4.45.2",
|
| 50 |
"use_cache": false,
|
| 51 |
"vocab_size": 128256
|
| 52 |
}
|
|
|
|
| 1 |
{
|
|
|
|
| 2 |
"architectures": [
|
| 3 |
"LlamaForCausalLM"
|
| 4 |
],
|
| 5 |
"attention_bias": false,
|
| 6 |
"attention_dropout": 0.0,
|
| 7 |
"bos_token_id": 128000,
|
| 8 |
+
"dtype": "float16",
|
| 9 |
"eos_token_id": 128009,
|
| 10 |
"head_dim": 128,
|
| 11 |
"hidden_act": "silu",
|
|
|
|
| 45 |
},
|
| 46 |
"rope_theta": 500000.0,
|
| 47 |
"tie_word_embeddings": false,
|
| 48 |
+
"transformers_version": "4.56.1",
|
|
|
|
| 49 |
"use_cache": false,
|
| 50 |
"vocab_size": 128256
|
| 51 |
}
|
generation_config.json
CHANGED
|
@@ -6,7 +6,8 @@
|
|
| 6 |
128008,
|
| 7 |
128009
|
| 8 |
],
|
|
|
|
| 9 |
"temperature": 0.6,
|
| 10 |
"top_p": 0.9,
|
| 11 |
-
"transformers_version": "4.
|
| 12 |
}
|
|
|
|
| 6 |
128008,
|
| 7 |
128009
|
| 8 |
],
|
| 9 |
+
"pad_token_id": 128009,
|
| 10 |
"temperature": 0.6,
|
| 11 |
"top_p": 0.9,
|
| 12 |
+
"transformers_version": "4.56.1"
|
| 13 |
}
|
model-00001-of-00002.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4977222696
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b15bb32e4274efc6299af199657e5a108a3a133afde7d6c8c286b1664cd94b4c
|
| 3 |
size 4977222696
|
model.safetensors.index.json
CHANGED
|
@@ -1,5 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
|
|
|
| 3 |
"total_size": 6027779904
|
| 4 |
},
|
| 5 |
"weight_map": {
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_parameters": 8030261248,
|
| 4 |
"total_size": 6027779904
|
| 5 |
},
|
| 6 |
"weight_map": {
|