Upload folder using huggingface_hub
Browse files- config.json +5 -5
- modeling_cloverlm.py +1 -1
config.json
CHANGED
|
@@ -12,6 +12,7 @@
|
|
| 12 |
]
|
| 13 |
},
|
| 14 |
"d_head": 128,
|
|
|
|
| 15 |
"heads": 28,
|
| 16 |
"hidden_size": 3584,
|
| 17 |
"intermediate_size": 14336,
|
|
@@ -22,15 +23,14 @@
|
|
| 22 |
"num_blocks": 29,
|
| 23 |
"num_hidden_layers": 29,
|
| 24 |
"num_key_value_heads": 7,
|
|
|
|
|
|
|
|
|
|
| 25 |
"quartet_2_impl": "pseudoquant",
|
| 26 |
"ratio": 4,
|
| 27 |
"scale_type": "1/sqrt(d)",
|
| 28 |
-
"head_dim": 128,
|
| 29 |
"tie_word_embeddings": true,
|
| 30 |
"transformers_version": "5.3.0",
|
| 31 |
"vocab_size": 32000,
|
| 32 |
-
"weight_tying": true
|
| 33 |
-
"quantization_config": {
|
| 34 |
-
"quant_method": "quartet2"
|
| 35 |
-
}
|
| 36 |
}
|
|
|
|
| 12 |
]
|
| 13 |
},
|
| 14 |
"d_head": 128,
|
| 15 |
+
"head_dim": 128,
|
| 16 |
"heads": 28,
|
| 17 |
"hidden_size": 3584,
|
| 18 |
"intermediate_size": 14336,
|
|
|
|
| 23 |
"num_blocks": 29,
|
| 24 |
"num_hidden_layers": 29,
|
| 25 |
"num_key_value_heads": 7,
|
| 26 |
+
"quantization_config": {
|
| 27 |
+
"quant_method": "quartet2"
|
| 28 |
+
},
|
| 29 |
"quartet_2_impl": "pseudoquant",
|
| 30 |
"ratio": 4,
|
| 31 |
"scale_type": "1/sqrt(d)",
|
|
|
|
| 32 |
"tie_word_embeddings": true,
|
| 33 |
"transformers_version": "5.3.0",
|
| 34 |
"vocab_size": 32000,
|
| 35 |
+
"weight_tying": true
|
|
|
|
|
|
|
|
|
|
| 36 |
}
|
modeling_cloverlm.py
CHANGED
|
@@ -246,4 +246,4 @@ class CloverLMForCausalLM(PreTrainedModel, GenerationMixin):
|
|
| 246 |
return {"input_ids": input_ids}
|
| 247 |
|
| 248 |
def _supports_default_dynamic_cache(self):
|
| 249 |
-
return False
|
|
|
|
| 246 |
return {"input_ids": input_ids}
|
| 247 |
|
| 248 |
def _supports_default_dynamic_cache(self):
|
| 249 |
+
return False
|