Upload config.json with huggingface_hub
Browse files- config.json +5 -19
config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "
|
| 3 |
"architectures": [
|
| 4 |
"T5ForConditionalGeneration"
|
| 5 |
],
|
|
@@ -35,26 +35,12 @@
|
|
| 35 |
"num_beams": 4,
|
| 36 |
"prefix": "summarize: "
|
| 37 |
},
|
| 38 |
-
"
|
| 39 |
-
"
|
| 40 |
-
"
|
| 41 |
-
"
|
| 42 |
-
"prefix": "translate English to German: "
|
| 43 |
-
},
|
| 44 |
-
"translation_en_to_fr": {
|
| 45 |
-
"early_stopping": true,
|
| 46 |
-
"max_length": 300,
|
| 47 |
-
"num_beams": 4,
|
| 48 |
-
"prefix": "translate English to French: "
|
| 49 |
-
},
|
| 50 |
-
"translation_en_to_ro": {
|
| 51 |
-
"early_stopping": true,
|
| 52 |
-
"max_length": 300,
|
| 53 |
-
"num_beams": 4,
|
| 54 |
-
"prefix": "translate English to Romanian: "
|
| 55 |
}
|
| 56 |
},
|
| 57 |
-
"torch_dtype": "float32",
|
| 58 |
"transformers_version": "4.42.4",
|
| 59 |
"use_cache": true,
|
| 60 |
"vocab_size": 32128
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "t5-small",
|
| 3 |
"architectures": [
|
| 4 |
"T5ForConditionalGeneration"
|
| 5 |
],
|
|
|
|
| 35 |
"num_beams": 4,
|
| 36 |
"prefix": "summarize: "
|
| 37 |
},
|
| 38 |
+
"table_qa": {
|
| 39 |
+
"max_length": 512,
|
| 40 |
+
"padding": "max_length",
|
| 41 |
+
"truncation": true
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
}
|
| 43 |
},
|
|
|
|
| 44 |
"transformers_version": "4.42.4",
|
| 45 |
"use_cache": true,
|
| 46 |
"vocab_size": 32128
|