Upload folder using huggingface_hub
#7
by
sharpenb
- opened
- config.json +24 -24
- model.safetensors +1 -1
- smash_config.json +1 -1
config.json
CHANGED
|
@@ -1,15 +1,15 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "/covalent/.cache/models/
|
| 3 |
"activation_function": "gelu_new",
|
| 4 |
"all_reduce_scores": {
|
| 5 |
"0": "NON_PARTICIPATING",
|
| 6 |
"1": "NON_PARTICIPATING",
|
| 7 |
-
"10": "
|
| 8 |
"100": "NON_PARTICIPATING",
|
| 9 |
"101": "NON_PARTICIPATING",
|
| 10 |
"102": "NON_PARTICIPATING",
|
| 11 |
"103": "NON_PARTICIPATING",
|
| 12 |
-
"104": "
|
| 13 |
"105": "NON_PARTICIPATING",
|
| 14 |
"106": "NON_PARTICIPATING",
|
| 15 |
"107": "NON_PARTICIPATING",
|
|
@@ -20,7 +20,7 @@
|
|
| 20 |
"111": "NON_PARTICIPATING",
|
| 21 |
"112": "NON_PARTICIPATING",
|
| 22 |
"113": "NON_PARTICIPATING",
|
| 23 |
-
"114": "
|
| 24 |
"115": "NON_PARTICIPATING",
|
| 25 |
"116": "NON_PARTICIPATING",
|
| 26 |
"117": "NON_PARTICIPATING",
|
|
@@ -36,28 +36,28 @@
|
|
| 36 |
"126": "NON_PARTICIPATING",
|
| 37 |
"127": "NON_PARTICIPATING",
|
| 38 |
"128": "NON_PARTICIPATING",
|
| 39 |
-
"129": "
|
| 40 |
"13": "NON_PARTICIPATING",
|
| 41 |
"130": "NON_PARTICIPATING",
|
| 42 |
"131": "NON_PARTICIPATING",
|
| 43 |
"132": "NON_PARTICIPATING",
|
| 44 |
"133": "NON_PARTICIPATING",
|
| 45 |
"134": "NON_PARTICIPATING",
|
| 46 |
-
"135": "
|
| 47 |
"136": "NON_PARTICIPATING",
|
| 48 |
"137": "NON_PARTICIPATING",
|
| 49 |
"138": "NON_PARTICIPATING",
|
| 50 |
"139": "NON_PARTICIPATING",
|
| 51 |
"14": "NON_PARTICIPATING",
|
| 52 |
-
"140": "
|
| 53 |
"141": "SUCCESS",
|
| 54 |
"142": "NON_PARTICIPATING",
|
| 55 |
-
"143": "
|
| 56 |
"144": "NON_PARTICIPATING",
|
| 57 |
"145": "NON_PARTICIPATING",
|
| 58 |
"146": "NON_PARTICIPATING",
|
| 59 |
"147": "NON_PARTICIPATING",
|
| 60 |
-
"148": "
|
| 61 |
"149": "NON_PARTICIPATING",
|
| 62 |
"15": "NON_PARTICIPATING",
|
| 63 |
"150": "NON_PARTICIPATING",
|
|
@@ -80,7 +80,7 @@
|
|
| 80 |
"166": "NON_PARTICIPATING",
|
| 81 |
"167": "NON_PARTICIPATING",
|
| 82 |
"168": "NON_PARTICIPATING",
|
| 83 |
-
"169": "
|
| 84 |
"17": "NON_PARTICIPATING",
|
| 85 |
"170": "NON_PARTICIPATING",
|
| 86 |
"171": "NON_PARTICIPATING",
|
|
@@ -90,7 +90,7 @@
|
|
| 90 |
"175": "NON_PARTICIPATING",
|
| 91 |
"176": "NON_PARTICIPATING",
|
| 92 |
"177": "NON_PARTICIPATING",
|
| 93 |
-
"178": "
|
| 94 |
"179": "SUCCESS",
|
| 95 |
"18": "NON_PARTICIPATING",
|
| 96 |
"180": "NON_PARTICIPATING",
|
|
@@ -106,7 +106,7 @@
|
|
| 106 |
"19": "NON_PARTICIPATING",
|
| 107 |
"190": "NON_PARTICIPATING",
|
| 108 |
"191": "NON_PARTICIPATING",
|
| 109 |
-
"192": "
|
| 110 |
"193": "NON_PARTICIPATING",
|
| 111 |
"194": "NON_PARTICIPATING",
|
| 112 |
"195": "NON_PARTICIPATING",
|
|
@@ -114,13 +114,13 @@
|
|
| 114 |
"197": "NON_PARTICIPATING",
|
| 115 |
"198": "NON_PARTICIPATING",
|
| 116 |
"199": "NON_PARTICIPATING",
|
| 117 |
-
"2": "
|
| 118 |
"20": "NON_PARTICIPATING",
|
| 119 |
"200": "NON_PARTICIPATING",
|
| 120 |
"201": "NON_PARTICIPATING",
|
| 121 |
"202": "NON_PARTICIPATING",
|
| 122 |
"203": "NON_PARTICIPATING",
|
| 123 |
-
"204": "
|
| 124 |
"205": "NON_PARTICIPATING",
|
| 125 |
"206": "NON_PARTICIPATING",
|
| 126 |
"207": "NON_PARTICIPATING",
|
|
@@ -133,7 +133,7 @@
|
|
| 133 |
"213": "SUCCESS",
|
| 134 |
"214": "NON_PARTICIPATING",
|
| 135 |
"215": "NON_PARTICIPATING",
|
| 136 |
-
"216": "
|
| 137 |
"217": "SUCCESS",
|
| 138 |
"218": "NON_PARTICIPATING",
|
| 139 |
"219": "NON_PARTICIPATING",
|
|
@@ -212,7 +212,7 @@
|
|
| 212 |
"55": "NON_PARTICIPATING",
|
| 213 |
"56": "NON_PARTICIPATING",
|
| 214 |
"57": "NON_PARTICIPATING",
|
| 215 |
-
"58": "
|
| 216 |
"59": "NON_PARTICIPATING",
|
| 217 |
"6": "NON_PARTICIPATING",
|
| 218 |
"60": "NON_PARTICIPATING",
|
|
@@ -220,9 +220,9 @@
|
|
| 220 |
"62": "NON_PARTICIPATING",
|
| 221 |
"63": "NON_PARTICIPATING",
|
| 222 |
"64": "NON_PARTICIPATING",
|
| 223 |
-
"65": "
|
| 224 |
"66": "NON_PARTICIPATING",
|
| 225 |
-
"67": "
|
| 226 |
"68": "NON_PARTICIPATING",
|
| 227 |
"69": "NON_PARTICIPATING",
|
| 228 |
"7": "SUCCESS",
|
|
@@ -230,16 +230,16 @@
|
|
| 230 |
"71": "NON_PARTICIPATING",
|
| 231 |
"72": "NON_PARTICIPATING",
|
| 232 |
"73": "NON_PARTICIPATING",
|
| 233 |
-
"74": "
|
| 234 |
-
"75": "
|
| 235 |
"76": "NON_PARTICIPATING",
|
| 236 |
-
"77": "
|
| 237 |
"78": "NON_PARTICIPATING",
|
| 238 |
"79": "NON_PARTICIPATING",
|
| 239 |
"8": "NON_PARTICIPATING",
|
| 240 |
"80": "NON_PARTICIPATING",
|
| 241 |
"81": "NON_PARTICIPATING",
|
| 242 |
-
"82": "
|
| 243 |
"83": "NON_PARTICIPATING",
|
| 244 |
"84": "NON_PARTICIPATING",
|
| 245 |
"85": "NON_PARTICIPATING",
|
|
@@ -254,9 +254,9 @@
|
|
| 254 |
"93": "NON_PARTICIPATING",
|
| 255 |
"94": "NON_PARTICIPATING",
|
| 256 |
"95": "SUCCESS",
|
| 257 |
-
"96": "
|
| 258 |
"97": "NON_PARTICIPATING",
|
| 259 |
-
"98": "
|
| 260 |
"99": "NON_PARTICIPATING"
|
| 261 |
},
|
| 262 |
"architectures": [
|
|
|
|
| 1 |
{
|
| 2 |
+
"_name_or_path": "/covalent/.cache/models/tmplxft50bgs7culthd",
|
| 3 |
"activation_function": "gelu_new",
|
| 4 |
"all_reduce_scores": {
|
| 5 |
"0": "NON_PARTICIPATING",
|
| 6 |
"1": "NON_PARTICIPATING",
|
| 7 |
+
"10": "SUCCESS",
|
| 8 |
"100": "NON_PARTICIPATING",
|
| 9 |
"101": "NON_PARTICIPATING",
|
| 10 |
"102": "NON_PARTICIPATING",
|
| 11 |
"103": "NON_PARTICIPATING",
|
| 12 |
+
"104": "SUCCESS",
|
| 13 |
"105": "NON_PARTICIPATING",
|
| 14 |
"106": "NON_PARTICIPATING",
|
| 15 |
"107": "NON_PARTICIPATING",
|
|
|
|
| 20 |
"111": "NON_PARTICIPATING",
|
| 21 |
"112": "NON_PARTICIPATING",
|
| 22 |
"113": "NON_PARTICIPATING",
|
| 23 |
+
"114": "SUCCESS",
|
| 24 |
"115": "NON_PARTICIPATING",
|
| 25 |
"116": "NON_PARTICIPATING",
|
| 26 |
"117": "NON_PARTICIPATING",
|
|
|
|
| 36 |
"126": "NON_PARTICIPATING",
|
| 37 |
"127": "NON_PARTICIPATING",
|
| 38 |
"128": "NON_PARTICIPATING",
|
| 39 |
+
"129": "SUCCESS",
|
| 40 |
"13": "NON_PARTICIPATING",
|
| 41 |
"130": "NON_PARTICIPATING",
|
| 42 |
"131": "NON_PARTICIPATING",
|
| 43 |
"132": "NON_PARTICIPATING",
|
| 44 |
"133": "NON_PARTICIPATING",
|
| 45 |
"134": "NON_PARTICIPATING",
|
| 46 |
+
"135": "SUCCESS",
|
| 47 |
"136": "NON_PARTICIPATING",
|
| 48 |
"137": "NON_PARTICIPATING",
|
| 49 |
"138": "NON_PARTICIPATING",
|
| 50 |
"139": "NON_PARTICIPATING",
|
| 51 |
"14": "NON_PARTICIPATING",
|
| 52 |
+
"140": "SUCCESS",
|
| 53 |
"141": "SUCCESS",
|
| 54 |
"142": "NON_PARTICIPATING",
|
| 55 |
+
"143": "NON_PARTICIPATING",
|
| 56 |
"144": "NON_PARTICIPATING",
|
| 57 |
"145": "NON_PARTICIPATING",
|
| 58 |
"146": "NON_PARTICIPATING",
|
| 59 |
"147": "NON_PARTICIPATING",
|
| 60 |
+
"148": "SUCCESS",
|
| 61 |
"149": "NON_PARTICIPATING",
|
| 62 |
"15": "NON_PARTICIPATING",
|
| 63 |
"150": "NON_PARTICIPATING",
|
|
|
|
| 80 |
"166": "NON_PARTICIPATING",
|
| 81 |
"167": "NON_PARTICIPATING",
|
| 82 |
"168": "NON_PARTICIPATING",
|
| 83 |
+
"169": "NON_PARTICIPATING",
|
| 84 |
"17": "NON_PARTICIPATING",
|
| 85 |
"170": "NON_PARTICIPATING",
|
| 86 |
"171": "NON_PARTICIPATING",
|
|
|
|
| 90 |
"175": "NON_PARTICIPATING",
|
| 91 |
"176": "NON_PARTICIPATING",
|
| 92 |
"177": "NON_PARTICIPATING",
|
| 93 |
+
"178": "SUCCESS",
|
| 94 |
"179": "SUCCESS",
|
| 95 |
"18": "NON_PARTICIPATING",
|
| 96 |
"180": "NON_PARTICIPATING",
|
|
|
|
| 106 |
"19": "NON_PARTICIPATING",
|
| 107 |
"190": "NON_PARTICIPATING",
|
| 108 |
"191": "NON_PARTICIPATING",
|
| 109 |
+
"192": "NON_PARTICIPATING",
|
| 110 |
"193": "NON_PARTICIPATING",
|
| 111 |
"194": "NON_PARTICIPATING",
|
| 112 |
"195": "NON_PARTICIPATING",
|
|
|
|
| 114 |
"197": "NON_PARTICIPATING",
|
| 115 |
"198": "NON_PARTICIPATING",
|
| 116 |
"199": "NON_PARTICIPATING",
|
| 117 |
+
"2": "NON_PARTICIPATING",
|
| 118 |
"20": "NON_PARTICIPATING",
|
| 119 |
"200": "NON_PARTICIPATING",
|
| 120 |
"201": "NON_PARTICIPATING",
|
| 121 |
"202": "NON_PARTICIPATING",
|
| 122 |
"203": "NON_PARTICIPATING",
|
| 123 |
+
"204": "SUCCESS",
|
| 124 |
"205": "NON_PARTICIPATING",
|
| 125 |
"206": "NON_PARTICIPATING",
|
| 126 |
"207": "NON_PARTICIPATING",
|
|
|
|
| 133 |
"213": "SUCCESS",
|
| 134 |
"214": "NON_PARTICIPATING",
|
| 135 |
"215": "NON_PARTICIPATING",
|
| 136 |
+
"216": "NON_PARTICIPATING",
|
| 137 |
"217": "SUCCESS",
|
| 138 |
"218": "NON_PARTICIPATING",
|
| 139 |
"219": "NON_PARTICIPATING",
|
|
|
|
| 212 |
"55": "NON_PARTICIPATING",
|
| 213 |
"56": "NON_PARTICIPATING",
|
| 214 |
"57": "NON_PARTICIPATING",
|
| 215 |
+
"58": "NON_PARTICIPATING",
|
| 216 |
"59": "NON_PARTICIPATING",
|
| 217 |
"6": "NON_PARTICIPATING",
|
| 218 |
"60": "NON_PARTICIPATING",
|
|
|
|
| 220 |
"62": "NON_PARTICIPATING",
|
| 221 |
"63": "NON_PARTICIPATING",
|
| 222 |
"64": "NON_PARTICIPATING",
|
| 223 |
+
"65": "SUCCESS",
|
| 224 |
"66": "NON_PARTICIPATING",
|
| 225 |
+
"67": "SUCCESS",
|
| 226 |
"68": "NON_PARTICIPATING",
|
| 227 |
"69": "NON_PARTICIPATING",
|
| 228 |
"7": "SUCCESS",
|
|
|
|
| 230 |
"71": "NON_PARTICIPATING",
|
| 231 |
"72": "NON_PARTICIPATING",
|
| 232 |
"73": "NON_PARTICIPATING",
|
| 233 |
+
"74": "NON_PARTICIPATING",
|
| 234 |
+
"75": "SUCCESS",
|
| 235 |
"76": "NON_PARTICIPATING",
|
| 236 |
+
"77": "SUCCESS",
|
| 237 |
"78": "NON_PARTICIPATING",
|
| 238 |
"79": "NON_PARTICIPATING",
|
| 239 |
"8": "NON_PARTICIPATING",
|
| 240 |
"80": "NON_PARTICIPATING",
|
| 241 |
"81": "NON_PARTICIPATING",
|
| 242 |
+
"82": "SUCCESS",
|
| 243 |
"83": "NON_PARTICIPATING",
|
| 244 |
"84": "NON_PARTICIPATING",
|
| 245 |
"85": "NON_PARTICIPATING",
|
|
|
|
| 254 |
"93": "NON_PARTICIPATING",
|
| 255 |
"94": "NON_PARTICIPATING",
|
| 256 |
"95": "SUCCESS",
|
| 257 |
+
"96": "SUCCESS",
|
| 258 |
"97": "NON_PARTICIPATING",
|
| 259 |
+
"98": "SUCCESS",
|
| 260 |
"99": "NON_PARTICIPATING"
|
| 261 |
},
|
| 262 |
"architectures": [
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1207575528
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b623ff5c15df43120192267328105e02e13c37dd169b4bc5e962907ea98af273
|
| 3 |
size 1207575528
|
smash_config.json
CHANGED
|
@@ -28,7 +28,7 @@
|
|
| 28 |
"quant_llm-int8_weight_bits": 8,
|
| 29 |
"max_batch_size": 1,
|
| 30 |
"device": "cuda",
|
| 31 |
-
"cache_dir": "/covalent/.cache/models/
|
| 32 |
"task": "",
|
| 33 |
"save_load_fn": "bitsandbytes",
|
| 34 |
"save_load_fn_args": {}
|
|
|
|
| 28 |
"quant_llm-int8_weight_bits": 8,
|
| 29 |
"max_batch_size": 1,
|
| 30 |
"device": "cuda",
|
| 31 |
+
"cache_dir": "/covalent/.cache/models/tmplxft50bg",
|
| 32 |
"task": "",
|
| 33 |
"save_load_fn": "bitsandbytes",
|
| 34 |
"save_load_fn_args": {}
|