Upload folder using huggingface_hub
Browse files
CodeDecoder.mlmodelc/analytics/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 243
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d52078ac1b879b055032ae3266b7cbd6a48a969c875c0f9835318feec20c18f4
|
| 3 |
size 243
|
CodeDecoder.mlmodelc/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 642
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:83a6da5ed37a7b16fd371db634a76997e0320764acd1d748c5acf11f62211751
|
| 3 |
size 642
|
CodeDecoder.mlmodelc/metadata.json
CHANGED
|
@@ -88,7 +88,7 @@
|
|
| 88 |
"name" : "MLModelType_mlProgram"
|
| 89 |
},
|
| 90 |
"userDefinedMetadata" : {
|
| 91 |
-
"com.github.apple.coremltools.conversion_date" : "2026-
|
| 92 |
"com.github.apple.coremltools.source" : "torch==2.10.0",
|
| 93 |
"com.github.apple.coremltools.version" : "9.0",
|
| 94 |
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
|
@@ -155,7 +155,7 @@
|
|
| 155 |
"type" : "MultiArray"
|
| 156 |
}
|
| 157 |
],
|
| 158 |
-
"generatedClassName" : "
|
| 159 |
"method" : "predict"
|
| 160 |
}
|
| 161 |
]
|
|
|
|
| 88 |
"name" : "MLModelType_mlProgram"
|
| 89 |
},
|
| 90 |
"userDefinedMetadata" : {
|
| 91 |
+
"com.github.apple.coremltools.conversion_date" : "2026-04-03",
|
| 92 |
"com.github.apple.coremltools.source" : "torch==2.10.0",
|
| 93 |
"com.github.apple.coremltools.version" : "9.0",
|
| 94 |
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
|
|
|
| 155 |
"type" : "MultiArray"
|
| 156 |
}
|
| 157 |
],
|
| 158 |
+
"generatedClassName" : "CodeDecoder_fixed",
|
| 159 |
"method" : "predict"
|
| 160 |
}
|
| 161 |
]
|
CodeDecoder.mlmodelc/model.mil
CHANGED
|
@@ -5236,23 +5236,23 @@ program(1.3)
|
|
| 5236 |
tensor<fp16, [1, 1, 1024]> input_cast_fp16 = mul(x = var_11363_cast_fp16, y = var_11366_cast_fp16)[name = string("input_cast_fp16")];
|
| 5237 |
tensor<fp16, [3072, 1024]> codec_head_weight_to_fp16_palettized = constexpr_lut_to_dense(indices = tensor<uint8, [3072, 1024]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(440678400))), lut = tensor<fp16, [1, 1, 256, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(443824192))))[name = string("codec_head_weight_to_fp16_palettized")];
|
| 5238 |
tensor<fp16, [1, 1, 3072]> logits = linear(bias = linear_4_bias_0_to_fp16, weight = codec_head_weight_to_fp16_palettized, x = input_cast_fp16)[name = string("linear_196_cast_fp16")];
|
| 5239 |
-
tensor<int32, [1]>
|
| 5240 |
-
tensor<fp16, [1, 1024]>
|
| 5241 |
-
tensor<int32, [1]>
|
| 5242 |
-
tensor<fp16, [1, 1024, 1]>
|
| 5243 |
-
tensor<int32, [1]>
|
| 5244 |
-
tensor<fp16, [1, 1024, 1, 1]> hidden_states = expand_dims(axes =
|
| 5245 |
-
int32 var_11378 = const()[name = string("op_11378"), val = int32(1)];
|
| 5246 |
-
bool new_kv_k_interleave_0 = const()[name = string("new_kv_k_interleave_0"), val = bool(false)];
|
| 5247 |
-
tensor<fp16, [1, 28672, 1, 1]> new_kv_k_cast_fp16 = concat(axis = var_11378, interleave = new_kv_k_interleave_0, values = (nk_1_cast_fp16, nk_3_cast_fp16, nk_5_cast_fp16, nk_7_cast_fp16, nk_9_cast_fp16, nk_11_cast_fp16, nk_13_cast_fp16, nk_15_cast_fp16, nk_17_cast_fp16, nk_19_cast_fp16, nk_21_cast_fp16, nk_23_cast_fp16, nk_25_cast_fp16, nk_27_cast_fp16, nk_29_cast_fp16, nk_31_cast_fp16, nk_33_cast_fp16, nk_35_cast_fp16, nk_37_cast_fp16, nk_39_cast_fp16, nk_41_cast_fp16, nk_43_cast_fp16, nk_45_cast_fp16, nk_47_cast_fp16, nk_49_cast_fp16, nk_51_cast_fp16, nk_53_cast_fp16, nk_cast_fp16))[name = string("new_kv_k_cast_fp16")];
|
| 5248 |
int32 var_11381 = const()[name = string("op_11381"), val = int32(1)];
|
|
|
|
|
|
|
|
|
|
| 5249 |
bool new_kv_v_interleave_0 = const()[name = string("new_kv_v_interleave_0"), val = bool(false)];
|
| 5250 |
-
tensor<fp16, [1, 28672, 1, 1]> new_kv_v_cast_fp16 = concat(axis =
|
| 5251 |
-
tensor<fp16, [1, 28672, 1, 256]>
|
| 5252 |
-
tensor<fp16, [1, 28672, 1, 256]>
|
| 5253 |
-
tensor<fp16, [1, 28672, 1, 256]> new_key_cache = add(x =
|
| 5254 |
-
tensor<fp16, [1, 28672, 1, 256]>
|
| 5255 |
-
tensor<fp16, [1, 28672, 1, 256]>
|
| 5256 |
-
tensor<fp16, [1, 28672, 1, 256]> new_value_cache = add(x =
|
| 5257 |
} -> (logits, hidden_states, new_key_cache, new_value_cache);
|
| 5258 |
}
|
|
|
|
| 5236 |
tensor<fp16, [1, 1, 1024]> input_cast_fp16 = mul(x = var_11363_cast_fp16, y = var_11366_cast_fp16)[name = string("input_cast_fp16")];
|
| 5237 |
tensor<fp16, [3072, 1024]> codec_head_weight_to_fp16_palettized = constexpr_lut_to_dense(indices = tensor<uint8, [3072, 1024]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(440678400))), lut = tensor<fp16, [1, 1, 256, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(443824192))))[name = string("codec_head_weight_to_fp16_palettized")];
|
| 5238 |
tensor<fp16, [1, 1, 3072]> logits = linear(bias = linear_4_bias_0_to_fp16, weight = codec_head_weight_to_fp16_palettized, x = input_cast_fp16)[name = string("linear_196_cast_fp16")];
|
| 5239 |
+
tensor<int32, [1]> var_11375_axes_0 = const()[name = string("op_11375_axes_0"), val = tensor<int32, [1]>([0])];
|
| 5240 |
+
tensor<fp16, [1, 1024]> var_11375_cast_fp16 = squeeze(axes = var_11375_axes_0, x = hidden_states_449_cast_fp16)[name = string("op_11375_cast_fp16")];
|
| 5241 |
+
tensor<int32, [1]> var_11377_axes_0 = const()[name = string("op_11377_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 5242 |
+
tensor<fp16, [1, 1024, 1]> var_11377_cast_fp16 = expand_dims(axes = var_11377_axes_0, x = var_11375_cast_fp16)[name = string("op_11377_cast_fp16")];
|
| 5243 |
+
tensor<int32, [1]> var_11379_axes_0 = const()[name = string("op_11379_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 5244 |
+
tensor<fp16, [1, 1024, 1, 1]> hidden_states = expand_dims(axes = var_11379_axes_0, x = var_11377_cast_fp16)[name = string("op_11379_cast_fp16")];
|
|
|
|
|
|
|
|
|
|
| 5245 |
int32 var_11381 = const()[name = string("op_11381"), val = int32(1)];
|
| 5246 |
+
bool new_kv_k_interleave_0 = const()[name = string("new_kv_k_interleave_0"), val = bool(false)];
|
| 5247 |
+
tensor<fp16, [1, 28672, 1, 1]> new_kv_k_cast_fp16 = concat(axis = var_11381, interleave = new_kv_k_interleave_0, values = (nk_1_cast_fp16, nk_3_cast_fp16, nk_5_cast_fp16, nk_7_cast_fp16, nk_9_cast_fp16, nk_11_cast_fp16, nk_13_cast_fp16, nk_15_cast_fp16, nk_17_cast_fp16, nk_19_cast_fp16, nk_21_cast_fp16, nk_23_cast_fp16, nk_25_cast_fp16, nk_27_cast_fp16, nk_29_cast_fp16, nk_31_cast_fp16, nk_33_cast_fp16, nk_35_cast_fp16, nk_37_cast_fp16, nk_39_cast_fp16, nk_41_cast_fp16, nk_43_cast_fp16, nk_45_cast_fp16, nk_47_cast_fp16, nk_49_cast_fp16, nk_51_cast_fp16, nk_53_cast_fp16, nk_cast_fp16))[name = string("new_kv_k_cast_fp16")];
|
| 5248 |
+
int32 var_11384 = const()[name = string("op_11384"), val = int32(1)];
|
| 5249 |
bool new_kv_v_interleave_0 = const()[name = string("new_kv_v_interleave_0"), val = bool(false)];
|
| 5250 |
+
tensor<fp16, [1, 28672, 1, 1]> new_kv_v_cast_fp16 = concat(axis = var_11384, interleave = new_kv_v_interleave_0, values = (nv_1_cast_fp16, nv_3_cast_fp16, nv_5_cast_fp16, nv_7_cast_fp16, nv_9_cast_fp16, nv_11_cast_fp16, nv_13_cast_fp16, nv_15_cast_fp16, nv_17_cast_fp16, nv_19_cast_fp16, nv_21_cast_fp16, nv_23_cast_fp16, nv_25_cast_fp16, nv_27_cast_fp16, nv_29_cast_fp16, nv_31_cast_fp16, nv_33_cast_fp16, nv_35_cast_fp16, nv_37_cast_fp16, nv_39_cast_fp16, nv_41_cast_fp16, nv_43_cast_fp16, nv_45_cast_fp16, nv_47_cast_fp16, nv_49_cast_fp16, nv_51_cast_fp16, nv_53_cast_fp16, nv_cast_fp16))[name = string("new_kv_v_cast_fp16")];
|
| 5251 |
+
tensor<fp16, [1, 28672, 1, 256]> var_11389_cast_fp16 = mul(x = key_cache, y = var_1203_cast_fp16)[name = string("op_11389_cast_fp16")];
|
| 5252 |
+
tensor<fp16, [1, 28672, 1, 256]> var_11390_cast_fp16 = mul(x = new_kv_k_cast_fp16, y = update_mask_cast_fp16)[name = string("op_11390_cast_fp16")];
|
| 5253 |
+
tensor<fp16, [1, 28672, 1, 256]> new_key_cache = add(x = var_11389_cast_fp16, y = var_11390_cast_fp16)[name = string("op_11392_cast_fp16")];
|
| 5254 |
+
tensor<fp16, [1, 28672, 1, 256]> var_11396_cast_fp16 = mul(x = value_cache, y = var_1203_cast_fp16)[name = string("op_11396_cast_fp16")];
|
| 5255 |
+
tensor<fp16, [1, 28672, 1, 256]> var_11397_cast_fp16 = mul(x = new_kv_v_cast_fp16, y = update_mask_cast_fp16)[name = string("op_11397_cast_fp16")];
|
| 5256 |
+
tensor<fp16, [1, 28672, 1, 256]> new_value_cache = add(x = var_11396_cast_fp16, y = var_11397_cast_fp16)[name = string("op_11399_cast_fp16")];
|
| 5257 |
} -> (logits, hidden_states, new_key_cache, new_value_cache);
|
| 5258 |
}
|