Commit ·
8a23912
1
Parent(s): 1f717f0
20251213
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- ggml-base-encoder.mlmodelc/analytics/coremldata.bin +1 -1
- ggml-base-encoder.mlmodelc/coremldata.bin +2 -2
- ggml-base-encoder.mlmodelc/metadata.json +22 -20
- ggml-base-encoder.mlmodelc/model.mil +0 -0
- ggml-base-encoder.mlmodelc/weights/weight.bin +1 -1
- ggml-large-v3-turbo-encoder.mlmodelc/model0/analytics/coremldata.bin → ggml-base-q8_0.bin +2 -2
- ggml-large-v2-encoder.mlmodelc/metadata.json +0 -66
- ggml-large-v2-encoder.mlmodelc/model.mil +0 -0
- ggml-large-v2-q8_0.bin +0 -3
- ggml-large-v3-encoder.mlmodelc/analytics/coremldata.bin +1 -1
- ggml-large-v3-encoder.mlmodelc/coremldata.bin +2 -2
- ggml-large-v3-encoder.mlmodelc/metadata.json +22 -20
- ggml-large-v3-encoder.mlmodelc/model.mil +0 -0
- ggml-large-v3-encoder.mlmodelc/weights/weight.bin +2 -2
- ggml-large-v3-turbo-encoder.mlmodelc/analytics/coremldata.bin +2 -2
- ggml-large-v3-turbo-encoder.mlmodelc/coremldata.bin +2 -2
- ggml-large-v3-turbo-encoder.mlmodelc/metadata.json +10 -16
- ggml-large-v3-turbo-encoder.mlmodelc/model.mil +0 -0
- ggml-large-v3-turbo-encoder.mlmodelc/model0/coremldata.bin +0 -3
- ggml-large-v3-turbo-encoder.mlmodelc/model0/model.mil +0 -0
- ggml-large-v3-turbo-encoder.mlmodelc/model0/weights/0-weight.bin +0 -3
- ggml-large-v3-turbo-encoder.mlmodelc/model1/analytics/coremldata.bin +0 -3
- ggml-large-v3-turbo-encoder.mlmodelc/model1/coremldata.bin +0 -3
- ggml-large-v3-turbo-encoder.mlmodelc/model1/model.mil +0 -0
- ggml-large-v3-turbo-encoder.mlmodelc/model1/weights/1-weight.bin +0 -3
- {ggml-large-v2-encoder.mlmodelc → ggml-large-v3-turbo-encoder.mlmodelc}/weights/weight.bin +2 -2
- ggml-medium-encoder.mlmodelc/analytics/coremldata.bin +1 -1
- ggml-medium-encoder.mlmodelc/coremldata.bin +2 -2
- ggml-medium-encoder.mlmodelc/metadata.json +22 -20
- ggml-medium-encoder.mlmodelc/model.mil +0 -0
- ggml-medium-encoder.mlmodelc/weights/weight.bin +1 -1
- ggml-base.bin → ggml-medium-q8_0.bin +2 -2
- ggml-medium.bin +0 -3
- ggml-small-encoder.mlmodelc/analytics/coremldata.bin +1 -1
- ggml-small-encoder.mlmodelc/coremldata.bin +2 -2
- ggml-small-encoder.mlmodelc/metadata.json +22 -20
- ggml-small-encoder.mlmodelc/model.mil +0 -0
- ggml-small-encoder.mlmodelc/weights/weight.bin +1 -1
- ggml-large-v2-encoder.mlmodelc/coremldata.bin → ggml-small-q8_0.bin +2 -2
- ggml-small.bin +0 -3
- ggml-tiny-encoder.mlmodelc/analytics/coremldata.bin +1 -1
- ggml-tiny-encoder.mlmodelc/coremldata.bin +2 -2
- ggml-tiny-encoder.mlmodelc/metadata.json +22 -20
- ggml-tiny-encoder.mlmodelc/model.mil +221 -265
- ggml-tiny-encoder.mlmodelc/weights/weight.bin +1 -1
- ggml-large-v2-encoder.mlmodelc/analytics/coremldata.bin → ggml-tiny-q8_0.bin +2 -2
- ggml-tiny.bin +0 -3
- index/base +1 -1
- index/large-v2 +0 -6
- index/large-v3-turbo +2 -8
ggml-base-encoder.mlmodelc/analytics/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 243
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ce0ee7deaec2f6b2c5927b6a124f866047ab9bfd04e92c30aadae7b9b45e0d7
|
| 3 |
size 243
|
ggml-base-encoder.mlmodelc/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98b532b5a0af2ecdf289928a9b20e6a5aa780da7fd5833fde5c4f90dfa8747ef
|
| 3 |
+
size 379
|
ggml-base-encoder.mlmodelc/metadata.json
CHANGED
|
@@ -17,36 +17,38 @@
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
-
"specificationVersion" :
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
-
"
|
| 23 |
-
"
|
| 24 |
-
"
|
| 25 |
-
"
|
| 26 |
-
"
|
| 27 |
-
"
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
-
"Ios17.mul" : 12,
|
| 31 |
-
"Ios17.transpose" : 25
|
| 32 |
},
|
| 33 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 34 |
"isUpdatable" : "0",
|
|
|
|
|
|
|
|
|
|
| 35 |
"availability" : {
|
| 36 |
-
"macOS" : "
|
| 37 |
-
"tvOS" : "
|
| 38 |
-
"visionOS" : "
|
| 39 |
-
"watchOS" : "
|
| 40 |
-
"iOS" : "
|
| 41 |
-
"macCatalyst" : "
|
| 42 |
},
|
| 43 |
"modelType" : {
|
| 44 |
"name" : "MLModelType_mlProgram"
|
| 45 |
},
|
| 46 |
"userDefinedMetadata" : {
|
| 47 |
-
"com.github.apple.coremltools.
|
| 48 |
-
"com.github.apple.coremltools.source" : "torch==2.
|
| 49 |
-
"com.github.apple.coremltools.version" : "
|
|
|
|
| 50 |
},
|
| 51 |
"inputSchema" : [
|
| 52 |
{
|
|
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
+
"specificationVersion" : 9,
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
+
"Ios18.scaledDotProductAttention" : 6,
|
| 23 |
+
"Ios18.linear" : 36,
|
| 24 |
+
"Ios18.gelu" : 8,
|
| 25 |
+
"Ios18.layerNorm" : 13,
|
| 26 |
+
"Ios18.transpose" : 25,
|
| 27 |
+
"Ios18.conv" : 2,
|
| 28 |
+
"Ios18.add" : 13,
|
| 29 |
+
"Ios18.reshape" : 24
|
|
|
|
|
|
|
| 30 |
},
|
| 31 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 32 |
"isUpdatable" : "0",
|
| 33 |
+
"stateSchema" : [
|
| 34 |
+
|
| 35 |
+
],
|
| 36 |
"availability" : {
|
| 37 |
+
"macOS" : "15.0",
|
| 38 |
+
"tvOS" : "18.0",
|
| 39 |
+
"visionOS" : "2.0",
|
| 40 |
+
"watchOS" : "11.0",
|
| 41 |
+
"iOS" : "18.0",
|
| 42 |
+
"macCatalyst" : "18.0"
|
| 43 |
},
|
| 44 |
"modelType" : {
|
| 45 |
"name" : "MLModelType_mlProgram"
|
| 46 |
},
|
| 47 |
"userDefinedMetadata" : {
|
| 48 |
+
"com.github.apple.coremltools.conversion_date" : "2025-12-13",
|
| 49 |
+
"com.github.apple.coremltools.source" : "torch==2.9.1",
|
| 50 |
+
"com.github.apple.coremltools.version" : "9.0",
|
| 51 |
+
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
| 52 |
},
|
| 53 |
"inputSchema" : [
|
| 54 |
{
|
ggml-base-encoder.mlmodelc/model.mil
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-base-encoder.mlmodelc/weights/weight.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 41188544
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c45bee989219532c4cec616d439c51f280ac9d7b04f7847c4b7d7daba1d47523
|
| 3 |
size 41188544
|
ggml-large-v3-turbo-encoder.mlmodelc/model0/analytics/coremldata.bin → ggml-base-q8_0.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1a911a31e5812ebffef4c517c19186627a0203b7e95b5dbb6db884a9dc446ef
|
| 3 |
+
size 56663708
|
ggml-large-v2-encoder.mlmodelc/metadata.json
DELETED
|
@@ -1,66 +0,0 @@
|
|
| 1 |
-
[
|
| 2 |
-
{
|
| 3 |
-
"metadataOutputVersion" : "3.0",
|
| 4 |
-
"storagePrecision" : "Float16",
|
| 5 |
-
"outputSchema" : [
|
| 6 |
-
{
|
| 7 |
-
"hasShapeFlexibility" : "0",
|
| 8 |
-
"isOptional" : "0",
|
| 9 |
-
"dataType" : "Float16",
|
| 10 |
-
"formattedType" : "MultiArray (Float16 1 × 1500 × 1280)",
|
| 11 |
-
"shortDescription" : "",
|
| 12 |
-
"shape" : "[1, 1500, 1280]",
|
| 13 |
-
"name" : "output",
|
| 14 |
-
"type" : "MultiArray"
|
| 15 |
-
}
|
| 16 |
-
],
|
| 17 |
-
"modelParameters" : [
|
| 18 |
-
|
| 19 |
-
],
|
| 20 |
-
"specificationVersion" : 8,
|
| 21 |
-
"mlProgramOperationTypeHistogram" : {
|
| 22 |
-
"Ios17.layerNorm" : 65,
|
| 23 |
-
"Ios17.reshape" : 128,
|
| 24 |
-
"Ios17.conv" : 2,
|
| 25 |
-
"Ios17.linear" : 192,
|
| 26 |
-
"Ios17.add" : 65,
|
| 27 |
-
"Ios17.matmul" : 64,
|
| 28 |
-
"Ios16.gelu" : 34,
|
| 29 |
-
"Ios16.softmax" : 32,
|
| 30 |
-
"Ios17.mul" : 64,
|
| 31 |
-
"Ios17.transpose" : 129
|
| 32 |
-
},
|
| 33 |
-
"computePrecision" : "Mixed (Float16, Int32)",
|
| 34 |
-
"isUpdatable" : "0",
|
| 35 |
-
"availability" : {
|
| 36 |
-
"macOS" : "14.0",
|
| 37 |
-
"tvOS" : "17.0",
|
| 38 |
-
"visionOS" : "1.0",
|
| 39 |
-
"watchOS" : "10.0",
|
| 40 |
-
"iOS" : "17.0",
|
| 41 |
-
"macCatalyst" : "17.0"
|
| 42 |
-
},
|
| 43 |
-
"modelType" : {
|
| 44 |
-
"name" : "MLModelType_mlProgram"
|
| 45 |
-
},
|
| 46 |
-
"userDefinedMetadata" : {
|
| 47 |
-
"com.github.apple.coremltools.source_dialect" : "TorchScript",
|
| 48 |
-
"com.github.apple.coremltools.source" : "torch==2.3.0",
|
| 49 |
-
"com.github.apple.coremltools.version" : "7.2"
|
| 50 |
-
},
|
| 51 |
-
"inputSchema" : [
|
| 52 |
-
{
|
| 53 |
-
"hasShapeFlexibility" : "0",
|
| 54 |
-
"isOptional" : "0",
|
| 55 |
-
"dataType" : "Float16",
|
| 56 |
-
"formattedType" : "MultiArray (Float16 1 × 80 × 3000)",
|
| 57 |
-
"shortDescription" : "",
|
| 58 |
-
"shape" : "[1, 80, 3000]",
|
| 59 |
-
"name" : "logmel_data",
|
| 60 |
-
"type" : "MultiArray"
|
| 61 |
-
}
|
| 62 |
-
],
|
| 63 |
-
"generatedClassName" : "ggml_large_v2_encoder",
|
| 64 |
-
"method" : "predict"
|
| 65 |
-
}
|
| 66 |
-
]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
ggml-large-v2-encoder.mlmodelc/model.mil
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-large-v2-q8_0.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:3a0b8cd189b25520643b628d08dd100a8f0c884c048c25950302bd01a259ae58
|
| 3 |
-
size 967527492
|
|
|
|
|
|
|
|
|
|
|
|
ggml-large-v3-encoder.mlmodelc/analytics/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 243
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2c012cd81c638ceccf50d4eb96a26f7a440a3a0c677a35120b927a047d1bccc5
|
| 3 |
size 243
|
ggml-large-v3-encoder.mlmodelc/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:435c32326d0bd2151a3fefa83c80af6c7bab4b3f797aeb6e5f18ab131dd2459b
|
| 3 |
+
size 380
|
ggml-large-v3-encoder.mlmodelc/metadata.json
CHANGED
|
@@ -17,36 +17,38 @@
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
-
"specificationVersion" :
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
-
"
|
| 23 |
-
"
|
| 24 |
-
"
|
| 25 |
-
"
|
| 26 |
-
"
|
| 27 |
-
"
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
-
"Ios17.mul" : 64,
|
| 31 |
-
"Ios17.transpose" : 129
|
| 32 |
},
|
| 33 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 34 |
"isUpdatable" : "0",
|
|
|
|
|
|
|
|
|
|
| 35 |
"availability" : {
|
| 36 |
-
"macOS" : "
|
| 37 |
-
"tvOS" : "
|
| 38 |
-
"visionOS" : "
|
| 39 |
-
"watchOS" : "
|
| 40 |
-
"iOS" : "
|
| 41 |
-
"macCatalyst" : "
|
| 42 |
},
|
| 43 |
"modelType" : {
|
| 44 |
"name" : "MLModelType_mlProgram"
|
| 45 |
},
|
| 46 |
"userDefinedMetadata" : {
|
| 47 |
-
"com.github.apple.coremltools.
|
| 48 |
-
"com.github.apple.coremltools.source" : "torch==2.
|
| 49 |
-
"com.github.apple.coremltools.version" : "
|
|
|
|
| 50 |
},
|
| 51 |
"inputSchema" : [
|
| 52 |
{
|
|
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
+
"specificationVersion" : 9,
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
+
"Ios18.scaledDotProductAttention" : 32,
|
| 23 |
+
"Ios18.linear" : 192,
|
| 24 |
+
"Ios18.gelu" : 34,
|
| 25 |
+
"Ios18.layerNorm" : 65,
|
| 26 |
+
"Ios18.transpose" : 129,
|
| 27 |
+
"Ios18.conv" : 2,
|
| 28 |
+
"Ios18.add" : 65,
|
| 29 |
+
"Ios18.reshape" : 128
|
|
|
|
|
|
|
| 30 |
},
|
| 31 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 32 |
"isUpdatable" : "0",
|
| 33 |
+
"stateSchema" : [
|
| 34 |
+
|
| 35 |
+
],
|
| 36 |
"availability" : {
|
| 37 |
+
"macOS" : "15.0",
|
| 38 |
+
"tvOS" : "18.0",
|
| 39 |
+
"visionOS" : "2.0",
|
| 40 |
+
"watchOS" : "11.0",
|
| 41 |
+
"iOS" : "18.0",
|
| 42 |
+
"macCatalyst" : "18.0"
|
| 43 |
},
|
| 44 |
"modelType" : {
|
| 45 |
"name" : "MLModelType_mlProgram"
|
| 46 |
},
|
| 47 |
"userDefinedMetadata" : {
|
| 48 |
+
"com.github.apple.coremltools.conversion_date" : "2025-12-13",
|
| 49 |
+
"com.github.apple.coremltools.source" : "torch==2.9.1",
|
| 50 |
+
"com.github.apple.coremltools.version" : "9.0",
|
| 51 |
+
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
| 52 |
},
|
| 53 |
"inputSchema" : [
|
| 54 |
{
|
ggml-large-v3-encoder.mlmodelc/model.mil
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-large-v3-encoder.mlmodelc/weights/weight.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7cdd15ff2f714bdca758058f70490c89617f764ddc1b25602a6c60eaa408856a
|
| 3 |
+
size 1273971776
|
ggml-large-v3-turbo-encoder.mlmodelc/analytics/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:521d0f794f87e8fea91108f66d373485c89fce52d861e3cca3e4a5a01b6734fd
|
| 3 |
+
size 243
|
ggml-large-v3-turbo-encoder.mlmodelc/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d188b1f3d2d1c8e15381f6c75bb2a2caa03a5094c30136fe897b6a1d91ce907c
|
| 3 |
+
size 380
|
ggml-large-v3-turbo-encoder.mlmodelc/metadata.json
CHANGED
|
@@ -19,17 +19,16 @@
|
|
| 19 |
],
|
| 20 |
"specificationVersion" : 9,
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
-
"Ios18.
|
| 23 |
"Ios18.linear" : 192,
|
| 24 |
"Ios18.gelu" : 34,
|
| 25 |
"Ios18.layerNorm" : 65,
|
| 26 |
"Ios18.transpose" : 129,
|
| 27 |
"Ios18.conv" : 2,
|
| 28 |
-
"Ios18.
|
| 29 |
-
"Ios18.
|
| 30 |
-
"Ios18.add" : 65
|
| 31 |
},
|
| 32 |
-
"computePrecision" : "Mixed (Float16,
|
| 33 |
"isUpdatable" : "0",
|
| 34 |
"stateSchema" : [
|
| 35 |
|
|
@@ -43,18 +42,13 @@
|
|
| 43 |
"macCatalyst" : "18.0"
|
| 44 |
},
|
| 45 |
"modelType" : {
|
| 46 |
-
"name" : "
|
| 47 |
-
"structure" : [
|
| 48 |
-
{
|
| 49 |
-
"name" : "MLModelType_mlProgram"
|
| 50 |
-
},
|
| 51 |
-
{
|
| 52 |
-
"name" : "MLModelType_mlProgram"
|
| 53 |
-
}
|
| 54 |
-
]
|
| 55 |
},
|
| 56 |
"userDefinedMetadata" : {
|
| 57 |
-
|
|
|
|
|
|
|
|
|
|
| 58 |
},
|
| 59 |
"inputSchema" : [
|
| 60 |
{
|
|
@@ -68,7 +62,7 @@
|
|
| 68 |
"type" : "MultiArray"
|
| 69 |
}
|
| 70 |
],
|
| 71 |
-
"generatedClassName" : "
|
| 72 |
"method" : "predict"
|
| 73 |
}
|
| 74 |
]
|
|
|
|
| 19 |
],
|
| 20 |
"specificationVersion" : 9,
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
+
"Ios18.scaledDotProductAttention" : 32,
|
| 23 |
"Ios18.linear" : 192,
|
| 24 |
"Ios18.gelu" : 34,
|
| 25 |
"Ios18.layerNorm" : 65,
|
| 26 |
"Ios18.transpose" : 129,
|
| 27 |
"Ios18.conv" : 2,
|
| 28 |
+
"Ios18.add" : 65,
|
| 29 |
+
"Ios18.reshape" : 128
|
|
|
|
| 30 |
},
|
| 31 |
+
"computePrecision" : "Mixed (Float16, Int32)",
|
| 32 |
"isUpdatable" : "0",
|
| 33 |
"stateSchema" : [
|
| 34 |
|
|
|
|
| 42 |
"macCatalyst" : "18.0"
|
| 43 |
},
|
| 44 |
"modelType" : {
|
| 45 |
+
"name" : "MLModelType_mlProgram"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
},
|
| 47 |
"userDefinedMetadata" : {
|
| 48 |
+
"com.github.apple.coremltools.conversion_date" : "2025-12-13",
|
| 49 |
+
"com.github.apple.coremltools.source" : "torch==2.9.1",
|
| 50 |
+
"com.github.apple.coremltools.version" : "9.0",
|
| 51 |
+
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
| 52 |
},
|
| 53 |
"inputSchema" : [
|
| 54 |
{
|
|
|
|
| 62 |
"type" : "MultiArray"
|
| 63 |
}
|
| 64 |
],
|
| 65 |
+
"generatedClassName" : "ggml_large_v3_turbo_encoder",
|
| 66 |
"method" : "predict"
|
| 67 |
}
|
| 68 |
]
|
ggml-large-v3-turbo-encoder.mlmodelc/model.mil
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-large-v3-turbo-encoder.mlmodelc/model0/coremldata.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:a2b0461e225831cc34e0017a300f867929784559e2ee471f01ddfd3452381076
|
| 3 |
-
size 201
|
|
|
|
|
|
|
|
|
|
|
|
ggml-large-v3-turbo-encoder.mlmodelc/model0/model.mil
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-large-v3-turbo-encoder.mlmodelc/model0/weights/0-weight.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:6d26bc07916a863ace6c1bb750aad546bdc7b16c12ebe198d3b511134b09eb68
|
| 3 |
-
size 644314048
|
|
|
|
|
|
|
|
|
|
|
|
ggml-large-v3-turbo-encoder.mlmodelc/model1/analytics/coremldata.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:5a8281049b2a65a3be541cfd9f949e84b8fe1c5251ce90e46da1626fed54e58a
|
| 3 |
-
size 108
|
|
|
|
|
|
|
|
|
|
|
|
ggml-large-v3-turbo-encoder.mlmodelc/model1/coremldata.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:ee5eef16c4adf0aac1e84f5676c8942aa2741fe919f65d72b1b23bd9dd417ab2
|
| 3 |
-
size 196
|
|
|
|
|
|
|
|
|
|
|
|
ggml-large-v3-turbo-encoder.mlmodelc/model1/model.mil
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-large-v3-turbo-encoder.mlmodelc/model1/weights/1-weight.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:28e9965c3b4ed7db29d226652d29ca6f09a35ee68d2cd61100e7b0f4d0a7095c
|
| 3 |
-
size 629660416
|
|
|
|
|
|
|
|
|
|
|
|
{ggml-large-v2-encoder.mlmodelc → ggml-large-v3-turbo-encoder.mlmodelc}/weights/weight.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e77bb855473f07c5e56a2343dcf1f3af9c80b6e61ce4247268d67f929e47a1c8
|
| 3 |
+
size 1273971776
|
ggml-medium-encoder.mlmodelc/analytics/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 243
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f3328e031445e499158cd4fc091b4ca7da84476e629db98372b1c52ed49e8f49
|
| 3 |
size 243
|
ggml-medium-encoder.mlmodelc/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1085b1ad3f7ed6de1553948e97404683f50b555a59ee26d06b8c425e8f8f30d1
|
| 3 |
+
size 379
|
ggml-medium-encoder.mlmodelc/metadata.json
CHANGED
|
@@ -17,36 +17,38 @@
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
-
"specificationVersion" :
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
-
"
|
| 23 |
-
"
|
| 24 |
-
"
|
| 25 |
-
"
|
| 26 |
-
"
|
| 27 |
-
"
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
-
"Ios17.mul" : 48,
|
| 31 |
-
"Ios17.transpose" : 97
|
| 32 |
},
|
| 33 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 34 |
"isUpdatable" : "0",
|
|
|
|
|
|
|
|
|
|
| 35 |
"availability" : {
|
| 36 |
-
"macOS" : "
|
| 37 |
-
"tvOS" : "
|
| 38 |
-
"visionOS" : "
|
| 39 |
-
"watchOS" : "
|
| 40 |
-
"iOS" : "
|
| 41 |
-
"macCatalyst" : "
|
| 42 |
},
|
| 43 |
"modelType" : {
|
| 44 |
"name" : "MLModelType_mlProgram"
|
| 45 |
},
|
| 46 |
"userDefinedMetadata" : {
|
| 47 |
-
"com.github.apple.coremltools.
|
| 48 |
-
"com.github.apple.coremltools.source" : "torch==2.
|
| 49 |
-
"com.github.apple.coremltools.version" : "
|
|
|
|
| 50 |
},
|
| 51 |
"inputSchema" : [
|
| 52 |
{
|
|
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
+
"specificationVersion" : 9,
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
+
"Ios18.scaledDotProductAttention" : 24,
|
| 23 |
+
"Ios18.linear" : 144,
|
| 24 |
+
"Ios18.gelu" : 26,
|
| 25 |
+
"Ios18.layerNorm" : 49,
|
| 26 |
+
"Ios18.transpose" : 97,
|
| 27 |
+
"Ios18.conv" : 2,
|
| 28 |
+
"Ios18.add" : 49,
|
| 29 |
+
"Ios18.reshape" : 96
|
|
|
|
|
|
|
| 30 |
},
|
| 31 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 32 |
"isUpdatable" : "0",
|
| 33 |
+
"stateSchema" : [
|
| 34 |
+
|
| 35 |
+
],
|
| 36 |
"availability" : {
|
| 37 |
+
"macOS" : "15.0",
|
| 38 |
+
"tvOS" : "18.0",
|
| 39 |
+
"visionOS" : "2.0",
|
| 40 |
+
"watchOS" : "11.0",
|
| 41 |
+
"iOS" : "18.0",
|
| 42 |
+
"macCatalyst" : "18.0"
|
| 43 |
},
|
| 44 |
"modelType" : {
|
| 45 |
"name" : "MLModelType_mlProgram"
|
| 46 |
},
|
| 47 |
"userDefinedMetadata" : {
|
| 48 |
+
"com.github.apple.coremltools.conversion_date" : "2025-12-13",
|
| 49 |
+
"com.github.apple.coremltools.source" : "torch==2.9.1",
|
| 50 |
+
"com.github.apple.coremltools.version" : "9.0",
|
| 51 |
+
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
| 52 |
},
|
| 53 |
"inputSchema" : [
|
| 54 |
{
|
ggml-medium-encoder.mlmodelc/model.mil
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-medium-encoder.mlmodelc/weights/weight.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 614458432
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2cd1baef4c7d8260ea817ea56705b3700155c01e8d3ea4bc8e364a8674a88d15
|
| 3 |
size 614458432
|
ggml-base.bin → ggml-medium-q8_0.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1562f96f3dd57b9517f286ea4b5843ed70759bce397035ffe84f0bb2d407d62b
|
| 3 |
+
size 488364756
|
ggml-medium.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:076f80d119389b32fb3ff7ee78330418de19970090183807dfa3da56a4f69fda
|
| 3 |
-
size 915642516
|
|
|
|
|
|
|
|
|
|
|
|
ggml-small-encoder.mlmodelc/analytics/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 243
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad9997a136336812138a5f02e4977a2a83c5112a5c2928beb7b43bafc55853df
|
| 3 |
size 243
|
ggml-small-encoder.mlmodelc/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c6253da76ccc8aefef97c4bc85a4f2a5687e4ed4db03a41c34efff4a7ec9de3e
|
| 3 |
+
size 379
|
ggml-small-encoder.mlmodelc/metadata.json
CHANGED
|
@@ -17,36 +17,38 @@
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
-
"specificationVersion" :
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
-
"
|
| 23 |
-
"
|
| 24 |
-
"
|
| 25 |
-
"
|
| 26 |
-
"
|
| 27 |
-
"
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
-
"Ios17.mul" : 24,
|
| 31 |
-
"Ios17.transpose" : 49
|
| 32 |
},
|
| 33 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 34 |
"isUpdatable" : "0",
|
|
|
|
|
|
|
|
|
|
| 35 |
"availability" : {
|
| 36 |
-
"macOS" : "
|
| 37 |
-
"tvOS" : "
|
| 38 |
-
"visionOS" : "
|
| 39 |
-
"watchOS" : "
|
| 40 |
-
"iOS" : "
|
| 41 |
-
"macCatalyst" : "
|
| 42 |
},
|
| 43 |
"modelType" : {
|
| 44 |
"name" : "MLModelType_mlProgram"
|
| 45 |
},
|
| 46 |
"userDefinedMetadata" : {
|
| 47 |
-
"com.github.apple.coremltools.
|
| 48 |
-
"com.github.apple.coremltools.source" : "torch==2.
|
| 49 |
-
"com.github.apple.coremltools.version" : "
|
|
|
|
| 50 |
},
|
| 51 |
"inputSchema" : [
|
| 52 |
{
|
|
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
+
"specificationVersion" : 9,
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
+
"Ios18.scaledDotProductAttention" : 12,
|
| 23 |
+
"Ios18.linear" : 72,
|
| 24 |
+
"Ios18.gelu" : 14,
|
| 25 |
+
"Ios18.layerNorm" : 25,
|
| 26 |
+
"Ios18.transpose" : 49,
|
| 27 |
+
"Ios18.conv" : 2,
|
| 28 |
+
"Ios18.add" : 25,
|
| 29 |
+
"Ios18.reshape" : 48
|
|
|
|
|
|
|
| 30 |
},
|
| 31 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 32 |
"isUpdatable" : "0",
|
| 33 |
+
"stateSchema" : [
|
| 34 |
+
|
| 35 |
+
],
|
| 36 |
"availability" : {
|
| 37 |
+
"macOS" : "15.0",
|
| 38 |
+
"tvOS" : "18.0",
|
| 39 |
+
"visionOS" : "2.0",
|
| 40 |
+
"watchOS" : "11.0",
|
| 41 |
+
"iOS" : "18.0",
|
| 42 |
+
"macCatalyst" : "18.0"
|
| 43 |
},
|
| 44 |
"modelType" : {
|
| 45 |
"name" : "MLModelType_mlProgram"
|
| 46 |
},
|
| 47 |
"userDefinedMetadata" : {
|
| 48 |
+
"com.github.apple.coremltools.conversion_date" : "2025-12-13",
|
| 49 |
+
"com.github.apple.coremltools.source" : "torch==2.9.1",
|
| 50 |
+
"com.github.apple.coremltools.version" : "9.0",
|
| 51 |
+
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
| 52 |
},
|
| 53 |
"inputSchema" : [
|
| 54 |
{
|
ggml-small-encoder.mlmodelc/model.mil
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ggml-small-encoder.mlmodelc/weights/weight.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 176321856
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d3ab676977d57b06993ee7ebc638fc8568a99ddb11cb7a445328ce50fbd8b36
|
| 3 |
size 176321856
|
ggml-large-v2-encoder.mlmodelc/coremldata.bin → ggml-small-q8_0.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2f9f98684c133761578bb32314e4b978c203bc6d4d1b6f5553c356104b9a7571
|
| 3 |
+
size 165242356
|
ggml-small.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:42ea1b1ba6ffa8cde9da77664db1ac5e24378f9d3d0363ca0104deebedac7732
|
| 3 |
-
size 308753476
|
|
|
|
|
|
|
|
|
|
|
|
ggml-tiny-encoder.mlmodelc/analytics/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 243
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e4caac0b38c9cca17df0f212e16a01d7046e2f454767565a9161988640492be8
|
| 3 |
size 243
|
ggml-tiny-encoder.mlmodelc/coremldata.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:de0977d50f2b62990f3fe458c5f94e6e05147e197dfce5ce9c6fadc370ea0e7f
|
| 3 |
+
size 379
|
ggml-tiny-encoder.mlmodelc/metadata.json
CHANGED
|
@@ -17,36 +17,38 @@
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
-
"specificationVersion" :
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
-
"
|
| 23 |
-
"
|
| 24 |
-
"
|
| 25 |
-
"
|
| 26 |
-
"
|
| 27 |
-
"
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
-
"Ios17.mul" : 8,
|
| 31 |
-
"Ios17.transpose" : 17
|
| 32 |
},
|
| 33 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 34 |
"isUpdatable" : "0",
|
|
|
|
|
|
|
|
|
|
| 35 |
"availability" : {
|
| 36 |
-
"macOS" : "
|
| 37 |
-
"tvOS" : "
|
| 38 |
-
"visionOS" : "
|
| 39 |
-
"watchOS" : "
|
| 40 |
-
"iOS" : "
|
| 41 |
-
"macCatalyst" : "
|
| 42 |
},
|
| 43 |
"modelType" : {
|
| 44 |
"name" : "MLModelType_mlProgram"
|
| 45 |
},
|
| 46 |
"userDefinedMetadata" : {
|
| 47 |
-
"com.github.apple.coremltools.
|
| 48 |
-
"com.github.apple.coremltools.source" : "torch==2.
|
| 49 |
-
"com.github.apple.coremltools.version" : "
|
|
|
|
| 50 |
},
|
| 51 |
"inputSchema" : [
|
| 52 |
{
|
|
|
|
| 17 |
"modelParameters" : [
|
| 18 |
|
| 19 |
],
|
| 20 |
+
"specificationVersion" : 9,
|
| 21 |
"mlProgramOperationTypeHistogram" : {
|
| 22 |
+
"Ios18.scaledDotProductAttention" : 4,
|
| 23 |
+
"Ios18.linear" : 24,
|
| 24 |
+
"Ios18.gelu" : 6,
|
| 25 |
+
"Ios18.layerNorm" : 9,
|
| 26 |
+
"Ios18.transpose" : 17,
|
| 27 |
+
"Ios18.conv" : 2,
|
| 28 |
+
"Ios18.add" : 9,
|
| 29 |
+
"Ios18.reshape" : 16
|
|
|
|
|
|
|
| 30 |
},
|
| 31 |
"computePrecision" : "Mixed (Float16, Int32)",
|
| 32 |
"isUpdatable" : "0",
|
| 33 |
+
"stateSchema" : [
|
| 34 |
+
|
| 35 |
+
],
|
| 36 |
"availability" : {
|
| 37 |
+
"macOS" : "15.0",
|
| 38 |
+
"tvOS" : "18.0",
|
| 39 |
+
"visionOS" : "2.0",
|
| 40 |
+
"watchOS" : "11.0",
|
| 41 |
+
"iOS" : "18.0",
|
| 42 |
+
"macCatalyst" : "18.0"
|
| 43 |
},
|
| 44 |
"modelType" : {
|
| 45 |
"name" : "MLModelType_mlProgram"
|
| 46 |
},
|
| 47 |
"userDefinedMetadata" : {
|
| 48 |
+
"com.github.apple.coremltools.conversion_date" : "2025-12-13",
|
| 49 |
+
"com.github.apple.coremltools.source" : "torch==2.9.1",
|
| 50 |
+
"com.github.apple.coremltools.version" : "9.0",
|
| 51 |
+
"com.github.apple.coremltools.source_dialect" : "TorchScript"
|
| 52 |
},
|
| 53 |
"inputSchema" : [
|
| 54 |
{
|
ggml-tiny-encoder.mlmodelc/model.mil
CHANGED
|
@@ -1,268 +1,224 @@
|
|
| 1 |
-
program(1.
|
| 2 |
-
[buildInfo = dict<
|
| 3 |
{
|
| 4 |
-
func main<
|
| 5 |
-
|
| 6 |
-
tensor<int32, [
|
| 7 |
-
tensor<int32, [1]>
|
| 8 |
-
tensor<
|
| 9 |
-
|
| 10 |
-
tensor<fp16, [384, 80, 3]>
|
| 11 |
-
tensor<fp16, [384]>
|
| 12 |
-
tensor<fp16, [1, 384, 3000]> var_28_cast_fp16 = conv(bias =
|
| 13 |
-
|
| 14 |
-
tensor<fp16, [1, 384, 3000]> input_1_cast_fp16 = gelu(mode = input_1_mode_0, x = var_28_cast_fp16)[name =
|
| 15 |
-
|
| 16 |
-
tensor<int32, [
|
| 17 |
-
tensor<int32, [1]>
|
| 18 |
-
tensor<
|
| 19 |
-
|
| 20 |
-
tensor<fp16, [384, 384, 3]>
|
| 21 |
-
tensor<fp16, [384]>
|
| 22 |
-
tensor<fp16, [1, 384, 1500]> var_46_cast_fp16 = conv(bias =
|
| 23 |
-
|
| 24 |
-
tensor<fp16, [1, 384, 1500]> x_3_cast_fp16 = gelu(mode = x_3_mode_0, x = var_46_cast_fp16)[name =
|
| 25 |
-
tensor<int32, [3]> var_52 = const()[name =
|
| 26 |
-
tensor<fp16, [1500, 384]> positional_embedding_to_fp16 = const()[name =
|
| 27 |
-
tensor<fp16, [1, 1500, 384]>
|
| 28 |
-
tensor<fp16, [1, 1500, 384]> var_55_cast_fp16 = add(x =
|
| 29 |
-
tensor<int32, []>
|
| 30 |
-
tensor<
|
| 31 |
-
tensor<fp16, [384]>
|
| 32 |
-
|
| 33 |
-
tensor<fp16, []>
|
| 34 |
-
tensor<fp16, [
|
| 35 |
-
tensor<fp16, [384
|
| 36 |
-
tensor<fp16, [384]>
|
| 37 |
-
tensor<fp16, [
|
| 38 |
-
tensor<fp16, [384
|
| 39 |
-
tensor<fp16, [384]>
|
| 40 |
-
tensor<fp16, [
|
| 41 |
-
tensor<fp16, [384
|
| 42 |
-
tensor<fp16, [384]>
|
| 43 |
-
tensor<
|
| 44 |
-
tensor<
|
| 45 |
-
tensor<
|
| 46 |
-
tensor<fp16, [1,
|
| 47 |
-
tensor<
|
| 48 |
-
tensor<
|
| 49 |
-
tensor<
|
| 50 |
-
tensor<
|
| 51 |
-
tensor<
|
| 52 |
-
tensor<
|
| 53 |
-
tensor<fp16, [1,
|
| 54 |
-
tensor<
|
| 55 |
-
tensor<
|
| 56 |
-
tensor<
|
| 57 |
-
tensor<int32, [
|
| 58 |
-
tensor<
|
| 59 |
-
tensor<fp16, [1,
|
| 60 |
-
tensor<fp16, [
|
| 61 |
-
tensor<fp16, [
|
| 62 |
-
tensor<fp16, [1,
|
| 63 |
-
tensor<
|
| 64 |
-
tensor<
|
| 65 |
-
tensor<fp16, [
|
| 66 |
-
tensor<fp16, [
|
| 67 |
-
tensor<
|
| 68 |
-
tensor<
|
| 69 |
-
tensor<fp16, [
|
| 70 |
-
tensor<fp16, [1, 1500,
|
| 71 |
-
|
| 72 |
-
tensor<fp16, [
|
| 73 |
-
tensor<fp16, [
|
| 74 |
-
tensor<fp16, [
|
| 75 |
-
tensor<
|
| 76 |
-
tensor<fp16, [384]>
|
| 77 |
-
tensor<
|
| 78 |
-
tensor<fp16, [
|
| 79 |
-
tensor<fp16, [
|
| 80 |
-
|
| 81 |
-
tensor<fp16, [1, 1500,
|
| 82 |
-
tensor<
|
| 83 |
-
tensor<fp16, [
|
| 84 |
-
tensor<fp16, [
|
| 85 |
-
tensor<fp16, [384]>
|
| 86 |
-
tensor<fp16, [1, 1500, 384]>
|
| 87 |
-
tensor<fp16, [
|
| 88 |
-
tensor<
|
| 89 |
-
tensor<
|
| 90 |
-
tensor<
|
| 91 |
-
tensor<fp16, [
|
| 92 |
-
tensor<
|
| 93 |
-
tensor<fp16, [1, 1500,
|
| 94 |
-
tensor<
|
| 95 |
-
tensor<fp16, [
|
| 96 |
-
tensor<
|
| 97 |
-
tensor<
|
| 98 |
-
tensor<
|
| 99 |
-
tensor<fp16, [
|
| 100 |
-
tensor<fp16, [
|
| 101 |
-
tensor<fp16, [1, 1500,
|
| 102 |
-
tensor<
|
| 103 |
-
tensor<
|
| 104 |
-
tensor<
|
| 105 |
-
tensor<fp16, [1, 1500, 6, 64]>
|
| 106 |
-
tensor<
|
| 107 |
-
tensor<fp16, [
|
| 108 |
-
tensor<fp16, [
|
| 109 |
-
tensor<fp16, [1, 1500,
|
| 110 |
-
tensor<
|
| 111 |
-
tensor<
|
| 112 |
-
tensor<
|
| 113 |
-
tensor<
|
| 114 |
-
tensor<
|
| 115 |
-
tensor<
|
| 116 |
-
tensor<
|
| 117 |
-
tensor<fp16, [1,
|
| 118 |
-
|
| 119 |
-
tensor<fp16, [1,
|
| 120 |
-
tensor<fp16, [
|
| 121 |
-
tensor<
|
| 122 |
-
tensor<
|
| 123 |
-
tensor<fp16, [1,
|
| 124 |
-
tensor<
|
| 125 |
-
tensor<
|
| 126 |
-
tensor<
|
| 127 |
-
|
| 128 |
-
tensor<fp16, [1, 1500, 384]>
|
| 129 |
-
tensor<fp16, [384, 384]>
|
| 130 |
-
tensor<fp16, [384]>
|
| 131 |
-
tensor<fp16, [1, 1500, 384]>
|
| 132 |
-
tensor<fp16, [
|
| 133 |
-
tensor<
|
| 134 |
-
tensor<fp16, [384]>
|
| 135 |
-
tensor<fp16, [384]>
|
| 136 |
-
tensor<fp16, [1, 1500, 384]>
|
| 137 |
-
tensor<
|
| 138 |
-
tensor<fp16, [
|
| 139 |
-
tensor<
|
| 140 |
-
tensor<
|
| 141 |
-
tensor<
|
| 142 |
-
tensor<fp16, [
|
| 143 |
-
tensor<
|
| 144 |
-
tensor<
|
| 145 |
-
tensor<
|
| 146 |
-
tensor<
|
| 147 |
-
tensor<
|
| 148 |
-
tensor<fp16, [
|
| 149 |
-
tensor<fp16, [
|
| 150 |
-
tensor<
|
| 151 |
-
tensor<
|
| 152 |
-
tensor<fp16, [
|
| 153 |
-
tensor<fp16, [384]>
|
| 154 |
-
tensor<fp16, [
|
| 155 |
-
tensor<fp16, [384
|
| 156 |
-
tensor<fp16, [1, 1500, 384]>
|
| 157 |
-
tensor<fp16, [
|
| 158 |
-
tensor<
|
| 159 |
-
tensor<fp16, [
|
| 160 |
-
tensor<
|
| 161 |
-
tensor<fp16, [1, 1500,
|
| 162 |
-
tensor<fp16, [
|
| 163 |
-
tensor<fp16, [
|
| 164 |
-
tensor<
|
| 165 |
-
|
| 166 |
-
tensor<fp16, [1,
|
| 167 |
-
tensor<fp16, [
|
| 168 |
-
tensor<
|
| 169 |
-
tensor<fp16, [1, 1500,
|
| 170 |
-
tensor<
|
| 171 |
-
tensor<
|
| 172 |
-
tensor<
|
| 173 |
-
tensor<
|
| 174 |
-
|
| 175 |
-
tensor<fp16, [1,
|
| 176 |
-
tensor<fp16, [
|
| 177 |
-
tensor<fp16, [
|
| 178 |
-
tensor<fp16, [1,
|
| 179 |
-
tensor<
|
| 180 |
-
tensor<
|
| 181 |
-
tensor<fp16, [
|
| 182 |
-
tensor<fp16, [
|
| 183 |
-
tensor<
|
| 184 |
-
tensor<int32, [
|
| 185 |
-
tensor<fp16, [1, 1500, 6, 64]>
|
| 186 |
-
tensor<
|
| 187 |
-
tensor<fp16, [
|
| 188 |
-
tensor<
|
| 189 |
-
tensor<fp16, [1, 1500,
|
| 190 |
-
tensor<
|
| 191 |
-
tensor<int32, [
|
| 192 |
-
tensor<
|
| 193 |
-
tensor<fp16, [
|
| 194 |
-
tensor<fp16, [1, 1500,
|
| 195 |
-
tensor<fp16, [
|
| 196 |
-
tensor<fp16, [
|
| 197 |
-
tensor<
|
| 198 |
-
tensor<
|
| 199 |
-
tensor<fp16, [1, 1500,
|
| 200 |
-
tensor<fp16, [
|
| 201 |
-
tensor<fp16, [384]>
|
| 202 |
-
tensor<fp16, [
|
| 203 |
-
tensor<fp16, [1, 1500, 384]>
|
| 204 |
-
tensor<
|
| 205 |
-
tensor<int32, [1]>
|
| 206 |
-
tensor<fp16, [384]>
|
| 207 |
-
tensor<fp16, [384]>
|
| 208 |
-
tensor<fp16, []>
|
| 209 |
-
tensor<fp16, [
|
| 210 |
-
tensor<fp16, [
|
| 211 |
-
tensor<fp16, [
|
| 212 |
-
|
| 213 |
-
tensor<fp16, [
|
| 214 |
-
tensor<fp16, [
|
| 215 |
-
tensor<fp16, [384
|
| 216 |
-
tensor<fp16, [384]>
|
| 217 |
-
tensor<fp16, [1, 1500, 384]>
|
| 218 |
-
tensor<int32, [
|
| 219 |
-
tensor<fp16, [
|
| 220 |
-
tensor<fp16, [
|
| 221 |
-
|
| 222 |
-
tensor<
|
| 223 |
-
tensor<fp16, [1, 1500, 6, 64]> var_434_cast_fp16 = reshape(shape = var_433, x = linear_19_cast_fp16)[name = tensor<string, []>("op_434_cast_fp16")];
|
| 224 |
-
tensor<fp16, [1, 1, 1, 1]> const_35_to_fp16 = const()[name = tensor<string, []>("const_35_to_fp16"), val = tensor<fp16, [1, 1, 1, 1]>([[[[0x1.6ap-2]]]])];
|
| 225 |
-
tensor<fp16, [1, 1500, 6, 64]> k_cast_fp16 = mul(x = var_434_cast_fp16, y = const_35_to_fp16)[name = tensor<string, []>("k_cast_fp16")];
|
| 226 |
-
tensor<int32, [4]> var_440 = const()[name = tensor<string, []>("op_440"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 227 |
-
tensor<fp16, [1, 1500, 6, 64]> var_441_cast_fp16 = reshape(shape = var_440, x = linear_20_cast_fp16)[name = tensor<string, []>("op_441_cast_fp16")];
|
| 228 |
-
tensor<int32, [4]> var_442 = const()[name = tensor<string, []>("op_442"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 229 |
-
tensor<bool, []> qk_transpose_x_0 = const()[name = tensor<string, []>("qk_transpose_x_0"), val = tensor<bool, []>(false)];
|
| 230 |
-
tensor<bool, []> qk_transpose_y_0 = const()[name = tensor<string, []>("qk_transpose_y_0"), val = tensor<bool, []>(false)];
|
| 231 |
-
tensor<int32, [4]> transpose_22_perm_0 = const()[name = tensor<string, []>("transpose_22_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
|
| 232 |
-
tensor<int32, [4]> transpose_23_perm_0 = const()[name = tensor<string, []>("transpose_23_perm_0"), val = tensor<int32, [4]>([0, 2, -1, -3])];
|
| 233 |
-
tensor<fp16, [1, 6, 64, 1500]> transpose_25 = transpose(perm = transpose_23_perm_0, x = k_cast_fp16)[name = tensor<string, []>("transpose_25")];
|
| 234 |
-
tensor<fp16, [1, 6, 1500, 64]> transpose_26 = transpose(perm = transpose_22_perm_0, x = q_cast_fp16)[name = tensor<string, []>("transpose_26")];
|
| 235 |
-
tensor<fp16, [1, 6, 1500, 1500]> qk_cast_fp16 = matmul(transpose_x = qk_transpose_x_0, transpose_y = qk_transpose_y_0, x = transpose_26, y = transpose_25)[name = tensor<string, []>("qk_cast_fp16")];
|
| 236 |
-
tensor<fp16, [1, 6, 1500, 1500]> var_446_cast_fp16 = softmax(axis = var_382, x = qk_cast_fp16)[name = tensor<string, []>("op_446_cast_fp16")];
|
| 237 |
-
tensor<bool, []> var_448_transpose_x_0 = const()[name = tensor<string, []>("op_448_transpose_x_0"), val = tensor<bool, []>(false)];
|
| 238 |
-
tensor<bool, []> var_448_transpose_y_0 = const()[name = tensor<string, []>("op_448_transpose_y_0"), val = tensor<bool, []>(false)];
|
| 239 |
-
tensor<fp16, [1, 6, 1500, 64]> transpose_27 = transpose(perm = var_442, x = var_441_cast_fp16)[name = tensor<string, []>("transpose_27")];
|
| 240 |
-
tensor<fp16, [1, 6, 1500, 64]> var_448_cast_fp16 = matmul(transpose_x = var_448_transpose_x_0, transpose_y = var_448_transpose_y_0, x = var_446_cast_fp16, y = transpose_27)[name = tensor<string, []>("op_448_cast_fp16")];
|
| 241 |
-
tensor<int32, [4]> var_449 = const()[name = tensor<string, []>("op_449"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 242 |
-
tensor<int32, [3]> concat_3 = const()[name = tensor<string, []>("concat_3"), val = tensor<int32, [3]>([1, 1500, 384])];
|
| 243 |
-
tensor<fp16, [1, 1500, 6, 64]> transpose_24 = transpose(perm = var_449, x = var_448_cast_fp16)[name = tensor<string, []>("transpose_24")];
|
| 244 |
-
tensor<fp16, [1, 1500, 384]> x_47_cast_fp16 = reshape(shape = concat_3, x = transpose_24)[name = tensor<string, []>("x_47_cast_fp16")];
|
| 245 |
-
tensor<fp16, [384, 384]> var_454_to_fp16 = const()[name = tensor<string, []>("op_454_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(13759424)))];
|
| 246 |
-
tensor<fp16, [384]> var_455_to_fp16 = const()[name = tensor<string, []>("op_455_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14054400)))];
|
| 247 |
-
tensor<fp16, [1, 1500, 384]> linear_21_cast_fp16 = linear(bias = var_455_to_fp16, weight = var_454_to_fp16, x = x_47_cast_fp16)[name = tensor<string, []>("linear_21_cast_fp16")];
|
| 248 |
-
tensor<fp16, [1, 1500, 384]> x_49_cast_fp16 = add(x = x_43_cast_fp16, y = linear_21_cast_fp16)[name = tensor<string, []>("x_49_cast_fp16")];
|
| 249 |
-
tensor<int32, [1]> var_462_axes_0 = const()[name = tensor<string, []>("op_462_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 250 |
-
tensor<fp16, [384]> blocks_3_mlp_ln_weight_to_fp16 = const()[name = tensor<string, []>("blocks_3_mlp_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14055232)))];
|
| 251 |
-
tensor<fp16, [384]> blocks_3_mlp_ln_bias_to_fp16 = const()[name = tensor<string, []>("blocks_3_mlp_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14056064)))];
|
| 252 |
-
tensor<fp16, [1, 1500, 384]> var_462_cast_fp16 = layer_norm(axes = var_462_axes_0, beta = blocks_3_mlp_ln_bias_to_fp16, epsilon = var_388_to_fp16, gamma = blocks_3_mlp_ln_weight_to_fp16, x = x_49_cast_fp16)[name = tensor<string, []>("op_462_cast_fp16")];
|
| 253 |
-
tensor<fp16, [1536, 384]> var_471_to_fp16 = const()[name = tensor<string, []>("op_471_to_fp16"), val = tensor<fp16, [1536, 384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(14056896)))];
|
| 254 |
-
tensor<fp16, [1536]> var_472_to_fp16 = const()[name = tensor<string, []>("op_472_to_fp16"), val = tensor<fp16, [1536]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15236608)))];
|
| 255 |
-
tensor<fp16, [1, 1500, 1536]> linear_22_cast_fp16 = linear(bias = var_472_to_fp16, weight = var_471_to_fp16, x = var_462_cast_fp16)[name = tensor<string, []>("linear_22_cast_fp16")];
|
| 256 |
-
tensor<string, []> x_53_mode_0 = const()[name = tensor<string, []>("x_53_mode_0"), val = tensor<string, []>("EXACT")];
|
| 257 |
-
tensor<fp16, [1, 1500, 1536]> x_53_cast_fp16 = gelu(mode = x_53_mode_0, x = linear_22_cast_fp16)[name = tensor<string, []>("x_53_cast_fp16")];
|
| 258 |
-
tensor<fp16, [384, 1536]> var_477_to_fp16 = const()[name = tensor<string, []>("op_477_to_fp16"), val = tensor<fp16, [384, 1536]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(15239744)))];
|
| 259 |
-
tensor<fp16, [384]> var_478_to_fp16 = const()[name = tensor<string, []>("op_478_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16419456)))];
|
| 260 |
-
tensor<fp16, [1, 1500, 384]> linear_23_cast_fp16 = linear(bias = var_478_to_fp16, weight = var_477_to_fp16, x = x_53_cast_fp16)[name = tensor<string, []>("linear_23_cast_fp16")];
|
| 261 |
-
tensor<fp16, [1, 1500, 384]> x_cast_fp16 = add(x = x_49_cast_fp16, y = linear_23_cast_fp16)[name = tensor<string, []>("x_cast_fp16")];
|
| 262 |
-
tensor<int32, [1]> var_491_axes_0 = const()[name = tensor<string, []>("op_491_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 263 |
-
tensor<fp16, [384]> ln_post_weight_to_fp16 = const()[name = tensor<string, []>("ln_post_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16420288)))];
|
| 264 |
-
tensor<fp16, [384]> ln_post_bias_to_fp16 = const()[name = tensor<string, []>("ln_post_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(16421120)))];
|
| 265 |
-
tensor<fp16, []> var_482_to_fp16 = const()[name = tensor<string, []>("op_482_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
|
| 266 |
-
tensor<fp16, [1, 1500, 384]> output = layer_norm(axes = var_491_axes_0, beta = ln_post_bias_to_fp16, epsilon = var_482_to_fp16, gamma = ln_post_weight_to_fp16, x = x_cast_fp16)[name = tensor<string, []>("op_491_cast_fp16")];
|
| 267 |
} -> (output);
|
| 268 |
}
|
|
|
|
| 1 |
+
program(1.3)
|
| 2 |
+
[buildInfo = dict<string, string>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.1"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0"}})]
|
| 3 |
{
|
| 4 |
+
func main<ios18>(tensor<fp16, [1, 80, 3000]> logmel_data) {
|
| 5 |
+
string var_28_pad_type_0 = const()[name = string("op_28_pad_type_0"), val = string("custom")];
|
| 6 |
+
tensor<int32, [2]> var_28_pad_0 = const()[name = string("op_28_pad_0"), val = tensor<int32, [2]>([1, 1])];
|
| 7 |
+
tensor<int32, [1]> var_28_strides_0 = const()[name = string("op_28_strides_0"), val = tensor<int32, [1]>([1])];
|
| 8 |
+
tensor<int32, [1]> var_28_dilations_0 = const()[name = string("op_28_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 9 |
+
int32 var_28_groups_0 = const()[name = string("op_28_groups_0"), val = int32(1)];
|
| 10 |
+
tensor<fp16, [384, 80, 3]> const_0_to_fp16 = const()[name = string("const_0_to_fp16"), val = tensor<fp16, [384, 80, 3]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
|
| 11 |
+
tensor<fp16, [384]> const_1_to_fp16 = const()[name = string("const_1_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(184448)))];
|
| 12 |
+
tensor<fp16, [1, 384, 3000]> var_28_cast_fp16 = conv(bias = const_1_to_fp16, dilations = var_28_dilations_0, groups = var_28_groups_0, pad = var_28_pad_0, pad_type = var_28_pad_type_0, strides = var_28_strides_0, weight = const_0_to_fp16, x = logmel_data)[name = string("op_28_cast_fp16")];
|
| 13 |
+
string input_1_mode_0 = const()[name = string("input_1_mode_0"), val = string("EXACT")];
|
| 14 |
+
tensor<fp16, [1, 384, 3000]> input_1_cast_fp16 = gelu(mode = input_1_mode_0, x = var_28_cast_fp16)[name = string("input_1_cast_fp16")];
|
| 15 |
+
string var_46_pad_type_0 = const()[name = string("op_46_pad_type_0"), val = string("custom")];
|
| 16 |
+
tensor<int32, [2]> var_46_pad_0 = const()[name = string("op_46_pad_0"), val = tensor<int32, [2]>([1, 1])];
|
| 17 |
+
tensor<int32, [1]> var_46_strides_0 = const()[name = string("op_46_strides_0"), val = tensor<int32, [1]>([2])];
|
| 18 |
+
tensor<int32, [1]> var_46_dilations_0 = const()[name = string("op_46_dilations_0"), val = tensor<int32, [1]>([1])];
|
| 19 |
+
int32 var_46_groups_0 = const()[name = string("op_46_groups_0"), val = int32(1)];
|
| 20 |
+
tensor<fp16, [384, 384, 3]> const_2_to_fp16 = const()[name = string("const_2_to_fp16"), val = tensor<fp16, [384, 384, 3]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(185280)))];
|
| 21 |
+
tensor<fp16, [384]> const_3_to_fp16 = const()[name = string("const_3_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1070080)))];
|
| 22 |
+
tensor<fp16, [1, 384, 1500]> var_46_cast_fp16 = conv(bias = const_3_to_fp16, dilations = var_46_dilations_0, groups = var_46_groups_0, pad = var_46_pad_0, pad_type = var_46_pad_type_0, strides = var_46_strides_0, weight = const_2_to_fp16, x = input_1_cast_fp16)[name = string("op_46_cast_fp16")];
|
| 23 |
+
string x_3_mode_0 = const()[name = string("x_3_mode_0"), val = string("EXACT")];
|
| 24 |
+
tensor<fp16, [1, 384, 1500]> x_3_cast_fp16 = gelu(mode = x_3_mode_0, x = var_46_cast_fp16)[name = string("x_3_cast_fp16")];
|
| 25 |
+
tensor<int32, [3]> var_52 = const()[name = string("op_52"), val = tensor<int32, [3]>([0, 2, 1])];
|
| 26 |
+
tensor<fp16, [1500, 384]> positional_embedding_to_fp16 = const()[name = string("positional_embedding_to_fp16"), val = tensor<fp16, [1500, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1070912)))];
|
| 27 |
+
tensor<fp16, [1, 1500, 384]> x_5_cast_fp16 = transpose(perm = var_52, x = x_3_cast_fp16)[name = string("transpose_52")];
|
| 28 |
+
tensor<fp16, [1, 1500, 384]> var_55_cast_fp16 = add(x = x_5_cast_fp16, y = positional_embedding_to_fp16)[name = string("op_55_cast_fp16")];
|
| 29 |
+
tensor<int32, [1]> var_82_axes_0 = const()[name = string("op_82_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 30 |
+
tensor<fp16, [384]> blocks_0_attn_ln_weight_to_fp16 = const()[name = string("blocks_0_attn_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2222976)))];
|
| 31 |
+
tensor<fp16, [384]> blocks_0_attn_ln_bias_to_fp16 = const()[name = string("blocks_0_attn_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2223808)))];
|
| 32 |
+
fp16 var_72_to_fp16 = const()[name = string("op_72_to_fp16"), val = fp16(0x1.5p-17)];
|
| 33 |
+
tensor<fp16, [1, 1500, 384]> var_82_cast_fp16 = layer_norm(axes = var_82_axes_0, beta = blocks_0_attn_ln_bias_to_fp16, epsilon = var_72_to_fp16, gamma = blocks_0_attn_ln_weight_to_fp16, x = var_55_cast_fp16)[name = string("op_82_cast_fp16")];
|
| 34 |
+
tensor<fp16, [384, 384]> const_4_to_fp16 = const()[name = string("const_4_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2224640)))];
|
| 35 |
+
tensor<fp16, [384]> const_5_to_fp16 = const()[name = string("const_5_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2519616)))];
|
| 36 |
+
tensor<fp16, [1, 1500, 384]> linear_0_cast_fp16 = linear(bias = const_5_to_fp16, weight = const_4_to_fp16, x = var_82_cast_fp16)[name = string("linear_0_cast_fp16")];
|
| 37 |
+
tensor<fp16, [384, 384]> const_6_to_fp16 = const()[name = string("const_6_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2520448)))];
|
| 38 |
+
tensor<fp16, [384]> linear_1_bias_0_to_fp16 = const()[name = string("linear_1_bias_0_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2815424)))];
|
| 39 |
+
tensor<fp16, [1, 1500, 384]> linear_1_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = const_6_to_fp16, x = var_82_cast_fp16)[name = string("linear_1_cast_fp16")];
|
| 40 |
+
tensor<fp16, [384, 384]> const_7_to_fp16 = const()[name = string("const_7_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(2816256)))];
|
| 41 |
+
tensor<fp16, [384]> const_8_to_fp16 = const()[name = string("const_8_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3111232)))];
|
| 42 |
+
tensor<fp16, [1, 1500, 384]> linear_2_cast_fp16 = linear(bias = const_8_to_fp16, weight = const_7_to_fp16, x = var_82_cast_fp16)[name = string("linear_2_cast_fp16")];
|
| 43 |
+
tensor<int32, [4]> var_106 = const()[name = string("op_106"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 44 |
+
tensor<fp16, [1, 1500, 6, 64]> var_107_cast_fp16 = reshape(shape = var_106, x = linear_0_cast_fp16)[name = string("op_107_cast_fp16")];
|
| 45 |
+
tensor<int32, [4]> var_112 = const()[name = string("op_112"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 46 |
+
tensor<fp16, [1, 1500, 6, 64]> var_113_cast_fp16 = reshape(shape = var_112, x = linear_1_cast_fp16)[name = string("op_113_cast_fp16")];
|
| 47 |
+
tensor<int32, [4]> var_118 = const()[name = string("op_118"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 48 |
+
tensor<fp16, [1, 1500, 6, 64]> var_119_cast_fp16 = reshape(shape = var_118, x = linear_2_cast_fp16)[name = string("op_119_cast_fp16")];
|
| 49 |
+
tensor<int32, [4]> transpose_24_perm_0 = const()[name = string("transpose_24_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 50 |
+
tensor<int32, [4]> transpose_25_perm_0 = const()[name = string("transpose_25_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 51 |
+
tensor<int32, [4]> transpose_26_perm_0 = const()[name = string("transpose_26_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 52 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_26 = transpose(perm = transpose_26_perm_0, x = var_119_cast_fp16)[name = string("transpose_49")];
|
| 53 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_25 = transpose(perm = transpose_25_perm_0, x = var_113_cast_fp16)[name = string("transpose_50")];
|
| 54 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_24 = transpose(perm = transpose_24_perm_0, x = var_107_cast_fp16)[name = string("transpose_51")];
|
| 55 |
+
tensor<fp16, [1, 6, 1500, 64]> a_1_cast_fp16 = scaled_dot_product_attention(key = transpose_25, query = transpose_24, value = transpose_26)[name = string("a_1_cast_fp16")];
|
| 56 |
+
tensor<int32, [4]> var_123 = const()[name = string("op_123"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 57 |
+
tensor<int32, [3]> concat_0 = const()[name = string("concat_0"), val = tensor<int32, [3]>([1, 1500, 384])];
|
| 58 |
+
tensor<fp16, [1, 1500, 6, 64]> var_124_cast_fp16 = transpose(perm = var_123, x = a_1_cast_fp16)[name = string("transpose_48")];
|
| 59 |
+
tensor<fp16, [1, 1500, 384]> x_11_cast_fp16 = reshape(shape = concat_0, x = var_124_cast_fp16)[name = string("x_11_cast_fp16")];
|
| 60 |
+
tensor<fp16, [384, 384]> const_15_to_fp16 = const()[name = string("const_15_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3112064)))];
|
| 61 |
+
tensor<fp16, [384]> const_16_to_fp16 = const()[name = string("const_16_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3407040)))];
|
| 62 |
+
tensor<fp16, [1, 1500, 384]> linear_3_cast_fp16 = linear(bias = const_16_to_fp16, weight = const_15_to_fp16, x = x_11_cast_fp16)[name = string("linear_3_cast_fp16")];
|
| 63 |
+
tensor<fp16, [1, 1500, 384]> x_13_cast_fp16 = add(x = var_55_cast_fp16, y = linear_3_cast_fp16)[name = string("x_13_cast_fp16")];
|
| 64 |
+
tensor<int32, [1]> var_136_axes_0 = const()[name = string("op_136_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 65 |
+
tensor<fp16, [384]> blocks_0_mlp_ln_weight_to_fp16 = const()[name = string("blocks_0_mlp_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3407872)))];
|
| 66 |
+
tensor<fp16, [384]> blocks_0_mlp_ln_bias_to_fp16 = const()[name = string("blocks_0_mlp_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3408704)))];
|
| 67 |
+
tensor<fp16, [1, 1500, 384]> var_136_cast_fp16 = layer_norm(axes = var_136_axes_0, beta = blocks_0_mlp_ln_bias_to_fp16, epsilon = var_72_to_fp16, gamma = blocks_0_mlp_ln_weight_to_fp16, x = x_13_cast_fp16)[name = string("op_136_cast_fp16")];
|
| 68 |
+
tensor<fp16, [1536, 384]> const_17_to_fp16 = const()[name = string("const_17_to_fp16"), val = tensor<fp16, [1536, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(3409536)))];
|
| 69 |
+
tensor<fp16, [1536]> const_18_to_fp16 = const()[name = string("const_18_to_fp16"), val = tensor<fp16, [1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(4589248)))];
|
| 70 |
+
tensor<fp16, [1, 1500, 1536]> linear_4_cast_fp16 = linear(bias = const_18_to_fp16, weight = const_17_to_fp16, x = var_136_cast_fp16)[name = string("linear_4_cast_fp16")];
|
| 71 |
+
string x_17_mode_0 = const()[name = string("x_17_mode_0"), val = string("EXACT")];
|
| 72 |
+
tensor<fp16, [1, 1500, 1536]> x_17_cast_fp16 = gelu(mode = x_17_mode_0, x = linear_4_cast_fp16)[name = string("x_17_cast_fp16")];
|
| 73 |
+
tensor<fp16, [384, 1536]> const_19_to_fp16 = const()[name = string("const_19_to_fp16"), val = tensor<fp16, [384, 1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(4592384)))];
|
| 74 |
+
tensor<fp16, [384]> const_20_to_fp16 = const()[name = string("const_20_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(5772096)))];
|
| 75 |
+
tensor<fp16, [1, 1500, 384]> linear_5_cast_fp16 = linear(bias = const_20_to_fp16, weight = const_19_to_fp16, x = x_17_cast_fp16)[name = string("linear_5_cast_fp16")];
|
| 76 |
+
tensor<fp16, [1, 1500, 384]> x_19_cast_fp16 = add(x = x_13_cast_fp16, y = linear_5_cast_fp16)[name = string("x_19_cast_fp16")];
|
| 77 |
+
tensor<int32, [1]> var_176_axes_0 = const()[name = string("op_176_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 78 |
+
tensor<fp16, [384]> blocks_1_attn_ln_weight_to_fp16 = const()[name = string("blocks_1_attn_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(5772928)))];
|
| 79 |
+
tensor<fp16, [384]> blocks_1_attn_ln_bias_to_fp16 = const()[name = string("blocks_1_attn_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(5773760)))];
|
| 80 |
+
fp16 var_166_to_fp16 = const()[name = string("op_166_to_fp16"), val = fp16(0x1.5p-17)];
|
| 81 |
+
tensor<fp16, [1, 1500, 384]> var_176_cast_fp16 = layer_norm(axes = var_176_axes_0, beta = blocks_1_attn_ln_bias_to_fp16, epsilon = var_166_to_fp16, gamma = blocks_1_attn_ln_weight_to_fp16, x = x_19_cast_fp16)[name = string("op_176_cast_fp16")];
|
| 82 |
+
tensor<fp16, [384, 384]> const_21_to_fp16 = const()[name = string("const_21_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(5774592)))];
|
| 83 |
+
tensor<fp16, [384]> const_22_to_fp16 = const()[name = string("const_22_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6069568)))];
|
| 84 |
+
tensor<fp16, [1, 1500, 384]> linear_6_cast_fp16 = linear(bias = const_22_to_fp16, weight = const_21_to_fp16, x = var_176_cast_fp16)[name = string("linear_6_cast_fp16")];
|
| 85 |
+
tensor<fp16, [384, 384]> const_23_to_fp16 = const()[name = string("const_23_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6070400)))];
|
| 86 |
+
tensor<fp16, [1, 1500, 384]> linear_7_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = const_23_to_fp16, x = var_176_cast_fp16)[name = string("linear_7_cast_fp16")];
|
| 87 |
+
tensor<fp16, [384, 384]> const_24_to_fp16 = const()[name = string("const_24_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6365376)))];
|
| 88 |
+
tensor<fp16, [384]> const_25_to_fp16 = const()[name = string("const_25_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6660352)))];
|
| 89 |
+
tensor<fp16, [1, 1500, 384]> linear_8_cast_fp16 = linear(bias = const_25_to_fp16, weight = const_24_to_fp16, x = var_176_cast_fp16)[name = string("linear_8_cast_fp16")];
|
| 90 |
+
tensor<int32, [4]> var_200 = const()[name = string("op_200"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 91 |
+
tensor<fp16, [1, 1500, 6, 64]> var_201_cast_fp16 = reshape(shape = var_200, x = linear_6_cast_fp16)[name = string("op_201_cast_fp16")];
|
| 92 |
+
tensor<int32, [4]> var_206 = const()[name = string("op_206"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 93 |
+
tensor<fp16, [1, 1500, 6, 64]> var_207_cast_fp16 = reshape(shape = var_206, x = linear_7_cast_fp16)[name = string("op_207_cast_fp16")];
|
| 94 |
+
tensor<int32, [4]> var_212 = const()[name = string("op_212"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 95 |
+
tensor<fp16, [1, 1500, 6, 64]> var_213_cast_fp16 = reshape(shape = var_212, x = linear_8_cast_fp16)[name = string("op_213_cast_fp16")];
|
| 96 |
+
tensor<int32, [4]> transpose_27_perm_0 = const()[name = string("transpose_27_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 97 |
+
tensor<int32, [4]> transpose_28_perm_0 = const()[name = string("transpose_28_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 98 |
+
tensor<int32, [4]> transpose_29_perm_0 = const()[name = string("transpose_29_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 99 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_29 = transpose(perm = transpose_29_perm_0, x = var_213_cast_fp16)[name = string("transpose_45")];
|
| 100 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_28 = transpose(perm = transpose_28_perm_0, x = var_207_cast_fp16)[name = string("transpose_46")];
|
| 101 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_27 = transpose(perm = transpose_27_perm_0, x = var_201_cast_fp16)[name = string("transpose_47")];
|
| 102 |
+
tensor<fp16, [1, 6, 1500, 64]> a_3_cast_fp16 = scaled_dot_product_attention(key = transpose_28, query = transpose_27, value = transpose_29)[name = string("a_3_cast_fp16")];
|
| 103 |
+
tensor<int32, [4]> var_217 = const()[name = string("op_217"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 104 |
+
tensor<int32, [3]> concat_1 = const()[name = string("concat_1"), val = tensor<int32, [3]>([1, 1500, 384])];
|
| 105 |
+
tensor<fp16, [1, 1500, 6, 64]> var_218_cast_fp16 = transpose(perm = var_217, x = a_3_cast_fp16)[name = string("transpose_44")];
|
| 106 |
+
tensor<fp16, [1, 1500, 384]> x_23_cast_fp16 = reshape(shape = concat_1, x = var_218_cast_fp16)[name = string("x_23_cast_fp16")];
|
| 107 |
+
tensor<fp16, [384, 384]> const_32_to_fp16 = const()[name = string("const_32_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6661184)))];
|
| 108 |
+
tensor<fp16, [384]> const_33_to_fp16 = const()[name = string("const_33_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6956160)))];
|
| 109 |
+
tensor<fp16, [1, 1500, 384]> linear_9_cast_fp16 = linear(bias = const_33_to_fp16, weight = const_32_to_fp16, x = x_23_cast_fp16)[name = string("linear_9_cast_fp16")];
|
| 110 |
+
tensor<fp16, [1, 1500, 384]> x_25_cast_fp16 = add(x = x_19_cast_fp16, y = linear_9_cast_fp16)[name = string("x_25_cast_fp16")];
|
| 111 |
+
tensor<int32, [1]> var_230_axes_0 = const()[name = string("op_230_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 112 |
+
tensor<fp16, [384]> blocks_1_mlp_ln_weight_to_fp16 = const()[name = string("blocks_1_mlp_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6956992)))];
|
| 113 |
+
tensor<fp16, [384]> blocks_1_mlp_ln_bias_to_fp16 = const()[name = string("blocks_1_mlp_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6957824)))];
|
| 114 |
+
tensor<fp16, [1, 1500, 384]> var_230_cast_fp16 = layer_norm(axes = var_230_axes_0, beta = blocks_1_mlp_ln_bias_to_fp16, epsilon = var_166_to_fp16, gamma = blocks_1_mlp_ln_weight_to_fp16, x = x_25_cast_fp16)[name = string("op_230_cast_fp16")];
|
| 115 |
+
tensor<fp16, [1536, 384]> const_34_to_fp16 = const()[name = string("const_34_to_fp16"), val = tensor<fp16, [1536, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(6958656)))];
|
| 116 |
+
tensor<fp16, [1536]> const_35_to_fp16 = const()[name = string("const_35_to_fp16"), val = tensor<fp16, [1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(8138368)))];
|
| 117 |
+
tensor<fp16, [1, 1500, 1536]> linear_10_cast_fp16 = linear(bias = const_35_to_fp16, weight = const_34_to_fp16, x = var_230_cast_fp16)[name = string("linear_10_cast_fp16")];
|
| 118 |
+
string x_29_mode_0 = const()[name = string("x_29_mode_0"), val = string("EXACT")];
|
| 119 |
+
tensor<fp16, [1, 1500, 1536]> x_29_cast_fp16 = gelu(mode = x_29_mode_0, x = linear_10_cast_fp16)[name = string("x_29_cast_fp16")];
|
| 120 |
+
tensor<fp16, [384, 1536]> const_36_to_fp16 = const()[name = string("const_36_to_fp16"), val = tensor<fp16, [384, 1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(8141504)))];
|
| 121 |
+
tensor<fp16, [384]> const_37_to_fp16 = const()[name = string("const_37_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(9321216)))];
|
| 122 |
+
tensor<fp16, [1, 1500, 384]> linear_11_cast_fp16 = linear(bias = const_37_to_fp16, weight = const_36_to_fp16, x = x_29_cast_fp16)[name = string("linear_11_cast_fp16")];
|
| 123 |
+
tensor<fp16, [1, 1500, 384]> x_31_cast_fp16 = add(x = x_25_cast_fp16, y = linear_11_cast_fp16)[name = string("x_31_cast_fp16")];
|
| 124 |
+
tensor<int32, [1]> var_270_axes_0 = const()[name = string("op_270_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 125 |
+
tensor<fp16, [384]> blocks_2_attn_ln_weight_to_fp16 = const()[name = string("blocks_2_attn_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(9322048)))];
|
| 126 |
+
tensor<fp16, [384]> blocks_2_attn_ln_bias_to_fp16 = const()[name = string("blocks_2_attn_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(9322880)))];
|
| 127 |
+
fp16 var_260_to_fp16 = const()[name = string("op_260_to_fp16"), val = fp16(0x1.5p-17)];
|
| 128 |
+
tensor<fp16, [1, 1500, 384]> var_270_cast_fp16 = layer_norm(axes = var_270_axes_0, beta = blocks_2_attn_ln_bias_to_fp16, epsilon = var_260_to_fp16, gamma = blocks_2_attn_ln_weight_to_fp16, x = x_31_cast_fp16)[name = string("op_270_cast_fp16")];
|
| 129 |
+
tensor<fp16, [384, 384]> const_38_to_fp16 = const()[name = string("const_38_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(9323712)))];
|
| 130 |
+
tensor<fp16, [384]> const_39_to_fp16 = const()[name = string("const_39_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(9618688)))];
|
| 131 |
+
tensor<fp16, [1, 1500, 384]> linear_12_cast_fp16 = linear(bias = const_39_to_fp16, weight = const_38_to_fp16, x = var_270_cast_fp16)[name = string("linear_12_cast_fp16")];
|
| 132 |
+
tensor<fp16, [384, 384]> const_40_to_fp16 = const()[name = string("const_40_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(9619520)))];
|
| 133 |
+
tensor<fp16, [1, 1500, 384]> linear_13_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = const_40_to_fp16, x = var_270_cast_fp16)[name = string("linear_13_cast_fp16")];
|
| 134 |
+
tensor<fp16, [384, 384]> const_41_to_fp16 = const()[name = string("const_41_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(9914496)))];
|
| 135 |
+
tensor<fp16, [384]> const_42_to_fp16 = const()[name = string("const_42_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(10209472)))];
|
| 136 |
+
tensor<fp16, [1, 1500, 384]> linear_14_cast_fp16 = linear(bias = const_42_to_fp16, weight = const_41_to_fp16, x = var_270_cast_fp16)[name = string("linear_14_cast_fp16")];
|
| 137 |
+
tensor<int32, [4]> var_294 = const()[name = string("op_294"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 138 |
+
tensor<fp16, [1, 1500, 6, 64]> var_295_cast_fp16 = reshape(shape = var_294, x = linear_12_cast_fp16)[name = string("op_295_cast_fp16")];
|
| 139 |
+
tensor<int32, [4]> var_300 = const()[name = string("op_300"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 140 |
+
tensor<fp16, [1, 1500, 6, 64]> var_301_cast_fp16 = reshape(shape = var_300, x = linear_13_cast_fp16)[name = string("op_301_cast_fp16")];
|
| 141 |
+
tensor<int32, [4]> var_306 = const()[name = string("op_306"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 142 |
+
tensor<fp16, [1, 1500, 6, 64]> var_307_cast_fp16 = reshape(shape = var_306, x = linear_14_cast_fp16)[name = string("op_307_cast_fp16")];
|
| 143 |
+
tensor<int32, [4]> transpose_30_perm_0 = const()[name = string("transpose_30_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 144 |
+
tensor<int32, [4]> transpose_31_perm_0 = const()[name = string("transpose_31_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 145 |
+
tensor<int32, [4]> transpose_32_perm_0 = const()[name = string("transpose_32_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 146 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_32 = transpose(perm = transpose_32_perm_0, x = var_307_cast_fp16)[name = string("transpose_41")];
|
| 147 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_31 = transpose(perm = transpose_31_perm_0, x = var_301_cast_fp16)[name = string("transpose_42")];
|
| 148 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_30 = transpose(perm = transpose_30_perm_0, x = var_295_cast_fp16)[name = string("transpose_43")];
|
| 149 |
+
tensor<fp16, [1, 6, 1500, 64]> a_5_cast_fp16 = scaled_dot_product_attention(key = transpose_31, query = transpose_30, value = transpose_32)[name = string("a_5_cast_fp16")];
|
| 150 |
+
tensor<int32, [4]> var_311 = const()[name = string("op_311"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 151 |
+
tensor<int32, [3]> concat_2 = const()[name = string("concat_2"), val = tensor<int32, [3]>([1, 1500, 384])];
|
| 152 |
+
tensor<fp16, [1, 1500, 6, 64]> var_312_cast_fp16 = transpose(perm = var_311, x = a_5_cast_fp16)[name = string("transpose_40")];
|
| 153 |
+
tensor<fp16, [1, 1500, 384]> x_35_cast_fp16 = reshape(shape = concat_2, x = var_312_cast_fp16)[name = string("x_35_cast_fp16")];
|
| 154 |
+
tensor<fp16, [384, 384]> const_49_to_fp16 = const()[name = string("const_49_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(10210304)))];
|
| 155 |
+
tensor<fp16, [384]> const_50_to_fp16 = const()[name = string("const_50_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(10505280)))];
|
| 156 |
+
tensor<fp16, [1, 1500, 384]> linear_15_cast_fp16 = linear(bias = const_50_to_fp16, weight = const_49_to_fp16, x = x_35_cast_fp16)[name = string("linear_15_cast_fp16")];
|
| 157 |
+
tensor<fp16, [1, 1500, 384]> x_37_cast_fp16 = add(x = x_31_cast_fp16, y = linear_15_cast_fp16)[name = string("x_37_cast_fp16")];
|
| 158 |
+
tensor<int32, [1]> var_324_axes_0 = const()[name = string("op_324_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 159 |
+
tensor<fp16, [384]> blocks_2_mlp_ln_weight_to_fp16 = const()[name = string("blocks_2_mlp_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(10506112)))];
|
| 160 |
+
tensor<fp16, [384]> blocks_2_mlp_ln_bias_to_fp16 = const()[name = string("blocks_2_mlp_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(10506944)))];
|
| 161 |
+
tensor<fp16, [1, 1500, 384]> var_324_cast_fp16 = layer_norm(axes = var_324_axes_0, beta = blocks_2_mlp_ln_bias_to_fp16, epsilon = var_260_to_fp16, gamma = blocks_2_mlp_ln_weight_to_fp16, x = x_37_cast_fp16)[name = string("op_324_cast_fp16")];
|
| 162 |
+
tensor<fp16, [1536, 384]> const_51_to_fp16 = const()[name = string("const_51_to_fp16"), val = tensor<fp16, [1536, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(10507776)))];
|
| 163 |
+
tensor<fp16, [1536]> const_52_to_fp16 = const()[name = string("const_52_to_fp16"), val = tensor<fp16, [1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(11687488)))];
|
| 164 |
+
tensor<fp16, [1, 1500, 1536]> linear_16_cast_fp16 = linear(bias = const_52_to_fp16, weight = const_51_to_fp16, x = var_324_cast_fp16)[name = string("linear_16_cast_fp16")];
|
| 165 |
+
string x_41_mode_0 = const()[name = string("x_41_mode_0"), val = string("EXACT")];
|
| 166 |
+
tensor<fp16, [1, 1500, 1536]> x_41_cast_fp16 = gelu(mode = x_41_mode_0, x = linear_16_cast_fp16)[name = string("x_41_cast_fp16")];
|
| 167 |
+
tensor<fp16, [384, 1536]> const_53_to_fp16 = const()[name = string("const_53_to_fp16"), val = tensor<fp16, [384, 1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(11690624)))];
|
| 168 |
+
tensor<fp16, [384]> const_54_to_fp16 = const()[name = string("const_54_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(12870336)))];
|
| 169 |
+
tensor<fp16, [1, 1500, 384]> linear_17_cast_fp16 = linear(bias = const_54_to_fp16, weight = const_53_to_fp16, x = x_41_cast_fp16)[name = string("linear_17_cast_fp16")];
|
| 170 |
+
tensor<fp16, [1, 1500, 384]> x_43_cast_fp16 = add(x = x_37_cast_fp16, y = linear_17_cast_fp16)[name = string("x_43_cast_fp16")];
|
| 171 |
+
tensor<int32, [1]> var_364_axes_0 = const()[name = string("op_364_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 172 |
+
tensor<fp16, [384]> blocks_3_attn_ln_weight_to_fp16 = const()[name = string("blocks_3_attn_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(12871168)))];
|
| 173 |
+
tensor<fp16, [384]> blocks_3_attn_ln_bias_to_fp16 = const()[name = string("blocks_3_attn_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(12872000)))];
|
| 174 |
+
fp16 var_354_to_fp16 = const()[name = string("op_354_to_fp16"), val = fp16(0x1.5p-17)];
|
| 175 |
+
tensor<fp16, [1, 1500, 384]> var_364_cast_fp16 = layer_norm(axes = var_364_axes_0, beta = blocks_3_attn_ln_bias_to_fp16, epsilon = var_354_to_fp16, gamma = blocks_3_attn_ln_weight_to_fp16, x = x_43_cast_fp16)[name = string("op_364_cast_fp16")];
|
| 176 |
+
tensor<fp16, [384, 384]> const_55_to_fp16 = const()[name = string("const_55_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(12872832)))];
|
| 177 |
+
tensor<fp16, [384]> const_56_to_fp16 = const()[name = string("const_56_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(13167808)))];
|
| 178 |
+
tensor<fp16, [1, 1500, 384]> linear_18_cast_fp16 = linear(bias = const_56_to_fp16, weight = const_55_to_fp16, x = var_364_cast_fp16)[name = string("linear_18_cast_fp16")];
|
| 179 |
+
tensor<fp16, [384, 384]> const_57_to_fp16 = const()[name = string("const_57_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(13168640)))];
|
| 180 |
+
tensor<fp16, [1, 1500, 384]> linear_19_cast_fp16 = linear(bias = linear_1_bias_0_to_fp16, weight = const_57_to_fp16, x = var_364_cast_fp16)[name = string("linear_19_cast_fp16")];
|
| 181 |
+
tensor<fp16, [384, 384]> const_58_to_fp16 = const()[name = string("const_58_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(13463616)))];
|
| 182 |
+
tensor<fp16, [384]> const_59_to_fp16 = const()[name = string("const_59_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(13758592)))];
|
| 183 |
+
tensor<fp16, [1, 1500, 384]> linear_20_cast_fp16 = linear(bias = const_59_to_fp16, weight = const_58_to_fp16, x = var_364_cast_fp16)[name = string("linear_20_cast_fp16")];
|
| 184 |
+
tensor<int32, [4]> var_388 = const()[name = string("op_388"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 185 |
+
tensor<fp16, [1, 1500, 6, 64]> var_389_cast_fp16 = reshape(shape = var_388, x = linear_18_cast_fp16)[name = string("op_389_cast_fp16")];
|
| 186 |
+
tensor<int32, [4]> var_394 = const()[name = string("op_394"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 187 |
+
tensor<fp16, [1, 1500, 6, 64]> var_395_cast_fp16 = reshape(shape = var_394, x = linear_19_cast_fp16)[name = string("op_395_cast_fp16")];
|
| 188 |
+
tensor<int32, [4]> var_400 = const()[name = string("op_400"), val = tensor<int32, [4]>([1, 1500, 6, -1])];
|
| 189 |
+
tensor<fp16, [1, 1500, 6, 64]> var_401_cast_fp16 = reshape(shape = var_400, x = linear_20_cast_fp16)[name = string("op_401_cast_fp16")];
|
| 190 |
+
tensor<int32, [4]> transpose_33_perm_0 = const()[name = string("transpose_33_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 191 |
+
tensor<int32, [4]> transpose_34_perm_0 = const()[name = string("transpose_34_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 192 |
+
tensor<int32, [4]> transpose_35_perm_0 = const()[name = string("transpose_35_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 193 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_35 = transpose(perm = transpose_35_perm_0, x = var_401_cast_fp16)[name = string("transpose_37")];
|
| 194 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_34 = transpose(perm = transpose_34_perm_0, x = var_395_cast_fp16)[name = string("transpose_38")];
|
| 195 |
+
tensor<fp16, [1, 6, 1500, 64]> transpose_33 = transpose(perm = transpose_33_perm_0, x = var_389_cast_fp16)[name = string("transpose_39")];
|
| 196 |
+
tensor<fp16, [1, 6, 1500, 64]> a_cast_fp16 = scaled_dot_product_attention(key = transpose_34, query = transpose_33, value = transpose_35)[name = string("a_cast_fp16")];
|
| 197 |
+
tensor<int32, [4]> var_405 = const()[name = string("op_405"), val = tensor<int32, [4]>([0, 2, 1, 3])];
|
| 198 |
+
tensor<int32, [3]> concat_3 = const()[name = string("concat_3"), val = tensor<int32, [3]>([1, 1500, 384])];
|
| 199 |
+
tensor<fp16, [1, 1500, 6, 64]> var_406_cast_fp16 = transpose(perm = var_405, x = a_cast_fp16)[name = string("transpose_36")];
|
| 200 |
+
tensor<fp16, [1, 1500, 384]> x_47_cast_fp16 = reshape(shape = concat_3, x = var_406_cast_fp16)[name = string("x_47_cast_fp16")];
|
| 201 |
+
tensor<fp16, [384, 384]> const_66_to_fp16 = const()[name = string("const_66_to_fp16"), val = tensor<fp16, [384, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(13759424)))];
|
| 202 |
+
tensor<fp16, [384]> const_67_to_fp16 = const()[name = string("const_67_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(14054400)))];
|
| 203 |
+
tensor<fp16, [1, 1500, 384]> linear_21_cast_fp16 = linear(bias = const_67_to_fp16, weight = const_66_to_fp16, x = x_47_cast_fp16)[name = string("linear_21_cast_fp16")];
|
| 204 |
+
tensor<fp16, [1, 1500, 384]> x_49_cast_fp16 = add(x = x_43_cast_fp16, y = linear_21_cast_fp16)[name = string("x_49_cast_fp16")];
|
| 205 |
+
tensor<int32, [1]> var_418_axes_0 = const()[name = string("op_418_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 206 |
+
tensor<fp16, [384]> blocks_3_mlp_ln_weight_to_fp16 = const()[name = string("blocks_3_mlp_ln_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(14055232)))];
|
| 207 |
+
tensor<fp16, [384]> blocks_3_mlp_ln_bias_to_fp16 = const()[name = string("blocks_3_mlp_ln_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(14056064)))];
|
| 208 |
+
tensor<fp16, [1, 1500, 384]> var_418_cast_fp16 = layer_norm(axes = var_418_axes_0, beta = blocks_3_mlp_ln_bias_to_fp16, epsilon = var_354_to_fp16, gamma = blocks_3_mlp_ln_weight_to_fp16, x = x_49_cast_fp16)[name = string("op_418_cast_fp16")];
|
| 209 |
+
tensor<fp16, [1536, 384]> const_68_to_fp16 = const()[name = string("const_68_to_fp16"), val = tensor<fp16, [1536, 384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(14056896)))];
|
| 210 |
+
tensor<fp16, [1536]> const_69_to_fp16 = const()[name = string("const_69_to_fp16"), val = tensor<fp16, [1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(15236608)))];
|
| 211 |
+
tensor<fp16, [1, 1500, 1536]> linear_22_cast_fp16 = linear(bias = const_69_to_fp16, weight = const_68_to_fp16, x = var_418_cast_fp16)[name = string("linear_22_cast_fp16")];
|
| 212 |
+
string x_53_mode_0 = const()[name = string("x_53_mode_0"), val = string("EXACT")];
|
| 213 |
+
tensor<fp16, [1, 1500, 1536]> x_53_cast_fp16 = gelu(mode = x_53_mode_0, x = linear_22_cast_fp16)[name = string("x_53_cast_fp16")];
|
| 214 |
+
tensor<fp16, [384, 1536]> const_70_to_fp16 = const()[name = string("const_70_to_fp16"), val = tensor<fp16, [384, 1536]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(15239744)))];
|
| 215 |
+
tensor<fp16, [384]> const_71_to_fp16 = const()[name = string("const_71_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(16419456)))];
|
| 216 |
+
tensor<fp16, [1, 1500, 384]> linear_23_cast_fp16 = linear(bias = const_71_to_fp16, weight = const_70_to_fp16, x = x_53_cast_fp16)[name = string("linear_23_cast_fp16")];
|
| 217 |
+
tensor<fp16, [1, 1500, 384]> x_cast_fp16 = add(x = x_49_cast_fp16, y = linear_23_cast_fp16)[name = string("x_cast_fp16")];
|
| 218 |
+
tensor<int32, [1]> var_447_axes_0 = const()[name = string("op_447_axes_0"), val = tensor<int32, [1]>([-1])];
|
| 219 |
+
tensor<fp16, [384]> ln_post_weight_to_fp16 = const()[name = string("ln_post_weight_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(16420288)))];
|
| 220 |
+
tensor<fp16, [384]> ln_post_bias_to_fp16 = const()[name = string("ln_post_bias_to_fp16"), val = tensor<fp16, [384]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(16421120)))];
|
| 221 |
+
fp16 var_438_to_fp16 = const()[name = string("op_438_to_fp16"), val = fp16(0x1.5p-17)];
|
| 222 |
+
tensor<fp16, [1, 1500, 384]> output = layer_norm(axes = var_447_axes_0, beta = ln_post_bias_to_fp16, epsilon = var_438_to_fp16, gamma = ln_post_weight_to_fp16, x = x_cast_fp16)[name = string("op_447_cast_fp16")];
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 223 |
} -> (output);
|
| 224 |
}
|
ggml-tiny-encoder.mlmodelc/weights/weight.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 16421952
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4efa9bb81afaf12ac6d7cf7a3a4ba1e6b92f05f96ae77fd55cf725e2ecd3a5fd
|
| 3 |
size 16421952
|
ggml-large-v2-encoder.mlmodelc/analytics/coremldata.bin → ggml-tiny-q8_0.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98bfdeb46504a115c4c0dcbaf14713b003152b2ce23f14f639d7f6d39dee043a
|
| 3 |
+
size 32561100
|
ggml-tiny.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:c7e4b084cefeebeed66fb9d096a29b836125edbb8456fea5a9c77b4efc085323
|
| 3 |
-
size 60079860
|
|
|
|
|
|
|
|
|
|
|
|
index/base
CHANGED
|
@@ -3,4 +3,4 @@ ggml-base-encoder.mlmodelc/metadata.json
|
|
| 3 |
ggml-base-encoder.mlmodelc/model.mil
|
| 4 |
ggml-base-encoder.mlmodelc/coremldata.bin
|
| 5 |
ggml-base-encoder.mlmodelc/analytics/coremldata.bin
|
| 6 |
-
ggml-base.bin
|
|
|
|
| 3 |
ggml-base-encoder.mlmodelc/model.mil
|
| 4 |
ggml-base-encoder.mlmodelc/coremldata.bin
|
| 5 |
ggml-base-encoder.mlmodelc/analytics/coremldata.bin
|
| 6 |
+
ggml-base-q8_0.bin
|
index/large-v2
DELETED
|
@@ -1,6 +0,0 @@
|
|
| 1 |
-
ggml-large-v2-encoder.mlmodelc/weights/weight.bin
|
| 2 |
-
ggml-large-v2-encoder.mlmodelc/metadata.json
|
| 3 |
-
ggml-large-v2-encoder.mlmodelc/model.mil
|
| 4 |
-
ggml-large-v2-encoder.mlmodelc/coremldata.bin
|
| 5 |
-
ggml-large-v2-encoder.mlmodelc/analytics/coremldata.bin
|
| 6 |
-
ggml-large-v2-q8_0.bin
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
index/large-v3-turbo
CHANGED
|
@@ -1,12 +1,6 @@
|
|
|
|
|
| 1 |
ggml-large-v3-turbo-encoder.mlmodelc/metadata.json
|
| 2 |
-
ggml-large-v3-turbo-encoder.mlmodelc/
|
| 3 |
-
ggml-large-v3-turbo-encoder.mlmodelc/model0/model.mil
|
| 4 |
-
ggml-large-v3-turbo-encoder.mlmodelc/model0/coremldata.bin
|
| 5 |
-
ggml-large-v3-turbo-encoder.mlmodelc/model0/analytics/coremldata.bin
|
| 6 |
-
ggml-large-v3-turbo-encoder.mlmodelc/model1/weights/1-weight.bin
|
| 7 |
-
ggml-large-v3-turbo-encoder.mlmodelc/model1/model.mil
|
| 8 |
-
ggml-large-v3-turbo-encoder.mlmodelc/model1/coremldata.bin
|
| 9 |
-
ggml-large-v3-turbo-encoder.mlmodelc/model1/analytics/coremldata.bin
|
| 10 |
ggml-large-v3-turbo-encoder.mlmodelc/coremldata.bin
|
| 11 |
ggml-large-v3-turbo-encoder.mlmodelc/analytics/coremldata.bin
|
| 12 |
ggml-large-v3-turbo-q8_0.bin
|
|
|
|
| 1 |
+
ggml-large-v3-turbo-encoder.mlmodelc/weights/weight.bin
|
| 2 |
ggml-large-v3-turbo-encoder.mlmodelc/metadata.json
|
| 3 |
+
ggml-large-v3-turbo-encoder.mlmodelc/model.mil
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
ggml-large-v3-turbo-encoder.mlmodelc/coremldata.bin
|
| 5 |
ggml-large-v3-turbo-encoder.mlmodelc/analytics/coremldata.bin
|
| 6 |
ggml-large-v3-turbo-q8_0.bin
|