tomAndJetty commited on
Commit
a0d335a
·
verified ·
1 Parent(s): 999eaaa

Delete openai_whisper-tiny

Browse files
Files changed (35) hide show
  1. openai_whisper-tiny/AudioEncoder.mlmodelc/analytics/coremldata.bin +0 -3
  2. openai_whisper-tiny/AudioEncoder.mlmodelc/coremldata.bin +0 -3
  3. openai_whisper-tiny/AudioEncoder.mlmodelc/metadata.json +0 -70
  4. openai_whisper-tiny/AudioEncoder.mlmodelc/model.mil +0 -0
  5. openai_whisper-tiny/AudioEncoder.mlmodelc/weights/weight.bin +0 -3
  6. openai_whisper-tiny/MelSpectrogram.mlmodelc/analytics/coremldata.bin +0 -3
  7. openai_whisper-tiny/MelSpectrogram.mlmodelc/coremldata.bin +0 -3
  8. openai_whisper-tiny/MelSpectrogram.mlmodelc/metadata.json +0 -74
  9. openai_whisper-tiny/MelSpectrogram.mlmodelc/model.mil +0 -66
  10. openai_whisper-tiny/MelSpectrogram.mlmodelc/weights/weight.bin +0 -3
  11. openai_whisper-tiny/TextDecoder.mlmodelc/analytics/coremldata.bin +0 -3
  12. openai_whisper-tiny/TextDecoder.mlmodelc/coremldata.bin +0 -3
  13. openai_whisper-tiny/TextDecoder.mlmodelc/metadata.json +0 -167
  14. openai_whisper-tiny/TextDecoder.mlmodelc/model.mil +0 -0
  15. openai_whisper-tiny/TextDecoder.mlmodelc/weights/weight.bin +0 -3
  16. openai_whisper-tiny/config.json +0 -144
  17. openai_whisper-tiny/generation_config.json +0 -248
  18. openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/analytics/coremldata.bin +0 -3
  19. openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/coremldata.bin +0 -3
  20. openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/metadata.json +0 -69
  21. openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/model.mil +0 -0
  22. openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/model.mlmodel +0 -3
  23. openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/weights/weight.bin +0 -3
  24. openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/analytics/coremldata.bin +0 -3
  25. openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/coremldata.bin +0 -3
  26. openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/metadata.json +0 -71
  27. openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/model.mil +0 -66
  28. openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/weights/weight.bin +0 -3
  29. openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/analytics/coremldata.bin +0 -3
  30. openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/coremldata.bin +0 -3
  31. openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/metadata.json +0 -165
  32. openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/model.mil +0 -0
  33. openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/model.mlmodel +0 -3
  34. openai_whisper-tiny/openai_whisper-small/config.json +0 -1
  35. openai_whisper-tiny/openai_whisper-small/generation_config.json +0 -1
openai_whisper-tiny/AudioEncoder.mlmodelc/analytics/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:37dbcb7943d1c5b92e0ae956102df6f69f343d1c6584f9f1511e2044fa8890f8
3
- size 243
 
 
 
 
openai_whisper-tiny/AudioEncoder.mlmodelc/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:abdbc66c120f462ae758502376c141c105e094780c537f997be62a8d577e775a
3
- size 347
 
 
 
 
openai_whisper-tiny/AudioEncoder.mlmodelc/metadata.json DELETED
@@ -1,70 +0,0 @@
1
- [
2
- {
3
- "metadataOutputVersion" : "3.0",
4
- "storagePrecision" : "Float16",
5
- "outputSchema" : [
6
- {
7
- "hasShapeFlexibility" : "0",
8
- "isOptional" : "0",
9
- "dataType" : "Float16",
10
- "formattedType" : "MultiArray (Float16 1 × 384 × 1 × 1500)",
11
- "shortDescription" : "",
12
- "shape" : "[1, 384, 1, 1500]",
13
- "name" : "encoder_output_embeds",
14
- "type" : "MultiArray"
15
- }
16
- ],
17
- "modelParameters" : [
18
-
19
- ],
20
- "specificationVersion" : 7,
21
- "mlProgramOperationTypeHistogram" : {
22
- "Concat" : 28,
23
- "Ios16.add" : 9,
24
- "Ios16.mul" : 144,
25
- "SliceByIndex" : 216,
26
- "Transpose" : 4,
27
- "Ios16.batchNorm" : 9,
28
- "Ios16.gelu" : 6,
29
- "Ios16.einsum" : 288,
30
- "Ios16.softmax" : 144,
31
- "Ios16.layerNorm" : 9,
32
- "Ios16.conv" : 26
33
- },
34
- "computePrecision" : "Mixed (Float16, Int32)",
35
- "isUpdatable" : "0",
36
- "stateSchema" : [
37
-
38
- ],
39
- "availability" : {
40
- "macOS" : "13.0",
41
- "tvOS" : "16.0",
42
- "visionOS" : "1.0",
43
- "watchOS" : "9.0",
44
- "iOS" : "16.0",
45
- "macCatalyst" : "16.0"
46
- },
47
- "modelType" : {
48
- "name" : "MLModelType_mlProgram"
49
- },
50
- "userDefinedMetadata" : {
51
- "com.github.apple.coremltools.source_dialect" : "TorchScript",
52
- "com.github.apple.coremltools.source" : "torch==2.6.0",
53
- "com.github.apple.coremltools.version" : "8.2"
54
- },
55
- "inputSchema" : [
56
- {
57
- "hasShapeFlexibility" : "0",
58
- "isOptional" : "0",
59
- "dataType" : "Float16",
60
- "formattedType" : "MultiArray (Float16 1 × 80 × 1 × 3000)",
61
- "shortDescription" : "",
62
- "shape" : "[1, 80, 1, 3000]",
63
- "name" : "melspectrogram_features",
64
- "type" : "MultiArray"
65
- }
66
- ],
67
- "generatedClassName" : "AudioEncoder",
68
- "method" : "predict"
69
- }
70
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/AudioEncoder.mlmodelc/model.mil DELETED
The diff for this file is too large to render. See raw diff
 
openai_whisper-tiny/AudioEncoder.mlmodelc/weights/weight.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:52267fceba5c3f1bacb93027ec8e6edf72f096030d6e2ade687a616d67aeceea
3
- size 16422784
 
 
 
 
openai_whisper-tiny/MelSpectrogram.mlmodelc/analytics/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4f367993f0198e9858a4d89fb054318982c91a9bb5946e29231421c2f1100b9
3
- size 243
 
 
 
 
openai_whisper-tiny/MelSpectrogram.mlmodelc/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:806321f1034184a10b04dc50816219dec8ae9789698712050c81edecb9bb5aa7
3
- size 328
 
 
 
 
openai_whisper-tiny/MelSpectrogram.mlmodelc/metadata.json DELETED
@@ -1,74 +0,0 @@
1
- [
2
- {
3
- "metadataOutputVersion" : "3.0",
4
- "storagePrecision" : "Float16",
5
- "outputSchema" : [
6
- {
7
- "hasShapeFlexibility" : "0",
8
- "isOptional" : "0",
9
- "dataType" : "Float16",
10
- "formattedType" : "MultiArray (Float16 1 × 80 × 1 × 3000)",
11
- "shortDescription" : "",
12
- "shape" : "[1, 80, 1, 3000]",
13
- "name" : "melspectrogram_features",
14
- "type" : "MultiArray"
15
- }
16
- ],
17
- "modelParameters" : [
18
-
19
- ],
20
- "specificationVersion" : 7,
21
- "mlProgramOperationTypeHistogram" : {
22
- "Ios16.reshape" : 2,
23
- "Ios16.mul" : 2,
24
- "SliceByIndex" : 1,
25
- "Ios16.sub" : 1,
26
- "Ios16.log" : 1,
27
- "Ios16.square" : 2,
28
- "Ios16.add" : 3,
29
- "Squeeze" : 2,
30
- "Ios16.matmul" : 1,
31
- "Ios16.conv" : 2,
32
- "Ios16.maximum" : 1,
33
- "ExpandDims" : 4,
34
- "Ios16.reduceMax" : 1,
35
- "Identity" : 1,
36
- "Pad" : 1
37
- },
38
- "computePrecision" : "Mixed (Float16, Int32)",
39
- "isUpdatable" : "0",
40
- "stateSchema" : [
41
-
42
- ],
43
- "availability" : {
44
- "macOS" : "13.0",
45
- "tvOS" : "16.0",
46
- "visionOS" : "1.0",
47
- "watchOS" : "9.0",
48
- "iOS" : "16.0",
49
- "macCatalyst" : "16.0"
50
- },
51
- "modelType" : {
52
- "name" : "MLModelType_mlProgram"
53
- },
54
- "userDefinedMetadata" : {
55
- "com.github.apple.coremltools.source_dialect" : "TorchScript",
56
- "com.github.apple.coremltools.source" : "torch==2.6.0",
57
- "com.github.apple.coremltools.version" : "8.2"
58
- },
59
- "inputSchema" : [
60
- {
61
- "hasShapeFlexibility" : "0",
62
- "isOptional" : "0",
63
- "dataType" : "Float16",
64
- "formattedType" : "MultiArray (Float16 480000)",
65
- "shortDescription" : "",
66
- "shape" : "[480000]",
67
- "name" : "audio",
68
- "type" : "MultiArray"
69
- }
70
- ],
71
- "generatedClassName" : "MelSpectrogram",
72
- "method" : "predict"
73
- }
74
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/MelSpectrogram.mlmodelc/model.mil DELETED
@@ -1,66 +0,0 @@
1
- program(1.0)
2
- [buildInfo = dict<tensor<string, []>, tensor<string, []>>({{"coremlc-component-MIL", "3402.3.2"}, {"coremlc-version", "3402.4.1"}, {"coremltools-component-torch", "2.6.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.2"}})]
3
- {
4
- func main<ios16>(tensor<fp16, [480000]> audio) {
5
- tensor<int32, [3]> var_10 = const()[name = tensor<string, []>("op_10"), val = tensor<int32, [3]>([1, 1, 480000])];
6
- tensor<fp16, [1, 1, 480000]> input_1_cast_fp16 = reshape(shape = var_10, x = audio)[name = tensor<string, []>("input_1_cast_fp16")];
7
- tensor<int32, [6]> input_3_pad_0 = const()[name = tensor<string, []>("input_3_pad_0"), val = tensor<int32, [6]>([0, 0, 0, 0, 200, 200])];
8
- tensor<string, []> input_3_mode_0 = const()[name = tensor<string, []>("input_3_mode_0"), val = tensor<string, []>("reflect")];
9
- tensor<fp16, []> const_1_to_fp16 = const()[name = tensor<string, []>("const_1_to_fp16"), val = tensor<fp16, []>(0x0p+0)];
10
- tensor<fp16, [1, 1, 480400]> input_3_cast_fp16 = pad(constant_val = const_1_to_fp16, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1_cast_fp16)[name = tensor<string, []>("input_3_cast_fp16")];
11
- tensor<int32, [1]> var_22 = const()[name = tensor<string, []>("op_22"), val = tensor<int32, [1]>([480400])];
12
- tensor<fp16, [480400]> input_cast_fp16 = reshape(shape = var_22, x = input_3_cast_fp16)[name = tensor<string, []>("input_cast_fp16")];
13
- tensor<int32, [1]> expand_dims_0_axes_0 = const()[name = tensor<string, []>("expand_dims_0_axes_0"), val = tensor<int32, [1]>([0])];
14
- tensor<fp16, [1, 480400]> expand_dims_0_cast_fp16 = expand_dims(axes = expand_dims_0_axes_0, x = input_cast_fp16)[name = tensor<string, []>("expand_dims_0_cast_fp16")];
15
- tensor<int32, [1]> expand_dims_3 = const()[name = tensor<string, []>("expand_dims_3"), val = tensor<int32, [1]>([160])];
16
- tensor<int32, [1]> expand_dims_4_axes_0 = const()[name = tensor<string, []>("expand_dims_4_axes_0"), val = tensor<int32, [1]>([1])];
17
- tensor<fp16, [1, 1, 480400]> expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = expand_dims_0_cast_fp16)[name = tensor<string, []>("expand_dims_4_cast_fp16")];
18
- tensor<string, []> conv_0_pad_type_0 = const()[name = tensor<string, []>("conv_0_pad_type_0"), val = tensor<string, []>("valid")];
19
- tensor<int32, [2]> conv_0_pad_0 = const()[name = tensor<string, []>("conv_0_pad_0"), val = tensor<int32, [2]>([0, 0])];
20
- tensor<int32, [1]> conv_0_dilations_0 = const()[name = tensor<string, []>("conv_0_dilations_0"), val = tensor<int32, [1]>([1])];
21
- tensor<int32, []> conv_0_groups_0 = const()[name = tensor<string, []>("conv_0_groups_0"), val = tensor<int32, []>(1)];
22
- tensor<fp16, [201, 1, 400]> expand_dims_1_to_fp16 = const()[name = tensor<string, []>("expand_dims_1_to_fp16"), val = tensor<fp16, [201, 1, 400]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(64)))];
23
- tensor<fp16, [1, 201, 3001]> conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor<string, []>("conv_0_cast_fp16")];
24
- tensor<string, []> conv_1_pad_type_0 = const()[name = tensor<string, []>("conv_1_pad_type_0"), val = tensor<string, []>("valid")];
25
- tensor<int32, [2]> conv_1_pad_0 = const()[name = tensor<string, []>("conv_1_pad_0"), val = tensor<int32, [2]>([0, 0])];
26
- tensor<int32, [1]> conv_1_dilations_0 = const()[name = tensor<string, []>("conv_1_dilations_0"), val = tensor<int32, [1]>([1])];
27
- tensor<int32, []> conv_1_groups_0 = const()[name = tensor<string, []>("conv_1_groups_0"), val = tensor<int32, []>(1)];
28
- tensor<fp16, [201, 1, 400]> expand_dims_2_to_fp16 = const()[name = tensor<string, []>("expand_dims_2_to_fp16"), val = tensor<fp16, [201, 1, 400]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(160960)))];
29
- tensor<fp16, [1, 201, 3001]> conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor<string, []>("conv_1_cast_fp16")];
30
- tensor<int32, [1]> squeeze_0_axes_0 = const()[name = tensor<string, []>("squeeze_0_axes_0"), val = tensor<int32, [1]>([0])];
31
- tensor<fp16, [201, 3001]> squeeze_0_cast_fp16 = squeeze(axes = squeeze_0_axes_0, x = conv_0_cast_fp16)[name = tensor<string, []>("squeeze_0_cast_fp16")];
32
- tensor<int32, [1]> squeeze_1_axes_0 = const()[name = tensor<string, []>("squeeze_1_axes_0"), val = tensor<int32, [1]>([0])];
33
- tensor<fp16, [201, 3001]> squeeze_1_cast_fp16 = squeeze(axes = squeeze_1_axes_0, x = conv_1_cast_fp16)[name = tensor<string, []>("squeeze_1_cast_fp16")];
34
- tensor<fp16, [201, 3001]> square_0_cast_fp16 = square(x = squeeze_0_cast_fp16)[name = tensor<string, []>("square_0_cast_fp16")];
35
- tensor<fp16, [201, 3001]> square_1_cast_fp16 = square(x = squeeze_1_cast_fp16)[name = tensor<string, []>("square_1_cast_fp16")];
36
- tensor<fp16, [201, 3001]> add_1_cast_fp16 = add(x = square_0_cast_fp16, y = square_1_cast_fp16)[name = tensor<string, []>("add_1_cast_fp16")];
37
- tensor<fp16, [201, 3001]> magnitudes_1_cast_fp16 = identity(x = add_1_cast_fp16)[name = tensor<string, []>("magnitudes_1_cast_fp16")];
38
- tensor<int32, [2]> magnitudes_begin_0 = const()[name = tensor<string, []>("magnitudes_begin_0"), val = tensor<int32, [2]>([0, 0])];
39
- tensor<int32, [2]> magnitudes_end_0 = const()[name = tensor<string, []>("magnitudes_end_0"), val = tensor<int32, [2]>([201, 3000])];
40
- tensor<bool, [2]> magnitudes_end_mask_0 = const()[name = tensor<string, []>("magnitudes_end_mask_0"), val = tensor<bool, [2]>([true, false])];
41
- tensor<fp16, [201, 3000]> magnitudes_cast_fp16 = slice_by_index(begin = magnitudes_begin_0, end = magnitudes_end_0, end_mask = magnitudes_end_mask_0, x = magnitudes_1_cast_fp16)[name = tensor<string, []>("magnitudes_cast_fp16")];
42
- tensor<bool, []> mel_spec_1_transpose_x_0 = const()[name = tensor<string, []>("mel_spec_1_transpose_x_0"), val = tensor<bool, []>(false)];
43
- tensor<bool, []> mel_spec_1_transpose_y_0 = const()[name = tensor<string, []>("mel_spec_1_transpose_y_0"), val = tensor<bool, []>(false)];
44
- tensor<fp16, [80, 201]> mel_filters_to_fp16 = const()[name = tensor<string, []>("mel_filters_to_fp16"), val = tensor<fp16, [80, 201]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(321856)))];
45
- tensor<fp16, [80, 3000]> mel_spec_1_cast_fp16 = matmul(transpose_x = mel_spec_1_transpose_x_0, transpose_y = mel_spec_1_transpose_y_0, x = mel_filters_to_fp16, y = magnitudes_cast_fp16)[name = tensor<string, []>("mel_spec_1_cast_fp16")];
46
- tensor<fp16, []> var_41_to_fp16 = const()[name = tensor<string, []>("op_41_to_fp16"), val = tensor<fp16, []>(0x1p-24)];
47
- tensor<fp16, [80, 3000]> mel_spec_cast_fp16 = add(x = mel_spec_1_cast_fp16, y = var_41_to_fp16)[name = tensor<string, []>("mel_spec_cast_fp16")];
48
- tensor<fp16, []> log_0_epsilon_0_to_fp16 = const()[name = tensor<string, []>("log_0_epsilon_0_to_fp16"), val = tensor<fp16, []>(0x0p+0)];
49
- tensor<fp16, [80, 3000]> log_0_cast_fp16 = log(epsilon = log_0_epsilon_0_to_fp16, x = mel_spec_cast_fp16)[name = tensor<string, []>("log_0_cast_fp16")];
50
- tensor<fp16, []> mul_0_y_0_to_fp16 = const()[name = tensor<string, []>("mul_0_y_0_to_fp16"), val = tensor<fp16, []>(0x1.bccp-2)];
51
- tensor<fp16, [80, 3000]> mul_0_cast_fp16 = mul(x = log_0_cast_fp16, y = mul_0_y_0_to_fp16)[name = tensor<string, []>("mul_0_cast_fp16")];
52
- tensor<bool, []> var_44_keep_dims_0 = const()[name = tensor<string, []>("op_44_keep_dims_0"), val = tensor<bool, []>(false)];
53
- tensor<fp16, []> var_44_cast_fp16 = reduce_max(keep_dims = var_44_keep_dims_0, x = mul_0_cast_fp16)[name = tensor<string, []>("op_44_cast_fp16")];
54
- tensor<fp16, []> var_46_to_fp16 = const()[name = tensor<string, []>("op_46_to_fp16"), val = tensor<fp16, []>(0x1p+3)];
55
- tensor<fp16, []> var_47_cast_fp16 = sub(x = var_44_cast_fp16, y = var_46_to_fp16)[name = tensor<string, []>("op_47_cast_fp16")];
56
- tensor<fp16, [80, 3000]> log_spec_3_cast_fp16 = maximum(x = mul_0_cast_fp16, y = var_47_cast_fp16)[name = tensor<string, []>("log_spec_3_cast_fp16")];
57
- tensor<fp16, []> var_50_to_fp16 = const()[name = tensor<string, []>("op_50_to_fp16"), val = tensor<fp16, []>(0x1p+2)];
58
- tensor<fp16, [80, 3000]> var_51_cast_fp16 = add(x = log_spec_3_cast_fp16, y = var_50_to_fp16)[name = tensor<string, []>("op_51_cast_fp16")];
59
- tensor<fp16, []> _inversed_log_spec_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_log_spec_y_0_to_fp16"), val = tensor<fp16, []>(0x1p-2)];
60
- tensor<fp16, [80, 3000]> _inversed_log_spec_cast_fp16 = mul(x = var_51_cast_fp16, y = _inversed_log_spec_y_0_to_fp16)[name = tensor<string, []>("_inversed_log_spec_cast_fp16")];
61
- tensor<int32, [1]> var_55_axes_0 = const()[name = tensor<string, []>("op_55_axes_0"), val = tensor<int32, [1]>([0])];
62
- tensor<fp16, [1, 80, 3000]> var_55_cast_fp16 = expand_dims(axes = var_55_axes_0, x = _inversed_log_spec_cast_fp16)[name = tensor<string, []>("op_55_cast_fp16")];
63
- tensor<int32, [1]> var_62_axes_0 = const()[name = tensor<string, []>("op_62_axes_0"), val = tensor<int32, [1]>([2])];
64
- tensor<fp16, [1, 80, 1, 3000]> melspectrogram_features = expand_dims(axes = var_62_axes_0, x = var_55_cast_fp16)[name = tensor<string, []>("op_62_cast_fp16")];
65
- } -> (melspectrogram_features);
66
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/MelSpectrogram.mlmodelc/weights/weight.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:801024dbc7a89c677be1f8b285de3409e35f7d1786c9c8d9d0d6842ac57a1c83
3
- size 354080
 
 
 
 
openai_whisper-tiny/TextDecoder.mlmodelc/analytics/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:70c25a4908f1b0082f2bd62989f2120eb08addebca634439c9207b6ee776154a
3
- size 243
 
 
 
 
openai_whisper-tiny/TextDecoder.mlmodelc/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:940113f3791f56fae5e918f1fce8102f6f903cbd7b8cb5a9e49b2d011d4239b5
3
- size 633
 
 
 
 
openai_whisper-tiny/TextDecoder.mlmodelc/metadata.json DELETED
@@ -1,167 +0,0 @@
1
- [
2
- {
3
- "metadataOutputVersion" : "3.0",
4
- "storagePrecision" : "Float16",
5
- "outputSchema" : [
6
- {
7
- "hasShapeFlexibility" : "0",
8
- "isOptional" : "0",
9
- "dataType" : "Float16",
10
- "formattedType" : "MultiArray (Float16 1 × 1 × 51865)",
11
- "shortDescription" : "",
12
- "shape" : "[1, 1, 51865]",
13
- "name" : "logits",
14
- "type" : "MultiArray"
15
- },
16
- {
17
- "hasShapeFlexibility" : "0",
18
- "isOptional" : "0",
19
- "dataType" : "Float16",
20
- "formattedType" : "MultiArray (Float16 1 × 1536 × 1 × 1)",
21
- "shortDescription" : "",
22
- "shape" : "[1, 1536, 1, 1]",
23
- "name" : "key_cache_updates",
24
- "type" : "MultiArray"
25
- },
26
- {
27
- "hasShapeFlexibility" : "0",
28
- "isOptional" : "0",
29
- "dataType" : "Float16",
30
- "formattedType" : "MultiArray (Float16 1 × 1536 × 1 × 1)",
31
- "shortDescription" : "",
32
- "shape" : "[1, 1536, 1, 1]",
33
- "name" : "value_cache_updates",
34
- "type" : "MultiArray"
35
- },
36
- {
37
- "hasShapeFlexibility" : "0",
38
- "isOptional" : "0",
39
- "dataType" : "Float16",
40
- "formattedType" : "MultiArray (Float16 1 × 1500)",
41
- "shortDescription" : "",
42
- "shape" : "[1, 1500]",
43
- "name" : "alignment_heads_weights",
44
- "type" : "MultiArray"
45
- }
46
- ],
47
- "modelParameters" : [
48
-
49
- ],
50
- "specificationVersion" : 7,
51
- "mlProgramOperationTypeHistogram" : {
52
- "Ios16.linear" : 1,
53
- "Concat" : 3,
54
- "Ios16.reduceMean" : 1,
55
- "Ios16.mul" : 16,
56
- "Ios16.layerNorm" : 13,
57
- "SliceByIndex" : 12,
58
- "Transpose" : 1,
59
- "Ios16.conv" : 40,
60
- "Ios16.add" : 25,
61
- "Squeeze" : 1,
62
- "Ios16.matmul" : 16,
63
- "Ios16.softmax" : 8,
64
- "Ios16.gelu" : 4,
65
- "ExpandDims" : 6,
66
- "Ios16.batchNorm" : 13,
67
- "Split" : 2,
68
- "Ios16.gather" : 2,
69
- "Ios16.reshape" : 32
70
- },
71
- "computePrecision" : "Mixed (Float16, Int32)",
72
- "isUpdatable" : "0",
73
- "stateSchema" : [
74
-
75
- ],
76
- "availability" : {
77
- "macOS" : "13.0",
78
- "tvOS" : "16.0",
79
- "visionOS" : "1.0",
80
- "watchOS" : "9.0",
81
- "iOS" : "16.0",
82
- "macCatalyst" : "16.0"
83
- },
84
- "modelType" : {
85
- "name" : "MLModelType_mlProgram"
86
- },
87
- "userDefinedMetadata" : {
88
- "com.github.apple.coremltools.source_dialect" : "TorchScript",
89
- "com.github.apple.coremltools.source" : "torch==2.6.0",
90
- "com.github.apple.coremltools.version" : "8.2"
91
- },
92
- "inputSchema" : [
93
- {
94
- "hasShapeFlexibility" : "0",
95
- "isOptional" : "0",
96
- "dataType" : "Int32",
97
- "formattedType" : "MultiArray (Int32 1)",
98
- "shortDescription" : "",
99
- "shape" : "[1]",
100
- "name" : "input_ids",
101
- "type" : "MultiArray"
102
- },
103
- {
104
- "hasShapeFlexibility" : "0",
105
- "isOptional" : "0",
106
- "dataType" : "Int32",
107
- "formattedType" : "MultiArray (Int32 1)",
108
- "shortDescription" : "",
109
- "shape" : "[1]",
110
- "name" : "cache_length",
111
- "type" : "MultiArray"
112
- },
113
- {
114
- "hasShapeFlexibility" : "0",
115
- "isOptional" : "0",
116
- "dataType" : "Float16",
117
- "formattedType" : "MultiArray (Float16 1 × 1536 × 1 × 448)",
118
- "shortDescription" : "",
119
- "shape" : "[1, 1536, 1, 448]",
120
- "name" : "key_cache",
121
- "type" : "MultiArray"
122
- },
123
- {
124
- "hasShapeFlexibility" : "0",
125
- "isOptional" : "0",
126
- "dataType" : "Float16",
127
- "formattedType" : "MultiArray (Float16 1 × 1536 × 1 × 448)",
128
- "shortDescription" : "",
129
- "shape" : "[1, 1536, 1, 448]",
130
- "name" : "value_cache",
131
- "type" : "MultiArray"
132
- },
133
- {
134
- "hasShapeFlexibility" : "0",
135
- "isOptional" : "0",
136
- "dataType" : "Float16",
137
- "formattedType" : "MultiArray (Float16 1 × 448)",
138
- "shortDescription" : "",
139
- "shape" : "[1, 448]",
140
- "name" : "kv_cache_update_mask",
141
- "type" : "MultiArray"
142
- },
143
- {
144
- "hasShapeFlexibility" : "0",
145
- "isOptional" : "0",
146
- "dataType" : "Float16",
147
- "formattedType" : "MultiArray (Float16 1 × 384 × 1 × 1500)",
148
- "shortDescription" : "",
149
- "shape" : "[1, 384, 1, 1500]",
150
- "name" : "encoder_output_embeds",
151
- "type" : "MultiArray"
152
- },
153
- {
154
- "hasShapeFlexibility" : "0",
155
- "isOptional" : "0",
156
- "dataType" : "Float16",
157
- "formattedType" : "MultiArray (Float16 1 × 448)",
158
- "shortDescription" : "",
159
- "shape" : "[1, 448]",
160
- "name" : "decoder_key_padding_mask",
161
- "type" : "MultiArray"
162
- }
163
- ],
164
- "generatedClassName" : "TextDecoder",
165
- "method" : "predict"
166
- }
167
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/TextDecoder.mlmodelc/model.mil DELETED
The diff for this file is too large to render. See raw diff
 
openai_whisper-tiny/TextDecoder.mlmodelc/weights/weight.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:778d39fa5b45d8a628ba68b2037c7be19f7d22d9116bf8cfc06013272c7f6def
3
- size 59216434
 
 
 
 
openai_whisper-tiny/config.json DELETED
@@ -1,144 +0,0 @@
1
- {
2
- "_name_or_path": "openai/whisper-tiny",
3
- "activation_dropout": 0.0,
4
- "activation_function": "gelu",
5
- "architectures": [
6
- "WhisperForConditionalGeneration"
7
- ],
8
- "attention_dropout": 0.0,
9
- "begin_suppress_tokens": [
10
- 220,
11
- 50257
12
- ],
13
- "bos_token_id": 50257,
14
- "d_model": 384,
15
- "decoder_attention_heads": 6,
16
- "decoder_ffn_dim": 1536,
17
- "decoder_layerdrop": 0.0,
18
- "decoder_layers": 4,
19
- "decoder_start_token_id": 50258,
20
- "dropout": 0.0,
21
- "encoder_attention_heads": 6,
22
- "encoder_ffn_dim": 1536,
23
- "encoder_layerdrop": 0.0,
24
- "encoder_layers": 4,
25
- "eos_token_id": 50257,
26
- "forced_decoder_ids": [
27
- [
28
- 1,
29
- 50259
30
- ],
31
- [
32
- 2,
33
- 50359
34
- ],
35
- [
36
- 3,
37
- 50363
38
- ]
39
- ],
40
- "init_std": 0.02,
41
- "is_encoder_decoder": true,
42
- "max_length": 448,
43
- "max_source_positions": 1500,
44
- "max_target_positions": 448,
45
- "model_type": "whisper",
46
- "num_hidden_layers": 4,
47
- "num_mel_bins": 80,
48
- "pad_token_id": 50257,
49
- "scale_embedding": false,
50
- "suppress_tokens": [
51
- 1,
52
- 2,
53
- 7,
54
- 8,
55
- 9,
56
- 10,
57
- 14,
58
- 25,
59
- 26,
60
- 27,
61
- 28,
62
- 29,
63
- 31,
64
- 58,
65
- 59,
66
- 60,
67
- 61,
68
- 62,
69
- 63,
70
- 90,
71
- 91,
72
- 92,
73
- 93,
74
- 359,
75
- 503,
76
- 522,
77
- 542,
78
- 873,
79
- 893,
80
- 902,
81
- 918,
82
- 922,
83
- 931,
84
- 1350,
85
- 1853,
86
- 1982,
87
- 2460,
88
- 2627,
89
- 3246,
90
- 3253,
91
- 3268,
92
- 3536,
93
- 3846,
94
- 3961,
95
- 4183,
96
- 4667,
97
- 6585,
98
- 6647,
99
- 7273,
100
- 9061,
101
- 9383,
102
- 10428,
103
- 10929,
104
- 11938,
105
- 12033,
106
- 12331,
107
- 12562,
108
- 13793,
109
- 14157,
110
- 14635,
111
- 15265,
112
- 15618,
113
- 16553,
114
- 16604,
115
- 18362,
116
- 18956,
117
- 20075,
118
- 21675,
119
- 22520,
120
- 26130,
121
- 26161,
122
- 26435,
123
- 28279,
124
- 29464,
125
- 31650,
126
- 32302,
127
- 32470,
128
- 36865,
129
- 42863,
130
- 47425,
131
- 49870,
132
- 50254,
133
- 50258,
134
- 50358,
135
- 50359,
136
- 50360,
137
- 50361,
138
- 50362
139
- ],
140
- "torch_dtype": "float32",
141
- "transformers_version": "4.27.0.dev0",
142
- "use_cache": true,
143
- "vocab_size": 51865
144
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/generation_config.json DELETED
@@ -1,248 +0,0 @@
1
- {
2
- "alignment_heads": [
3
- [
4
- 2,
5
- 2
6
- ],
7
- [
8
- 3,
9
- 0
10
- ],
11
- [
12
- 3,
13
- 2
14
- ],
15
- [
16
- 3,
17
- 3
18
- ],
19
- [
20
- 3,
21
- 4
22
- ],
23
- [
24
- 3,
25
- 5
26
- ]
27
- ],
28
- "begin_suppress_tokens": [
29
- 220,
30
- 50257
31
- ],
32
- "bos_token_id": 50257,
33
- "decoder_start_token_id": 50258,
34
- "eos_token_id": 50257,
35
- "forced_decoder_ids": [
36
- [
37
- 1,
38
- null
39
- ],
40
- [
41
- 2,
42
- 50359
43
- ]
44
- ],
45
- "is_multilingual": true,
46
- "lang_to_id": {
47
- "<|af|>": 50327,
48
- "<|am|>": 50334,
49
- "<|ar|>": 50272,
50
- "<|as|>": 50350,
51
- "<|az|>": 50304,
52
- "<|ba|>": 50355,
53
- "<|be|>": 50330,
54
- "<|bg|>": 50292,
55
- "<|bn|>": 50302,
56
- "<|bo|>": 50347,
57
- "<|br|>": 50309,
58
- "<|bs|>": 50315,
59
- "<|ca|>": 50270,
60
- "<|cs|>": 50283,
61
- "<|cy|>": 50297,
62
- "<|da|>": 50285,
63
- "<|de|>": 50261,
64
- "<|el|>": 50281,
65
- "<|en|>": 50259,
66
- "<|es|>": 50262,
67
- "<|et|>": 50307,
68
- "<|eu|>": 50310,
69
- "<|fa|>": 50300,
70
- "<|fi|>": 50277,
71
- "<|fo|>": 50338,
72
- "<|fr|>": 50265,
73
- "<|gl|>": 50319,
74
- "<|gu|>": 50333,
75
- "<|haw|>": 50352,
76
- "<|ha|>": 50354,
77
- "<|he|>": 50279,
78
- "<|hi|>": 50276,
79
- "<|hr|>": 50291,
80
- "<|ht|>": 50339,
81
- "<|hu|>": 50286,
82
- "<|hy|>": 50312,
83
- "<|id|>": 50275,
84
- "<|is|>": 50311,
85
- "<|it|>": 50274,
86
- "<|ja|>": 50266,
87
- "<|jw|>": 50356,
88
- "<|ka|>": 50329,
89
- "<|kk|>": 50316,
90
- "<|km|>": 50323,
91
- "<|kn|>": 50306,
92
- "<|ko|>": 50264,
93
- "<|la|>": 50294,
94
- "<|lb|>": 50345,
95
- "<|ln|>": 50353,
96
- "<|lo|>": 50336,
97
- "<|lt|>": 50293,
98
- "<|lv|>": 50301,
99
- "<|mg|>": 50349,
100
- "<|mi|>": 50295,
101
- "<|mk|>": 50308,
102
- "<|ml|>": 50296,
103
- "<|mn|>": 50314,
104
- "<|mr|>": 50320,
105
- "<|ms|>": 50282,
106
- "<|mt|>": 50343,
107
- "<|my|>": 50346,
108
- "<|ne|>": 50313,
109
- "<|nl|>": 50271,
110
- "<|nn|>": 50342,
111
- "<|no|>": 50288,
112
- "<|oc|>": 50328,
113
- "<|pa|>": 50321,
114
- "<|pl|>": 50269,
115
- "<|ps|>": 50340,
116
- "<|pt|>": 50267,
117
- "<|ro|>": 50284,
118
- "<|ru|>": 50263,
119
- "<|sa|>": 50344,
120
- "<|sd|>": 50332,
121
- "<|si|>": 50322,
122
- "<|sk|>": 50298,
123
- "<|sl|>": 50305,
124
- "<|sn|>": 50324,
125
- "<|so|>": 50326,
126
- "<|sq|>": 50317,
127
- "<|sr|>": 50303,
128
- "<|su|>": 50357,
129
- "<|sv|>": 50273,
130
- "<|sw|>": 50318,
131
- "<|ta|>": 50287,
132
- "<|te|>": 50299,
133
- "<|tg|>": 50331,
134
- "<|th|>": 50289,
135
- "<|tk|>": 50341,
136
- "<|tl|>": 50348,
137
- "<|tr|>": 50268,
138
- "<|tt|>": 50351,
139
- "<|uk|>": 50280,
140
- "<|ur|>": 50290,
141
- "<|uz|>": 50337,
142
- "<|vi|>": 50278,
143
- "<|yi|>": 50335,
144
- "<|yo|>": 50325,
145
- "<|zh|>": 50260
146
- },
147
- "max_initial_timestamp_index": 50,
148
- "max_length": 448,
149
- "no_timestamps_token_id": 50363,
150
- "pad_token_id": 50257,
151
- "prev_sot_token_id": 50361,
152
- "return_timestamps": false,
153
- "suppress_tokens": [
154
- 1,
155
- 2,
156
- 7,
157
- 8,
158
- 9,
159
- 10,
160
- 14,
161
- 25,
162
- 26,
163
- 27,
164
- 28,
165
- 29,
166
- 31,
167
- 58,
168
- 59,
169
- 60,
170
- 61,
171
- 62,
172
- 63,
173
- 90,
174
- 91,
175
- 92,
176
- 93,
177
- 359,
178
- 503,
179
- 522,
180
- 542,
181
- 873,
182
- 893,
183
- 902,
184
- 918,
185
- 922,
186
- 931,
187
- 1350,
188
- 1853,
189
- 1982,
190
- 2460,
191
- 2627,
192
- 3246,
193
- 3253,
194
- 3268,
195
- 3536,
196
- 3846,
197
- 3961,
198
- 4183,
199
- 4667,
200
- 6585,
201
- 6647,
202
- 7273,
203
- 9061,
204
- 9383,
205
- 10428,
206
- 10929,
207
- 11938,
208
- 12033,
209
- 12331,
210
- 12562,
211
- 13793,
212
- 14157,
213
- 14635,
214
- 15265,
215
- 15618,
216
- 16553,
217
- 16604,
218
- 18362,
219
- 18956,
220
- 20075,
221
- 21675,
222
- 22520,
223
- 26130,
224
- 26161,
225
- 26435,
226
- 28279,
227
- 29464,
228
- 31650,
229
- 32302,
230
- 32470,
231
- 36865,
232
- 42863,
233
- 47425,
234
- 49870,
235
- 50254,
236
- 50258,
237
- 50358,
238
- 50359,
239
- 50360,
240
- 50361,
241
- 50362
242
- ],
243
- "task_to_id": {
244
- "transcribe": 50359,
245
- "translate": 50358
246
- },
247
- "transformers_version": "4.31.0.dev0"
248
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/analytics/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:211457b92a0ced67bb8625efe39799a0030c4fc71eb87d7284ea81043caccde7
3
- size 243
 
 
 
 
openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d68f152b6573ac55203a3dc8383730e6ecde685c7d2a88815b89820c88e35371
3
- size 347
 
 
 
 
openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/metadata.json DELETED
@@ -1,69 +0,0 @@
1
- [
2
- {
3
- "metadataOutputVersion" : "3.0",
4
- "storagePrecision" : "Float16",
5
- "outputSchema" : [
6
- {
7
- "hasShapeFlexibility" : "0",
8
- "isOptional" : "0",
9
- "dataType" : "Float16",
10
- "formattedType" : "MultiArray (Float16 1 × 768 × 1 × 1500)",
11
- "shortDescription" : "",
12
- "shape" : "[1, 768, 1, 1500]",
13
- "name" : "encoder_output_embeds",
14
- "type" : "MultiArray"
15
- }
16
- ],
17
- "modelParameters" : [
18
-
19
- ],
20
- "specificationVersion" : 7,
21
- "mlProgramOperationTypeHistogram" : {
22
- "Concat" : 156,
23
- "Ios16.rsqrt" : 25,
24
- "Ios16.mul" : 626,
25
- "SliceByIndex" : 1008,
26
- "Ios16.sub" : 25,
27
- "Transpose" : 12,
28
- "Ios16.einsum" : 1152,
29
- "Ios16.conv" : 74,
30
- "Ios16.add" : 50,
31
- "Ios16.reduceMean" : 50,
32
- "Ios16.softmax" : 576,
33
- "Ios16.gelu" : 14,
34
- "Ios16.batchNorm" : 25
35
- },
36
- "computePrecision" : "Mixed (Float16, Int32)",
37
- "isUpdatable" : "0",
38
- "availability" : {
39
- "macOS" : "13.0",
40
- "tvOS" : "16.0",
41
- "visionOS" : "1.0",
42
- "watchOS" : "9.0",
43
- "iOS" : "16.0",
44
- "macCatalyst" : "16.0"
45
- },
46
- "modelType" : {
47
- "name" : "MLModelType_mlProgram"
48
- },
49
- "userDefinedMetadata" : {
50
- "com.github.apple.coremltools.source_dialect" : "TorchScript",
51
- "com.github.apple.coremltools.source" : "torch==2.2.1",
52
- "com.github.apple.coremltools.version" : "7.1"
53
- },
54
- "inputSchema" : [
55
- {
56
- "hasShapeFlexibility" : "0",
57
- "isOptional" : "0",
58
- "dataType" : "Float16",
59
- "formattedType" : "MultiArray (Float16 1 × 80 × 1 × 3000)",
60
- "shortDescription" : "",
61
- "shape" : "[1, 80, 1, 3000]",
62
- "name" : "melspectrogram_features",
63
- "type" : "MultiArray"
64
- }
65
- ],
66
- "generatedClassName" : "AudioEncoder",
67
- "method" : "predict"
68
- }
69
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/model.mil DELETED
The diff for this file is too large to render. See raw diff
 
openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/model.mlmodel DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:68ca04660b8b050c68ca54c27d97c47e4133bc591422cb7009de8922d56fb8c9
3
- size 155271
 
 
 
 
openai_whisper-tiny/openai_whisper-small/AudioEncoder.mlmodelc/weights/weight.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe35cef2c9406993a635639b16f373f6debb0215ac115b7bf93fa03c8e10310b
3
- size 176323456
 
 
 
 
openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/analytics/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f77e6457285248f99cd7aa3fd4cc2efbb17733e63e7023ac53abe1f95785d07
3
- size 243
 
 
 
 
openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:dabdc5aa69f6ef4d97dc9499f5c30514e00e96b53b750b33a5a6471363c71662
3
- size 328
 
 
 
 
openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/metadata.json DELETED
@@ -1,71 +0,0 @@
1
- [
2
- {
3
- "metadataOutputVersion" : "3.0",
4
- "storagePrecision" : "Float16",
5
- "outputSchema" : [
6
- {
7
- "hasShapeFlexibility" : "0",
8
- "isOptional" : "0",
9
- "dataType" : "Float16",
10
- "formattedType" : "MultiArray (Float16 1 × 80 × 1 × 3000)",
11
- "shortDescription" : "",
12
- "shape" : "[1, 80, 1, 3000]",
13
- "name" : "melspectrogram_features",
14
- "type" : "MultiArray"
15
- }
16
- ],
17
- "modelParameters" : [
18
-
19
- ],
20
- "specificationVersion" : 7,
21
- "mlProgramOperationTypeHistogram" : {
22
- "Pad" : 1,
23
- "Ios16.mul" : 2,
24
- "SliceByIndex" : 1,
25
- "Ios16.sub" : 1,
26
- "Ios16.log" : 1,
27
- "Ios16.conv" : 2,
28
- "Ios16.add" : 3,
29
- "Ios16.square" : 2,
30
- "Ios16.matmul" : 1,
31
- "Squeeze" : 2,
32
- "Ios16.maximum" : 1,
33
- "ExpandDims" : 4,
34
- "Ios16.reduceMax" : 1,
35
- "Identity" : 1,
36
- "Ios16.reshape" : 2
37
- },
38
- "computePrecision" : "Mixed (Float16, Int32)",
39
- "isUpdatable" : "0",
40
- "availability" : {
41
- "macOS" : "13.0",
42
- "tvOS" : "16.0",
43
- "visionOS" : "1.0",
44
- "watchOS" : "9.0",
45
- "iOS" : "16.0",
46
- "macCatalyst" : "16.0"
47
- },
48
- "modelType" : {
49
- "name" : "MLModelType_mlProgram"
50
- },
51
- "userDefinedMetadata" : {
52
- "com.github.apple.coremltools.source_dialect" : "TorchScript",
53
- "com.github.apple.coremltools.source" : "torch==2.2.1",
54
- "com.github.apple.coremltools.version" : "7.1"
55
- },
56
- "inputSchema" : [
57
- {
58
- "hasShapeFlexibility" : "0",
59
- "isOptional" : "0",
60
- "dataType" : "Float16",
61
- "formattedType" : "MultiArray (Float16 480000)",
62
- "shortDescription" : "",
63
- "shape" : "[480000]",
64
- "name" : "audio",
65
- "type" : "MultiArray"
66
- }
67
- ],
68
- "generatedClassName" : "MelSpectrogram",
69
- "method" : "predict"
70
- }
71
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/model.mil DELETED
@@ -1,66 +0,0 @@
1
- program(1.0)
2
- [buildInfo = dict<tensor<string, []>, tensor<string, []>>({{"coremlc-component-MIL", "5.33.5"}, {"coremlc-version", "1877.40.3"}, {"coremltools-component-torch", "2.2.1"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "7.1"}})]
3
- {
4
- func main<ios16>(tensor<fp16, [480000]> audio) {
5
- tensor<int32, [3]> var_10 = const()[name = tensor<string, []>("op_10"), val = tensor<int32, [3]>([1, 1, 480000])];
6
- tensor<fp16, [1, 1, 480000]> input_1_cast_fp16 = reshape(shape = var_10, x = audio)[name = tensor<string, []>("input_1_cast_fp16")];
7
- tensor<int32, [6]> input_3_pad_0 = const()[name = tensor<string, []>("input_3_pad_0"), val = tensor<int32, [6]>([0, 0, 0, 0, 200, 200])];
8
- tensor<string, []> input_3_mode_0 = const()[name = tensor<string, []>("input_3_mode_0"), val = tensor<string, []>("reflect")];
9
- tensor<fp16, []> input_3_constant_val_0_to_fp16 = const()[name = tensor<string, []>("input_3_constant_val_0_to_fp16"), val = tensor<fp16, []>(0x0p+0)];
10
- tensor<fp16, [1, 1, 480400]> input_3_cast_fp16 = pad(constant_val = input_3_constant_val_0_to_fp16, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1_cast_fp16)[name = tensor<string, []>("input_3_cast_fp16")];
11
- tensor<int32, [1]> var_22 = const()[name = tensor<string, []>("op_22"), val = tensor<int32, [1]>([480400])];
12
- tensor<fp16, [480400]> input_cast_fp16 = reshape(shape = var_22, x = input_3_cast_fp16)[name = tensor<string, []>("input_cast_fp16")];
13
- tensor<int32, [1]> expand_dims_0_axes_0 = const()[name = tensor<string, []>("expand_dims_0_axes_0"), val = tensor<int32, [1]>([0])];
14
- tensor<fp16, [1, 480400]> expand_dims_0_cast_fp16 = expand_dims(axes = expand_dims_0_axes_0, x = input_cast_fp16)[name = tensor<string, []>("expand_dims_0_cast_fp16")];
15
- tensor<int32, [1]> expand_dims_3 = const()[name = tensor<string, []>("expand_dims_3"), val = tensor<int32, [1]>([160])];
16
- tensor<int32, [1]> expand_dims_4_axes_0 = const()[name = tensor<string, []>("expand_dims_4_axes_0"), val = tensor<int32, [1]>([1])];
17
- tensor<fp16, [1, 1, 480400]> expand_dims_4_cast_fp16 = expand_dims(axes = expand_dims_4_axes_0, x = expand_dims_0_cast_fp16)[name = tensor<string, []>("expand_dims_4_cast_fp16")];
18
- tensor<string, []> conv_0_pad_type_0 = const()[name = tensor<string, []>("conv_0_pad_type_0"), val = tensor<string, []>("valid")];
19
- tensor<int32, [2]> conv_0_pad_0 = const()[name = tensor<string, []>("conv_0_pad_0"), val = tensor<int32, [2]>([0, 0])];
20
- tensor<int32, [1]> conv_0_dilations_0 = const()[name = tensor<string, []>("conv_0_dilations_0"), val = tensor<int32, [1]>([1])];
21
- tensor<int32, []> conv_0_groups_0 = const()[name = tensor<string, []>("conv_0_groups_0"), val = tensor<int32, []>(1)];
22
- tensor<fp16, [201, 1, 400]> expand_dims_1_to_fp16 = const()[name = tensor<string, []>("expand_dims_1_to_fp16"), val = tensor<fp16, [201, 1, 400]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(64)))];
23
- tensor<fp16, [1, 201, 3001]> conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_3, weight = expand_dims_1_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor<string, []>("conv_0_cast_fp16")];
24
- tensor<string, []> conv_1_pad_type_0 = const()[name = tensor<string, []>("conv_1_pad_type_0"), val = tensor<string, []>("valid")];
25
- tensor<int32, [2]> conv_1_pad_0 = const()[name = tensor<string, []>("conv_1_pad_0"), val = tensor<int32, [2]>([0, 0])];
26
- tensor<int32, [1]> conv_1_dilations_0 = const()[name = tensor<string, []>("conv_1_dilations_0"), val = tensor<int32, [1]>([1])];
27
- tensor<int32, []> conv_1_groups_0 = const()[name = tensor<string, []>("conv_1_groups_0"), val = tensor<int32, []>(1)];
28
- tensor<fp16, [201, 1, 400]> expand_dims_2_to_fp16 = const()[name = tensor<string, []>("expand_dims_2_to_fp16"), val = tensor<fp16, [201, 1, 400]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(160960)))];
29
- tensor<fp16, [1, 201, 3001]> conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_3, weight = expand_dims_2_to_fp16, x = expand_dims_4_cast_fp16)[name = tensor<string, []>("conv_1_cast_fp16")];
30
- tensor<int32, [1]> squeeze_0_axes_0 = const()[name = tensor<string, []>("squeeze_0_axes_0"), val = tensor<int32, [1]>([0])];
31
- tensor<fp16, [201, 3001]> squeeze_0_cast_fp16 = squeeze(axes = squeeze_0_axes_0, x = conv_0_cast_fp16)[name = tensor<string, []>("squeeze_0_cast_fp16")];
32
- tensor<int32, [1]> squeeze_1_axes_0 = const()[name = tensor<string, []>("squeeze_1_axes_0"), val = tensor<int32, [1]>([0])];
33
- tensor<fp16, [201, 3001]> squeeze_1_cast_fp16 = squeeze(axes = squeeze_1_axes_0, x = conv_1_cast_fp16)[name = tensor<string, []>("squeeze_1_cast_fp16")];
34
- tensor<fp16, [201, 3001]> square_0_cast_fp16 = square(x = squeeze_0_cast_fp16)[name = tensor<string, []>("square_0_cast_fp16")];
35
- tensor<fp16, [201, 3001]> square_1_cast_fp16 = square(x = squeeze_1_cast_fp16)[name = tensor<string, []>("square_1_cast_fp16")];
36
- tensor<fp16, [201, 3001]> add_1_cast_fp16 = add(x = square_0_cast_fp16, y = square_1_cast_fp16)[name = tensor<string, []>("add_1_cast_fp16")];
37
- tensor<fp16, [201, 3001]> magnitudes_1_cast_fp16 = identity(x = add_1_cast_fp16)[name = tensor<string, []>("magnitudes_1_cast_fp16")];
38
- tensor<int32, [2]> magnitudes_begin_0 = const()[name = tensor<string, []>("magnitudes_begin_0"), val = tensor<int32, [2]>([0, 0])];
39
- tensor<int32, [2]> magnitudes_end_0 = const()[name = tensor<string, []>("magnitudes_end_0"), val = tensor<int32, [2]>([201, 3000])];
40
- tensor<bool, [2]> magnitudes_end_mask_0 = const()[name = tensor<string, []>("magnitudes_end_mask_0"), val = tensor<bool, [2]>([true, false])];
41
- tensor<fp16, [201, 3000]> magnitudes_cast_fp16 = slice_by_index(begin = magnitudes_begin_0, end = magnitudes_end_0, end_mask = magnitudes_end_mask_0, x = magnitudes_1_cast_fp16)[name = tensor<string, []>("magnitudes_cast_fp16")];
42
- tensor<bool, []> mel_spec_1_transpose_x_0 = const()[name = tensor<string, []>("mel_spec_1_transpose_x_0"), val = tensor<bool, []>(false)];
43
- tensor<bool, []> mel_spec_1_transpose_y_0 = const()[name = tensor<string, []>("mel_spec_1_transpose_y_0"), val = tensor<bool, []>(false)];
44
- tensor<fp16, [80, 201]> mel_filters_to_fp16 = const()[name = tensor<string, []>("mel_filters_to_fp16"), val = tensor<fp16, [80, 201]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(321856)))];
45
- tensor<fp16, [80, 3000]> mel_spec_1_cast_fp16 = matmul(transpose_x = mel_spec_1_transpose_x_0, transpose_y = mel_spec_1_transpose_y_0, x = mel_filters_to_fp16, y = magnitudes_cast_fp16)[name = tensor<string, []>("mel_spec_1_cast_fp16")];
46
- tensor<fp16, []> var_41_to_fp16 = const()[name = tensor<string, []>("op_41_to_fp16"), val = tensor<fp16, []>(0x1p-24)];
47
- tensor<fp16, [80, 3000]> mel_spec_cast_fp16 = add(x = mel_spec_1_cast_fp16, y = var_41_to_fp16)[name = tensor<string, []>("mel_spec_cast_fp16")];
48
- tensor<fp16, []> log_0_epsilon_0_to_fp16 = const()[name = tensor<string, []>("log_0_epsilon_0_to_fp16"), val = tensor<fp16, []>(0x0p+0)];
49
- tensor<fp16, [80, 3000]> log_0_cast_fp16 = log(epsilon = log_0_epsilon_0_to_fp16, x = mel_spec_cast_fp16)[name = tensor<string, []>("log_0_cast_fp16")];
50
- tensor<fp16, []> mul_0_y_0_to_fp16 = const()[name = tensor<string, []>("mul_0_y_0_to_fp16"), val = tensor<fp16, []>(0x1.bccp-2)];
51
- tensor<fp16, [80, 3000]> mul_0_cast_fp16 = mul(x = log_0_cast_fp16, y = mul_0_y_0_to_fp16)[name = tensor<string, []>("mul_0_cast_fp16")];
52
- tensor<bool, []> var_44_keep_dims_0 = const()[name = tensor<string, []>("op_44_keep_dims_0"), val = tensor<bool, []>(false)];
53
- tensor<fp16, []> var_44_cast_fp16 = reduce_max(keep_dims = var_44_keep_dims_0, x = mul_0_cast_fp16)[name = tensor<string, []>("op_44_cast_fp16")];
54
- tensor<fp16, []> var_46_to_fp16 = const()[name = tensor<string, []>("op_46_to_fp16"), val = tensor<fp16, []>(0x1p+3)];
55
- tensor<fp16, []> var_47_cast_fp16 = sub(x = var_44_cast_fp16, y = var_46_to_fp16)[name = tensor<string, []>("op_47_cast_fp16")];
56
- tensor<fp16, [80, 3000]> log_spec_3_cast_fp16 = maximum(x = mul_0_cast_fp16, y = var_47_cast_fp16)[name = tensor<string, []>("log_spec_3_cast_fp16")];
57
- tensor<fp16, []> var_50_to_fp16 = const()[name = tensor<string, []>("op_50_to_fp16"), val = tensor<fp16, []>(0x1p+2)];
58
- tensor<fp16, [80, 3000]> var_51_cast_fp16 = add(x = log_spec_3_cast_fp16, y = var_50_to_fp16)[name = tensor<string, []>("op_51_cast_fp16")];
59
- tensor<fp16, []> _inversed_log_spec_y_0_to_fp16 = const()[name = tensor<string, []>("_inversed_log_spec_y_0_to_fp16"), val = tensor<fp16, []>(0x1p-2)];
60
- tensor<fp16, [80, 3000]> _inversed_log_spec_cast_fp16 = mul(x = var_51_cast_fp16, y = _inversed_log_spec_y_0_to_fp16)[name = tensor<string, []>("_inversed_log_spec_cast_fp16")];
61
- tensor<int32, [1]> var_55_axes_0 = const()[name = tensor<string, []>("op_55_axes_0"), val = tensor<int32, [1]>([0])];
62
- tensor<fp16, [1, 80, 3000]> var_55_cast_fp16 = expand_dims(axes = var_55_axes_0, x = _inversed_log_spec_cast_fp16)[name = tensor<string, []>("op_55_cast_fp16")];
63
- tensor<int32, [1]> var_62_axes_0 = const()[name = tensor<string, []>("op_62_axes_0"), val = tensor<int32, [1]>([2])];
64
- tensor<fp16, [1, 80, 1, 3000]> melspectrogram_features = expand_dims(axes = var_62_axes_0, x = var_55_cast_fp16)[name = tensor<string, []>("op_62_cast_fp16")];
65
- } -> (melspectrogram_features);
66
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/openai_whisper-small/MelSpectrogram.mlmodelc/weights/weight.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:267017e533b5f542d195fd9a775f2ba649075128283ce8e86c63a2ec20de5b07
3
- size 354080
 
 
 
 
openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/analytics/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:39c0d6d55353bc61ef8071081bb958dd1ab7b0b7f2a3338a797f1a64211e084c
3
- size 243
 
 
 
 
openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/coremldata.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2ccd0b8920701386ab9554f7db47b43e55ee07863280ee5d829d5272839adc2
3
- size 633
 
 
 
 
openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/metadata.json DELETED
@@ -1,165 +0,0 @@
1
- [
2
- {
3
- "metadataOutputVersion" : "3.0",
4
- "storagePrecision" : "Float16",
5
- "outputSchema" : [
6
- {
7
- "hasShapeFlexibility" : "0",
8
- "isOptional" : "0",
9
- "dataType" : "Float16",
10
- "formattedType" : "MultiArray (Float16 1 × 1 × 51865)",
11
- "shortDescription" : "",
12
- "shape" : "[1, 1, 51865]",
13
- "name" : "logits",
14
- "type" : "MultiArray"
15
- },
16
- {
17
- "hasShapeFlexibility" : "0",
18
- "isOptional" : "0",
19
- "dataType" : "Float16",
20
- "formattedType" : "MultiArray (Float16 1 × 9216 × 1 × 1)",
21
- "shortDescription" : "",
22
- "shape" : "[1, 9216, 1, 1]",
23
- "name" : "key_cache_updates",
24
- "type" : "MultiArray"
25
- },
26
- {
27
- "hasShapeFlexibility" : "0",
28
- "isOptional" : "0",
29
- "dataType" : "Float16",
30
- "formattedType" : "MultiArray (Float16 1 × 9216 × 1 × 1)",
31
- "shortDescription" : "",
32
- "shape" : "[1, 9216, 1, 1]",
33
- "name" : "value_cache_updates",
34
- "type" : "MultiArray"
35
- },
36
- {
37
- "hasShapeFlexibility" : "0",
38
- "isOptional" : "0",
39
- "dataType" : "Float16",
40
- "formattedType" : "MultiArray (Float16 1 × 1500)",
41
- "shortDescription" : "",
42
- "shape" : "[1, 1500]",
43
- "name" : "alignment_heads_weights",
44
- "type" : "MultiArray"
45
- }
46
- ],
47
- "modelParameters" : [
48
-
49
- ],
50
- "specificationVersion" : 7,
51
- "mlProgramOperationTypeHistogram" : {
52
- "Split" : 2,
53
- "Concat" : 3,
54
- "Ios16.rsqrt" : 37,
55
- "Ios16.mul" : 146,
56
- "Squeeze" : 1,
57
- "SliceByIndex" : 20,
58
- "Ios16.sub" : 38,
59
- "Transpose" : 1,
60
- "Ios16.conv" : 120,
61
- "Ios16.add" : 110,
62
- "Ios16.linear" : 1,
63
- "Ios16.matmul" : 48,
64
- "Ios16.gelu" : 12,
65
- "Ios16.reduceMean" : 75,
66
- "ExpandDims" : 6,
67
- "Ios16.batchNorm" : 37,
68
- "Ios16.gather" : 2,
69
- "Ios16.reshape" : 96,
70
- "Ios16.softmax" : 24
71
- },
72
- "computePrecision" : "Mixed (Float16, Int32)",
73
- "isUpdatable" : "0",
74
- "availability" : {
75
- "macOS" : "13.0",
76
- "tvOS" : "16.0",
77
- "visionOS" : "1.0",
78
- "watchOS" : "9.0",
79
- "iOS" : "16.0",
80
- "macCatalyst" : "16.0"
81
- },
82
- "modelType" : {
83
- "name" : "MLModelType_mlProgram"
84
- },
85
- "userDefinedMetadata" : {
86
- "com.github.apple.coremltools.source_dialect" : "TorchScript",
87
- "com.github.apple.coremltools.source" : "torch==2.2.1",
88
- "com.github.apple.coremltools.version" : "7.1"
89
- },
90
- "inputSchema" : [
91
- {
92
- "hasShapeFlexibility" : "0",
93
- "isOptional" : "0",
94
- "dataType" : "Int32",
95
- "formattedType" : "MultiArray (Int32 1)",
96
- "shortDescription" : "",
97
- "shape" : "[1]",
98
- "name" : "input_ids",
99
- "type" : "MultiArray"
100
- },
101
- {
102
- "hasShapeFlexibility" : "0",
103
- "isOptional" : "0",
104
- "dataType" : "Int32",
105
- "formattedType" : "MultiArray (Int32 1)",
106
- "shortDescription" : "",
107
- "shape" : "[1]",
108
- "name" : "cache_length",
109
- "type" : "MultiArray"
110
- },
111
- {
112
- "hasShapeFlexibility" : "0",
113
- "isOptional" : "0",
114
- "dataType" : "Float16",
115
- "formattedType" : "MultiArray (Float16 1 × 9216 × 1 × 224)",
116
- "shortDescription" : "",
117
- "shape" : "[1, 9216, 1, 224]",
118
- "name" : "key_cache",
119
- "type" : "MultiArray"
120
- },
121
- {
122
- "hasShapeFlexibility" : "0",
123
- "isOptional" : "0",
124
- "dataType" : "Float16",
125
- "formattedType" : "MultiArray (Float16 1 × 9216 × 1 × 224)",
126
- "shortDescription" : "",
127
- "shape" : "[1, 9216, 1, 224]",
128
- "name" : "value_cache",
129
- "type" : "MultiArray"
130
- },
131
- {
132
- "hasShapeFlexibility" : "0",
133
- "isOptional" : "0",
134
- "dataType" : "Float16",
135
- "formattedType" : "MultiArray (Float16 1 × 224)",
136
- "shortDescription" : "",
137
- "shape" : "[1, 224]",
138
- "name" : "kv_cache_update_mask",
139
- "type" : "MultiArray"
140
- },
141
- {
142
- "hasShapeFlexibility" : "0",
143
- "isOptional" : "0",
144
- "dataType" : "Float16",
145
- "formattedType" : "MultiArray (Float16 1 × 768 × 1 × 1500)",
146
- "shortDescription" : "",
147
- "shape" : "[1, 768, 1, 1500]",
148
- "name" : "encoder_output_embeds",
149
- "type" : "MultiArray"
150
- },
151
- {
152
- "hasShapeFlexibility" : "0",
153
- "isOptional" : "0",
154
- "dataType" : "Float16",
155
- "formattedType" : "MultiArray (Float16 1 × 224)",
156
- "shortDescription" : "",
157
- "shape" : "[1, 224]",
158
- "name" : "decoder_key_padding_mask",
159
- "type" : "MultiArray"
160
- }
161
- ],
162
- "generatedClassName" : "TextDecoder",
163
- "method" : "predict"
164
- }
165
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/model.mil DELETED
The diff for this file is too large to render. See raw diff
 
openai_whisper-tiny/openai_whisper-small/TextDecoder.mlmodelc/model.mlmodel DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ea861c6dfdd866ed0f2e7fe0c3df7459daa44481cb25236e03698dd6d259391
3
- size 313629
 
 
 
 
openai_whisper-tiny/openai_whisper-small/config.json DELETED
@@ -1 +0,0 @@
1
- {"_name_or_path": "openai/whisper-small", "activation_dropout": 0.0, "activation_function": "gelu", "architectures": ["WhisperForConditionalGeneration"], "attention_dropout": 0.0, "begin_suppress_tokens": [220, 50257], "bos_token_id": 50257, "d_model": 768, "decoder_attention_heads": 12, "decoder_ffn_dim": 3072, "decoder_layerdrop": 0.0, "decoder_layers": 12, "decoder_start_token_id": 50258, "dropout": 0.0, "encoder_attention_heads": 12, "encoder_ffn_dim": 3072, "encoder_layerdrop": 0.0, "encoder_layers": 12, "eos_token_id": 50257, "forced_decoder_ids": [[1, 50259], [2, 50359], [3, 50363]], "init_std": 0.02, "is_encoder_decoder": true, "max_length": 448, "max_source_positions": 1500, "max_target_positions": 448, "model_type": "whisper", "num_hidden_layers": 12, "num_mel_bins": 80, "pad_token_id": 50257, "scale_embedding": false, "suppress_tokens": [1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50360, 50361, 50362], "torch_dtype": "float32", "transformers_version": "4.27.0.dev0", "use_cache": true, "vocab_size": 51865}
 
 
openai_whisper-tiny/openai_whisper-small/generation_config.json DELETED
@@ -1 +0,0 @@
1
- {"alignment_heads": [[5, 3], [5, 9], [8, 0], [8, 4], [8, 7], [8, 8], [9, 0], [9, 7], [9, 9], [10, 5]], "begin_suppress_tokens": [220, 50257], "bos_token_id": 50257, "decoder_start_token_id": 50258, "eos_token_id": 50257, "forced_decoder_ids": [[1, null], [2, 50359]], "is_multilingual": true, "lang_to_id": {"<|af|>": 50327, "<|am|>": 50334, "<|ar|>": 50272, "<|as|>": 50350, "<|az|>": 50304, "<|ba|>": 50355, "<|be|>": 50330, "<|bg|>": 50292, "<|bn|>": 50302, "<|bo|>": 50347, "<|br|>": 50309, "<|bs|>": 50315, "<|ca|>": 50270, "<|cs|>": 50283, "<|cy|>": 50297, "<|da|>": 50285, "<|de|>": 50261, "<|el|>": 50281, "<|en|>": 50259, "<|es|>": 50262, "<|et|>": 50307, "<|eu|>": 50310, "<|fa|>": 50300, "<|fi|>": 50277, "<|fo|>": 50338, "<|fr|>": 50265, "<|gl|>": 50319, "<|gu|>": 50333, "<|haw|>": 50352, "<|ha|>": 50354, "<|he|>": 50279, "<|hi|>": 50276, "<|hr|>": 50291, "<|ht|>": 50339, "<|hu|>": 50286, "<|hy|>": 50312, "<|id|>": 50275, "<|is|>": 50311, "<|it|>": 50274, "<|ja|>": 50266, "<|jw|>": 50356, "<|ka|>": 50329, "<|kk|>": 50316, "<|km|>": 50323, "<|kn|>": 50306, "<|ko|>": 50264, "<|la|>": 50294, "<|lb|>": 50345, "<|ln|>": 50353, "<|lo|>": 50336, "<|lt|>": 50293, "<|lv|>": 50301, "<|mg|>": 50349, "<|mi|>": 50295, "<|mk|>": 50308, "<|ml|>": 50296, "<|mn|>": 50314, "<|mr|>": 50320, "<|ms|>": 50282, "<|mt|>": 50343, "<|my|>": 50346, "<|ne|>": 50313, "<|nl|>": 50271, "<|nn|>": 50342, "<|no|>": 50288, "<|oc|>": 50328, "<|pa|>": 50321, "<|pl|>": 50269, "<|ps|>": 50340, "<|pt|>": 50267, "<|ro|>": 50284, "<|ru|>": 50263, "<|sa|>": 50344, "<|sd|>": 50332, "<|si|>": 50322, "<|sk|>": 50298, "<|sl|>": 50305, "<|sn|>": 50324, "<|so|>": 50326, "<|sq|>": 50317, "<|sr|>": 50303, "<|su|>": 50357, "<|sv|>": 50273, "<|sw|>": 50318, "<|ta|>": 50287, "<|te|>": 50299, "<|tg|>": 50331, "<|th|>": 50289, "<|tk|>": 50341, "<|tl|>": 50348, "<|tr|>": 50268, "<|tt|>": 50351, "<|uk|>": 50280, "<|ur|>": 50290, "<|uz|>": 50337, "<|vi|>": 50278, "<|yi|>": 50335, "<|yo|>": 50325, "<|zh|>": 50260}, "max_initial_timestamp_index": 50, "max_length": 448, "no_timestamps_token_id": 50363, "pad_token_id": 50257, "prev_sot_token_id": 50361, "return_timestamps": false, "suppress_tokens": [1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 359, 503, 522, 542, 873, 893, 902, 918, 922, 931, 1350, 1853, 1982, 2460, 2627, 3246, 3253, 3268, 3536, 3846, 3961, 4183, 4667, 6585, 6647, 7273, 9061, 9383, 10428, 10929, 11938, 12033, 12331, 12562, 13793, 14157, 14635, 15265, 15618, 16553, 16604, 18362, 18956, 20075, 21675, 22520, 26130, 26161, 26435, 28279, 29464, 31650, 32302, 32470, 36865, 42863, 47425, 49870, 50254, 50258, 50358, 50359, 50360, 50361, 50362], "task_to_id": {"transcribe": 50359, "translate": 50358}, "transformers_version": "4.31.0.dev0"}