dacorvo HF Staff commited on
Commit
614cbff
·
verified ·
1 Parent(s): 3dcc8d4

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +36 -0
  2. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.2.dev1/qwen3/Qwen/Qwen3-0.6B/4ddf27fb5dfd4909a319.json +87 -0
  3. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.5.dev1/60feecaa0c4c075e2f3e46a3f55d9a273f0ddd75a0ecf64e4ae27352e0819506/ee71309c308c0bf312e7.json +87 -0
  4. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.5.dev1/qwen3/Qwen/Qwen3-Embedding-0.6B/ee71309c308c0bf312e7.json +87 -0
  5. neuronxcc-2.21.18209.0+043b1bf7/MODULE_9d35248a93d4142a3cf1+24129607/model.hlo_module.pb +1 -1
  6. neuronxcc-2.21.18209.0+043b1bf7/MODULE_9d35248a93d4142a3cf1+24129607/model.neff +1 -1
  7. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/compile_flags.json +1 -0
  8. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/model.done +0 -0
  9. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/model.hlo_module.pb +3 -0
  10. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/model.neff +3 -0
  11. neuronxcc-2.21.18209.0+043b1bf7/MODULE_c2b3afa51c57d431a332+a02c3a36/model.neff +1 -1
  12. neuronxcc-2.21.18209.0+043b1bf7/MODULE_c2b3afa51c57d431a332+a02c3a36/wrapped_neff.hlo +1 -1
  13. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/compile_flags.json +1 -0
  14. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/model.done +0 -0
  15. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/model.hlo_module.pb +3 -0
  16. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/model.neff +3 -0
  17. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/compile_flags.json +1 -0
  18. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/model.done +0 -0
  19. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/model.hlo_module.pb +3 -0
  20. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/model.neff +3 -0
  21. neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/wrapped_neff.hlo +3 -0
  22. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1/43f30d107729db6b760b2a90833f43ca44f68e6468ef64e131f71e2fa9f5b23f/b76d86cc7c31e0e29c99.json +104 -0
  23. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1/43f30d107729db6b760b2a90833f43ca44f68e6468ef64e131f71e2fa9f5b23f/f71907dd62c8b40e471f.json +104 -0
  24. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1/gpt_oss/openai/gpt-oss-20b/b76d86cc7c31e0e29c99.json +104 -0
  25. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/026fe44014d3f650a32e.json +95 -0
  26. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/22e7cd94e28b353a796b.json +95 -0
  27. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/535e7795fb656220e11a.json +95 -0
  28. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/644bf9550db2b89192aa.json +95 -0
  29. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/6501938b047f7d373cdd.json +95 -0
  30. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/7bfaacc5ae3961c38f6c.json +95 -0
  31. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/912ca3353189929b7b15.json +95 -0
  32. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/b2f13b35b2326e133272.json +95 -0
  33. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/b4a6b1d49ffe1fb09037.json +95 -0
  34. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/b52dd6b442b63b75e7b7.json +95 -0
  35. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/bba41446ac8406c873ee.json +95 -0
  36. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/c67e5e7f18f7dbf76e0c.json +95 -0
  37. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/d27d49077cbb0bf50eb9.json +95 -0
  38. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/d59c183509c2a322fd8b.json +95 -0
  39. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/f4ff220867f67d4c7b95.json +95 -0
  40. neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/f56d8d1a0bac69205ae6.json +95 -0
  41. neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/compile_flags.json +1 -0
  42. neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/model.done +0 -0
  43. neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/model.hlo_module.pb +3 -0
  44. neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/model.neff +3 -0
  45. neuronxcc-2.21.33363.0+82129205/MODULE_0fed968161ee53fb93ea+fb4cc044/compile_flags.json +1 -0
  46. neuronxcc-2.21.33363.0+82129205/MODULE_0fed968161ee53fb93ea+fb4cc044/model.hlo_module.pb +3 -0
  47. neuronxcc-2.21.33363.0+82129205/MODULE_0fed968161ee53fb93ea+fb4cc044/model.hlo_module.pb.lock +0 -0
  48. neuronxcc-2.21.33363.0+82129205/MODULE_12461911462967760525+e30acd3a/compile_flags.json +1 -0
  49. neuronxcc-2.21.33363.0+82129205/MODULE_12461911462967760525+e30acd3a/model.done +0 -0
  50. neuronxcc-2.21.33363.0+82129205/MODULE_12461911462967760525+e30acd3a/model.hlo_module.pb +3 -0
.gitattributes CHANGED
@@ -16234,3 +16234,39 @@ neuronxcc-2.21.33363.0+82129205/MODULE_bb7697643c8e00d54022+ac10809c/model.neff
16234
  neuronxcc-2.21.33363.0+82129205/MODULE_bb7697643c8e00d54022+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
16235
  neuronxcc-2.21.33363.0+82129205/MODULE_bc3563e123000ea21cfc+f40230a8/model.neff filter=lfs diff=lfs merge=lfs -text
16236
  neuronxcc-2.21.33363.0+82129205/MODULE_bc3563e123000ea21cfc+f40230a8/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16234
  neuronxcc-2.21.33363.0+82129205/MODULE_bb7697643c8e00d54022+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
16235
  neuronxcc-2.21.33363.0+82129205/MODULE_bc3563e123000ea21cfc+f40230a8/model.neff filter=lfs diff=lfs merge=lfs -text
16236
  neuronxcc-2.21.33363.0+82129205/MODULE_bc3563e123000ea21cfc+f40230a8/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
16237
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16238
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/model.neff filter=lfs diff=lfs merge=lfs -text
16239
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/model.neff filter=lfs diff=lfs merge=lfs -text
16240
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
16241
+ neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16242
+ neuronxcc-2.21.33363.0+82129205/MODULE_12461911462967760525+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16243
+ neuronxcc-2.21.33363.0+82129205/MODULE_12a28df006c59e3576b1+a02c3a36/model.neff filter=lfs diff=lfs merge=lfs -text
16244
+ neuronxcc-2.21.33363.0+82129205/MODULE_12a28df006c59e3576b1+a02c3a36/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
16245
+ neuronxcc-2.21.33363.0+82129205/MODULE_14192315105465839210+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16246
+ neuronxcc-2.21.33363.0+82129205/MODULE_1582447fecd752c400bc+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16247
+ neuronxcc-2.21.33363.0+82129205/MODULE_1605570184959188488+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16248
+ neuronxcc-2.21.33363.0+82129205/MODULE_17516957483816756456+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16249
+ neuronxcc-2.21.33363.0+82129205/MODULE_17567634454620665439+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16250
+ neuronxcc-2.21.33363.0+82129205/MODULE_249689642401186199+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16251
+ neuronxcc-2.21.33363.0+82129205/MODULE_2749861b5474cc5ff87c+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16252
+ neuronxcc-2.21.33363.0+82129205/MODULE_28ddfe81fe61b9dbddf0+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16253
+ neuronxcc-2.21.33363.0+82129205/MODULE_3a22f3795f95b4bc7eef+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16254
+ neuronxcc-2.21.33363.0+82129205/MODULE_446d3822c1ba9fbe18ba+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16255
+ neuronxcc-2.21.33363.0+82129205/MODULE_4f6bc8a0a1a483ba7b55+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16256
+ neuronxcc-2.21.33363.0+82129205/MODULE_541e319c815dc74f5f11+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16257
+ neuronxcc-2.21.33363.0+82129205/MODULE_6f7d67366590b6d1215b+a02c3a36/model.neff filter=lfs diff=lfs merge=lfs -text
16258
+ neuronxcc-2.21.33363.0+82129205/MODULE_6f7d67366590b6d1215b+a02c3a36/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
16259
+ neuronxcc-2.21.33363.0+82129205/MODULE_75cb0c08a4502feae194+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16260
+ neuronxcc-2.21.33363.0+82129205/MODULE_7eba564503196965e26c+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16261
+ neuronxcc-2.21.33363.0+82129205/MODULE_813f4ca5979c12f89e82+24129607/model.neff filter=lfs diff=lfs merge=lfs -text
16262
+ neuronxcc-2.21.33363.0+82129205/MODULE_8cdf0668abc5ec3574a0+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16263
+ neuronxcc-2.21.33363.0+82129205/MODULE_93667ce6adf087dd27b0+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16264
+ neuronxcc-2.21.33363.0+82129205/MODULE_9644019637695497943+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16265
+ neuronxcc-2.21.33363.0+82129205/MODULE_9847227195640834860+e30acd3a/model.neff filter=lfs diff=lfs merge=lfs -text
16266
+ neuronxcc-2.21.33363.0+82129205/MODULE_992f9d164a8d61e4a43a+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16267
+ neuronxcc-2.21.33363.0+82129205/MODULE_aace4e3c5cf3d468cf2e+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16268
+ neuronxcc-2.21.33363.0+82129205/MODULE_b3ac1cf6845fcd156f52+fb4cc044/model.neff filter=lfs diff=lfs merge=lfs -text
16269
+ neuronxcc-2.21.33363.0+82129205/MODULE_bd2f2fd229df31e1d6fc+24129607/model.neff filter=lfs diff=lfs merge=lfs -text
16270
+ neuronxcc-2.21.33363.0+82129205/MODULE_d4b74a4b48943996d251+a02c3a36/model.neff filter=lfs diff=lfs merge=lfs -text
16271
+ neuronxcc-2.21.33363.0+82129205/MODULE_d4b74a4b48943996d251+a02c3a36/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
16272
+ neuronxcc-2.21.33363.0+82129205/MODULE_fc8e09ebb5e540fef7a3+24129607/model.neff filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.2.dev1/qwen3/Qwen/Qwen3-0.6B/4ddf27fb5dfd4909a319.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-0.6B",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 1024,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention"
45
+ ],
46
+ "max_position_embeddings": 40960,
47
+ "max_window_layers": 28,
48
+ "model_type": "qwen3",
49
+ "neuron": {
50
+ "_serialized_key": "NxDNeuronConfig",
51
+ "batch_size": 2,
52
+ "capacity_factor": null,
53
+ "checkpoint_id": "Qwen/Qwen3-0.6B",
54
+ "checkpoint_revision": "c1899de289a04d12100db370d81485cdf75e47ca",
55
+ "continuous_batching": true,
56
+ "ep_degree": 1,
57
+ "fused_qkv": true,
58
+ "glu_mlp": true,
59
+ "local_ranks_size": 2,
60
+ "max_batch_size": 2,
61
+ "max_context_length": 2048,
62
+ "max_topk": 256,
63
+ "n_active_tokens": 2048,
64
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
65
+ "on_device_sampling": false,
66
+ "optimum_neuron_version": "0.4.2.dev1",
67
+ "output_logits": false,
68
+ "pp_degree": 1,
69
+ "sequence_length": 2048,
70
+ "speculation_length": 0,
71
+ "start_rank_id": 0,
72
+ "target": "trn1",
73
+ "torch_dtype": "bfloat16",
74
+ "tp_degree": 2
75
+ },
76
+ "num_attention_heads": 16,
77
+ "num_hidden_layers": 28,
78
+ "num_key_value_heads": 8,
79
+ "rms_norm_eps": 1e-06,
80
+ "rope_scaling": null,
81
+ "rope_theta": 1000000,
82
+ "sliding_window": null,
83
+ "tie_word_embeddings": true,
84
+ "use_cache": true,
85
+ "use_sliding_window": false,
86
+ "vocab_size": 151936
87
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.5.dev1/60feecaa0c4c075e2f3e46a3f55d9a273f0ddd75a0ecf64e4ae27352e0819506/ee71309c308c0bf312e7.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-0.6B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 1024,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention"
45
+ ],
46
+ "max_position_embeddings": 32768,
47
+ "max_window_layers": 28,
48
+ "model_type": "qwen3",
49
+ "neuron": {
50
+ "_serialized_key": "NxDNeuronConfig",
51
+ "batch_size": 6,
52
+ "capacity_factor": null,
53
+ "checkpoint_id": "Qwen/Qwen3-Embedding-0.6B",
54
+ "checkpoint_revision": "c54f2e6e80b2d7b7de06f51cec4959f6b3e03418",
55
+ "continuous_batching": false,
56
+ "ep_degree": 1,
57
+ "fused_qkv": true,
58
+ "glu_mlp": true,
59
+ "local_ranks_size": 2,
60
+ "max_batch_size": 6,
61
+ "max_context_length": 8192,
62
+ "max_topk": 256,
63
+ "n_active_tokens": 8192,
64
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
65
+ "on_device_sampling": false,
66
+ "optimum_neuron_version": "0.4.5.dev1",
67
+ "output_logits": false,
68
+ "pp_degree": 1,
69
+ "sequence_length": 8192,
70
+ "speculation_length": 0,
71
+ "start_rank_id": 0,
72
+ "target": "trn1",
73
+ "torch_dtype": "bfloat16",
74
+ "tp_degree": 2
75
+ },
76
+ "num_attention_heads": 16,
77
+ "num_hidden_layers": 28,
78
+ "num_key_value_heads": 8,
79
+ "rms_norm_eps": 1e-06,
80
+ "rope_scaling": null,
81
+ "rope_theta": 1000000,
82
+ "sliding_window": null,
83
+ "tie_word_embeddings": true,
84
+ "use_cache": true,
85
+ "use_sliding_window": false,
86
+ "vocab_size": 151669
87
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.5.dev1/qwen3/Qwen/Qwen3-Embedding-0.6B/ee71309c308c0bf312e7.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-0.6B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 1024,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention"
45
+ ],
46
+ "max_position_embeddings": 32768,
47
+ "max_window_layers": 28,
48
+ "model_type": "qwen3",
49
+ "neuron": {
50
+ "_serialized_key": "NxDNeuronConfig",
51
+ "batch_size": 6,
52
+ "capacity_factor": null,
53
+ "checkpoint_id": "Qwen/Qwen3-Embedding-0.6B",
54
+ "checkpoint_revision": "c54f2e6e80b2d7b7de06f51cec4959f6b3e03418",
55
+ "continuous_batching": false,
56
+ "ep_degree": 1,
57
+ "fused_qkv": true,
58
+ "glu_mlp": true,
59
+ "local_ranks_size": 2,
60
+ "max_batch_size": 6,
61
+ "max_context_length": 8192,
62
+ "max_topk": 256,
63
+ "n_active_tokens": 8192,
64
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
65
+ "on_device_sampling": false,
66
+ "optimum_neuron_version": "0.4.5.dev1",
67
+ "output_logits": false,
68
+ "pp_degree": 1,
69
+ "sequence_length": 8192,
70
+ "speculation_length": 0,
71
+ "start_rank_id": 0,
72
+ "target": "trn1",
73
+ "torch_dtype": "bfloat16",
74
+ "tp_degree": 2
75
+ },
76
+ "num_attention_heads": 16,
77
+ "num_hidden_layers": 28,
78
+ "num_key_value_heads": 8,
79
+ "rms_norm_eps": 1e-06,
80
+ "rope_scaling": null,
81
+ "rope_theta": 1000000,
82
+ "sliding_window": null,
83
+ "tie_word_embeddings": true,
84
+ "use_cache": true,
85
+ "use_sliding_window": false,
86
+ "vocab_size": 151669
87
+ }
neuronxcc-2.21.18209.0+043b1bf7/MODULE_9d35248a93d4142a3cf1+24129607/model.hlo_module.pb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:770f59a15ab233bea70cb16360c196b23b356f67051422593116053095e3db38
3
  size 739558
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c0311f0ae811c00f59cb813f4ee06008d0d18e5bdf0534ecf04e14c6eadcad8
3
  size 739558
neuronxcc-2.21.18209.0+043b1bf7/MODULE_9d35248a93d4142a3cf1+24129607/model.neff CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:58017c0e14cfb2aa2617054665041a111f19a8658e3856ce7f80963501abfa68
3
  size 26133504
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e80290bb0de9fba2d791707189cfe81ec5896082ad2fc95838ce7c687a2f6ed4
3
  size 26133504
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccd8dd07c208eb87030d40b6fe53666dbd98b0ad6fa3cf5fa46bd1e5d023563f
3
+ size 662166
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a9ea3bcc615b10517bb2+fb4cc044/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cd272e40e65fb42ee0c8247f443e277f005acb0ff4230bdc0d2f6c5dfdbe901
3
+ size 27689984
neuronxcc-2.21.18209.0+043b1bf7/MODULE_c2b3afa51c57d431a332+a02c3a36/model.neff CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4fb43f4ea0d29f851cd0c0d3575ab21545fe6863d03f53111b381301e3eee665
3
  size 1926144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2835fba37931b9c58c33d25543bd60d85d61d996e61ac3b6e78e5b4c95bfb85
3
  size 1926144
neuronxcc-2.21.18209.0+043b1bf7/MODULE_c2b3afa51c57d431a332+a02c3a36/wrapped_neff.hlo CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2cfc00602dde4b251f7ee5e382c599874555cf1857694aab39c3a65c89e133d4
3
  size 2082478
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6af3d77381b3fde23da8966bdbf8f343010c5869ecb7fef585166f1178547f33
3
  size 2082478
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf35cc5188092a5ed97a2fd36b936a577ef7dc5f8e103e38714748065ecdd328
3
+ size 693532
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d1e28d83b0b75daadd33+24129607/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0bbf7638bc657c8f8d72fdd57dcfbbc3d7ef74e645f02f878fc8178e8b3971f
3
+ size 11265024
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/token_generation/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dd18126105abb5dddc19543d22a99b11b3ae776516a3845ea8956ffaadb15a0
3
+ size 702870
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:503663d1e9e83515a143dbe1a2fa6d4151c04250e2d0a53da66ea3d1a2e75847
3
+ size 2059264
neuronxcc-2.21.18209.0+043b1bf7/MODULE_d4fe40d7566b309ac86a+a02c3a36/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:848c9cf439fa7e607aff664a598fefc3dee96cfabe42dba3cdfbcb7241202510
3
+ size 2196458
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1/43f30d107729db6b760b2a90833f43ca44f68e6468ef64e131f71e2fa9f5b23f/b76d86cc7c31e0e29c99.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "openai/gpt-oss-20b",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "GptOssForCausalLM"
7
+ ],
8
+ "attention_bias": true,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "experts_per_token": 4,
12
+ "head_dim": 64,
13
+ "hidden_act": "silu",
14
+ "hidden_size": 2880,
15
+ "initial_context_length": 4096,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 2880,
18
+ "layer_types": [
19
+ "sliding_attention",
20
+ "full_attention",
21
+ "sliding_attention",
22
+ "full_attention",
23
+ "sliding_attention",
24
+ "full_attention",
25
+ "sliding_attention",
26
+ "full_attention",
27
+ "sliding_attention",
28
+ "full_attention",
29
+ "sliding_attention",
30
+ "full_attention",
31
+ "sliding_attention",
32
+ "full_attention",
33
+ "sliding_attention",
34
+ "full_attention",
35
+ "sliding_attention",
36
+ "full_attention",
37
+ "sliding_attention",
38
+ "full_attention",
39
+ "sliding_attention",
40
+ "full_attention",
41
+ "sliding_attention",
42
+ "full_attention"
43
+ ],
44
+ "max_position_embeddings": 131072,
45
+ "model_type": "gpt_oss",
46
+ "neuron": {
47
+ "_serialized_key": "NxDNeuronConfig",
48
+ "batch_size": 1,
49
+ "capacity_factor": null,
50
+ "checkpoint_id": "openai/gpt-oss-20b",
51
+ "checkpoint_revision": "6cee5e81ee83917806bbde320786a8fb61efebee",
52
+ "continuous_batching": false,
53
+ "ep_degree": 1,
54
+ "fused_qkv": true,
55
+ "glu_mlp": true,
56
+ "local_ranks_size": 8,
57
+ "max_batch_size": 1,
58
+ "max_context_length": 4096,
59
+ "max_topk": 256,
60
+ "n_active_tokens": 4096,
61
+ "neuronxcc_version": "2.21.33363.0+82129205",
62
+ "on_device_sampling": true,
63
+ "optimum_neuron_version": "0.4.5.dev1",
64
+ "output_logits": false,
65
+ "pp_degree": 1,
66
+ "sequence_length": 4096,
67
+ "speculation_length": 0,
68
+ "start_rank_id": 0,
69
+ "target": "trn1",
70
+ "torch_dtype": "bfloat16",
71
+ "tp_degree": 8
72
+ },
73
+ "num_attention_heads": 64,
74
+ "num_experts_per_tok": 4,
75
+ "num_hidden_layers": 24,
76
+ "num_key_value_heads": 8,
77
+ "num_local_experts": 32,
78
+ "output_router_logits": false,
79
+ "quantization_config": {
80
+ "modules_to_not_convert": [
81
+ "model.layers.*.self_attn",
82
+ "model.layers.*.mlp.router",
83
+ "model.embed_tokens",
84
+ "lm_head"
85
+ ],
86
+ "quant_method": "mxfp4"
87
+ },
88
+ "rms_norm_eps": 1e-05,
89
+ "rope_scaling": {
90
+ "beta_fast": 32.0,
91
+ "beta_slow": 1.0,
92
+ "factor": 32.0,
93
+ "original_max_position_embeddings": 4096,
94
+ "rope_type": "yarn",
95
+ "truncate": false
96
+ },
97
+ "rope_theta": 150000,
98
+ "router_aux_loss_coef": 0.9,
99
+ "sliding_window": 128,
100
+ "swiglu_limit": 7.0,
101
+ "tie_word_embeddings": false,
102
+ "use_cache": true,
103
+ "vocab_size": 201088
104
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1/43f30d107729db6b760b2a90833f43ca44f68e6468ef64e131f71e2fa9f5b23f/f71907dd62c8b40e471f.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "openai/gpt-oss-20b",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "GptOssForCausalLM"
7
+ ],
8
+ "attention_bias": true,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "experts_per_token": 4,
12
+ "head_dim": 64,
13
+ "hidden_act": "silu",
14
+ "hidden_size": 2880,
15
+ "initial_context_length": 4096,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 2880,
18
+ "layer_types": [
19
+ "sliding_attention",
20
+ "full_attention",
21
+ "sliding_attention",
22
+ "full_attention",
23
+ "sliding_attention",
24
+ "full_attention",
25
+ "sliding_attention",
26
+ "full_attention",
27
+ "sliding_attention",
28
+ "full_attention",
29
+ "sliding_attention",
30
+ "full_attention",
31
+ "sliding_attention",
32
+ "full_attention",
33
+ "sliding_attention",
34
+ "full_attention",
35
+ "sliding_attention",
36
+ "full_attention",
37
+ "sliding_attention",
38
+ "full_attention",
39
+ "sliding_attention",
40
+ "full_attention",
41
+ "sliding_attention",
42
+ "full_attention"
43
+ ],
44
+ "max_position_embeddings": 131072,
45
+ "model_type": "gpt_oss",
46
+ "neuron": {
47
+ "_serialized_key": "NxDNeuronConfig",
48
+ "batch_size": 1,
49
+ "capacity_factor": null,
50
+ "checkpoint_id": "openai/gpt-oss-20b",
51
+ "checkpoint_revision": "6cee5e81ee83917806bbde320786a8fb61efebee",
52
+ "continuous_batching": false,
53
+ "ep_degree": 1,
54
+ "fused_qkv": false,
55
+ "glu_mlp": true,
56
+ "local_ranks_size": 8,
57
+ "max_batch_size": 1,
58
+ "max_context_length": 4096,
59
+ "max_topk": 256,
60
+ "n_active_tokens": 4096,
61
+ "neuronxcc_version": "2.21.33363.0+82129205",
62
+ "on_device_sampling": false,
63
+ "optimum_neuron_version": "0.4.5.dev1",
64
+ "output_logits": false,
65
+ "pp_degree": 1,
66
+ "sequence_length": 4096,
67
+ "speculation_length": 0,
68
+ "start_rank_id": 0,
69
+ "target": "trn1",
70
+ "torch_dtype": "bfloat16",
71
+ "tp_degree": 8
72
+ },
73
+ "num_attention_heads": 64,
74
+ "num_experts_per_tok": 4,
75
+ "num_hidden_layers": 24,
76
+ "num_key_value_heads": 8,
77
+ "num_local_experts": 32,
78
+ "output_router_logits": false,
79
+ "quantization_config": {
80
+ "modules_to_not_convert": [
81
+ "model.layers.*.self_attn",
82
+ "model.layers.*.mlp.router",
83
+ "model.embed_tokens",
84
+ "lm_head"
85
+ ],
86
+ "quant_method": "mxfp4"
87
+ },
88
+ "rms_norm_eps": 1e-05,
89
+ "rope_scaling": {
90
+ "beta_fast": 32.0,
91
+ "beta_slow": 1.0,
92
+ "factor": 32.0,
93
+ "original_max_position_embeddings": 4096,
94
+ "rope_type": "yarn",
95
+ "truncate": false
96
+ },
97
+ "rope_theta": 150000,
98
+ "router_aux_loss_coef": 0.9,
99
+ "sliding_window": 128,
100
+ "swiglu_limit": 7.0,
101
+ "tie_word_embeddings": false,
102
+ "use_cache": true,
103
+ "vocab_size": 201088
104
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1/gpt_oss/openai/gpt-oss-20b/b76d86cc7c31e0e29c99.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "openai/gpt-oss-20b",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "GptOssForCausalLM"
7
+ ],
8
+ "attention_bias": true,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "experts_per_token": 4,
12
+ "head_dim": 64,
13
+ "hidden_act": "silu",
14
+ "hidden_size": 2880,
15
+ "initial_context_length": 4096,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 2880,
18
+ "layer_types": [
19
+ "sliding_attention",
20
+ "full_attention",
21
+ "sliding_attention",
22
+ "full_attention",
23
+ "sliding_attention",
24
+ "full_attention",
25
+ "sliding_attention",
26
+ "full_attention",
27
+ "sliding_attention",
28
+ "full_attention",
29
+ "sliding_attention",
30
+ "full_attention",
31
+ "sliding_attention",
32
+ "full_attention",
33
+ "sliding_attention",
34
+ "full_attention",
35
+ "sliding_attention",
36
+ "full_attention",
37
+ "sliding_attention",
38
+ "full_attention",
39
+ "sliding_attention",
40
+ "full_attention",
41
+ "sliding_attention",
42
+ "full_attention"
43
+ ],
44
+ "max_position_embeddings": 131072,
45
+ "model_type": "gpt_oss",
46
+ "neuron": {
47
+ "_serialized_key": "NxDNeuronConfig",
48
+ "batch_size": 1,
49
+ "capacity_factor": null,
50
+ "checkpoint_id": "openai/gpt-oss-20b",
51
+ "checkpoint_revision": "6cee5e81ee83917806bbde320786a8fb61efebee",
52
+ "continuous_batching": false,
53
+ "ep_degree": 1,
54
+ "fused_qkv": true,
55
+ "glu_mlp": true,
56
+ "local_ranks_size": 8,
57
+ "max_batch_size": 1,
58
+ "max_context_length": 4096,
59
+ "max_topk": 256,
60
+ "n_active_tokens": 4096,
61
+ "neuronxcc_version": "2.21.33363.0+82129205",
62
+ "on_device_sampling": true,
63
+ "optimum_neuron_version": "0.4.5.dev1",
64
+ "output_logits": false,
65
+ "pp_degree": 1,
66
+ "sequence_length": 4096,
67
+ "speculation_length": 0,
68
+ "start_rank_id": 0,
69
+ "target": "trn1",
70
+ "torch_dtype": "bfloat16",
71
+ "tp_degree": 8
72
+ },
73
+ "num_attention_heads": 64,
74
+ "num_experts_per_tok": 4,
75
+ "num_hidden_layers": 24,
76
+ "num_key_value_heads": 8,
77
+ "num_local_experts": 32,
78
+ "output_router_logits": false,
79
+ "quantization_config": {
80
+ "modules_to_not_convert": [
81
+ "model.layers.*.self_attn",
82
+ "model.layers.*.mlp.router",
83
+ "model.embed_tokens",
84
+ "lm_head"
85
+ ],
86
+ "quant_method": "mxfp4"
87
+ },
88
+ "rms_norm_eps": 1e-05,
89
+ "rope_scaling": {
90
+ "beta_fast": 32.0,
91
+ "beta_slow": 1.0,
92
+ "factor": 32.0,
93
+ "original_max_position_embeddings": 4096,
94
+ "rope_type": "yarn",
95
+ "truncate": false
96
+ },
97
+ "rope_theta": 150000,
98
+ "router_aux_loss_coef": 0.9,
99
+ "sliding_window": 128,
100
+ "swiglu_limit": 7.0,
101
+ "tie_word_embeddings": false,
102
+ "use_cache": true,
103
+ "vocab_size": 201088
104
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/026fe44014d3f650a32e.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 16,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 16,
69
+ "max_context_length": 1024,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 1024,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 1024,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/22e7cd94e28b353a796b.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 8,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 8,
69
+ "max_context_length": 8192,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 8192,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 8192,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/535e7795fb656220e11a.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 1,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 1,
69
+ "max_context_length": 16384,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 16384,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 16384,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/644bf9550db2b89192aa.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 8,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 8,
69
+ "max_context_length": 1024,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 1024,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 1024,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/6501938b047f7d373cdd.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 1,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 1,
69
+ "max_context_length": 8192,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 8192,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 8192,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/7bfaacc5ae3961c38f6c.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 1,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 1,
69
+ "max_context_length": 2048,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 2048,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 2048,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/912ca3353189929b7b15.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 1,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 1,
69
+ "max_context_length": 1024,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 1024,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 1024,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/b2f13b35b2326e133272.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 8,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 8,
69
+ "max_context_length": 2048,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 2048,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 2048,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/b4a6b1d49ffe1fb09037.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 4,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 4,
69
+ "max_context_length": 2048,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 2048,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 2048,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/b52dd6b442b63b75e7b7.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 4,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 4,
69
+ "max_context_length": 1024,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 1024,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 1024,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/bba41446ac8406c873ee.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 1,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 1,
69
+ "max_context_length": 32768,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 32768,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 32768,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/c67e5e7f18f7dbf76e0c.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 4,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 4,
69
+ "max_context_length": 16384,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 16384,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 16384,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/d27d49077cbb0bf50eb9.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 1,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 1,
69
+ "max_context_length": 4096,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 4096,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 4096,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/d59c183509c2a322fd8b.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 8,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 8,
69
+ "max_context_length": 4096,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 4096,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 4096,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/f4ff220867f67d4c7b95.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 4,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 4,
69
+ "max_context_length": 8192,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 8192,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 8192,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev2/af58eb15d8e02338dc2f2e880e9c6ec803a98278914b3606acdcc252e7e18429/f56d8d1a0bac69205ae6.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Embedding-8B",
4
+ "_task": "feature-extraction",
5
+ "architectures": [
6
+ "Qwen3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "dtype": "bfloat16",
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 12288,
16
+ "layer_types": [
17
+ "full_attention",
18
+ "full_attention",
19
+ "full_attention",
20
+ "full_attention",
21
+ "full_attention",
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention"
53
+ ],
54
+ "max_position_embeddings": 40960,
55
+ "max_window_layers": 36,
56
+ "model_type": "qwen3",
57
+ "neuron": {
58
+ "_serialized_key": "NxDNeuronConfig",
59
+ "batch_size": 4,
60
+ "capacity_factor": null,
61
+ "checkpoint_id": "Qwen/Qwen3-Embedding-8B",
62
+ "checkpoint_revision": "1d8ad4ca9b3dd8059ad90a75d4983776a23d44af",
63
+ "continuous_batching": false,
64
+ "ep_degree": 1,
65
+ "fused_qkv": true,
66
+ "glu_mlp": true,
67
+ "local_ranks_size": 2,
68
+ "max_batch_size": 4,
69
+ "max_context_length": 4096,
70
+ "max_topk": 256,
71
+ "n_active_tokens": 4096,
72
+ "neuronxcc_version": "2.21.33363.0+82129205",
73
+ "on_device_sampling": false,
74
+ "optimum_neuron_version": "0.4.5.dev2",
75
+ "output_logits": false,
76
+ "pp_degree": 1,
77
+ "sequence_length": 4096,
78
+ "speculation_length": 0,
79
+ "start_rank_id": 0,
80
+ "target": "trn1",
81
+ "torch_dtype": "bfloat16",
82
+ "tp_degree": 2
83
+ },
84
+ "num_attention_heads": 32,
85
+ "num_hidden_layers": 36,
86
+ "num_key_value_heads": 8,
87
+ "rms_norm_eps": 1e-06,
88
+ "rope_scaling": null,
89
+ "rope_theta": 1000000,
90
+ "sliding_window": null,
91
+ "tie_word_embeddings": false,
92
+ "use_cache": true,
93
+ "use_sliding_window": false,
94
+ "vocab_size": 151665
95
+ }
neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/model.done ADDED
File without changes
neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff76ba7d15f95d31c450a8344c01924810ab53f588d7ea80b0bb22ed28f06609
3
+ size 628810
neuronxcc-2.21.33363.0+82129205/MODULE_00fa01e7f21f195c4444+fb4cc044/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2499d61a87073ae61b4e0906914f9d27a68f55aab1b28e6df1428c6cd6800449
3
+ size 2591744
neuronxcc-2.21.33363.0+82129205/MODULE_0fed968161ee53fb93ea+fb4cc044/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.33363.0+82129205/MODULE_0fed968161ee53fb93ea+fb4cc044/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90610188fc62b33403b5d73edcf29cbc5b6d5cbec9a20d45bab4b5bbb232c7bb
3
+ size 628841
neuronxcc-2.21.33363.0+82129205/MODULE_0fed968161ee53fb93ea+fb4cc044/model.hlo_module.pb.lock ADDED
File without changes
neuronxcc-2.21.33363.0+82129205/MODULE_12461911462967760525+e30acd3a/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1"]
neuronxcc-2.21.33363.0+82129205/MODULE_12461911462967760525+e30acd3a/model.done ADDED
File without changes
neuronxcc-2.21.33363.0+82129205/MODULE_12461911462967760525+e30acd3a/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef78b6be207458b4d706f2b8a8e4ceaf2c439a683b520e876a672295b9e1e70
3
+ size 6660