optimum-internal-testing-user commited on
Commit
eb86d8a
·
verified ·
1 Parent(s): 5474bc7

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +15 -0
  2. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/granite/hf-internal-testing/tiny-random-GraniteForCausalLM/79d6a2644c0fb994afba.json +58 -0
  3. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/llama/llamafactory/tiny-random-Llama-3/19972825cc5b89209b02.json +62 -0
  4. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/llama4_text/tiny-random/llama-4/4f6687486daab74bc033.json +81 -0
  5. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/mixtral/dacorvo/Mixtral-tiny/cb323601e0aa10c70f9c.json +58 -0
  6. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/phi3/yujiepan/phi-4-tiny-random/5ff5bff574e21c1ee9b0.json +59 -0
  7. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/qwen2/yujiepan/qwen2.5-128k-tiny-random/eac4336716688c682e8c.json +64 -0
  8. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/qwen3_moe/optimum-internal-testing/tiny-random-qwen3_moe/a6f02f34734d58205a85.json +65 -0
  9. neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/compile_flags.json +1 -0
  10. neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/model.done +0 -0
  11. neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/model.hlo_module.pb +3 -0
  12. neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/model.neff +3 -0
  13. neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/wrapped_neff.hlo +3 -0
  14. neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/compile_flags.json +1 -0
  15. neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/model.done +0 -0
  16. neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/model.hlo_module.pb +3 -0
  17. neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/model.neff +3 -0
  18. neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/wrapped_neff.hlo +3 -0
  19. neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/compile_flags.json +1 -0
  20. neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/model.done +0 -0
  21. neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/model.hlo_module.pb +3 -0
  22. neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/model.neff +3 -0
  23. neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/compile_flags.json +1 -0
  24. neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/model.done +0 -0
  25. neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/model.hlo_module.pb +3 -0
  26. neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/model.neff +3 -0
  27. neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/compile_flags.json +1 -0
  28. neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/model.done +0 -0
  29. neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/model.hlo_module.pb +3 -0
  30. neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/model.neff +3 -0
  31. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/compile_flags.json +1 -0
  32. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/model.done +0 -0
  33. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/model.hlo_module.pb +3 -0
  34. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/model.neff +3 -0
  35. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/wrapped_neff.hlo +3 -0
  36. neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/compile_flags.json +1 -0
  37. neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/model.done +0 -0
  38. neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/model.hlo_module.pb +3 -0
  39. neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/model.neff +3 -0
  40. neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/wrapped_neff.hlo +3 -0
  41. neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/compile_flags.json +1 -0
  42. neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/model.done +0 -0
  43. neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/model.hlo_module.pb +3 -0
  44. neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/model.neff +3 -0
  45. neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/wrapped_neff.hlo +3 -0
  46. neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/compile_flags.json +1 -0
  47. neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/model.done +0 -0
  48. neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/model.hlo_module.pb +3 -0
  49. neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/model.neff +3 -0
  50. neuronxcc-2.21.18209.0+043b1bf7/MODULE_fef991cead1748c4d101+ed72d204/compile_flags.json +1 -0
.gitattributes CHANGED
@@ -5111,3 +5111,18 @@ neuronxcc-2.21.18209.0+043b1bf7/MODULE_cbf256b61a3a1c5f1f7a+677eeb9d/model.neff
5111
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f753c5295db2f87f81a1+a9d440f5/model.neff filter=lfs diff=lfs merge=lfs -text
5112
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f753c5295db2f87f81a1+a9d440f5/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
5113
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_b29b6228ab434f989a80+ed72d204/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5111
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f753c5295db2f87f81a1+a9d440f5/model.neff filter=lfs diff=lfs merge=lfs -text
5112
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f753c5295db2f87f81a1+a9d440f5/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
5113
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_b29b6228ab434f989a80+ed72d204/model.neff filter=lfs diff=lfs merge=lfs -text
5114
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/model.neff filter=lfs diff=lfs merge=lfs -text
5115
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
5116
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/model.neff filter=lfs diff=lfs merge=lfs -text
5117
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
5118
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/model.neff filter=lfs diff=lfs merge=lfs -text
5119
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/model.neff filter=lfs diff=lfs merge=lfs -text
5120
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/model.neff filter=lfs diff=lfs merge=lfs -text
5121
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/model.neff filter=lfs diff=lfs merge=lfs -text
5122
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
5123
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/model.neff filter=lfs diff=lfs merge=lfs -text
5124
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
5125
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/model.neff filter=lfs diff=lfs merge=lfs -text
5126
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
5127
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/model.neff filter=lfs diff=lfs merge=lfs -text
5128
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_fef991cead1748c4d101+ed72d204/model.neff filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/granite/hf-internal-testing/tiny-random-GraniteForCausalLM/79d6a2644c0fb994afba.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "hf-internal-testing/tiny-random-GraniteForCausalLM",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "GraniteForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "attention_multiplier": 1.0,
11
+ "embedding_multiplier": 1.0,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 32,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 64,
16
+ "logits_scaling": 1.0,
17
+ "max_position_embeddings": 2048,
18
+ "mlp_bias": false,
19
+ "model_type": "granite",
20
+ "neuron": {
21
+ "_serialized_key": "NxDNeuronConfig",
22
+ "batch_size": 1,
23
+ "capacity_factor": null,
24
+ "checkpoint_id": "hf-internal-testing/tiny-random-GraniteForCausalLM",
25
+ "checkpoint_revision": "c3074ebc0ac2fe545305f5e5f6cce2cc9b2aa0c5",
26
+ "continuous_batching": false,
27
+ "enable_bucketing": false,
28
+ "ep_degree": 1,
29
+ "fused_qkv": true,
30
+ "glu_mlp": true,
31
+ "local_ranks_size": 2,
32
+ "max_batch_size": 1,
33
+ "max_context_length": 1024,
34
+ "max_topk": 256,
35
+ "n_active_tokens": 1024,
36
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
37
+ "on_device_sampling": true,
38
+ "optimum_neuron_version": "0.4.0.dev0",
39
+ "output_logits": false,
40
+ "pp_degree": 1,
41
+ "sequence_length": 1024,
42
+ "speculation_length": 0,
43
+ "start_rank_id": 0,
44
+ "target": "trn1",
45
+ "torch_dtype": "bfloat16",
46
+ "tp_degree": 2
47
+ },
48
+ "num_attention_heads": 4,
49
+ "num_hidden_layers": 2,
50
+ "num_key_value_heads": 4,
51
+ "residual_multiplier": 1.0,
52
+ "rms_norm_eps": 1e-06,
53
+ "rope_scaling": null,
54
+ "rope_theta": 10000.0,
55
+ "tie_word_embeddings": false,
56
+ "use_cache": true,
57
+ "vocab_size": 49152
58
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/llama/llamafactory/tiny-random-Llama-3/19972825cc5b89209b02.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "llamafactory/tiny-random-Llama-3",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "LlamaForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "head_dim": 4,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 16,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 64,
15
+ "max_position_embeddings": 131072,
16
+ "mlp_bias": false,
17
+ "model_type": "llama",
18
+ "neuron": {
19
+ "_serialized_key": "NxDNeuronConfig",
20
+ "batch_size": 1,
21
+ "capacity_factor": null,
22
+ "checkpoint_id": "llamafactory/tiny-random-Llama-3",
23
+ "checkpoint_revision": "bf2a2e3bf199ad2ee96f02a3c00246c608db22a8",
24
+ "continuous_batching": false,
25
+ "enable_bucketing": false,
26
+ "ep_degree": 1,
27
+ "fused_qkv": true,
28
+ "glu_mlp": true,
29
+ "local_ranks_size": 2,
30
+ "max_batch_size": 1,
31
+ "max_context_length": 1024,
32
+ "max_topk": 256,
33
+ "n_active_tokens": 1024,
34
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
35
+ "on_device_sampling": true,
36
+ "optimum_neuron_version": "0.4.0.dev0",
37
+ "output_logits": false,
38
+ "pp_degree": 1,
39
+ "sequence_length": 1024,
40
+ "speculation_length": 0,
41
+ "start_rank_id": 0,
42
+ "target": "trn1",
43
+ "torch_dtype": "bfloat16",
44
+ "tp_degree": 2
45
+ },
46
+ "num_attention_heads": 4,
47
+ "num_hidden_layers": 2,
48
+ "num_key_value_heads": 4,
49
+ "pretraining_tp": 1,
50
+ "rms_norm_eps": 1e-05,
51
+ "rope_scaling": {
52
+ "factor": 8.0,
53
+ "high_freq_factor": 4.0,
54
+ "low_freq_factor": 1.0,
55
+ "original_max_position_embeddings": 8192,
56
+ "rope_type": "llama3"
57
+ },
58
+ "rope_theta": 500000.0,
59
+ "tie_word_embeddings": false,
60
+ "use_cache": true,
61
+ "vocab_size": 128256
62
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/llama4_text/tiny-random/llama-4/4f6687486daab74bc033.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "tiny-random/llama-4",
4
+ "_task": "text-generation",
5
+ "attention_bias": false,
6
+ "attention_chunk_size": 128,
7
+ "attention_dropout": 0.0,
8
+ "attn_scale": 0.1,
9
+ "attn_temperature_tuning": 4,
10
+ "cache_implementation": "hybrid",
11
+ "floor_scale": 8192,
12
+ "for_llm_compressor": false,
13
+ "head_dim": 32,
14
+ "hidden_act": "silu",
15
+ "hidden_size": 32,
16
+ "initializer_range": 0.02,
17
+ "interleave_moe_layer_step": 2,
18
+ "intermediate_size": 64,
19
+ "intermediate_size_mlp": 128,
20
+ "layer_types": [
21
+ "chunked_attention",
22
+ "chunked_attention",
23
+ "chunked_attention",
24
+ "full_attention"
25
+ ],
26
+ "max_position_embeddings": 1048576,
27
+ "model_type": "llama4_text",
28
+ "moe_layers": [
29
+ 1,
30
+ 3
31
+ ],
32
+ "neuron": {
33
+ "_serialized_key": "NxDNeuronConfig",
34
+ "batch_size": 1,
35
+ "capacity_factor": null,
36
+ "checkpoint_id": "tiny-random/llama-4",
37
+ "checkpoint_revision": "9e716f5d4d1ffe0a44a15f46f4a12b840439aba4",
38
+ "continuous_batching": false,
39
+ "enable_bucketing": false,
40
+ "ep_degree": 1,
41
+ "fused_qkv": false,
42
+ "glu_mlp": true,
43
+ "local_ranks_size": 2,
44
+ "max_batch_size": 1,
45
+ "max_context_length": 1024,
46
+ "max_topk": 256,
47
+ "n_active_tokens": 1024,
48
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
49
+ "on_device_sampling": true,
50
+ "optimum_neuron_version": "0.4.0.dev0",
51
+ "output_logits": false,
52
+ "pp_degree": 1,
53
+ "sequence_length": 1024,
54
+ "speculation_length": 0,
55
+ "start_rank_id": 0,
56
+ "target": "trn1",
57
+ "torch_dtype": "bfloat16",
58
+ "tp_degree": 2
59
+ },
60
+ "no_rope_layers": [
61
+ 1,
62
+ 1,
63
+ 1,
64
+ 0
65
+ ],
66
+ "num_attention_heads": 1,
67
+ "num_experts_per_tok": 1,
68
+ "num_hidden_layers": 4,
69
+ "num_key_value_heads": 1,
70
+ "num_local_experts": 8,
71
+ "output_router_logits": false,
72
+ "rms_norm_eps": 1e-05,
73
+ "rope_scaling": null,
74
+ "rope_theta": 500000.0,
75
+ "router_aux_loss_coef": 0.001,
76
+ "router_jitter_noise": 0.0,
77
+ "tie_word_embeddings": true,
78
+ "use_cache": true,
79
+ "use_qk_norm": true,
80
+ "vocab_size": 202048
81
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/mixtral/dacorvo/Mixtral-tiny/cb323601e0aa10c70f9c.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "dacorvo/Mixtral-tiny",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "MixtralForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "head_dim": 32,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1024,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3584,
14
+ "max_position_embeddings": 1024,
15
+ "model_type": "mixtral",
16
+ "neuron": {
17
+ "_serialized_key": "NxDNeuronConfig",
18
+ "batch_size": 1,
19
+ "capacity_factor": null,
20
+ "checkpoint_id": "dacorvo/Mixtral-tiny",
21
+ "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6",
22
+ "continuous_batching": false,
23
+ "enable_bucketing": false,
24
+ "ep_degree": 1,
25
+ "fused_qkv": false,
26
+ "glu_mlp": true,
27
+ "local_ranks_size": 2,
28
+ "max_batch_size": 1,
29
+ "max_context_length": 1024,
30
+ "max_topk": 256,
31
+ "n_active_tokens": 1024,
32
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
33
+ "on_device_sampling": false,
34
+ "optimum_neuron_version": "0.4.0.dev0",
35
+ "output_logits": false,
36
+ "pp_degree": 1,
37
+ "sequence_length": 1024,
38
+ "speculation_length": 0,
39
+ "start_rank_id": 0,
40
+ "target": "trn1",
41
+ "torch_dtype": "bfloat16",
42
+ "tp_degree": 2
43
+ },
44
+ "num_attention_heads": 32,
45
+ "num_experts_per_tok": 2,
46
+ "num_hidden_layers": 2,
47
+ "num_key_value_heads": 8,
48
+ "num_local_experts": 8,
49
+ "output_router_logits": false,
50
+ "rms_norm_eps": 1e-05,
51
+ "rope_theta": 10000.0,
52
+ "router_aux_loss_coef": 0.001,
53
+ "router_jitter_noise": 0.0,
54
+ "sliding_window": 4096,
55
+ "tie_word_embeddings": false,
56
+ "use_cache": true,
57
+ "vocab_size": 32000
58
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/phi3/yujiepan/phi-4-tiny-random/5ff5bff574e21c1ee9b0.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "yujiepan/phi-4-tiny-random",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "Phi3ForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "auto_map": {},
11
+ "embd_pdrop": 0.0,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 16,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 32,
16
+ "max_position_embeddings": 16384,
17
+ "model_type": "phi3",
18
+ "neuron": {
19
+ "_serialized_key": "NxDNeuronConfig",
20
+ "batch_size": 1,
21
+ "capacity_factor": null,
22
+ "checkpoint_id": "yujiepan/phi-4-tiny-random",
23
+ "checkpoint_revision": "18a9a1168dc97ac6d128f811925670c275610f5a",
24
+ "continuous_batching": false,
25
+ "enable_bucketing": false,
26
+ "ep_degree": 1,
27
+ "fused_qkv": true,
28
+ "glu_mlp": true,
29
+ "local_ranks_size": 2,
30
+ "max_batch_size": 1,
31
+ "max_context_length": 1024,
32
+ "max_topk": 256,
33
+ "n_active_tokens": 1024,
34
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
35
+ "on_device_sampling": true,
36
+ "optimum_neuron_version": "0.4.0.dev0",
37
+ "output_logits": false,
38
+ "pp_degree": 1,
39
+ "sequence_length": 1024,
40
+ "speculation_length": 0,
41
+ "start_rank_id": 0,
42
+ "target": "trn1",
43
+ "torch_dtype": "bfloat16",
44
+ "tp_degree": 2
45
+ },
46
+ "num_attention_heads": 2,
47
+ "num_hidden_layers": 2,
48
+ "num_key_value_heads": 1,
49
+ "original_max_position_embeddings": 16384,
50
+ "partial_rotary_factor": 1.0,
51
+ "resid_pdrop": 0.0,
52
+ "rms_norm_eps": 1e-05,
53
+ "rope_scaling": null,
54
+ "rope_theta": 250000,
55
+ "sliding_window": null,
56
+ "tie_word_embeddings": false,
57
+ "use_cache": true,
58
+ "vocab_size": 100352
59
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/qwen2/yujiepan/qwen2.5-128k-tiny-random/eac4336716688c682e8c.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "yujiepan/qwen2.5-128k-tiny-random",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "Qwen2ForCausalLM"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 8,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 16,
13
+ "layer_types": [
14
+ "full_attention",
15
+ "full_attention"
16
+ ],
17
+ "max_position_embeddings": 32768,
18
+ "max_window_layers": 1,
19
+ "model_type": "qwen2",
20
+ "neuron": {
21
+ "_serialized_key": "NxDNeuronConfig",
22
+ "batch_size": 1,
23
+ "capacity_factor": null,
24
+ "checkpoint_id": "yujiepan/qwen2.5-128k-tiny-random",
25
+ "checkpoint_revision": "c8296d4ca3f87782876d2382fbb6481d1beb8ef0",
26
+ "continuous_batching": false,
27
+ "enable_bucketing": false,
28
+ "ep_degree": 1,
29
+ "fused_qkv": false,
30
+ "glu_mlp": true,
31
+ "local_ranks_size": 2,
32
+ "max_batch_size": 1,
33
+ "max_context_length": 1024,
34
+ "max_topk": 256,
35
+ "n_active_tokens": 1024,
36
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
37
+ "on_device_sampling": true,
38
+ "optimum_neuron_version": "0.4.0.dev0",
39
+ "output_logits": false,
40
+ "pp_degree": 1,
41
+ "sequence_length": 1024,
42
+ "speculation_length": 0,
43
+ "start_rank_id": 0,
44
+ "target": "trn1",
45
+ "torch_dtype": "bfloat16",
46
+ "tp_degree": 2
47
+ },
48
+ "num_attention_heads": 4,
49
+ "num_hidden_layers": 2,
50
+ "num_key_value_heads": 2,
51
+ "rms_norm_eps": 1e-06,
52
+ "rope_scaling": {
53
+ "factor": 4.0,
54
+ "original_max_position_embeddings": 32768,
55
+ "rope_type": "yarn",
56
+ "type": "yarn"
57
+ },
58
+ "rope_theta": 1000000.0,
59
+ "sliding_window": null,
60
+ "tie_word_embeddings": false,
61
+ "use_cache": true,
62
+ "use_sliding_window": false,
63
+ "vocab_size": 152064
64
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.0.dev0/qwen3_moe/optimum-internal-testing/tiny-random-qwen3_moe/a6f02f34734d58205a85.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "optimum-internal-testing/tiny-random-qwen3_moe",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "Qwen3MoeForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "decoder_sparse_step": 2,
11
+ "head_dim": 32,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 64,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 128,
16
+ "max_position_embeddings": 40960,
17
+ "max_window_layers": 1,
18
+ "mlp_only_layers": [],
19
+ "model_type": "qwen3_moe",
20
+ "moe_intermediate_size": 128,
21
+ "neuron": {
22
+ "_serialized_key": "NxDNeuronConfig",
23
+ "batch_size": 1,
24
+ "capacity_factor": null,
25
+ "checkpoint_id": "optimum-internal-testing/tiny-random-qwen3_moe",
26
+ "checkpoint_revision": "e0230be2839556b44b7400a233c73c74b4abb7af",
27
+ "continuous_batching": false,
28
+ "enable_bucketing": false,
29
+ "ep_degree": 1,
30
+ "fused_qkv": false,
31
+ "glu_mlp": true,
32
+ "local_ranks_size": 2,
33
+ "max_batch_size": 1,
34
+ "max_context_length": 1024,
35
+ "max_topk": 256,
36
+ "n_active_tokens": 1024,
37
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
38
+ "on_device_sampling": true,
39
+ "optimum_neuron_version": "0.4.0.dev0",
40
+ "output_logits": false,
41
+ "pp_degree": 1,
42
+ "sequence_length": 1024,
43
+ "speculation_length": 0,
44
+ "start_rank_id": 0,
45
+ "target": "trn1",
46
+ "torch_dtype": "bfloat16",
47
+ "tp_degree": 2
48
+ },
49
+ "norm_topk_prob": true,
50
+ "num_attention_heads": 2,
51
+ "num_experts": 8,
52
+ "num_experts_per_tok": 2,
53
+ "num_hidden_layers": 2,
54
+ "num_key_value_heads": 1,
55
+ "output_router_logits": false,
56
+ "rms_norm_eps": 1e-06,
57
+ "rope_scaling": null,
58
+ "rope_theta": 1000000.0,
59
+ "router_aux_loss_coef": 0.001,
60
+ "sliding_window": null,
61
+ "tie_word_embeddings": true,
62
+ "use_cache": true,
63
+ "use_sliding_window": false,
64
+ "vocab_size": 151936
65
+ }
neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23fff2222c80e609ede15fd7899202e3dcfec823c20d5693d35391a2f351de45
3
+ size 82753
neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25d5794197cc1c440a87ceb511f5c1be579e938e1804d40a69b8ef1de5520571
3
+ size 277504
neuronxcc-2.21.18209.0+043b1bf7/MODULE_10d60f8fad55e1974521+a9d440f5/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f620b1d7b37b2896c0fbf86202063926fb1a18d68cc3d20eecca1c10090bc877
3
+ size 285854
neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20ed0858162f8fc2c58bc4c23d0925dbeda5897dae0889f87758206676394f4e
3
+ size 81843
neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9f08146dd5aa669a7a5ede165f2203a5a1d493ea0f8ffbada902aabcc925731
3
+ size 246784
neuronxcc-2.21.18209.0+043b1bf7/MODULE_13aa568fe38435861b30+a9d440f5/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b9cd23cd69de2f90c806808492a0ba59374d106b61dc84a5da694f7eaa6544e
3
+ size 255104
neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/context_encoding_model/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc7da2288fa0065f5732bfc84bd11a1f2628bca9861071457605d3d1e2fe916a
3
+ size 90382
neuronxcc-2.21.18209.0+043b1bf7/MODULE_1e1261ac026ac7064c12+ed72d204/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c0b29e79eb8a89f0c7bd1af01320c2b64606cb55036efd8000daf55aad89acc
3
+ size 359424
neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/context_encoding_model/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60a30064c095e66d36b82f0f119f8c90e2d8f9250729a0fc969fdda597569500
3
+ size 86123
neuronxcc-2.21.18209.0+043b1bf7/MODULE_585fd41ddb2ec3219380+ed72d204/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3b8f03e2bb9f3ae70c214bc950dea144994836574886708e96d52119507dd07
3
+ size 257024
neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/context_encoding_model/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24624bc45be46654efb5eb289cc62aae8e0818687d6378e93a7244553928e7d7
3
+ size 83504
neuronxcc-2.21.18209.0+043b1bf7/MODULE_607905c97e8a375ad8f1+ed72d204/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1def2aa1595ee48cf7ddf6dfa4a7d2c7ffa27f4a0ec104c45bce6bd29fc6238f
3
+ size 328704
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0902b9467e36affd529236be99c466bb326eb59e56e6b6414b2bde17d7913ac8
3
+ size 676474
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5fc99e7ec1985f69435ca23e19dce386cbd8c525fe8d2312b9868c81fae5716
3
+ size 543744
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a6ac71f9922f4f01abf5+a9d440f5/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6d898c25998f845a5baa83037e9802318d822e4e70a3c4ecdfc708db5d74f19
3
+ size 563380
neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db7ebbb9e2aa4d3c3d4667f098e18faa9ed0a231c9831e73c8118bfad21264ad
3
+ size 91147
neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4ea59696b130915819c0ac02b7eb02b02772c1e64866eb6aa4ebf4398a3c5d1
3
+ size 277504
neuronxcc-2.21.18209.0+043b1bf7/MODULE_c0a2583211556c96a6a3+a9d440f5/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:557828feb4ce2b868dc358d87c8adebb3fe5c6597abdd6a1f8788c3d40aa9f1c
3
+ size 289031
neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:833b82cfd5de929b110eae7f2597028fd2aba27feb903dc4a6389317ec3d4c73
3
+ size 85236
neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7107fd17fab62ea880b3fc40e65d9322170bdc32b2f0b47ee90bd468ca184858
3
+ size 216064
neuronxcc-2.21.18209.0+043b1bf7/MODULE_e51f8e75d0439314e217+a9d440f5/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38beb98efa2cceb12ac6bf7c5384bf134a28598bab758ba4a38b5ee0cab4d21b
3
+ size 224412
neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/context_encoding_model/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb9d3c9ec74f1ca339872d2af14bf39e9ac08891d8cbb125730abe59da953421
3
+ size 82772
neuronxcc-2.21.18209.0+043b1bf7/MODULE_fc37bb64f7e86324075c+ed72d204/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c40c4e13ba0555414340c531f5a55db4a6134efed627548ed0bf43d3f6680170
3
+ size 267264
neuronxcc-2.21.18209.0+043b1bf7/MODULE_fef991cead1748c4d101+ed72d204/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/context_encoding_model/_tp0_bk0/log-neuron-cc.txt"]