dacorvo HF Staff commited on
Commit
5714ed3
·
verified ·
1 Parent(s): 0a189a8

Synchronizing local compiler cache.

Browse files
.gitattributes CHANGED
@@ -11894,3 +11894,6 @@ neuronxcc-2.21.18209.0+043b1bf7/MODULE_14928681760864605802+f7f529f3/model.neff
11894
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_2092161936687880602+f7f529f3/model.neff filter=lfs diff=lfs merge=lfs -text
11895
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_2927013901654163041+f7f529f3/model.neff filter=lfs diff=lfs merge=lfs -text
11896
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_2962232050317550424+f7f529f3/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
 
11894
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_2092161936687880602+f7f529f3/model.neff filter=lfs diff=lfs merge=lfs -text
11895
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_2927013901654163041+f7f529f3/model.neff filter=lfs diff=lfs merge=lfs -text
11896
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_2962232050317550424+f7f529f3/model.neff filter=lfs diff=lfs merge=lfs -text
11897
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_669f33fc567910d306ab+564b7b5b/model.neff filter=lfs diff=lfs merge=lfs -text
11898
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_b85364df3fc31288dc1c+d0d57c8a/model.neff filter=lfs diff=lfs merge=lfs -text
11899
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_b85364df3fc31288dc1c+d0d57c8a/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.1.dev0/qwen3_moe/Qwen/Qwen3-Coder-30B-A3B-Instruct/1215d01d0b52e0c80683.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "Qwen/Qwen3-Coder-30B-A3B-Instruct",
4
+ "_task": "text-generation",
5
+ "architectures": [
6
+ "Qwen3MoeForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "decoder_sparse_step": 1,
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 2048,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 5472,
16
+ "max_position_embeddings": 262144,
17
+ "max_window_layers": 28,
18
+ "mlp_only_layers": [],
19
+ "model_type": "qwen3_moe",
20
+ "moe_intermediate_size": 768,
21
+ "neuron": {
22
+ "_serialized_key": "NxDNeuronConfig",
23
+ "batch_size": 16,
24
+ "capacity_factor": null,
25
+ "checkpoint_id": "Qwen/Qwen3-Coder-30B-A3B-Instruct",
26
+ "checkpoint_revision": "573fa3901e5799703b1e60825b0ec024a4c0f1d3",
27
+ "continuous_batching": true,
28
+ "enable_bucketing": false,
29
+ "ep_degree": 1,
30
+ "fused_qkv": false,
31
+ "glu_mlp": true,
32
+ "local_ranks_size": 4,
33
+ "max_batch_size": 16,
34
+ "max_context_length": 4096,
35
+ "max_topk": 256,
36
+ "n_active_tokens": 4096,
37
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
38
+ "on_device_sampling": true,
39
+ "optimum_neuron_version": "0.4.1.dev0",
40
+ "output_logits": false,
41
+ "pp_degree": 1,
42
+ "sequence_length": 4096,
43
+ "speculation_length": 0,
44
+ "start_rank_id": 0,
45
+ "target": "trn2",
46
+ "torch_dtype": "bfloat16",
47
+ "tp_degree": 4
48
+ },
49
+ "norm_topk_prob": true,
50
+ "num_attention_heads": 32,
51
+ "num_experts": 128,
52
+ "num_experts_per_tok": 8,
53
+ "num_hidden_layers": 48,
54
+ "num_key_value_heads": 4,
55
+ "output_router_logits": false,
56
+ "qkv_bias": false,
57
+ "rms_norm_eps": 1e-06,
58
+ "rope_scaling": null,
59
+ "rope_theta": 10000000,
60
+ "router_aux_loss_coef": 0.0,
61
+ "shared_expert_intermediate_size": 0,
62
+ "sliding_window": null,
63
+ "tie_word_embeddings": false,
64
+ "use_cache": true,
65
+ "use_qk_norm": true,
66
+ "use_sliding_window": false,
67
+ "vocab_size": 151936
68
+ }
neuronxcc-2.21.18209.0+043b1bf7/MODULE_2a3d222b1e460d753301+d0d57c8a/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--enable-saturate-infinity", "--enable-mixed-precision-accumulation", "--model-type", "transformer", "-O1", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2", "--auto-cast=none", "--internal-enable-dge-levels", "vector_dynamic_offsets", "--internal-hlo2tensorizer-options=--verify-hlo=true", "--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_2a3d222b1e460d753301+d0d57c8a/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e0800ecd46f0d680b0c9bf5514c1b241a43ff41648d510d053ed27d3d3b0fa8
3
+ size 14468174
neuronxcc-2.21.18209.0+043b1bf7/MODULE_2a3d222b1e460d753301+d0d57c8a/model.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ Failed compilation with ['neuronx-cc', 'compile', '--framework=XLA', '/tmp/nxd_model/token_generation_model/_tp0_bk0/model.MODULE_2a3d222b1e460d753301+d0d57c8a.hlo_module.pb', '--output', '/tmp/nxd_model/token_generation_model/_tp0_bk0/model.MODULE_2a3d222b1e460d753301+d0d57c8a.neff', '--target=trn2', '--enable-saturate-infinity', '--enable-mixed-precision-accumulation', '--model-type', 'transformer', '-O1', '--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2', '--auto-cast=none', '--internal-enable-dge-levels', 'vector_dynamic_offsets', '--internal-hlo2tensorizer-options=--verify-hlo=true', '--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt', '--enable-internal-neff-wrapper', '--verbose=35']: 2025-10-16T11:15:30Z 2025-10-16 11:15:30.307267: F hilo/hlo_passes/NeuronHloVerifier.cc:504] [ERROR] [NCC_VRF009] Memory requirement exceeds target architecture's HBM limit. Needed 18507002752 bytes (17 GB) vs. available 17179869184 bytes (16 GB). TIP: Consider using smaller batches or applying model parallelism
2
+
neuronxcc-2.21.18209.0+043b1bf7/MODULE_669f33fc567910d306ab+564b7b5b/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--enable-saturate-infinity", "--enable-mixed-precision-accumulation", "--model-type", "transformer", "-O1", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2", "--auto-cast=none", "--internal-enable-dge-levels", "vector_dynamic_offsets", "--internal-hlo2tensorizer-options=--verify-hlo=true", "--logfile=/tmp/nxd_model/context_encoding_model/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_669f33fc567910d306ab+564b7b5b/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_669f33fc567910d306ab+564b7b5b/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f8551db25d6c000741567bc5ef73f556581b39bb73b58a4b4580ed042cb0023
3
+ size 10995696
neuronxcc-2.21.18209.0+043b1bf7/MODULE_669f33fc567910d306ab+564b7b5b/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc379dc0f4e63d796eced8a100b901e4de0141de37d9cb52834aacb00835921a
3
+ size 16159744
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b85364df3fc31288dc1c+d0d57c8a/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--enable-saturate-infinity", "--enable-mixed-precision-accumulation", "--model-type", "transformer", "-O1", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2", "--auto-cast=none", "--internal-enable-dge-levels", "vector_dynamic_offsets", "--internal-hlo2tensorizer-options=--verify-hlo=true", "--logfile=/tmp/nxd_model/token_generation_model/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b85364df3fc31288dc1c+d0d57c8a/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b85364df3fc31288dc1c+d0d57c8a/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abcac81baad6aef1db291146b7601567b89bcb5db34f7676b18942692a0de687
3
+ size 12051954
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b85364df3fc31288dc1c+d0d57c8a/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3a41cf2fa4c52126a3d319bd14e3d84c3a8a2d8c415b52f94eb7bd3c607ccef
3
+ size 3984384
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b85364df3fc31288dc1c+d0d57c8a/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7bdbd9d86286b4d2d40a110747321080624d5ea42200084e2278c84abf095d8
3
+ size 4257451