dacorvo HF Staff commited on
Commit
cc9aeb8
·
verified ·
1 Parent(s): 30e1497

Synchronizing local compiler cache.

Browse files
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.6.dev2/441269935591cad8d370e512c0b93cdd2fce6247c40e5a4866d872ee5338b0de/468443f97476debe433b.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "unsloth/gemma-3-270m-it",
4
+ "_sliding_window_pattern": 6,
5
+ "_task": "text-generation",
6
+ "architectures": [
7
+ "Gemma3ForCausalLM"
8
+ ],
9
+ "attention_bias": false,
10
+ "attention_dropout": 0.0,
11
+ "attn_logit_softcapping": null,
12
+ "dtype": "bfloat16",
13
+ "final_logit_softcapping": null,
14
+ "head_dim": 256,
15
+ "hidden_activation": "gelu_pytorch_tanh",
16
+ "hidden_size": 640,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 2048,
19
+ "layer_types": [
20
+ "sliding_attention",
21
+ "sliding_attention",
22
+ "sliding_attention",
23
+ "sliding_attention",
24
+ "sliding_attention",
25
+ "full_attention",
26
+ "sliding_attention",
27
+ "sliding_attention",
28
+ "sliding_attention",
29
+ "sliding_attention",
30
+ "sliding_attention",
31
+ "full_attention",
32
+ "sliding_attention",
33
+ "sliding_attention",
34
+ "sliding_attention",
35
+ "sliding_attention",
36
+ "sliding_attention",
37
+ "full_attention"
38
+ ],
39
+ "max_position_embeddings": 32768,
40
+ "model_type": "gemma3_text",
41
+ "neuron": {
42
+ "_serialized_key": "NxDNeuronConfig",
43
+ "batch_size": 1,
44
+ "capacity_factor": null,
45
+ "checkpoint_id": "unsloth/gemma-3-270m-it",
46
+ "checkpoint_revision": "23cf460f6bb16954176b3ddcc8d4f250501458a9",
47
+ "continuous_batching": false,
48
+ "ep_degree": 1,
49
+ "fused_qkv": true,
50
+ "glu_mlp": true,
51
+ "local_ranks_size": 2,
52
+ "max_batch_size": 1,
53
+ "max_context_length": 8192,
54
+ "max_topk": 256,
55
+ "n_active_tokens": 8192,
56
+ "neuronxcc_version": "2.21.33363.0+82129205",
57
+ "on_device_sampling": true,
58
+ "optimum_neuron_version": "0.4.6.dev2",
59
+ "output_logits": false,
60
+ "pp_degree": 1,
61
+ "prefill_chunk_size": 0,
62
+ "sequence_length": 8192,
63
+ "speculation_length": 0,
64
+ "start_rank_id": 0,
65
+ "target": "trn1",
66
+ "torch_dtype": "bfloat16",
67
+ "tp_degree": 2
68
+ },
69
+ "num_attention_heads": 4,
70
+ "num_hidden_layers": 18,
71
+ "num_key_value_heads": 1,
72
+ "query_pre_attn_scalar": 256,
73
+ "rms_norm_eps": 1e-06,
74
+ "rope_local_base_freq": 10000.0,
75
+ "rope_scaling": null,
76
+ "rope_theta": 1000000.0,
77
+ "sliding_window": 512,
78
+ "unsloth_fixed": true,
79
+ "use_bidirectional_attention": false,
80
+ "use_cache": true,
81
+ "vocab_size": 262144
82
+ }
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.6.dev2/gemma3_text/unsloth/gemma-3-270m-it/468443f97476debe433b.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "unsloth/gemma-3-270m-it",
4
+ "_sliding_window_pattern": 6,
5
+ "_task": "text-generation",
6
+ "architectures": [
7
+ "Gemma3ForCausalLM"
8
+ ],
9
+ "attention_bias": false,
10
+ "attention_dropout": 0.0,
11
+ "attn_logit_softcapping": null,
12
+ "dtype": "bfloat16",
13
+ "final_logit_softcapping": null,
14
+ "head_dim": 256,
15
+ "hidden_activation": "gelu_pytorch_tanh",
16
+ "hidden_size": 640,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 2048,
19
+ "layer_types": [
20
+ "sliding_attention",
21
+ "sliding_attention",
22
+ "sliding_attention",
23
+ "sliding_attention",
24
+ "sliding_attention",
25
+ "full_attention",
26
+ "sliding_attention",
27
+ "sliding_attention",
28
+ "sliding_attention",
29
+ "sliding_attention",
30
+ "sliding_attention",
31
+ "full_attention",
32
+ "sliding_attention",
33
+ "sliding_attention",
34
+ "sliding_attention",
35
+ "sliding_attention",
36
+ "sliding_attention",
37
+ "full_attention"
38
+ ],
39
+ "max_position_embeddings": 32768,
40
+ "model_type": "gemma3_text",
41
+ "neuron": {
42
+ "_serialized_key": "NxDNeuronConfig",
43
+ "batch_size": 1,
44
+ "capacity_factor": null,
45
+ "checkpoint_id": "unsloth/gemma-3-270m-it",
46
+ "checkpoint_revision": "23cf460f6bb16954176b3ddcc8d4f250501458a9",
47
+ "continuous_batching": false,
48
+ "ep_degree": 1,
49
+ "fused_qkv": true,
50
+ "glu_mlp": true,
51
+ "local_ranks_size": 2,
52
+ "max_batch_size": 1,
53
+ "max_context_length": 8192,
54
+ "max_topk": 256,
55
+ "n_active_tokens": 8192,
56
+ "neuronxcc_version": "2.21.33363.0+82129205",
57
+ "on_device_sampling": true,
58
+ "optimum_neuron_version": "0.4.6.dev2",
59
+ "output_logits": false,
60
+ "pp_degree": 1,
61
+ "prefill_chunk_size": 0,
62
+ "sequence_length": 8192,
63
+ "speculation_length": 0,
64
+ "start_rank_id": 0,
65
+ "target": "trn1",
66
+ "torch_dtype": "bfloat16",
67
+ "tp_degree": 2
68
+ },
69
+ "num_attention_heads": 4,
70
+ "num_hidden_layers": 18,
71
+ "num_key_value_heads": 1,
72
+ "query_pre_attn_scalar": 256,
73
+ "rms_norm_eps": 1e-06,
74
+ "rope_local_base_freq": 10000.0,
75
+ "rope_scaling": null,
76
+ "rope_theta": 1000000.0,
77
+ "sliding_window": 512,
78
+ "unsloth_fixed": true,
79
+ "use_bidirectional_attention": false,
80
+ "use_cache": true,
81
+ "vocab_size": 262144
82
+ }
neuronxcc-2.21.33363.0+82129205/MODULE_a26fd792225a238c6638+a02c3a36/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=1", "--logfile=/tmp/nxd_model/token_generation/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.33363.0+82129205/MODULE_a26fd792225a238c6638+a02c3a36/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef830748873b5c5612aa5dcb37309fb909dadd8660ccda725a681089f93d3c4
3
+ size 420625
neuronxcc-2.21.33363.0+82129205/MODULE_c69743a3b026cb8c0f9e+a02c3a36/model.hlo_module.pb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:726697ae6d69106a35fc6caab871c35bccbf129873b887f75986cfbdd687e625
3
  size 588724
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19abd02589b7e770cc7d22a41832a1d7c1b18e5db1cdbd0f57e95939951b2311
3
  size 588724
neuronxcc-2.21.33363.0+82129205/MODULE_c69743a3b026cb8c0f9e+a02c3a36/model.neff CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7c258ba5a15e6c435b0719ac65ac7aed62119f5d3d3dc497bc8557452b5afc7
3
  size 1659904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e9a16db24435f3344d68819315ac43e7ec035fdf41c68ba517491e1cc7db394
3
  size 1659904
neuronxcc-2.21.33363.0+82129205/MODULE_c69743a3b026cb8c0f9e+a02c3a36/wrapped_neff.hlo CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd30f302fc03f3e2cbf90294931da3ae0f5faa5a5f0cf021340988ef20175088
3
  size 1782293
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee0ea35ca5a7f2dda442d215bb13c572d23bc55b448a956b0f57522e111c3d11
3
  size 1782293