dacorvo HF Staff commited on
Commit
88d85d1
·
verified ·
1 Parent(s): 1ca5b87

Synchronizing local compiler cache.

Browse files
.gitattributes CHANGED
@@ -1927,3 +1927,5 @@ neuronxcc-2.16.372.0+4a9b2326/MODULE_f65495564d03f1c90e3a+613edded/model.neff fi
1927
  neuronxcc-2.16.372.0+4a9b2326/MODULE_f914d922823d8ceace3c+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
1928
  neuronxcc-2.16.372.0+4a9b2326/MODULE_177e7d4bfa5feb0f544b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
1929
  neuronxcc-2.16.372.0+4a9b2326/MODULE_a95326667095c4d75db0+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
 
 
 
1927
  neuronxcc-2.16.372.0+4a9b2326/MODULE_f914d922823d8ceace3c+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
1928
  neuronxcc-2.16.372.0+4a9b2326/MODULE_177e7d4bfa5feb0f544b+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
1929
  neuronxcc-2.16.372.0+4a9b2326/MODULE_a95326667095c4d75db0+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
1930
+ neuronxcc-2.16.372.0+4a9b2326/MODULE_396f866b25e26394047c+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
1931
+ neuronxcc-2.16.372.0+4a9b2326/MODULE_f91cd98a64f373af274d+613edded/model.neff filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.16.372.0+4a9b2326/0_REGISTRY/0.1.0.dev2/inference/llama/unsloth/Llama-3.2-1B-Instruct/49d61e18927b113552d1.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128009, "head_dim": 64, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "unsloth/Llama-3.2-1B-Instruct", "checkpoint_revision": "9b58d4a36161a1e49ecf0a69d20b2736fef8e438", "compiler_type": "neuronx-cc", "compiler_version": "2.16.372.0+4a9b2326", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pad_token_id": 128004, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": true, "torch_dtype": "bfloat16", "unsloth_fixed": true, "use_cache": true, "vocab_size": 128256}
neuronxcc-2.16.372.0+4a9b2326/MODULE_396f866b25e26394047c+613edded/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
neuronxcc-2.16.372.0+4a9b2326/MODULE_396f866b25e26394047c+613edded/model.done ADDED
File without changes
neuronxcc-2.16.372.0+4a9b2326/MODULE_396f866b25e26394047c+613edded/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8210594bad3d116ba0d73d29ba21d8197083302ce1bd7e605a5b9a7654c4d90e
3
+ size 222605
neuronxcc-2.16.372.0+4a9b2326/MODULE_396f866b25e26394047c+613edded/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cc37d213fe1c4dc9702aa25d5b78d8c6e4b057da4aa86bf0d99943729f8f81f
3
+ size 27792384
neuronxcc-2.16.372.0+4a9b2326/MODULE_f91cd98a64f373af274d+613edded/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn1", "--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
neuronxcc-2.16.372.0+4a9b2326/MODULE_f91cd98a64f373af274d+613edded/model.done ADDED
File without changes
neuronxcc-2.16.372.0+4a9b2326/MODULE_f91cd98a64f373af274d+613edded/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc998a517871f9971580382edc47f8e45377a0dab229bff749798ab51c992928
3
+ size 181367
neuronxcc-2.16.372.0+4a9b2326/MODULE_f91cd98a64f373af274d+613edded/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77dd9e4543e1ac818e4fd323945cf7f7f810e11a47115dfc4350a4fe240d7326
3
+ size 2366464