dacorvo HF Staff commited on
Commit
7091183
·
verified ·
1 Parent(s): 0887fcc

Delete neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1

Browse files
neuronxcc-2.21.33363.0+82129205/0_REGISTRY/0.4.5.dev1/cf6b9a360dcf294104671106bae2adbd9fd291823bb60a351883163684073231/3310d05484c3f0d11dcf.json DELETED
@@ -1,63 +0,0 @@
1
- {
2
- "_entry_class": "SingleModelCacheEntry",
3
- "_model_id": "unsloth/Llama-3.2-1B-Instruct",
4
- "_task": "text-generation",
5
- "architectures": [
6
- "LlamaForCausalLM"
7
- ],
8
- "attention_bias": false,
9
- "attention_dropout": 0.0,
10
- "dtype": "bfloat16",
11
- "head_dim": 64,
12
- "hidden_act": "silu",
13
- "hidden_size": 2048,
14
- "initializer_range": 0.02,
15
- "intermediate_size": 8192,
16
- "max_position_embeddings": 131072,
17
- "mlp_bias": false,
18
- "model_type": "llama",
19
- "neuron": {
20
- "_serialized_key": "NxDNeuronConfig",
21
- "batch_size": 1,
22
- "capacity_factor": null,
23
- "checkpoint_id": "unsloth/Llama-3.2-1B-Instruct",
24
- "checkpoint_revision": "5a8abab4a5d6f164389b1079fb721cfab8d7126c",
25
- "continuous_batching": false,
26
- "ep_degree": 1,
27
- "fused_qkv": true,
28
- "glu_mlp": true,
29
- "local_ranks_size": 1,
30
- "max_batch_size": 1,
31
- "max_context_length": 4096,
32
- "max_topk": 256,
33
- "n_active_tokens": 4096,
34
- "neuronxcc_version": "2.21.33363.0+82129205",
35
- "on_device_sampling": true,
36
- "optimum_neuron_version": "0.4.5.dev1",
37
- "output_logits": false,
38
- "pp_degree": 1,
39
- "sequence_length": 4096,
40
- "speculation_length": 0,
41
- "start_rank_id": 0,
42
- "target": "trn1",
43
- "torch_dtype": "bfloat16",
44
- "tp_degree": 1
45
- },
46
- "num_attention_heads": 32,
47
- "num_hidden_layers": 16,
48
- "num_key_value_heads": 8,
49
- "pretraining_tp": 1,
50
- "rms_norm_eps": 1e-05,
51
- "rope_scaling": {
52
- "factor": 32.0,
53
- "high_freq_factor": 4.0,
54
- "low_freq_factor": 1.0,
55
- "original_max_position_embeddings": 8192,
56
- "rope_type": "llama3"
57
- },
58
- "rope_theta": 500000.0,
59
- "tie_word_embeddings": true,
60
- "unsloth_fixed": true,
61
- "use_cache": true,
62
- "vocab_size": 128256
63
- }