Delete neuronxcc-2.16.303.0+d9f03cda
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/11cebcde130c1f5b8a5d.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/481519fdaa82a5e9ca3c.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/5276a012f2eefa9af6da.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/gpt2/aa40852fa5208b294329.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/4b98629fe202b2140a00.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/af13e5d873398d00de9e.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/d0a6fa1b6956a4d680fa.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-70B-Instruct/8382ce744316c5e96168.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-70B-Instruct/f78fb1b51c0a86c73fe1.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/0a04e054893aa9fc9527.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/1b405fdcd7a404de447f.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/5a26e9e32d1a95deb7e9.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/a168d15b4566c85e9fc8.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/b828518bd74cc3b11dc0.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/d4873490f07d6364f226.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/4811debc00fce09ac124.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/70df192a162dc76a8ba2.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/b6900574c407686bc850.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/b708ec372a7fc766d289.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/d8df4e1c37e866f90233.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/e852034e61abed3bc0ba.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/0dd9463a56ca665e0009.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/21ab9e45cd41b65fb4e0.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ec3b5393c3977096c001.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mixtral/dacorvo/Mixtral-tiny/63d71adb49398ef365fe.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mixtral/dacorvo/Mixtral-tiny/84e0c54d67bb0c2ff590.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mixtral/dacorvo/Mixtral-tiny/dfa5935d58cd292c0422.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-72B-Instruct/31e20d081c7e41aaca2f.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-72B-Instruct/8d0c44cbd5e8aec22406.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-72B-Instruct/cd41d41826e43d4cf0ba.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-7B-Instruct/33e2b8582ee983e39510.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-7B-Instruct/5cca9f20b9c6707b6477.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-7B-Instruct/ae4504091d356b47e96a.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/compile_flags.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/model.done +0 -0
- neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/model.hlo_module.pb +0 -3
- neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/model.neff +0 -3
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/compile_flags.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/model.done +0 -0
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/model.hlo_module.pb +0 -3
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/model.neff +0 -0
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/compile_flags.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/model.done +0 -0
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/model.hlo_module.pb +0 -3
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/model.neff +0 -3
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/compile_flags.json +0 -1
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/model.done +0 -0
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/model.hlo_module.pb +0 -3
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/model.neff +0 -3
- neuronxcc-2.16.303.0+d9f03cda/MODULE_0d95eb3b3153bf26363d+613edded/compile_flags.json +0 -1
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/11cebcde130c1f5b8a5d.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/481519fdaa82a5e9ca3c.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/5276a012f2eefa9af6da.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/gpt2/aa40852fa5208b294329.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/4b98629fe202b2140a00.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/af13e5d873398d00de9e.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/d0a6fa1b6956a4d680fa.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-70B-Instruct/8382ce744316c5e96168.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 28672, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "meta-llama/Llama-3.1-70B-Instruct", "checkpoint_revision": "1605565b47bb9346c5515c34102e054115b4f98b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 64, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-70B-Instruct/f78fb1b51c0a86c73fe1.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 28672, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "meta-llama/Llama-3.1-70B-Instruct", "checkpoint_revision": "1605565b47bb9346c5515c34102e054115b4f98b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 64, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/0a04e054893aa9fc9527.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "meta-llama/Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 16, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/1b405fdcd7a404de447f.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 64, "checkpoint_id": "meta-llama/Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 64, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/5a26e9e32d1a95deb7e9.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 128, "checkpoint_id": "meta-llama/Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 64, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/a168d15b4566c85e9fc8.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 64, "checkpoint_id": "meta-llama/Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 32, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/b828518bd74cc3b11dc0.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "meta-llama/Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 32, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Llama-3.1-8B-Instruct/d4873490f07d6364f226.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": [128001, 128008, 128009], "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "meta-llama/Llama-3.1-8B-Instruct", "checkpoint_revision": "0e9e39f249a16976918f6564b8830bc894c89659", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 1024, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/4811debc00fce09ac124.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "meta-llama/Meta-Llama-3.1-8B", "checkpoint_revision": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/70df192a162dc76a8ba2.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 1, "checkpoint_id": "meta-llama/Meta-Llama-3.1-8B", "checkpoint_revision": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/b6900574c407686bc850.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 8, "checkpoint_id": "meta-llama/Meta-Llama-3.1-8B", "checkpoint_revision": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/b708ec372a7fc766d289.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 128, "checkpoint_id": "meta-llama/Meta-Llama-3.1-8B", "checkpoint_revision": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/d8df4e1c37e866f90233.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "meta-llama/Meta-Llama-3.1-8B", "checkpoint_revision": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/llama/meta-llama/Meta-Llama-3.1-8B/e852034e61abed3bc0ba.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 131072, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "meta-llama/Meta-Llama-3.1-8B", "checkpoint_revision": "d04e592bb4f6aa9cfee91e2e20afa771667e1d4b", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 8.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/0dd9463a56ca665e0009.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/21ab9e45cd41b65fb4e0.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/ec3b5393c3977096c001.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "head_dim": 8, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mixtral/dacorvo/Mixtral-tiny/63d71adb49398ef365fe.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mixtral/dacorvo/Mixtral-tiny/84e0c54d67bb0c2ff590.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/mixtral/dacorvo/Mixtral-tiny/dfa5935d58cd292c0422.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-72B-Instruct/31e20d081c7e41aaca2f.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 29568, "max_position_embeddings": 32768, "max_window_layers": 70, "model_type": "qwen2", "neuron": {"auto_cast_type": "bf16", "batch_size": 32, "checkpoint_id": "Qwen/Qwen2.5-72B-Instruct", "checkpoint_revision": "d3d951150c1e5848237cd6a7ad11df4836aee842", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 64, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 8, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-72B-Instruct/8d0c44cbd5e8aec22406.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 29568, "max_position_embeddings": 32768, "max_window_layers": 70, "model_type": "qwen2", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "Qwen/Qwen2.5-72B-Instruct", "checkpoint_revision": "d3d951150c1e5848237cd6a7ad11df4836aee842", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 64, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 8, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-72B-Instruct/cd41d41826e43d4cf0ba.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 29568, "max_position_embeddings": 32768, "max_window_layers": 70, "model_type": "qwen2", "neuron": {"auto_cast_type": "bf16", "batch_size": 64, "checkpoint_id": "Qwen/Qwen2.5-72B-Instruct", "checkpoint_revision": "d3d951150c1e5848237cd6a7ad11df4836aee842", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 64, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 8, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-7B-Instruct/33e2b8582ee983e39510.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 18944, "max_position_embeddings": 32768, "max_window_layers": 28, "model_type": "qwen2", "neuron": {"auto_cast_type": "bf16", "batch_size": 8, "checkpoint_id": "Qwen/Qwen2.5-7B-Instruct", "checkpoint_revision": "bb46c15ee4bb56c5b63245ef50fd7637234d6f75", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 1, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 28, "num_hidden_layers": 28, "num_key_value_heads": 4, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-7B-Instruct/5cca9f20b9c6707b6477.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 18944, "max_position_embeddings": 32768, "max_window_layers": 28, "model_type": "qwen2", "neuron": {"auto_cast_type": "bf16", "batch_size": 16, "checkpoint_id": "Qwen/Qwen2.5-7B-Instruct", "checkpoint_revision": "bb46c15ee4bb56c5b63245ef50fd7637234d6f75", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 1, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 28, "num_hidden_layers": 28, "num_key_value_heads": 4, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/0_REGISTRY/0.0.28.dev0/inference/qwen2/Qwen/Qwen2.5-7B-Instruct/ae4504091d356b47e96a.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"architectures": ["Qwen2ForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 151643, "eos_token_id": 151645, "hidden_act": "silu", "hidden_size": 3584, "initializer_range": 0.02, "intermediate_size": 18944, "max_position_embeddings": 32768, "max_window_layers": 28, "model_type": "qwen2", "neuron": {"auto_cast_type": "bf16", "batch_size": 4, "checkpoint_id": "Qwen/Qwen2.5-7B-Instruct", "checkpoint_revision": "bb46c15ee4bb56c5b63245ef50fd7637234d6f75", "compiler_type": "neuronx-cc", "compiler_version": "2.16.303.0+d9f03cda", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 28, "num_hidden_layers": 28, "num_key_value_heads": 4, "rms_norm_eps": 1e-06, "rope_theta": 1000000.0, "sliding_window": null, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "use_sliding_window": false, "vocab_size": 152064}
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/compile_flags.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
["--target=trn2", "--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/model.done
DELETED
|
File without changes
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/model.hlo_module.pb
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:fae6bba5e76799bce81338669eab39b6f67989cebf3e30d0ed3bff2a98afc556
|
| 3 |
-
size 331842
|
|
|
|
|
|
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_01c540090ff8bced389b+7ac4dbae/model.neff
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:ae241934369b4c671b6a1e3b5286933f850c24642730687b8c06a0bf6bda1ad1
|
| 3 |
-
size 3206144
|
|
|
|
|
|
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/compile_flags.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
["--target=trn1", "--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/model.done
DELETED
|
File without changes
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/model.hlo_module.pb
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:ec9373f814ab03ecf1f250e2f08fa06127361ee356b1d62b3b34d4fc688b5261
|
| 3 |
-
size 52661
|
|
|
|
|
|
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0912ea9cf6fe9e1af454+613edded/model.neff
DELETED
|
Binary file (82.9 kB)
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/compile_flags.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
["--target=trn2", "--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/model.done
DELETED
|
File without changes
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/model.hlo_module.pb
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:5ad3238f27b4a9d135b388a179b33523dda19b51d168f88a5fa6fad2ed4e7c07
|
| 3 |
-
size 438733
|
|
|
|
|
|
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0924c0dfe6c0bf06cc17+7ac4dbae/model.neff
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:3609eae16ab9a907e3a082ce670b0224793f54c36846051970277550243cc58e
|
| 3 |
-
size 20675584
|
|
|
|
|
|
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/compile_flags.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
["--target=trn2", "--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/model.done
DELETED
|
File without changes
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/model.hlo_module.pb
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:376bbe53219bb85008e5cd4702ed2cb62dc11a22ce499d582735a227184c8101
|
| 3 |
-
size 331842
|
|
|
|
|
|
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0b08b19b4706a50917ba+7ac4dbae/model.neff
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:d6a6d15af0cf4a6fa0cca4b0306385d0f594cae08274973cf1c963f731827ece
|
| 3 |
-
size 3154944
|
|
|
|
|
|
|
|
|
|
|
|
neuronxcc-2.16.303.0+d9f03cda/MODULE_0d95eb3b3153bf26363d+613edded/compile_flags.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
["--target=trn1", "--model-type=transformer", "--auto-cast=none", "--execute-repetition=1"]
|
|
|
|
|
|