dacorvo HF Staff commited on
Commit
21b0b60
·
verified ·
1 Parent(s): 6855239

Delete neuronxcc-2.14.227.0+2d4f85be

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/2c73925e46a18751b152.json +0 -1
  2. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7b3393a7c5fd984a34d9.json +0 -1
  3. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/c997564eba3d6777671f.json +0 -1
  4. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/720f4c07d5f83b2ff9d4.json +0 -1
  5. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/f54e0d6fd7a8f4a15970.json +0 -1
  6. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/1fb2d1d6f793fd1c9744.json +0 -1
  7. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/65314683f5d624301ac7.json +0 -1
  8. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/6fa0619f83aff7ea628e.json +0 -1
  9. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/ad50cdf65ebac5fec89c.json +0 -1
  10. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/f32b4f1b111de9eb757b.json +0 -1
  11. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/openai-community/gpt2/1249d63d4d5c2e285546.json +0 -1
  12. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/openai-community/gpt2/6bfc4f92bf3397ee076c.json +0 -1
  13. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/openai-community/gpt2/f6a990533107b53bd088.json +0 -1
  14. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/07ca11781d2a782a310d.json +0 -1
  15. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/102403e8b14ba6cfddfc.json +0 -1
  16. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/4dbc1484f4f8fac732b3.json +0 -1
  17. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/7132382858e4979ff08d.json +0 -1
  18. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/89becbc133a202838e89.json +0 -1
  19. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/95ba6fff05cf856e8b58.json +0 -1
  20. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/LargeWorldModel/LWM-Text-Chat-1M/1a5e6cb020ffbd366c9d.json +0 -1
  21. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/LargeWorldModel/LWM-Text-Chat-1M/85cee1378c49e9bd9345.json +0 -1
  22. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/4ce991f77dcad86e8b22.json +0 -1
  23. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/abacusai/Smaug-72B-v0.1/0d8e57049c33c5b5c6d9.json +0 -1
  24. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/abacusai/Smaug-72B-v0.1/364906a5de98225ddc2a.json +0 -1
  25. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/1caf4d7e91344cb4edd4.json +0 -1
  26. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/6896790e87a085e9f268.json +0 -1
  27. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/a5d3d8806cd364c7a404.json +0 -1
  28. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/cc9c0a9006808dddfa27.json +0 -1
  29. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/defog/sqlcoder-7b-2/2a7f28749bf930679a90.json +0 -1
  30. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/defog/sqlcoder-7b-2/d93fe38049979b3a11ec.json +0 -1
  31. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/gorilla-llm/gorilla-openfunctions-v2/1675f856c527911cc075.json +0 -1
  32. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/gorilla-llm/gorilla-openfunctions-v2/8919f202fa2c453a5b01.json +0 -1
  33. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/ibm/labradorite-13b/0249f8755f1b6b4f9758.json +0 -1
  34. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/ibm/labradorite-13b/5bea66ac968792320916.json +0 -1
  35. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/ibm/labradorite-13b/cd69690e4daa0d46b110.json +0 -1
  36. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/ChatMusician/587fc1fb5c309060ab0a.json +0 -1
  37. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/ChatMusician/9a81805ff4067a45e60a.json +0 -1
  38. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/OpenCodeInterpreter-DS-6.7B/76b3c6d150157aecf7dc.json +0 -1
  39. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/OpenCodeInterpreter-DS-6.7B/cada4d393bbfdb636df1.json +0 -1
  40. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/1f1367c0cc366a15a9ed.json +0 -1
  41. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/670388f2bc4b70f03ab8.json +0 -1
  42. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/6768cae045972320134e.json +0 -1
  43. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/6b851dcb4093a8a0aa2c.json +0 -1
  44. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/768707faa39666e82e14.json +0 -1
  45. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/861050381645b81661a7.json +0 -1
  46. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/acadfa72a5b2bd2b8d55.json +0 -1
  47. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/d45199ec3e6e802b71c2.json +0 -1
  48. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/dfb30cdcb69c459f0b9c.json +0 -1
  49. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/e3998f3a732380eae75e.json +0 -1
  50. neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-70b-chat-hf/044ec5e8a8f4bdfa3487.json +0 -1
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/2c73925e46a18751b152.json DELETED
@@ -1 +0,0 @@
1
- {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/7b3393a7c5fd984a34d9.json DELETED
@@ -1 +0,0 @@
1
- {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/c997564eba3d6777671f.json DELETED
@@ -1 +0,0 @@
1
- {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/720f4c07d5f83b2ff9d4.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/f54e0d6fd7a8f4a15970.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 128, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/1fb2d1d6f793fd1c9744.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/65314683f5d624301ac7.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/6fa0619f83aff7ea628e.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 128, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/ad50cdf65ebac5fec89c.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 512, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/f32b4f1b111de9eb757b.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/openai-community/gpt2/1249d63d4d5c2e285546.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 16, "checkpoint_id": "openai-community/gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/openai-community/gpt2/6bfc4f92bf3397ee076c.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "openai-community/gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/gpt2/openai-community/gpt2/f6a990533107b53bd088.json DELETED
@@ -1 +0,0 @@
1
- {"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "openai-community/gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/07ca11781d2a782a310d.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 7168, "initializer_range": 0.02, "intermediate_size": 20480, "max_position_embeddings": 200000, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "01-ai/Yi-34B-200K", "checkpoint_revision": "868e4b3e60493abd3478abe667ea5e9440d90317", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 56, "num_hidden_layers": 60, "num_key_value_heads": 8, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 64000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/102403e8b14ba6cfddfc.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 7168, "initializer_range": 0.02, "intermediate_size": 20480, "max_position_embeddings": 200000, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "01-ai/Yi-34B-200K", "checkpoint_revision": "7c27c5cab0f5675edc0d9910769d0a6695ade7c7", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 56, "num_hidden_layers": 60, "num_key_value_heads": 8, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 64000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/4dbc1484f4f8fac732b3.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 7168, "initializer_range": 0.02, "intermediate_size": 20480, "max_position_embeddings": 200000, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "01-ai/Yi-34B-200K", "checkpoint_revision": "7c27c5cab0f5675edc0d9910769d0a6695ade7c7", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 56, "num_hidden_layers": 60, "num_key_value_heads": 8, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 64000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/7132382858e4979ff08d.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 7168, "initializer_range": 0.02, "intermediate_size": 20480, "max_position_embeddings": 200000, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "01-ai/Yi-34B-200K", "checkpoint_revision": "1068d547bdb1d013f9fd0315f9472327841af3b7", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 56, "num_hidden_layers": 60, "num_key_value_heads": 8, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 64000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/89becbc133a202838e89.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 7168, "initializer_range": 0.02, "intermediate_size": 20480, "max_position_embeddings": 200000, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "01-ai/Yi-34B-200K", "checkpoint_revision": "868e4b3e60493abd3478abe667ea5e9440d90317", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 56, "num_hidden_layers": 60, "num_key_value_heads": 8, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 64000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/01-ai/Yi-34B-200K/95ba6fff05cf856e8b58.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 7168, "initializer_range": 0.02, "intermediate_size": 20480, "max_position_embeddings": 200000, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "01-ai/Yi-34B-200K", "checkpoint_revision": "1068d547bdb1d013f9fd0315f9472327841af3b7", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 56, "num_hidden_layers": 60, "num_key_value_heads": 8, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 64000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/LargeWorldModel/LWM-Text-Chat-1M/1a5e6cb020ffbd366c9d.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 1048576, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "LargeWorldModel/LWM-Text-Chat-1M", "checkpoint_revision": "0598c443b02aeb1a1f9f6788e9af85ea762a452d", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 50000000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/LargeWorldModel/LWM-Text-Chat-1M/85cee1378c49e9bd9345.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 1048576, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "LargeWorldModel/LWM-Text-Chat-1M", "checkpoint_revision": "0598c443b02aeb1a1f9f6788e9af85ea762a452d", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 50000000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/NousResearch/Hermes-2-Theta-Llama-3-8B/4ce991f77dcad86e8b22.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 128000, "eos_token_id": 128003, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 14336, "max_position_embeddings": 8192, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "NousResearch/Hermes-2-Theta-Llama-3-8B", "checkpoint_revision": "81fc25d402859151f907ad479b18070a3c5a3c11", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 128256}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/abacusai/Smaug-72B-v0.1/0d8e57049c33c5b5c6d9.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": true, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 151643, "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 24576, "max_position_embeddings": 32768, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "abacusai/Smaug-72B-v0.1", "checkpoint_revision": "a1d657156f82c24b670158406378648233487011", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 64, "pad_token_id": 151643, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 1000000, "seq_length": 32768, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 152064}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/abacusai/Smaug-72B-v0.1/364906a5de98225ddc2a.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": true, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 151643, "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 24576, "max_position_embeddings": 32768, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "abacusai/Smaug-72B-v0.1", "checkpoint_revision": "a1d657156f82c24b670158406378648233487011", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 64, "pad_token_id": 151643, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 1000000, "seq_length": 32768, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 152064}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/1caf4d7e91344cb4edd4.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/6896790e87a085e9f268.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 128, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/a5d3d8806cd364c7a404.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/cc9c0a9006808dddfa27.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/defog/sqlcoder-7b-2/2a7f28749bf930679a90.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 16384, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "defog/sqlcoder-7b-2", "checkpoint_revision": "7e5b6f7981c0aa7d143f6bec6fa26625bdfcbe66", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 1000000, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32016}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/defog/sqlcoder-7b-2/d93fe38049979b3a11ec.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 16384, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "defog/sqlcoder-7b-2", "checkpoint_revision": "7e5b6f7981c0aa7d143f6bec6fa26625bdfcbe66", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 1000000, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32016}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/gorilla-llm/gorilla-openfunctions-v2/1675f856c527911cc075.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100000, "eos_token_id": 100015, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "gorilla-llm/gorilla-openfunctions-v2", "checkpoint_revision": "1f6ac3b8bb09e347a875f5ff08cd66798706341e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 30, "num_key_value_heads": 32, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": false, "vocab_size": 102400}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/gorilla-llm/gorilla-openfunctions-v2/8919f202fa2c453a5b01.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 100000, "eos_token_id": 100015, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gorilla-llm/gorilla-openfunctions-v2", "checkpoint_revision": "1f6ac3b8bb09e347a875f5ff08cd66798706341e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 30, "num_key_value_heads": 32, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": false, "vocab_size": 102400}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/ibm/labradorite-13b/0249f8755f1b6b4f9758.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "end_token_id": 32000, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "ibm/labradorite-13b", "checkpoint_revision": "26921cf1d63c9c657e8260d08eaacc14fca43da0", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pad_token_id": 32001, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32008}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/ibm/labradorite-13b/5bea66ac968792320916.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "end_token_id": 32000, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 8, "checkpoint_id": "ibm/labradorite-13b", "checkpoint_revision": "26921cf1d63c9c657e8260d08eaacc14fca43da0", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pad_token_id": 32001, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32008}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/ibm/labradorite-13b/cd69690e4daa0d46b110.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "end_token_id": 32000, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "ibm/labradorite-13b", "checkpoint_revision": "26921cf1d63c9c657e8260d08eaacc14fca43da0", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 8, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pad_token_id": 32001, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": true, "vocab_size": 32008}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/ChatMusician/587fc1fb5c309060ab0a.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 2048, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "m-a-p/ChatMusician", "checkpoint_revision": "cd8064911f26cd6701561eb733319adfa88383df", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/ChatMusician/9a81805ff4067a45e60a.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 2048, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "m-a-p/ChatMusician", "checkpoint_revision": "cd8064911f26cd6701561eb733319adfa88383df", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pad_token_id": 0, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/OpenCodeInterpreter-DS-6.7B/76b3c6d150157aecf7dc.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 32013, "eos_token_id": 32021, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 16384, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "m-a-p/OpenCodeInterpreter-DS-6.7B", "checkpoint_revision": "60b89884df814590abd76757a6db4a527cbdfc91", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": {"factor": 4.0, "type": "linear"}, "rope_theta": 100000, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": false, "vocab_size": 32256}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/m-a-p/OpenCodeInterpreter-DS-6.7B/cada4d393bbfdb636df1.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 32013, "eos_token_id": 32021, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 16384, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "m-a-p/OpenCodeInterpreter-DS-6.7B", "checkpoint_revision": "60b89884df814590abd76757a6db4a527cbdfc91", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 2, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 32, "num_hidden_layers": 32, "num_key_value_heads": 32, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": {"factor": 4.0, "type": "linear"}, "rope_theta": 100000, "tie_word_embeddings": false, "torch_dtype": "bfloat16", "use_cache": false, "vocab_size": 32256}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/1f1367c0cc366a15a9ed.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 16, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/670388f2bc4b70f03ab8.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 32, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/6768cae045972320134e.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 8, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/6b851dcb4093a8a0aa2c.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 8, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 12, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/768707faa39666e82e14.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 32, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 12, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/861050381645b81661a7.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 16, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 12, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/acadfa72a5b2bd2b8d55.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/d45199ec3e6e802b71c2.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/dfb30cdcb69c459f0b9c.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 12, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-13b-chat-hf/e3998f3a732380eae75e.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 5120, "initializer_range": 0.02, "intermediate_size": 13824, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-2-13b-chat-hf", "checkpoint_revision": "a2cb7a712bb6e5e736ca7f8cd98167f81a0b5bd8", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 12, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 40, "num_hidden_layers": 40, "num_key_value_heads": 40, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
 
 
neuronxcc-2.14.227.0+2d4f85be/0_REGISTRY/0.0.24.dev0/inference/llama/meta-llama/Llama-2-70b-chat-hf/044ec5e8a8f4bdfa3487.json DELETED
@@ -1 +0,0 @@
1
- {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 8192, "initializer_range": 0.02, "intermediate_size": 28672, "max_position_embeddings": 4096, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "meta-llama/Llama-2-70b-chat-hf", "checkpoint_revision": "e9149a12809580e8602995856f8098ce973d1080", "compiler_type": "neuronx-cc", "compiler_version": "2.14.227.0+2d4f85be", "num_cores": 24, "sequence_length": 4096, "task": "text-generation"}, "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 8, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}