Text Generation
Transformers
Safetensors
PyTorch
English
nvidia
suhara commited on
Commit
8f26a60
·
verified ·
1 Parent(s): eb78bd0

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +4 -4
config.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "architectures": [
3
- "Nemotron5ForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "attention_head_dim": 128,
8
  "auto_map": {
9
- "AutoConfig": "configuration_nemotron5.Nemotron5Config",
10
- "AutoModelForCausalLM": "modeling_nemotron5.Nemotron5ForCausalLM"
11
  },
12
  "bos_token_id": 1,
13
  "chunk_size": 256,
@@ -27,7 +27,7 @@
27
  "max_position_embeddings": 8192,
28
  "mlp_bias": false,
29
  "mlp_hidden_act": "relu2",
30
- "model_type": "nemotron5",
31
  "n_groups": 8,
32
  "num_attention_heads": 32,
33
  "num_hidden_layers": 52,
 
1
  {
2
  "architectures": [
3
+ "NemotronHForCausalLM"
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "attention_head_dim": 128,
8
  "auto_map": {
9
+ "AutoConfig": "configuration_nemotron_h.NemotronHConfig",
10
+ "AutoModelForCausalLM": "modeling_nemotron_h.NemotronHForCausalLM"
11
  },
12
  "bos_token_id": 1,
13
  "chunk_size": 256,
 
27
  "max_position_embeddings": 8192,
28
  "mlp_bias": false,
29
  "mlp_hidden_act": "relu2",
30
+ "model_type": "nemotron_h",
31
  "n_groups": 8,
32
  "num_attention_heads": 32,
33
  "num_hidden_layers": 52,