CrashOverrideX commited on
Commit
6b16a2d
·
verified ·
1 Parent(s): af5dc29

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +2 -4
config.json CHANGED
@@ -10,12 +10,10 @@
10
  "intermediate_size": 4096,
11
  "max_position_embeddings": 1024,
12
  "rms_norm_eps": 1e-05,
 
13
  "rope_theta": 10000.0,
14
  "tie_word_embeddings": true,
15
  "use_cache": true,
16
  "torch_dtype": "float16",
17
- "transformers_version": "4.36.0",
18
- "auto_map": {
19
- "AutoModelForCausalLM": "modeling_llama.LlamaForCausalLM"
20
- }
21
  }
 
10
  "intermediate_size": 4096,
11
  "max_position_embeddings": 1024,
12
  "rms_norm_eps": 1e-05,
13
+ "rope_scaling": null,
14
  "rope_theta": 10000.0,
15
  "tie_word_embeddings": true,
16
  "use_cache": true,
17
  "torch_dtype": "float16",
18
+ "transformers_version": "4.36.0"
 
 
 
19
  }