jaeyong2 commited on
Commit
e91d5e8
·
verified ·
1 Parent(s): 4621722

Upload Qwen3ForCausalLM

Browse files
Files changed (3) hide show
  1. config.json +1 -1
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -56,7 +56,7 @@
56
  "spmd_fsdp_sharding": false,
57
  "tie_word_embeddings": true,
58
  "torch_dtype": "bfloat16",
59
- "transformers_version": "4.55.2",
60
  "use_cache": false,
61
  "use_sliding_window": false,
62
  "vocab_size": 151936
 
56
  "spmd_fsdp_sharding": false,
57
  "tie_word_embeddings": true,
58
  "torch_dtype": "bfloat16",
59
+ "transformers_version": "4.55.3",
60
  "use_cache": false,
61
  "use_sliding_window": false,
62
  "vocab_size": 151936
generation_config.json CHANGED
@@ -9,5 +9,5 @@
9
  "temperature": 0.6,
10
  "top_k": 20,
11
  "top_p": 0.95,
12
- "transformers_version": "4.55.2"
13
  }
 
9
  "temperature": 0.6,
10
  "top_k": 20,
11
  "top_p": 0.95,
12
+ "transformers_version": "4.55.3"
13
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3f07d731013e0a458115d1ffc5c1f7a1bc78620648ab3146746c735097390a4
3
  size 1192135096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4f18be0638489b5ee36900a61ee5712d9e324b72889c914bbd4ed8a51aa1501
3
  size 1192135096