jerryzh168 commited on
Commit
997d19c
·
verified ·
1 Parent(s): d0167a7

Upload OPTForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +1 -1
  2. generation_config.json +1 -1
config.json CHANGED
@@ -74,7 +74,7 @@
74
  "untie_embedding_weights": false
75
  },
76
  "torch_dtype": "bfloat16",
77
- "transformers_version": "4.53.0",
78
  "use_cache": true,
79
  "vocab_size": 50272,
80
  "word_embed_proj_dim": 768
 
74
  "untie_embedding_weights": false
75
  },
76
  "torch_dtype": "bfloat16",
77
+ "transformers_version": "4.53.3",
78
  "use_cache": true,
79
  "vocab_size": 50272,
80
  "word_embed_proj_dim": 768
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 2,
4
  "eos_token_id": 2,
5
  "pad_token_id": 1,
6
- "transformers_version": "4.53.0"
7
  }
 
3
  "bos_token_id": 2,
4
  "eos_token_id": 2,
5
  "pad_token_id": 1,
6
+ "transformers_version": "4.53.3"
7
  }