b3x0m commited on
Commit
0a573db
·
verified ·
1 Parent(s): b9ed7fd

Upload XomdichForConditionalGeneration

Browse files
Files changed (2) hide show
  1. config.json +1 -1
  2. pytorch_model.bin +1 -1
config.json CHANGED
@@ -7,7 +7,7 @@
7
  "eos_token_id": 2,
8
  "hidden_size": 512,
9
  "max_sequence_length": 512,
10
- "model_type": "Xomdich",
11
  "num_attention_heads": 8,
12
  "num_key_value_heads": 4,
13
  "num_layers": 12,
 
7
  "eos_token_id": 2,
8
  "hidden_size": 512,
9
  "max_sequence_length": 512,
10
+ "model_type": "Hyper-Xomdich",
11
  "num_attention_heads": 8,
12
  "num_key_value_heads": 4,
13
  "num_layers": 12,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90f668ec283bc9fb2b9f243141d3839d57b334d2e0f84427cb4989fa13ad6b4e
3
  size 471757350
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2833061f759a4a886c5f0e0dd35a95da5776baebf51b9b6414ad942d61a88da3
3
  size 471757350