jadicorn commited on
Commit
c103c35
·
verified ·
1 Parent(s): c69305b

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -4,10 +4,10 @@
4
  "GPT2LMHeadModel"
5
  ],
6
  "attn_pdrop": 0.1,
7
- "bos_token_id": 50256,
8
  "dtype": "float32",
9
  "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
  "initializer_range": 0.02,
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
@@ -27,5 +27,5 @@
27
  "summary_use_proj": true,
28
  "transformers_version": "4.57.1",
29
  "use_cache": true,
30
- "vocab_size": 50257
31
  }
 
4
  "GPT2LMHeadModel"
5
  ],
6
  "attn_pdrop": 0.1,
7
+ "bos_token_id": 1,
8
  "dtype": "float32",
9
  "embd_pdrop": 0.1,
10
+ "eos_token_id": 2,
11
  "initializer_range": 0.02,
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
 
27
  "summary_use_proj": true,
28
  "transformers_version": "4.57.1",
29
  "use_cache": true,
30
+ "vocab_size": 8192
31
  }
generation_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_from_model_config": true,
3
- "bos_token_id": 50256,
4
- "eos_token_id": 50256,
5
  "transformers_version": "4.57.1"
6
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
  "transformers_version": "4.57.1"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:956420d3c573d7f77036b8e51f2e2e187332857a8526c01f5337e3daf2923708
3
- size 65155056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d7b3526349a299a4d159e77115825b1662172e60c405de4adb3f6370ab8f652
3
+ size 22080496
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a04c860d08c80039ffe0996267290961732a27df61b34bf8940813a5e38cc71
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65acd1d7f2d53e8ff4c06aadbb970382b48ebb477d6374b52df831557a9f9f81
3
  size 5496