kyu5787 commited on
Commit
46b189e
·
verified ·
1 Parent(s): 91e6896

End of training

Browse files
adapter_config.json CHANGED
@@ -26,8 +26,8 @@
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
- "k_proj",
30
- "q_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "q_proj",
30
+ "k_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:faf732716528ba13a0684bb11ab11b0da7db9010324ef06015dceef29f98e196
3
  size 64011808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b2f8baa2a2a2452a00664f3d836de75ae4f818b569b391b1a148b656749af2e
3
  size 64011808
tokenizer.json CHANGED
@@ -1,7 +1,14 @@
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
+ "padding": {
5
+ "strategy": "BatchLongest",
6
+ "direction": "Right",
7
+ "pad_to_multiple_of": null,
8
+ "pad_id": 2,
9
+ "pad_type_id": 0,
10
+ "pad_token": "<|im_end|>"
11
+ },
12
  "added_tokens": [
13
  {
14
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd9adbd06ed07b22cea978413da345df024faff246e2c47890a15aa4d7f6481f
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8798cfb5073eb9cb47859431cb58848a43999274c33c5247698ea927f0d4eb93
3
  size 5496