lvcalucioli commited on
Commit
9b349f3
·
verified ·
1 Parent(s): b2c7cd9

ca-finetuned-phi-2

Browse files
README.md CHANGED
@@ -1,8 +1,11 @@
1
  ---
2
  license: mit
3
- base_model: microsoft/phi-2
4
  tags:
 
 
5
  - generated_from_trainer
 
6
  model-index:
7
  - name: ca-finetuned-phi-2
8
  results: []
@@ -49,7 +52,8 @@ The following hyperparameters were used during training:
49
 
50
  ### Framework versions
51
 
52
- - Transformers 4.31.0
53
- - Pytorch 2.1.0+cu121
 
54
  - Datasets 2.16.1
55
- - Tokenizers 0.13.3
 
1
  ---
2
  license: mit
3
+ library_name: peft
4
  tags:
5
+ - trl
6
+ - sft
7
  - generated_from_trainer
8
+ base_model: microsoft/phi-2
9
  model-index:
10
  - name: ca-finetuned-phi-2
11
  results: []
 
52
 
53
  ### Framework versions
54
 
55
+ - PEFT 0.8.2
56
+ - Transformers 4.38.0.dev0
57
+ - Pytorch 2.0.1+cu117
58
  - Datasets 2.16.1
59
+ - Tokenizers 0.15.1
adapter_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "auto_mapping": null,
3
  "base_model_name_or_path": "microsoft/phi-2",
4
  "bias": "none",
@@ -7,16 +8,21 @@
7
  "init_lora_weights": true,
8
  "layers_pattern": null,
9
  "layers_to_transform": null,
10
- "lora_alpha": 32,
 
11
  "lora_dropout": 0.1,
 
 
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
- "r": 16,
 
15
  "revision": null,
16
  "target_modules": [
17
- "Wqkv",
18
  "fc1",
 
19
  "fc2"
20
  ],
21
- "task_type": "CAUSAL_LM"
 
22
  }
 
1
  {
2
+ "alpha_pattern": {},
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "microsoft/phi-2",
5
  "bias": "none",
 
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
+ "loftq_config": {},
12
+ "lora_alpha": 64,
13
  "lora_dropout": 0.1,
14
+ "megatron_config": null,
15
+ "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 32,
19
+ "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "fc1",
23
+ "Wqkv",
24
  "fc2"
25
  ],
26
+ "task_type": "CAUSAL_LM",
27
+ "use_rslora": false
28
  }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7257da06b5e8654f0df08cdc88fb9681af51f95aada1ad3257c8abfbfb085e0
3
+ size 146825352
special_tokens_map.json CHANGED
@@ -1,6 +1,24 @@
1
  {
2
- "bos_token": "<|endoftext|>",
3
- "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
4
  "pad_token": "<|endoftext|>",
5
- "unk_token": "<|endoftext|>"
 
 
 
 
 
 
6
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
  "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
  }
tokenizer_config.json CHANGED
@@ -318,6 +318,7 @@
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
 
321
  "tokenizer_class": "CodeGenTokenizer",
322
  "unk_token": "<|endoftext|>"
323
  }
 
318
  "clean_up_tokenization_spaces": true,
319
  "eos_token": "<|endoftext|>",
320
  "model_max_length": 2048,
321
+ "pad_token": "<|endoftext|>",
322
  "tokenizer_class": "CodeGenTokenizer",
323
  "unk_token": "<|endoftext|>"
324
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68f50dc5733f5be25e3e8a26c6ed9fc0142e3da61aa5159248df7e6b541d54a2
3
- size 4408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02f954991a87e54b4627a2b72a1bb5be59c0d3173606f9ee5a05b41cc21b7c44
3
+ size 4283