Hitisha commited on
Commit
bc139a1
·
verified ·
1 Parent(s): 7300642

Model save

Browse files
README.md CHANGED
@@ -17,19 +17,6 @@ should probably proofread and complete it, then remove this comment. -->
17
  # orpo-phi3
18
 
19
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on the None dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 4.4137
22
- - Rewards/chosen: -0.4068
23
- - Rewards/rejected: -0.4068
24
- - Rewards/accuracies: 0.0
25
- - Rewards/margins: 0.0
26
- - Logps/rejected: -4.0684
27
- - Logps/chosen: -4.0684
28
- - Logits/rejected: 22.3128
29
- - Logits/chosen: 22.3128
30
- - Nll Loss: 4.3444
31
- - Log Odds Ratio: -0.6931
32
- - Log Odds Chosen: 0.0
33
 
34
  ## Model description
35
 
@@ -59,17 +46,10 @@ The following hyperparameters were used during training:
59
  - lr_scheduler_warmup_steps: 10
60
  - num_epochs: 1
61
 
62
- ### Training results
63
-
64
- | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen | Nll Loss | Log Odds Ratio | Log Odds Chosen |
65
- |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|:--------:|:--------------:|:---------------:|
66
- | 4.1766 | 0.8 | 1 | 4.4137 | -0.4068 | -0.4068 | 0.0 | 0.0 | -4.0684 | -4.0684 | 22.3128 | 22.3128 | 4.3444 | -0.6931 | 0.0 |
67
-
68
-
69
  ### Framework versions
70
 
71
  - PEFT 0.11.1
72
- - Transformers 4.41.2
73
- - Pytorch 2.3.0+cu121
74
  - Datasets 2.20.0
75
  - Tokenizers 0.19.1
 
17
  # orpo-phi3
18
 
19
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on the None dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  ## Model description
22
 
 
46
  - lr_scheduler_warmup_steps: 10
47
  - num_epochs: 1
48
 
 
 
 
 
 
 
 
49
  ### Framework versions
50
 
51
  - PEFT 0.11.1
52
+ - Transformers 4.42.3
53
+ - Pytorch 2.1.2
54
  - Datasets 2.20.0
55
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "gate_up_proj",
24
  "o_proj",
 
25
  "qkv_proj",
26
  "down_proj"
27
  ],
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "o_proj",
24
+ "gate_up_proj",
25
  "qkv_proj",
26
  "down_proj"
27
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:30a595c549d8d93ddaccc29e80a93b2a02720016cf695a50c46ea2f82f673459
3
  size 887450008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd0f5bb1237bf549fa73ac2c283393d35d064f7c3c8ca6170b9632bc644ce2cd
3
  size 887450008
runs/Jul15_06-59-30_08422e6f1a8e/events.out.tfevents.1721026772.08422e6f1a8e.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b0e7088509cb81f021f4876412413687e9d5decfce922f492bd0ea27843a8f5
3
+ size 19541
tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_bos_token": false,
3
  "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
 
1
  {
2
  "add_bos_token": false,
3
  "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:996082b3e31474d0f885e97d42c2162e4ad600cfee32e088aa509e8dfebdd7ac
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a168b42dff42511590c6f2518b832bc993667bc81459de5d369f89df13b5f8d
3
  size 5432