mecoaoge2 commited on
Commit
5c5c6c1
·
1 Parent(s): 9ee0d5e

End of training

Browse files
README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - trl
6
+ - dpo
7
+ - generated_from_trainer
8
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
9
+ model-index:
10
+ - name: fununun
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # fununun
18
+
19
+ This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.6908
22
+ - Rewards/chosen: 0.0015
23
+ - Rewards/rejected: -0.0032
24
+ - Rewards/accuracies: 0.7176
25
+ - Rewards/margins: 0.0047
26
+ - Logps/rejected: -197.2385
27
+ - Logps/chosen: -235.0630
28
+ - Logits/rejected: -3.0691
29
+ - Logits/chosen: -3.1037
30
+
31
+ ## Model description
32
+
33
+ More information needed
34
+
35
+ ## Intended uses & limitations
36
+
37
+ More information needed
38
+
39
+ ## Training and evaluation data
40
+
41
+ More information needed
42
+
43
+ ## Training procedure
44
+
45
+ ### Training hyperparameters
46
+
47
+ The following hyperparameters were used during training:
48
+ - learning_rate: 5e-07
49
+ - train_batch_size: 2
50
+ - eval_batch_size: 2
51
+ - seed: 42
52
+ - gradient_accumulation_steps: 16
53
+ - total_train_batch_size: 32
54
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
+ - lr_scheduler_type: linear
56
+ - lr_scheduler_warmup_steps: 20
57
+ - training_steps: 100
58
+
59
+ ### Training results
60
+
61
+ | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
62
+ |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
63
+ | 0.693 | 0.04 | 20 | 0.6927 | 0.0001 | -0.0007 | 0.5614 | 0.0009 | -197.2139 | -235.0765 | -3.0688 | -3.1035 |
64
+ | 0.6922 | 0.07 | 40 | 0.6919 | 0.0007 | -0.0017 | 0.6440 | 0.0024 | -197.2236 | -235.0704 | -3.0690 | -3.1036 |
65
+ | 0.6913 | 0.11 | 60 | 0.6913 | 0.0011 | -0.0025 | 0.6886 | 0.0037 | -197.2319 | -235.0664 | -3.0691 | -3.1037 |
66
+ | 0.6909 | 0.15 | 80 | 0.6909 | 0.0014 | -0.0030 | 0.7098 | 0.0044 | -197.2367 | -235.0639 | -3.0691 | -3.1037 |
67
+ | 0.6906 | 0.19 | 100 | 0.6908 | 0.0015 | -0.0032 | 0.7176 | 0.0047 | -197.2385 | -235.0630 | -3.0691 | -3.1037 |
68
+
69
+
70
+ ### Framework versions
71
+
72
+ - PEFT 0.7.1
73
+ - Transformers 4.36.2
74
+ - Pytorch 2.1.0+cu121
75
+ - Datasets 2.16.1
76
+ - Tokenizers 0.15.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f79288f6df2addc359d37611b2ea6803f2a9fd04486b552a9dbadff2b8f0b623
3
  size 50503544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fdb0e6a7f2b8dbb74a544f65375ce519674e7c64c38aa5cce19cb6f83e5027c
3
  size 50503544
runs/Jan10_06-19-39_da9d9452362f/events.out.tfevents.1704867739.da9d9452362f.217.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91e11faa7c06d1957ee18c6aaf1150223a819850067fc0460a7692f566eca3a6
3
- size 11332
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc000860a7ef75c22818d48d5c484e018aac8b676427a71c00ac5d1169ed10f9
3
+ size 15003