Panga-Azazia commited on
Commit
7ee1490
·
verified ·
1 Parent(s): 7877066

End of training

Browse files
README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: reward-model-1
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ # reward-model-1
14
+
15
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Loss: 0.0609
18
+ - Mse: 0.0609
19
+ - R2: 0.5447
20
+ - Pearson: 0.7406
21
+
22
+ ## Model description
23
+
24
+ More information needed
25
+
26
+ ## Intended uses & limitations
27
+
28
+ More information needed
29
+
30
+ ## Training and evaluation data
31
+
32
+ More information needed
33
+
34
+ ## Training procedure
35
+
36
+ ### Training hyperparameters
37
+
38
+ The following hyperparameters were used during training:
39
+ - learning_rate: 0.0001
40
+ - train_batch_size: 8
41
+ - eval_batch_size: 8
42
+ - seed: 42
43
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
44
+ - lr_scheduler_type: cosine
45
+ - num_epochs: 10
46
+
47
+ ### Training results
48
+
49
+ | Training Loss | Epoch | Step | Validation Loss | Mse | R2 | Pearson |
50
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:-------:|
51
+ | 0.1237 | 0.8 | 100 | 0.1100 | 0.1100 | 0.1781 | 0.5916 |
52
+ | 0.0675 | 1.6 | 200 | 0.0723 | 0.0723 | 0.4597 | 0.6906 |
53
+ | 0.0562 | 2.4 | 300 | 0.0684 | 0.0684 | 0.4890 | 0.7094 |
54
+ | 0.0625 | 3.2 | 400 | 0.0650 | 0.0650 | 0.5145 | 0.7175 |
55
+ | 0.0563 | 4.0 | 500 | 0.0662 | 0.0662 | 0.5055 | 0.7120 |
56
+ | 0.0478 | 4.8 | 600 | 0.0616 | 0.0616 | 0.5396 | 0.7398 |
57
+ | 0.0454 | 5.6 | 700 | 0.0634 | 0.0634 | 0.5266 | 0.7264 |
58
+ | 0.0429 | 6.4 | 800 | 0.0607 | 0.0607 | 0.5467 | 0.7404 |
59
+ | 0.0422 | 7.2 | 900 | 0.0615 | 0.0615 | 0.5405 | 0.7429 |
60
+ | 0.0421 | 8.0 | 1000 | 0.0622 | 0.0622 | 0.5353 | 0.7338 |
61
+ | 0.0423 | 8.8 | 1100 | 0.0610 | 0.0610 | 0.5446 | 0.7424 |
62
+ | 0.0485 | 9.6 | 1200 | 0.0610 | 0.0610 | 0.5445 | 0.7416 |
63
+
64
+
65
+ ### Framework versions
66
+
67
+ - Transformers 4.53.3
68
+ - Pytorch 2.9.0+cu128
69
+ - Datasets 3.3.2
70
+ - Tokenizers 0.21.4
all_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "eval_loss": 0.060941558331251144,
4
+ "eval_mse": 0.060941558331251144,
5
+ "eval_pearson": 0.7406131625175476,
6
+ "eval_r2": 0.5447108745574951,
7
+ "eval_runtime": 3.9474,
8
+ "eval_samples_per_second": 50.667,
9
+ "eval_steps_per_second": 6.333
10
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f61a8a1394e9321be93513550c0b8a8bb6245230616671b4217b22d28fb31a9e
3
  size 3593828
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f45d244b4f6c68bfc1bc872993d1a78c00c08eb1fc5d15aad6b8d42b0e672063
3
  size 3593828
runs/Nov18_20-34-35_azazia/events.out.tfevents.1763498075.azazia.69531.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b80481b596862de24dc40fee0c71c57a4b518b5f0700f6aabc365b046e90c99
3
- size 11766
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6d097e25ce3005be6fb8d9352fbf7354a8f2bf32d9c3db2aabcd159c7ba5139
3
+ size 15679
runs/Nov18_20-34-35_azazia/events.out.tfevents.1763498395.azazia.69531.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce962ee5d26073ef785631016e1d59a20732dd0ac3c3c35c53a332f5180e5a9b
3
+ size 918
test_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "eval_loss": 0.060941558331251144,
4
+ "eval_mse": 0.060941558331251144,
5
+ "eval_pearson": 0.7406131625175476,
6
+ "eval_r2": 0.5447108745574951,
7
+ "eval_runtime": 3.9474,
8
+ "eval_samples_per_second": 50.667,
9
+ "eval_steps_per_second": 6.333
10
+ }