ChrisWhiteQMUL commited on
Commit
1fdbe4f
·
verified ·
1 Parent(s): 5f7c373

trained_sentiment

Browse files
README.md CHANGED
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.6323
22
 
23
  ## Model description
24
 
@@ -51,17 +51,17 @@ The following hyperparameters were used during training:
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss |
55
- |:-------------:|:------:|:----:|:---------------:|
56
- | 1.4554 | 0.9944 | 155 | 1.4685 |
57
- | 1.1957 | 1.9952 | 311 | 1.4965 |
58
- | 0.8349 | 2.9832 | 465 | 1.6323 |
59
 
60
 
61
  ### Framework versions
62
 
63
- - PEFT 0.11.1
64
- - Transformers 4.42.3
65
- - Pytorch 2.1.2
66
- - Datasets 2.20.0
67
  - Tokenizers 0.19.1
 
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 1.8211
22
 
23
  ## Model description
24
 
 
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss |
55
+ |:-------------:|:-----:|:----:|:---------------:|
56
+ | 1.5093 | 1.0 | 491 | 1.6536 |
57
+ | 1.1833 | 2.0 | 982 | 1.6768 |
58
+ | 0.8405 | 3.0 | 1473 | 1.8211 |
59
 
60
 
61
  ### Framework versions
62
 
63
+ - PEFT 0.12.0
64
+ - Transformers 4.44.0
65
+ - Pytorch 2.4.0
66
+ - Datasets 2.21.0
67
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "up_proj",
24
  "v_proj",
25
  "o_proj",
 
26
  "q_proj",
27
- "k_proj",
28
- "down_proj",
29
- "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "k_proj",
24
  "up_proj",
25
  "v_proj",
26
  "o_proj",
27
+ "gate_proj",
28
  "q_proj",
29
+ "down_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34a05f13213775cb104190a181b2113ed30b2cc23177371d7f3c40217b07f100
3
  size 671149168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efc11587817c99a14e37e37e10aea341c71f10924ee0e80c77f5622c2980ca6b
3
  size 671149168
runs/Sep05_09-04-16_b5c5cac84dd7/events.out.tfevents.1725527063.b5c5cac84dd7.36.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1876c0e32854a12392097c1eee6760e07c397cf8edd147d2e3f2a66e106025ac
3
+ size 19236
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fdd23b45d841263f7f3b48244046d2a75da29add7542f463dfd1cc990fa90478
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02407104359185f38047ae92f1530f30f89a09007915a37c292ceb7b619082c5
3
+ size 5496