colinrdavisMBA commited on
Commit
cb90c2f
·
verified ·
1 Parent(s): 6d2de3c

colinrdavisMBA/mistral-7binstruct-summary-100s

Browse files
README.md CHANGED
@@ -20,7 +20,7 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.4646
24
 
25
  ## Model description
26
 
@@ -52,14 +52,14 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:----:|:---------------:|
55
- | 1.6134 | 0.22 | 25 | 1.5514 |
56
- | 1.5949 | 0.43 | 50 | 1.4646 |
57
 
58
 
59
  ### Framework versions
60
 
61
- - PEFT 0.8.2
62
- - Transformers 4.38.1
63
  - Pytorch 2.1.0+cu121
64
- - Datasets 2.17.1
65
  - Tokenizers 0.15.2
 
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.4703
24
 
25
  ## Model description
26
 
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:----:|:---------------:|
55
+ | 1.7445 | 0.22 | 25 | 1.5450 |
56
+ | 1.557 | 0.43 | 50 | 1.4703 |
57
 
58
 
59
  ### Framework versions
60
 
61
+ - PEFT 0.9.0
62
+ - Transformers 4.38.2
63
  - Pytorch 2.1.0+cu121
64
+ - Datasets 2.18.0
65
  - Tokenizers 0.15.2
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -19,9 +19,10 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "v_proj",
23
- "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
 
26
  "use_rslora": false
27
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "q_proj",
23
+ "v_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
26
+ "use_dora": false,
27
  "use_rslora": false
28
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea3ca41896d5ecc7110c87eeabc151aafc4b078b63c57d00a7b8b585a62d1f82
3
- size 27280152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c48e0a972c1bedd6cafa5d1546efd158a9536992b2ae42f471fc4556842a7be2
3
+ size 27282328
runs/Mar02_00-53-12_cdf356d6afcd/events.out.tfevents.1709340802.cdf356d6afcd.2237.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4424ee8774af3e39ab6139b5dc07d5e8740a9b13e29217143a2a0165655b7f70
3
+ size 6966
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d475bae46eee9ea64e29ebe797b95ba15738cb726e56099e9fd09fccfc18f598
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff135a3b01ab8b460b20c8f554b3e8cd1dd02d18ea438edb3ef3224d89de63d2
3
  size 4920