IoakeimE commited on
Commit
850deeb
·
verified ·
1 Parent(s): 244bf88

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -4,9 +4,9 @@ library_name: transformers
4
  model_name: kto_simplification_balanced
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
- - unsloth
9
  - kto
 
 
10
  licence: license
11
  ---
12
 
@@ -28,18 +28,18 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ioakeime-aristotle-university-of-thessaloniki/kto_smiplification_balanced/runs/q9oqzg79)
32
 
33
 
34
  This model was trained with KTO, a method introduced in [KTO: Model Alignment as Prospect Theoretic Optimization](https://huggingface.co/papers/2402.01306).
35
 
36
  ### Framework versions
37
 
38
- - TRL: 0.19.0
39
- - Transformers: 4.53.0
40
- - Pytorch: 2.7.0+cu128
41
- - Datasets: 3.6.0
42
- - Tokenizers: 0.21.2
43
 
44
  ## Citations
45
 
 
4
  model_name: kto_simplification_balanced
5
  tags:
6
  - generated_from_trainer
 
 
7
  - kto
8
+ - unsloth
9
+ - trl
10
  licence: license
11
  ---
12
 
 
28
 
29
  ## Training procedure
30
 
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ioakeime-aristotle-university-of-thessaloniki/kto_smiplification_balanced/runs/bpfxv7y5)
32
 
33
 
34
  This model was trained with KTO, a method introduced in [KTO: Model Alignment as Prospect Theoretic Optimization](https://huggingface.co/papers/2402.01306).
35
 
36
  ### Framework versions
37
 
38
+ - TRL: 0.24.0
39
+ - Transformers: 4.57.3
40
+ - Pytorch: 2.9.0
41
+ - Datasets: 4.3.0
42
+ - Tokenizers: 0.22.1
43
 
44
  ## Citations
45
 
adapter_config.json CHANGED
@@ -1,9 +1,16 @@
1
  {
 
2
  "alpha_pattern": {},
3
- "auto_mapping": null,
 
 
 
 
 
4
  "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
5
  "bias": "none",
6
  "corda_config": null,
 
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
@@ -15,22 +22,23 @@
15
  "loftq_config": {},
16
  "lora_alpha": 16,
17
  "lora_bias": false,
18
- "lora_dropout": 0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
 
23
  "qalora_group_size": 16,
24
  "r": 16,
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
28
  "down_proj",
29
  "k_proj",
30
- "o_proj",
31
  "q_proj",
32
- "v_proj",
33
  "up_proj",
 
34
  "gate_proj"
35
  ],
36
  "target_parameters": null,
 
1
  {
2
+ "alora_invocation_tokens": null,
3
  "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "MistralForCausalLM",
7
+ "parent_library": "transformers.models.mistral.modeling_mistral",
8
+ "unsloth_fixed": true
9
+ },
10
  "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
11
  "bias": "none",
12
  "corda_config": null,
13
+ "ensure_weight_tying": false,
14
  "eva_config": null,
15
  "exclude_modules": null,
16
  "fan_in_fan_out": false,
 
22
  "loftq_config": {},
23
  "lora_alpha": 16,
24
  "lora_bias": false,
25
+ "lora_dropout": 0.0,
26
  "megatron_config": null,
27
  "megatron_core": "megatron.core",
28
  "modules_to_save": null,
29
  "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
  "qalora_group_size": 16,
32
  "r": 16,
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
36
+ "v_proj",
37
  "down_proj",
38
  "k_proj",
 
39
  "q_proj",
 
40
  "up_proj",
41
+ "o_proj",
42
  "gate_proj"
43
  ],
44
  "target_parameters": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:234c57c0171045ee31f77b8494fa6eadf223c485501ff38ee6d067effa6a0a09
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:141a5935c1dd57689e8ae413b2597fbe9fd66ea2c2ae31f9514ea84ecc074a31
3
  size 167832240
tokenizer.json CHANGED
@@ -6964,6 +6964,12 @@
6964
  "id": "A",
6965
  "type_id": 0
6966
  }
 
 
 
 
 
 
6967
  }
6968
  ],
6969
  "pair": [
@@ -6979,6 +6985,12 @@
6979
  "type_id": 0
6980
  }
6981
  },
 
 
 
 
 
 
6982
  {
6983
  "SpecialToken": {
6984
  "id": "<s>",
@@ -6990,9 +7002,24 @@
6990
  "id": "B",
6991
  "type_id": 1
6992
  }
 
 
 
 
 
 
6993
  }
6994
  ],
6995
  "special_tokens": {
 
 
 
 
 
 
 
 
 
6996
  "<s>": {
6997
  "id": "<s>",
6998
  "ids": [
 
6964
  "id": "A",
6965
  "type_id": 0
6966
  }
6967
+ },
6968
+ {
6969
+ "SpecialToken": {
6970
+ "id": "</s>",
6971
+ "type_id": 0
6972
+ }
6973
  }
6974
  ],
6975
  "pair": [
 
6985
  "type_id": 0
6986
  }
6987
  },
6988
+ {
6989
+ "SpecialToken": {
6990
+ "id": "</s>",
6991
+ "type_id": 0
6992
+ }
6993
+ },
6994
  {
6995
  "SpecialToken": {
6996
  "id": "<s>",
 
7002
  "id": "B",
7003
  "type_id": 1
7004
  }
7005
+ },
7006
+ {
7007
+ "SpecialToken": {
7008
+ "id": "</s>",
7009
+ "type_id": 1
7010
+ }
7011
  }
7012
  ],
7013
  "special_tokens": {
7014
+ "</s>": {
7015
+ "id": "</s>",
7016
+ "ids": [
7017
+ 2
7018
+ ],
7019
+ "tokens": [
7020
+ "</s>"
7021
+ ]
7022
+ },
7023
  "<s>": {
7024
  "id": "<s>",
7025
  "ids": [
tokenizer_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "add_bos_token": true,
3
- "add_eos_token": false,
4
  "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
 
1
  {
2
  "add_bos_token": true,
3
+ "add_eos_token": true,
4
  "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91aa7e4b24a73bf63f76da8b8a36dc9a5ef4061cb2771c5a599bf78a8a1546f2
3
- size 6417
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74a34c14f6c99eb65f536192e3711f3533aca183ff32856608c59f7a6bc3188e
3
+ size 6481