IoakeimE commited on
Commit
4d54a0e
·
verified ·
1 Parent(s): b241f6e

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -28,17 +28,17 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ioakeime-aristotle-university-of-thessaloniki/dpo_smiplification/runs/ztfwit0b)
32
 
33
 
34
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
35
 
36
  ### Framework versions
37
 
38
- - TRL: 0.23.0
39
- - Transformers: 4.56.2
40
- - Pytorch: 2.8.0
41
- - Datasets: 3.6.0
42
  - Tokenizers: 0.22.1
43
 
44
  ## Citations
 
28
 
29
  ## Training procedure
30
 
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/ioakeime-aristotle-university-of-thessaloniki/dpo_smiplification/runs/tyu12m4x)
32
 
33
 
34
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
35
 
36
  ### Framework versions
37
 
38
+ - TRL: 0.24.0
39
+ - Transformers: 4.57.3
40
+ - Pytorch: 2.9.0
41
+ - Datasets: 4.3.0
42
  - Tokenizers: 0.22.1
43
 
44
  ## Citations
adapter_config.json CHANGED
@@ -1,9 +1,16 @@
1
  {
 
2
  "alpha_pattern": {},
3
- "auto_mapping": null,
 
 
 
 
 
4
  "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
5
  "bias": "none",
6
  "corda_config": null,
 
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
@@ -20,18 +27,19 @@
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
 
23
  "qalora_group_size": 16,
24
  "r": 16,
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "v_proj",
29
- "q_proj",
30
- "up_proj",
31
- "k_proj",
32
  "gate_proj",
 
33
  "down_proj",
34
- "o_proj"
 
 
 
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
 
1
  {
2
+ "alora_invocation_tokens": null,
3
  "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "MistralForCausalLM",
7
+ "parent_library": "transformers.models.mistral.modeling_mistral",
8
+ "unsloth_fixed": true
9
+ },
10
  "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
11
  "bias": "none",
12
  "corda_config": null,
13
+ "ensure_weight_tying": false,
14
  "eva_config": null,
15
  "exclude_modules": null,
16
  "fan_in_fan_out": false,
 
27
  "megatron_core": "megatron.core",
28
  "modules_to_save": null,
29
  "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
  "qalora_group_size": 16,
32
  "r": 16,
33
  "rank_pattern": {},
34
  "revision": null,
35
  "target_modules": [
 
 
 
 
36
  "gate_proj",
37
+ "q_proj",
38
  "down_proj",
39
+ "v_proj",
40
+ "o_proj",
41
+ "up_proj",
42
+ "k_proj"
43
  ],
44
  "target_parameters": null,
45
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c260b48c0a7d488b06d4d0666290048d457d733f659055895a3863763bfae3e
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74065212c9d3bf7b824618e60b16b77e0656ae9a78d5f13554bd30f5edc9d200
3
  size 167832240
tokenizer.json CHANGED
@@ -6964,6 +6964,12 @@
6964
  "id": "A",
6965
  "type_id": 0
6966
  }
 
 
 
 
 
 
6967
  }
6968
  ],
6969
  "pair": [
@@ -6979,6 +6985,12 @@
6979
  "type_id": 0
6980
  }
6981
  },
 
 
 
 
 
 
6982
  {
6983
  "SpecialToken": {
6984
  "id": "<s>",
@@ -6990,9 +7002,24 @@
6990
  "id": "B",
6991
  "type_id": 1
6992
  }
 
 
 
 
 
 
6993
  }
6994
  ],
6995
  "special_tokens": {
 
 
 
 
 
 
 
 
 
6996
  "<s>": {
6997
  "id": "<s>",
6998
  "ids": [
 
6964
  "id": "A",
6965
  "type_id": 0
6966
  }
6967
+ },
6968
+ {
6969
+ "SpecialToken": {
6970
+ "id": "</s>",
6971
+ "type_id": 0
6972
+ }
6973
  }
6974
  ],
6975
  "pair": [
 
6985
  "type_id": 0
6986
  }
6987
  },
6988
+ {
6989
+ "SpecialToken": {
6990
+ "id": "</s>",
6991
+ "type_id": 0
6992
+ }
6993
+ },
6994
  {
6995
  "SpecialToken": {
6996
  "id": "<s>",
 
7002
  "id": "B",
7003
  "type_id": 1
7004
  }
7005
+ },
7006
+ {
7007
+ "SpecialToken": {
7008
+ "id": "</s>",
7009
+ "type_id": 1
7010
+ }
7011
  }
7012
  ],
7013
  "special_tokens": {
7014
+ "</s>": {
7015
+ "id": "</s>",
7016
+ "ids": [
7017
+ 2
7018
+ ],
7019
+ "tokens": [
7020
+ "</s>"
7021
+ ]
7022
+ },
7023
  "<s>": {
7024
  "id": "<s>",
7025
  "ids": [
tokenizer_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "add_bos_token": true,
3
- "add_eos_token": false,
4
  "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
 
1
  {
2
  "add_bos_token": true,
3
+ "add_eos_token": true,
4
  "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d11c0521246a68bdb6a8335ed968d814e898bd344f90101460a310eed56b3895
3
- size 6865
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3431d0c4d717553177a1a69fd77ec6e3d1ece5bbb135a5a8575cb75f7b6cda3
3
+ size 6929