bitshrine commited on
Commit
6d960bd
·
verified ·
1 Parent(s): 008dcfc

gemma-2-2B-hermes-function-calling-adapter

Browse files
README.md CHANGED
@@ -1,7 +1,6 @@
1
  ---
2
- base_model: unsloth/gemma-2-2b-it-bnb-4bit
3
  library_name: peft
4
- license: gemma
5
  tags:
6
  - trl
7
  - sft
@@ -17,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # gemma-2-2B-hermes-function-calling-adapter
19
 
20
- This model is a fine-tuned version of [unsloth/gemma-2-2b-it-bnb-4bit](https://huggingface.co/unsloth/gemma-2-2b-it-bnb-4bit) on an unknown dataset.
21
 
22
  ## Model description
23
 
@@ -45,16 +44,12 @@ The following hyperparameters were used during training:
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 5
48
- - training_steps: 180
49
  - mixed_precision_training: Native AMP
50
 
51
- ### Training results
52
-
53
-
54
-
55
  ### Framework versions
56
 
57
- - PEFT 0.12.0
58
  - Transformers 4.44.2
59
  - Pytorch 2.4.1+cu121
60
  - Datasets 3.0.0
 
1
  ---
2
+ base_model: bitshrine/gemma-2-2B-it-hermes-json-mode
3
  library_name: peft
 
4
  tags:
5
  - trl
6
  - sft
 
16
 
17
  # gemma-2-2B-hermes-function-calling-adapter
18
 
19
+ This model is a fine-tuned version of [bitshrine/gemma-2-2B-it-hermes-json-mode](https://huggingface.co/bitshrine/gemma-2-2B-it-hermes-json-mode) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 5
47
+ - training_steps: 1262
48
  - mixed_precision_training: Native AMP
49
 
 
 
 
 
50
  ### Framework versions
51
 
52
+ - PEFT 0.13.0
53
  - Transformers 4.44.2
54
  - Pytorch 2.4.1+cu121
55
  - Datasets 3.0.0
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "unsloth/gemma-2-2b-it-bnb-4bit",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -10,25 +10,25 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 16,
14
  "lora_dropout": 0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
24
- "k_proj",
25
- "up_proj",
26
  "q_proj",
27
- "v_proj",
28
  "o_proj",
29
- "gate_proj"
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
33
- "use_rslora": false
34
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "bitshrine/gemma-2-2B-it-hermes-json-mode",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 32,
14
  "lora_dropout": 0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 32,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "q_proj",
 
24
  "o_proj",
25
+ "v_proj",
26
+ "up_proj",
27
+ "down_proj",
28
+ "gate_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
33
+ "use_rslora": true
34
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0645cbf3b72d27961ca56f1f50a32f8200bff47811d71479462b782ef78b91a5
3
- size 2442466784
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b7c3d46e3836f421b55996ce3b41f737cfbd441019c6b68b113c76428d99d43
3
+ size 2525534008
runs/Sep26_06-47-30_d62f90f8f9d3/events.out.tfevents.1727333276.d62f90f8f9d3.422.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7d94a36551c13286f558617477ca54ad6609d98f879139f0dcd48d75152641a
3
+ size 111229
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe6497e601cf2fb129e59a2d7925e04e03dddce9487aee92d124a426fcc8f6d0
3
- size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8c78a30101788d8ccdc3c26c408302563baeec811ff389eb4fcb635898afc66
3
+ size 5560