voidful commited on
Commit
129939e
·
verified ·
1 Parent(s): c656cd3

Model save

Browse files
README.md CHANGED
@@ -2,7 +2,10 @@
2
  library_name: peft
3
  base_model: voidful/llm-codec
4
  tags:
5
- - generated_from_trainer
 
 
 
6
  model-index:
7
  - name: llm-codec-librispeech
8
  results: []
@@ -15,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [voidful/llm-codec](https://huggingface.co/voidful/llm-codec) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 8.4240
19
 
20
  ## Model description
21
 
@@ -49,15 +52,15 @@ The following hyperparameters were used during training:
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:------:|:---------------:|
52
- | 8.7083 | 1.0 | 35156 | 8.7066 |
53
- | 8.3461 | 2.0 | 70312 | 8.4832 |
54
- | 8.3644 | 3.0 | 105468 | 8.4240 |
55
 
56
 
57
  ### Framework versions
58
 
59
- - PEFT 0.15.0
60
  - Transformers 4.57.3
61
- - Pytorch 2.8.0+cu126
62
- - Datasets 4.0.0
63
- - Tokenizers 0.22.0
 
2
  library_name: peft
3
  base_model: voidful/llm-codec
4
  tags:
5
+ - base_model:adapter:voidful/llm-codec
6
+ - lora
7
+ - transformers
8
+ pipeline_tag: text-generation
9
  model-index:
10
  - name: llm-codec-librispeech
11
  results: []
 
18
 
19
  This model is a fine-tuned version of [voidful/llm-codec](https://huggingface.co/voidful/llm-codec) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 8.4274
22
 
23
  ## Model description
24
 
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:------:|:---------------:|
55
+ | 8.7753 | 1.0 | 35156 | 8.7088 |
56
+ | 8.3352 | 2.0 | 70312 | 8.4871 |
57
+ | 8.376 | 3.0 | 105468 | 8.4274 |
58
 
59
 
60
  ### Framework versions
61
 
62
+ - PEFT 0.18.0
63
  - Transformers 4.57.3
64
+ - Pytorch 2.9.0+cu129
65
+ - Datasets 4.4.1
66
+ - Tokenizers 0.22.1
final/README.md CHANGED
@@ -1,6 +1,11 @@
1
  ---
2
  base_model: voidful/llm-codec
3
  library_name: peft
 
 
 
 
 
4
  ---
5
 
6
  # Model Card for Model ID
@@ -199,4 +204,4 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
199
  [More Information Needed]
200
  ### Framework versions
201
 
202
- - PEFT 0.15.0
 
1
  ---
2
  base_model: voidful/llm-codec
3
  library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:voidful/llm-codec
7
+ - lora
8
+ - transformers
9
  ---
10
 
11
  # Model Card for Model ID
 
204
  [More Information Needed]
205
  ### Framework versions
206
 
207
+ - PEFT 0.18.0
final/adapter_config.json CHANGED
@@ -1,9 +1,12 @@
1
  {
 
2
  "alpha_pattern": {},
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "voidful/llm-codec",
5
  "bias": "none",
6
  "corda_config": null,
 
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
@@ -20,20 +23,24 @@
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
 
 
23
  "r": 64,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "gate_proj",
28
- "k_proj",
29
  "q_proj",
30
- "up_proj",
31
  "o_proj",
 
 
32
  "v_proj",
33
- "down_proj"
 
34
  ],
 
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
37
  "use_dora": false,
 
38
  "use_rslora": false
39
  }
 
1
  {
2
+ "alora_invocation_tokens": null,
3
  "alpha_pattern": {},
4
+ "arrow_config": null,
5
  "auto_mapping": null,
6
  "base_model_name_or_path": "voidful/llm-codec",
7
  "bias": "none",
8
  "corda_config": null,
9
+ "ensure_weight_tying": false,
10
  "eva_config": null,
11
  "exclude_modules": null,
12
  "fan_in_fan_out": false,
 
23
  "megatron_core": "megatron.core",
24
  "modules_to_save": null,
25
  "peft_type": "LORA",
26
+ "peft_version": "0.18.0",
27
+ "qalora_group_size": 16,
28
  "r": 64,
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
 
32
  "q_proj",
 
33
  "o_proj",
34
+ "k_proj",
35
+ "down_proj",
36
  "v_proj",
37
+ "up_proj",
38
+ "gate_proj"
39
  ],
40
+ "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
42
  "trainable_token_indices": null,
43
  "use_dora": false,
44
+ "use_qalora": false,
45
  "use_rslora": false
46
  }
final/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ec670faef160266af110492dd83512f76fd0c9d4d739ebbee32a75847acc420
3
  size 528550256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a16703691beb3920653d04fcd565151a9415fcd15b74c1c014b7c25cbad0bbf
3
  size 528550256
final/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09ef793003aa3f31e0fa9d57c9998371626de0241ae516edb6c9795266860010
3
  size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:086a4fbc7538a577fb092ba511f7e1a4d6b3867e138da19ed05f63656337f331
3
  size 5841