EYEDOL commited on
Commit
6cec3ad
·
verified ·
1 Parent(s): 9391472

EYEDOL/llama-3.1-8b-telecom-lora

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: llama-3.1-8b-telecom-lora
5
  tags:
6
  - generated_from_trainer
7
- - sft
8
  - trl
 
9
  licence: license
10
  ---
11
 
@@ -34,11 +34,11 @@ This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.25.0
38
- - Transformers: 4.57.1
39
- - Pytorch: 2.8.0+cu126
40
  - Datasets: 4.0.0
41
- - Tokenizers: 0.22.1
42
 
43
  ## Citations
44
 
 
4
  model_name: llama-3.1-8b-telecom-lora
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
 
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.17.0
38
+ - Transformers: 4.49.0
39
+ - Pytorch: 2.6.0
40
  - Datasets: 4.0.0
41
+ - Tokenizers: 0.21.4
42
 
43
  ## Citations
44
 
adapter_config.json CHANGED
@@ -3,9 +3,6 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "meta-llama/Llama-3.2-1B",
5
  "bias": "none",
6
- "corda_config": null,
7
- "eva_config": null,
8
- "exclude_modules": null,
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
11
  "init_lora_weights": true,
@@ -14,7 +11,6 @@
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
  "lora_alpha": 128,
17
- "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
@@ -23,23 +19,19 @@
23
  "lm_head"
24
  ],
25
  "peft_type": "LORA",
26
- "qalora_group_size": 16,
27
  "r": 64,
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
- "up_proj",
32
- "o_proj",
33
- "v_proj",
34
  "k_proj",
35
- "gate_proj",
 
36
  "down_proj",
37
- "q_proj"
 
 
38
  ],
39
- "target_parameters": null,
40
  "task_type": "CAUSAL_LM",
41
- "trainable_token_indices": null,
42
  "use_dora": false,
43
- "use_qalora": false,
44
  "use_rslora": false
45
  }
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "meta-llama/Llama-3.2-1B",
5
  "bias": "none",
 
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 128,
 
14
  "lora_dropout": 0.05,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
 
19
  "lm_head"
20
  ],
21
  "peft_type": "LORA",
 
22
  "r": 64,
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
 
26
  "k_proj",
27
+ "up_proj",
28
+ "q_proj",
29
  "down_proj",
30
+ "v_proj",
31
+ "o_proj",
32
+ "gate_proj"
33
  ],
 
34
  "task_type": "CAUSAL_LM",
 
35
  "use_dora": false,
 
36
  "use_rslora": false
37
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:366454c1da120daf28dfa6107f0182b864df6b41424596cd868a2c27ad386b47
3
  size 1231058392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ae95f3a38946351dc0994ba6ee1faf53fc9b3109c20e909df3737090706cf54
3
  size 1231058392
special_tokens_map.json CHANGED
@@ -12,12 +12,5 @@
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<|end_of_text|>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
  }
23
  }
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
 
 
 
 
 
 
 
15
  }
16
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76c0277e503a081af45eb9f198fbcd3833dbb13e0401312d2e565179a9cde5da
3
- size 17210191
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
tokenizer_config.json CHANGED
@@ -2058,6 +2058,5 @@
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
2061
- "pad_token": "<|end_of_text|>",
2062
- "tokenizer_class": "PreTrainedTokenizerFast"
2063
  }
 
2058
  "attention_mask"
2059
  ],
2060
  "model_max_length": 131072,
2061
+ "tokenizer_class": "PreTrainedTokenizer"
 
2062
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75d65cbc9d40515c8f049ec74b609fe189ba370093cb1a6fe48f9da0ccb8acad
3
- size 6161
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22d29aa07aa6c7dcca7c510a5b5427245556318d309934d2bd33089c29aeb0ed
3
+ size 5688