Brandoko commited on
Commit
1f008c4
·
1 Parent(s): 2ed40d4

Training in progress, epoch 1

Browse files
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - quant_method: bitsandbytes
9
+ - load_in_8bit: True
10
+ - load_in_4bit: False
11
+ - llm_int8_threshold: 6.0
12
+ - llm_int8_skip_modules: None
13
+ - llm_int8_enable_fp32_cpu_offload: False
14
+ - llm_int8_has_fp16_weight: False
15
+ - bnb_4bit_quant_type: fp4
16
+ - bnb_4bit_use_double_quant: False
17
+ - bnb_4bit_compute_dtype: float32
18
+ ### Framework versions
19
+
20
+
21
+ - PEFT 0.6.0.dev0
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "codellama/CodeLlama-7b-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": null,
7
  "inference_mode": true,
@@ -16,13 +16,13 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "down_proj",
20
- "v_proj",
21
  "gate_proj",
22
- "q_proj",
23
- "k_proj",
 
24
  "o_proj",
25
- "up_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "codellama/CodeLlama-7b-Instruct-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": null,
7
  "inference_mode": true,
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
 
19
  "gate_proj",
20
+ "v_proj",
21
+ "down_proj",
22
+ "up_proj",
23
  "o_proj",
24
+ "q_proj",
25
+ "k_proj"
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:252e5bb28e7b71f5caba598d8296e60ba463497fb5ca5f5625ebd1a32123168b
3
  size 319977229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c980182b1e1588556688b6a652a130642d20cf6bc0718d63b9b9ac36370b7a0b
3
  size 319977229
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "codellama/CodeLlama-7b-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "codellama/CodeLlama-7b-Instruct-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
tokenizer_config.json CHANGED
@@ -79,7 +79,7 @@
79
  "suffix_first": false,
80
  "suffix_token": "▁<SUF>",
81
  "tokenizer_class": "CodeLlamaTokenizer",
82
- "tokenizer_file": "/root/.cache/huggingface/hub/models--codellama--CodeLlama-7b-hf/snapshots/3773f63b4511b9e47a9a7ffc765eed7eb0169486/tokenizer.json",
83
  "trust_remote_code": false,
84
  "unk_token": "<unk>",
85
  "use_default_system_prompt": false,
 
79
  "suffix_first": false,
80
  "suffix_token": "▁<SUF>",
81
  "tokenizer_class": "CodeLlamaTokenizer",
82
+ "tokenizer_file": "/root/.cache/huggingface/hub/models--codellama--CodeLlama-7b-Instruct-hf/snapshots/6114dd1e16f69e0765ccbd7a64d33d04b265fbd2/tokenizer.json",
83
  "trust_remote_code": false,
84
  "unk_token": "<unk>",
85
  "use_default_system_prompt": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e829299a78f4b1dd4d8794df3b53ccce91314f8df1687084a34163608cc66743
3
  size 4475
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a2680f6024f84d722c28decac8dad6e0feddc6f82978a329380e710c8027a4
3
  size 4475