hailsbop commited on
Commit
7d73b02
·
verified ·
1 Parent(s): 814cf5d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -1
README.md CHANGED
@@ -1,3 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  # Mistral + LoRA Fine-Tuning
2
 
3
  Lightweight fine-tuning setup for adapting a Mistral-architecture model using LoRA.
@@ -50,4 +106,4 @@ print(tok.decode(out[0]))
50
  - `tokenizer_config.json` / `special_tokens_map.json` – tokenizer metadata
51
 
52
  ## License
53
- MIT
 
1
+ ---
2
+ model_name: mistral-base
3
+ tokenizer_name: mistral-base
4
+ lora:
5
+ r: 32
6
+ alpha: 64
7
+ dropout: 0.05
8
+ target_modules:
9
+ - q_proj
10
+ - k_proj
11
+ - v_proj
12
+ - o_proj
13
+ - gate_proj
14
+ - up_proj
15
+ - down_proj
16
+ training:
17
+ output_dir: ./lora-out
18
+ batch_size: 4
19
+ micro_batch_size: 1
20
+ gradient_accumulation_steps: 4
21
+ learning_rate: 0.0002
22
+ weight_decay: 0
23
+ warmup_ratio: 0.03
24
+ num_epochs: 3
25
+ max_grad_norm: 1
26
+ fp16: true
27
+ dataset:
28
+ path: ./data
29
+ field: text
30
+ shuffle: true
31
+ num_workers: 4
32
+ generation:
33
+ bos_token_id: 1
34
+ eos_token_id: 2
35
+ max_new_tokens: 256
36
+ temperature: 0.7
37
+ top_p: 0.9
38
+ logging:
39
+ steps: 10
40
+ save_steps: 200
41
+ license: apache-2.0
42
+ language:
43
+ - en
44
+ base_model:
45
+ - mistralai/Mistral-7B-Instruct-v0.2
46
+ pipeline_tag: text-generation
47
+ tags:
48
+ - merge
49
+ - lora
50
+ ---
51
+
52
+
53
+
54
+
55
+
56
+
57
  # Mistral + LoRA Fine-Tuning
58
 
59
  Lightweight fine-tuning setup for adapting a Mistral-architecture model using LoRA.
 
106
  - `tokenizer_config.json` / `special_tokens_map.json` – tokenizer metadata
107
 
108
  ## License
109
+ MIT