kosmylo1992 commited on
Commit
c0bb125
·
verified ·
1 Parent(s): 3833207

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -3
README.md CHANGED
@@ -1,3 +1,54 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Command-R 35B — SFT (Supervised Fine-Tuning)
2
+
3
+ **Model type:** Causal Language Model
4
+ **Base model:** [CohereLabs/c4ai-command-r-v01](https://huggingface.co/CohereLabs/c4ai-command-r-v01)
5
+ **License:** Apache 2.0
6
+ **Framework:** [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl)
7
+
8
+ ---
9
+
10
+ ## Overview
11
+
12
+ `commandr-SFT` is a **supervised fine-tuned** variant of Cohere’s Command-R 35B model.
13
+ Fine-tuning was performed on a high-quality instruction-following dataset using LoRA adapters, enabling improved conversational reasoning and question answering.
14
+
15
+ ---
16
+
17
+ ## Training Setup
18
+
19
+ **Objective:** Supervised fine-tuning (instruction following)
20
+ **Adapter type:** LoRA
21
+ **Precision:** bfloat16
22
+ **Hardware:** 8 nodes × 2 × NVIDIA A100 64GB GPUs
23
+ **Training duration:** ~6 hours
24
+ **Framework:** DeepSpeed ZeRO-1, Axolotl, PyTorch 2.5.1+cu121
25
+
26
+ ---
27
+
28
+ ## Dataset
29
+
30
+ **Name:** `axolotl_deduplicated_synthetic_qa.jsonl`
31
+ **Type:** Instruction-following synthetic QA dataset
32
+ **Split:** 70% train / 30% validation
33
+
34
+ Each sample follows a QA/chat format used in the `alpaca_chat.load_qa` schema.
35
+
36
+ ---
37
+
38
+ ## Hyperparameters
39
+
40
+ | Parameter | Value |
41
+ |------------|-------|
42
+ | Sequence length | 2048 |
43
+ | Micro batch size | 2 |
44
+ | Gradient accumulation | 2 |
45
+ | Learning rate | 2e-4 |
46
+ | LR scheduler | cosine |
47
+ | Optimizer | AdamW (8-bit) |
48
+ | LoRA rank (r) | 16 |
49
+ | LoRA alpha | 32 |
50
+ | LoRA dropout | 0.05 |
51
+ | Target modules | q_proj, v_proj, k_proj, o_proj |
52
+ | Epochs | 1 |
53
+ | Warmup steps | 10 |
54
+ | Weight decay | 0.0 |