File size: 765 Bytes
9a603b6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
{
  "model": "logos23-gemma2_2b",
  "family": "gemma2_2b",
  "family_name": "Gemma 2 2B",
  "base_model": "google/gemma-2-2b",
  "base_model_quantized": "google/gemma-2-2b",
  "method": "LoRA (bf16)",
  "framework": "unsloth",
  "lora_rank": 64,
  "lora_alpha": 64,
  "lora_target_modules": [
    "q_proj",
    "k_proj",
    "v_proj",
    "o_proj",
    "gate_proj",
    "up_proj",
    "down_proj"
  ],
  "epochs": 3,
  "effective_batch_size": 16,
  "load_in_4bit": false,
  "learning_rate": 0.0002,
  "lr_scheduler": "cosine",
  "max_seq_length": 2048,
  "dataset": "logos22_nothink.jsonl",
  "dataset_size": 895,
  "train_on_responses_only": true,
  "think_blocks": "stripped (no-think variant)",
  "final_loss": 1.2898975720717794,
  "runtime_seconds": 221.1209
}