TimberGu commited on
Commit
33d8063
·
verified ·
1 Parent(s): 6a99be9

Upload training_config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. training_config.json +38 -0
training_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model": "meta-llama/Meta-Llama-3.1-8B-Instruct",
3
+ "dataset": "Josephgflowers/Finance-Instruct-500k",
4
+ "dataset_config": "default",
5
+ "training_config": {
6
+ "model_name": "meta-llama/Meta-Llama-3.1-8B-Instruct",
7
+ "dataset_name": "Josephgflowers/Finance-Instruct-500k",
8
+ "dataset_config": "default",
9
+ "output_dir": "/content/drive/MyDrive/financial_llama_models/checkpoints",
10
+ "save_dir": "/content/drive/MyDrive/financial_llama_models/final_model",
11
+ "learning_rate": 0.0002,
12
+ "num_epochs": 2,
13
+ "warmup_ratio": 0.05,
14
+ "quantization": "8bit",
15
+ "attn_impl": "flash",
16
+ "align_save_with_eval": true,
17
+ "dataloader_num_workers": 4,
18
+ "optim": "paged_adamw_8bit",
19
+ "train_batch_size": 4,
20
+ "eval_batch_size": 8,
21
+ "gradient_accumulation_steps": 4,
22
+ "max_length": 512,
23
+ "eval_steps": 500,
24
+ "logging_steps": 100,
25
+ "max_train_samples": 25000,
26
+ "max_val_samples": 2500,
27
+ "gradient_checkpointing": false,
28
+ "lora_r": 16,
29
+ "lora_alpha": 32,
30
+ "lora_dropout": 0.1
31
+ },
32
+ "lora_config": {
33
+ "r": 16,
34
+ "alpha": 32,
35
+ "dropout": 0.1
36
+ },
37
+ "training_date": "2025-08-08T04:45:28.020772"
38
+ }