cagataydev commited on
Commit
aafea7c
·
verified ·
1 Parent(s): 5504640

Upload folder using huggingface_hub

Browse files
0000100_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:733d34934908187d60060aeb0a8c13a31f9a0c265c4519a3596d912cd5afe86c
3
+ size 19947012
0000200_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5acb0abdd589aec63732c9affe03f610ef9f9212a6cbf52d7f40dbd2d80bd88
3
+ size 19947012
adapter_config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_path": "/Users/cagatay/strands-mlx/adapter_105_conservative",
3
+ "batch_size": 1,
4
+ "config": null,
5
+ "data": "/Users/cagatay/strands-mlx/strands_training",
6
+ "fine_tune_type": "lora",
7
+ "grad_accumulation_steps": 1,
8
+ "grad_checkpoint": true,
9
+ "iters": 200,
10
+ "learning_rate": 1e-05,
11
+ "lora_parameters": {
12
+ "rank": 8,
13
+ "scale": 16.0,
14
+ "dropout": 0.0
15
+ },
16
+ "lr_schedule": null,
17
+ "mask_prompt": false,
18
+ "max_seq_length": 2048,
19
+ "model": "mlx-community/Qwen3-1.7B-4bit",
20
+ "num_layers": 16,
21
+ "optimizer": "adam",
22
+ "optimizer_config": {
23
+ "adam": {},
24
+ "adamw": {},
25
+ "muon": {},
26
+ "sgd": {},
27
+ "adafactor": {}
28
+ },
29
+ "project_name": null,
30
+ "report_to": null,
31
+ "resume_adapter_file": null,
32
+ "save_every": 100,
33
+ "seed": 0,
34
+ "steps_per_eval": 50,
35
+ "steps_per_report": 20,
36
+ "steps_per_save": 100,
37
+ "test": false,
38
+ "test_batches": 500,
39
+ "train": true,
40
+ "val_batches": 25
41
+ }
adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5acb0abdd589aec63732c9affe03f610ef9f9212a6cbf52d7f40dbd2d80bd88
3
+ size 19947012