Planuze commited on
Commit
dade7ed
·
verified ·
1 Parent(s): 69ece4d

Upload LoRA adapters

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ adapters/wandb/run-20260203_224646-so9y0g75/run-so9y0g75.wandb filter=lfs diff=lfs merge=lfs -text
adapters/adapter_config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_path": "adapters_mlx",
3
+ "batch_size": 2,
4
+ "config": "adapters_mlx/train_config.yaml",
5
+ "data": "data/processed",
6
+ "fine_tune_type": "lora",
7
+ "grad_accumulation_steps": 1,
8
+ "grad_checkpoint": true,
9
+ "iters": 20,
10
+ "learning_rate": 1e-05,
11
+ "lora": {
12
+ "alpha": 16,
13
+ "dropout": 0.0,
14
+ "r": 16
15
+ },
16
+ "lora_parameters": {
17
+ "rank": 8,
18
+ "dropout": 0.0,
19
+ "scale": 20.0
20
+ },
21
+ "lr_schedule": null,
22
+ "mask_prompt": false,
23
+ "max_seq_length": 2048,
24
+ "model": "mlx-community/Qwen2.5-14B-Instruct-4bit",
25
+ "num_layers": 16,
26
+ "optimizer": "adam",
27
+ "optimizer_config": {
28
+ "adam": {},
29
+ "adamw": {},
30
+ "muon": {},
31
+ "sgd": {},
32
+ "adafactor": {}
33
+ },
34
+ "project_name": "planus-llm",
35
+ "report_to": "wandb",
36
+ "resume_adapter_file": null,
37
+ "save_every": 100,
38
+ "seed": 3407,
39
+ "steps_per_eval": 100,
40
+ "steps_per_report": 10,
41
+ "test": false,
42
+ "test_batches": 500,
43
+ "train": true,
44
+ "val_batches": 25
45
+ }
adapters/adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e78777f05e602bc21b4c87430df0e83c3304c734ca04dca85f9d94d36f2f08c0
3
+ size 45899454
adapters/train_config.yaml ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adapter_path: adapters_mlx
2
+ batch_size: 2
3
+ data: data/processed
4
+ grad_checkpoint: true
5
+ iters: 20
6
+ learning_rate: 1.0e-05
7
+ lora:
8
+ alpha: 16
9
+ dropout: 0.0
10
+ r: 16
11
+ max_seq_length: 2048
12
+ model: mlx-community/Qwen2.5-14B-Instruct-4bit
13
+ save_every: 100
14
+ seed: 3407
15
+ steps_per_eval: 100
16
+ train: true
17
+ val_batches: 25
adapters/wandb/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-02-03T22:46:46.649174-03:00","level":"INFO","msg":"stream: starting","core version":"0.24.1"}
2
+ {"time":"2026-02-03T22:46:46.978134-03:00","level":"INFO","msg":"stream: created new stream","id":"so9y0g75"}
3
+ {"time":"2026-02-03T22:46:46.978191-03:00","level":"INFO","msg":"handler: started","stream_id":"so9y0g75"}
4
+ {"time":"2026-02-03T22:46:46.978488-03:00","level":"INFO","msg":"stream: started","id":"so9y0g75"}
5
+ {"time":"2026-02-03T22:46:46.978504-03:00","level":"INFO","msg":"writer: started","stream_id":"so9y0g75"}
6
+ {"time":"2026-02-03T22:46:46.978504-03:00","level":"INFO","msg":"sender: started","stream_id":"so9y0g75"}
7
+ {"time":"2026-02-03T23:05:33.65201-03:00","level":"INFO","msg":"stream: closing","id":"so9y0g75"}
8
+ {"time":"2026-02-03T23:05:35.080416-03:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-02-03T23:05:35.538389-03:00","level":"INFO","msg":"handler: closed","stream_id":"so9y0g75"}
10
+ {"time":"2026-02-03T23:05:35.538562-03:00","level":"INFO","msg":"sender: closed","stream_id":"so9y0g75"}
11
+ {"time":"2026-02-03T23:05:35.538598-03:00","level":"INFO","msg":"stream: closed","id":"so9y0g75"}
adapters/wandb/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_setup.py:_flush():81] Current SDK version is 0.24.1
2
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_setup.py:_flush():81] Configure stats pid to 48222
3
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:setup_run_log_directory():717] Logging user logs to adapters_mlx/wandb/run-20260203_224646-so9y0g75/logs/debug.log
5
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to adapters_mlx/wandb/run-20260203_224646-so9y0g75/logs/debug-internal.log
6
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:init():844] calling init triggers
7
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': 'mlx-community/Qwen2.5-14B-Instruct-4bit', 'train': True, 'data': 'data/processed', 'fine_tune_type': 'lora', 'optimizer': 'adam', 'mask_prompt': False, 'num_layers': 16, 'batch_size': 2, 'iters': 20, 'val_batches': 25, 'learning_rate': 1e-05, 'steps_per_report': 10, 'steps_per_eval': 100, 'grad_accumulation_steps': 1, 'resume_adapter_file': None, 'adapter_path': 'adapters_mlx', 'save_every': 100, 'test': False, 'test_batches': 500, 'max_seq_length': 2048, 'config': 'adapters_mlx/train_config.yaml', 'grad_checkpoint': True, 'report_to': 'wandb', 'project_name': 'planus-llm', 'seed': 3407, 'lora': {'alpha': 16, 'dropout': 0.0, 'r': 16}, 'optimizer_config': {'adam': {}, 'adamw': {}, 'muon': {}, 'sgd': {}, 'adafactor': {}}, 'lr_schedule': None, 'lora_parameters': {'rank': 8, 'dropout': 0.0, 'scale': 20.0}, '_wandb': {}}
9
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:init():892] starting backend
10
+ 2026-02-03 22:46:46,636 INFO MainThread:48222 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-02-03 22:46:46,648 INFO MainThread:48222 [wandb_init.py:init():903] backend started and connected
12
+ 2026-02-03 22:46:46,650 INFO MainThread:48222 [wandb_init.py:init():973] updated telemetry
13
+ 2026-02-03 22:46:46,667 INFO MainThread:48222 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-02-03 22:46:47,484 INFO MainThread:48222 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_console_start():2529] atexit reg
16
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_redirect():2377] redirect: wrap_raw
17
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_redirect():2446] Wrapping output streams.
18
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_redirect():2469] Redirects installed.
19
+ 2026-02-03 22:46:47,540 INFO MainThread:48222 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-02-03 23:05:33,649 INFO wandb-AsyncioManager-main:48222 [service_client.py:_forward_responses():94] Reached EOF.
21
+ 2026-02-03 23:05:33,650 INFO wandb-AsyncioManager-main:48222 [mailbox.py:close():154] Closing mailbox, abandoning 1 handles.
adapters/wandb/run-20260203_224646-so9y0g75/files/config.yaml ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.24.1
4
+ e:
5
+ 9wwcluo2ck30v81k0w3kns8xd71az26r:
6
+ apple:
7
+ ecpuCores: 2
8
+ gpuCores: 32
9
+ memoryGb: 32
10
+ name: Apple M1 Max
11
+ pcpuCores: 8
12
+ ramTotalBytes: "34359738368"
13
+ swapTotalBytes: "5368709120"
14
+ args:
15
+ - --config
16
+ - adapters_mlx/train_config.yaml
17
+ - --report-to
18
+ - wandb
19
+ - --project-name
20
+ - planus-llm
21
+ cpu_count: 10
22
+ cpu_count_logical: 10
23
+ disk:
24
+ /:
25
+ total: "994662584320"
26
+ used: "867762769920"
27
+ email: planuze@gmail.com
28
+ executable: /Users/rezende/Desktop/Planuze/planuze-llm/.venv/bin/python
29
+ git:
30
+ commit: 2180d8de1a20f6becde1d28262c43677022a9fbb
31
+ remote: https://github.com/Sr-Igor/planuze-llm.git
32
+ host: Igors-MacBook-Pro.local
33
+ memory:
34
+ total: "34359738368"
35
+ os: macOS-15.6-arm64-arm-64bit
36
+ program: -m mlx_lm.lora
37
+ python: CPython 3.9.6
38
+ root: adapters_mlx
39
+ startedAt: "2026-02-04T01:46:46.387687Z"
40
+ writerId: 9wwcluo2ck30v81k0w3kns8xd71az26r
41
+ m: []
42
+ python_version: 3.9.6
43
+ t:
44
+ "1":
45
+ - 11
46
+ - 49
47
+ "2":
48
+ - 11
49
+ - 49
50
+ "3":
51
+ - 13
52
+ - 16
53
+ - 61
54
+ "4": 3.9.6
55
+ "5": 0.24.1
56
+ "6": 4.57.6
57
+ "12": 0.24.1
58
+ "13": darwin-arm64
59
+ adapter_path:
60
+ value: adapters_mlx
61
+ batch_size:
62
+ value: 2
63
+ config:
64
+ value: adapters_mlx/train_config.yaml
65
+ data:
66
+ value: data/processed
67
+ fine_tune_type:
68
+ value: lora
69
+ grad_accumulation_steps:
70
+ value: 1
71
+ grad_checkpoint:
72
+ value: true
73
+ iters:
74
+ value: 20
75
+ learning_rate:
76
+ value: 1e-05
77
+ lora:
78
+ value:
79
+ alpha: 16
80
+ dropout: 0
81
+ r: 16
82
+ lora_parameters:
83
+ value:
84
+ dropout: 0
85
+ rank: 8
86
+ scale: 20
87
+ lr_schedule:
88
+ value: null
89
+ mask_prompt:
90
+ value: false
91
+ max_seq_length:
92
+ value: 2048
93
+ model:
94
+ value: mlx-community/Qwen2.5-14B-Instruct-4bit
95
+ num_layers:
96
+ value: 16
97
+ optimizer:
98
+ value: adam
99
+ project_name:
100
+ value: planus-llm
101
+ report_to:
102
+ value: wandb
103
+ resume_adapter_file:
104
+ value: null
105
+ save_every:
106
+ value: 100
107
+ seed:
108
+ value: 3407
109
+ steps_per_eval:
110
+ value: 100
111
+ steps_per_report:
112
+ value: 10
113
+ test:
114
+ value: false
115
+ test_batches:
116
+ value: 500
117
+ train:
118
+ value: true
119
+ val_batches:
120
+ value: 25
adapters/wandb/run-20260203_224646-so9y0g75/files/output.log ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Loading pretrained model
2
+ Fetching 10 files: 100%|██████████| 10/10 [00:00<00:00, 28493.91it/s]
3
+ Loading datasets
4
+ Training
5
+ Trainable parameters: 0.078% (11.469M/14770.034M)
6
+ Starting training..., iters: 20
7
+ Calculating loss...: 100%|██████████| 22/22 [04:57<00:00, 13.51s/it]
8
+ Iter 1: Val loss 1.780, Val took 297.162s
9
+ Iter 10: Train loss 1.393, Learning Rate 1.000e-05, It/sec 0.041, Tokens/sec 133.910, Trained Tokens 32580, Peak mem 15.545 GB
10
+ Calculating loss...: 100%|██████████| 22/22 [05:26<00:00, 14.85s/it]
11
+ Iter 20: Val loss 0.170, Val took 326.722s
12
+ Iter 20: Train loss 0.430, Learning Rate 1.000e-05, It/sec 0.039, Tokens/sec 125.126, Trained Tokens 64586, Peak mem 15.545 GB
13
+ Saved final weights to adapters_mlx/adapters.safetensors.
adapters/wandb/run-20260203_224646-so9y0g75/files/requirements.txt ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ anyio==4.12.1
2
+ cint==1.0.0
3
+ groq==1.0.0
4
+ kaitaistruct==0.11
5
+ anthropic==0.77.1
6
+ mlx-metal==0.29.3
7
+ threadpoolctl==3.6.0
8
+ docstring_parser==0.17.0
9
+ async-timeout==5.0.1
10
+ requests==2.32.5
11
+ aiohttp==3.13.3
12
+ intervaltree==3.2.1
13
+ exceptiongroup==1.3.1
14
+ multidict==6.7.1
15
+ pyarrow==21.0.0
16
+ polyfile-weave==0.5.9
17
+ diskcache==5.6.3
18
+ tzdata==2025.3
19
+ rich==14.3.2
20
+ pytz==2025.2
21
+ scipy==1.13.1
22
+ sentry-sdk==2.51.0
23
+ certifi==2026.1.4
24
+ urllib3==1.26.20
25
+ platformdirs==4.4.0
26
+ pydantic==2.12.5
27
+ pypdf==6.6.2
28
+ datasets==4.5.0
29
+ propcache==0.4.1
30
+ mlx-lm==0.29.1
31
+ wandb==0.24.1
32
+ protobuf==6.33.5
33
+ python-dotenv==1.2.1
34
+ tenacity==9.1.2
35
+ markdown-it-py==3.0.0
36
+ filelock==3.19.1
37
+ ruff==0.15.0
38
+ idna==3.11
39
+ openai==2.16.0
40
+ scikit-learn==1.6.1
41
+ click==8.1.8
42
+ httpcore==1.0.9
43
+ rpds-py==0.27.1
44
+ numpy==2.0.2
45
+ fsspec==2025.10.0
46
+ stdlib-list==0.11.1
47
+ charset-normalizer==3.4.4
48
+ graphql-core==3.2.7
49
+ eval_type_backport==0.3.1
50
+ referencing==0.36.2
51
+ h11==0.16.0
52
+ gitdb==4.0.12
53
+ sniffio==1.3.1
54
+ sentencepiece==0.2.1
55
+ regex==2026.1.15
56
+ mdurl==0.1.2
57
+ python-dateutil==2.9.0.post0
58
+ PyYAML==6.0.3
59
+ mlx==0.29.3
60
+ sortedcontainers==2.4.0
61
+ abnf==2.2.0
62
+ multiprocess==0.70.18
63
+ transformers==4.57.6
64
+ chardet==5.2.0
65
+ pydantic_core==2.41.5
66
+ frozenlist==1.8.0
67
+ typing_extensions==4.15.0
68
+ aiosignal==1.4.0
69
+ joblib==1.5.3
70
+ huggingface_hub==0.36.1
71
+ hf-xet==1.2.0
72
+ Jinja2==3.1.6
73
+ fickling==0.1.7
74
+ pandas==2.3.3
75
+ jiter==0.13.0
76
+ MarkupSafe==3.0.3
77
+ dill==0.4.0
78
+ distro==1.9.0
79
+ GitPython==3.1.46
80
+ graphviz==0.21
81
+ Pygments==2.19.2
82
+ aiohappyeyeballs==2.6.1
83
+ ollama==0.6.1
84
+ attrs==25.4.0
85
+ requests-toolbelt==1.0.0
86
+ networkx==3.2.1
87
+ packaging==26.0
88
+ httpx==0.28.1
89
+ smmap==5.0.2
90
+ tqdm==4.67.3
91
+ pillow==11.3.0
92
+ backoff==2.2.1
93
+ safetensors==0.7.0
94
+ tokenizers==0.22.2
95
+ psutil==7.2.2
96
+ typing-inspection==0.4.2
97
+ annotated-types==0.7.0
98
+ setuptools==58.0.4
99
+ six==1.17.0
100
+ yarl==1.22.0
101
+ xxhash==3.6.0
102
+ pip==26.0
adapters/wandb/run-20260203_224646-so9y0g75/files/wandb-metadata.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "macOS-15.6-arm64-arm-64bit",
3
+ "python": "CPython 3.9.6",
4
+ "startedAt": "2026-02-04T01:46:46.387687Z",
5
+ "args": [
6
+ "--config",
7
+ "adapters_mlx/train_config.yaml",
8
+ "--report-to",
9
+ "wandb",
10
+ "--project-name",
11
+ "planus-llm"
12
+ ],
13
+ "program": "-m mlx_lm.lora",
14
+ "git": {
15
+ "remote": "https://github.com/Sr-Igor/planuze-llm.git",
16
+ "commit": "2180d8de1a20f6becde1d28262c43677022a9fbb"
17
+ },
18
+ "email": "planuze@gmail.com",
19
+ "root": "adapters_mlx",
20
+ "host": "Igors-MacBook-Pro.local",
21
+ "executable": "/Users/rezende/Desktop/Planuze/planuze-llm/.venv/bin/python",
22
+ "cpu_count": 10,
23
+ "cpu_count_logical": 10,
24
+ "disk": {
25
+ "/": {
26
+ "total": "994662584320",
27
+ "used": "867762769920"
28
+ }
29
+ },
30
+ "memory": {
31
+ "total": "34359738368"
32
+ },
33
+ "apple": {
34
+ "name": "Apple M1 Max",
35
+ "ecpuCores": 2,
36
+ "pcpuCores": 8,
37
+ "gpuCores": 32,
38
+ "memoryGb": 32,
39
+ "swapTotalBytes": "5368709120",
40
+ "ramTotalBytes": "34359738368"
41
+ },
42
+ "writerId": "9wwcluo2ck30v81k0w3kns8xd71az26r"
43
+ }
adapters/wandb/run-20260203_224646-so9y0g75/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"learning_rate":9.999999747378752e-06,"iterations_per_second":0.039094455657632525,"_wandb":{"runtime":1126},"_timestamp":1.770170733439965e+09,"train_loss":0.429797887802124,"trained_tokens":64586,"iteration":20,"peak_memory":15.544893248,"_runtime":1126.162495084,"val_time":326.7221642909999,"tokens_per_second":125.12571477781867,"_step":20,"val_loss":0.16993679106235504}
adapters/wandb/run-20260203_224646-so9y0g75/logs/debug-core.log ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-02-03T22:46:46.441996-03:00","level":"INFO","msg":"main: starting server","port-filename":"/var/folders/k1/wm7466x54kj00f5pydt0yc0h0000gn/T/tmpnp1tq4le/port-48222.txt","pid":48222,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2026-02-03T22:46:46.44254-03:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":48222}
3
+ {"time":"2026-02-03T22:46:46.442511-03:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/var/folders/k1/wm7466x54kj00f5pydt0yc0h0000gn/T/wandb-48222-48231-4091335480/socket","Net":"unix"}}
4
+ {"time":"2026-02-03T22:46:46.635222-03:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1"}
5
+ {"time":"2026-02-03T22:46:46.649041-03:00","level":"INFO","msg":"handleInformInit: received","streamId":"so9y0g75","id":"1"}
6
+ {"time":"2026-02-03T22:46:46.978502-03:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"so9y0g75","id":"1"}
7
+ {"time":"2026-02-03T22:46:52.540246-03:00","level":"INFO","msg":"connection: cancelling request","id":"1","requestId":"x28rnn36al4u"}
8
+ {"time":"2026-02-03T23:05:33.651135-03:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1"}
9
+ {"time":"2026-02-03T23:05:33.651572-03:00","level":"INFO","msg":"server is shutting down"}
10
+ {"time":"2026-02-03T23:05:33.651559-03:00","level":"INFO","msg":"connection: closing","id":"1"}
11
+ {"time":"2026-02-03T23:05:33.651839-03:00","level":"INFO","msg":"connection: closed successfully","id":"1"}
12
+ {"time":"2026-02-03T23:05:33.652386-03:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/var/folders/k1/wm7466x54kj00f5pydt0yc0h0000gn/T/wandb-48222-48231-4091335480/socket","Net":"unix"}}
13
+ {"time":"2026-02-03T23:05:35.543824-03:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1"}
14
+ {"time":"2026-02-03T23:05:35.543884-03:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1"}
15
+ {"time":"2026-02-03T23:05:35.544193-03:00","level":"INFO","msg":"server is closed"}
adapters/wandb/run-20260203_224646-so9y0g75/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2026-02-03T22:46:46.649174-03:00","level":"INFO","msg":"stream: starting","core version":"0.24.1"}
2
+ {"time":"2026-02-03T22:46:46.978134-03:00","level":"INFO","msg":"stream: created new stream","id":"so9y0g75"}
3
+ {"time":"2026-02-03T22:46:46.978191-03:00","level":"INFO","msg":"handler: started","stream_id":"so9y0g75"}
4
+ {"time":"2026-02-03T22:46:46.978488-03:00","level":"INFO","msg":"stream: started","id":"so9y0g75"}
5
+ {"time":"2026-02-03T22:46:46.978504-03:00","level":"INFO","msg":"writer: started","stream_id":"so9y0g75"}
6
+ {"time":"2026-02-03T22:46:46.978504-03:00","level":"INFO","msg":"sender: started","stream_id":"so9y0g75"}
7
+ {"time":"2026-02-03T23:05:33.65201-03:00","level":"INFO","msg":"stream: closing","id":"so9y0g75"}
8
+ {"time":"2026-02-03T23:05:35.080416-03:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2026-02-03T23:05:35.538389-03:00","level":"INFO","msg":"handler: closed","stream_id":"so9y0g75"}
10
+ {"time":"2026-02-03T23:05:35.538562-03:00","level":"INFO","msg":"sender: closed","stream_id":"so9y0g75"}
11
+ {"time":"2026-02-03T23:05:35.538598-03:00","level":"INFO","msg":"stream: closed","id":"so9y0g75"}
adapters/wandb/run-20260203_224646-so9y0g75/logs/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_setup.py:_flush():81] Current SDK version is 0.24.1
2
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_setup.py:_flush():81] Configure stats pid to 48222
3
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_setup.py:_flush():81] Loading settings from environment variables
4
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:setup_run_log_directory():717] Logging user logs to adapters_mlx/wandb/run-20260203_224646-so9y0g75/logs/debug.log
5
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:setup_run_log_directory():718] Logging internal logs to adapters_mlx/wandb/run-20260203_224646-so9y0g75/logs/debug-internal.log
6
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:init():844] calling init triggers
7
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:init():849] wandb.init called with sweep_config: {}
8
+ config: {'model': 'mlx-community/Qwen2.5-14B-Instruct-4bit', 'train': True, 'data': 'data/processed', 'fine_tune_type': 'lora', 'optimizer': 'adam', 'mask_prompt': False, 'num_layers': 16, 'batch_size': 2, 'iters': 20, 'val_batches': 25, 'learning_rate': 1e-05, 'steps_per_report': 10, 'steps_per_eval': 100, 'grad_accumulation_steps': 1, 'resume_adapter_file': None, 'adapter_path': 'adapters_mlx', 'save_every': 100, 'test': False, 'test_batches': 500, 'max_seq_length': 2048, 'config': 'adapters_mlx/train_config.yaml', 'grad_checkpoint': True, 'report_to': 'wandb', 'project_name': 'planus-llm', 'seed': 3407, 'lora': {'alpha': 16, 'dropout': 0.0, 'r': 16}, 'optimizer_config': {'adam': {}, 'adamw': {}, 'muon': {}, 'sgd': {}, 'adafactor': {}}, 'lr_schedule': None, 'lora_parameters': {'rank': 8, 'dropout': 0.0, 'scale': 20.0}, '_wandb': {}}
9
+ 2026-02-03 22:46:46,390 INFO MainThread:48222 [wandb_init.py:init():892] starting backend
10
+ 2026-02-03 22:46:46,636 INFO MainThread:48222 [wandb_init.py:init():895] sending inform_init request
11
+ 2026-02-03 22:46:46,648 INFO MainThread:48222 [wandb_init.py:init():903] backend started and connected
12
+ 2026-02-03 22:46:46,650 INFO MainThread:48222 [wandb_init.py:init():973] updated telemetry
13
+ 2026-02-03 22:46:46,667 INFO MainThread:48222 [wandb_init.py:init():997] communicating run to backend with 90.0 second timeout
14
+ 2026-02-03 22:46:47,484 INFO MainThread:48222 [wandb_init.py:init():1042] starting run threads in backend
15
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_console_start():2529] atexit reg
16
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_redirect():2377] redirect: wrap_raw
17
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_redirect():2446] Wrapping output streams.
18
+ 2026-02-03 22:46:47,538 INFO MainThread:48222 [wandb_run.py:_redirect():2469] Redirects installed.
19
+ 2026-02-03 22:46:47,540 INFO MainThread:48222 [wandb_init.py:init():1082] run started, returning control to user process
20
+ 2026-02-03 23:05:33,649 INFO wandb-AsyncioManager-main:48222 [service_client.py:_forward_responses():94] Reached EOF.
21
+ 2026-02-03 23:05:33,650 INFO wandb-AsyncioManager-main:48222 [mailbox.py:close():154] Closing mailbox, abandoning 1 handles.
adapters/wandb/run-20260203_224646-so9y0g75/run-so9y0g75.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:717e62690fdece0fd8978bd029896c9ae156d1c0d63a5256e104f64d9b77f708
3
+ size 110846