piotrjanik commited on
Commit
459a98a
·
verified ·
1 Parent(s): 17893e6

Upload folder using huggingface_hub

Browse files
0000200_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54344b95471da867774e2667b5f058750fd4213dd9f056ce13171e01c37d71ac
3
+ size 46161566
0000400_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65e15679fae571bdc480bf0a99d91b74ea0fe17f484886fc9882ab16ff8c03e3
3
+ size 46161566
0000600_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e66f568b437464e6c0690b5870aeef4591c19318230b1767241568afe535179f
3
+ size 46161566
0000800_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f51c73e4c85f71e6ff46610aee62b248fdae7c272dc5bce08d9b2e600c6f9c6f
3
+ size 46161566
0001000_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f33565f4f7926e9f58963de97e0bc08f58ed33283b3db2e5b8b919157be80ea
3
+ size 46161566
README.md ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mlx-community/Qwen2.5-Coder-32B-Instruct-4bit
3
+ library_name: mlx
4
+ tags: [mlx, lora, ocm, oci, go]
5
+ ---
6
+
7
+ # ocm-coder
8
+
9
+ LoRA adapters fine-tuned on the Open Component Model (OCM) and OCI specification ecosystem.
10
+
11
+ **Base model:** [mlx-community/Qwen2.5-Coder-32B-Instruct-4bit](https://huggingface.co/mlx-community/Qwen2.5-Coder-32B-Instruct-4bit)
12
+
13
+ ## Usage
14
+
15
+ ```python
16
+ from mlx_lm import load, generate
17
+
18
+ model, tokenizer = load(
19
+ "mlx-community/Qwen2.5-Coder-32B-Instruct-4bit",
20
+ adapter_path="piotrjanik/ocm-coder",
21
+ )
22
+ ```
adapter_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_path": "output/adapters",
3
+ "batch_size": 4,
4
+ "config": "output/lora_train_config.yaml",
5
+ "data": "output/mlx_data",
6
+ "fine_tune_type": "lora",
7
+ "grad_accumulation_steps": 1,
8
+ "grad_checkpoint": false,
9
+ "iters": 1000,
10
+ "learning_rate": 1e-05,
11
+ "lora_parameters": {
12
+ "dropout": 0.0,
13
+ "rank": 8,
14
+ "scale": 10.0
15
+ },
16
+ "lr_schedule": null,
17
+ "mask_prompt": false,
18
+ "max_seq_length": 4096,
19
+ "model": "mlx-community/Qwen2.5-Coder-7B-Instruct-4bit",
20
+ "num_layers": 16,
21
+ "optimizer": "adam",
22
+ "optimizer_config": {
23
+ "adam": {},
24
+ "adamw": {},
25
+ "muon": {},
26
+ "sgd": {},
27
+ "adafactor": {}
28
+ },
29
+ "project_name": null,
30
+ "report_to": null,
31
+ "resume_adapter_file": null,
32
+ "save_every": 200,
33
+ "seed": 42,
34
+ "steps_per_eval": 200,
35
+ "steps_per_report": 10,
36
+ "test": false,
37
+ "test_batches": 500,
38
+ "train": true,
39
+ "val_batches": 25
40
+ }
adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f33565f4f7926e9f58963de97e0bc08f58ed33283b3db2e5b8b919157be80ea
3
+ size 46161566