jake-watkins commited on
Commit
bdbad5c
·
verified ·
1 Parent(s): 9d3f94c

Upload folder using huggingface_hub

Browse files
0000100_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b227021c30ba0f267e74f14de2ca09586ffb9d41cf3947f85010a86e1cfb0608
3
+ size 25179794
0000200_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d1efbe3abf5974a8f9e4ccc4cf9c5654e98f2ed4ddc26fa0e7e0894b26ce822
3
+ size 25179794
0000300_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:953bb40440ecd0992f8fee54409246a891d4724f4c16457b396075072e1f5d02
3
+ size 25179794
0000400_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:492ee273c5c330ed324a8e37d89b16e851a93b3f64392ba9876b0c5feb8ff91c
3
+ size 25179794
0000500_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:420762fbba79e665a73b1e4040f9e8f8fee934166bdf762ffe6ee908e79997a0
3
+ size 25179794
0000600_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c50414dd3ddbbb9153188e09aaeeb0f62ccf40dfcd5ef5613b4780158139778
3
+ size 25179794
adapter_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_path": "/Users/jakewatkins/source/projects/local-ai/trained-models/email-classifier",
3
+ "batch_size": 1,
4
+ "config": null,
5
+ "data": "/Users/jakewatkins/source/projects/local-ai/training-data/email-classification",
6
+ "fine_tune_type": "lora",
7
+ "grad_accumulation_steps": 1,
8
+ "grad_checkpoint": false,
9
+ "iters": 699,
10
+ "learning_rate": 1e-05,
11
+ "lora_parameters": {
12
+ "rank": 8,
13
+ "dropout": 0.0,
14
+ "scale": 20.0
15
+ },
16
+ "lr_schedule": null,
17
+ "mask_prompt": false,
18
+ "max_seq_length": 512,
19
+ "model": "microsoft/Phi-3-mini-4k-instruct",
20
+ "num_layers": 16,
21
+ "optimizer": "adam",
22
+ "optimizer_config": {
23
+ "adam": {},
24
+ "adamw": {},
25
+ "muon": {},
26
+ "sgd": {},
27
+ "adafactor": {}
28
+ },
29
+ "project_name": null,
30
+ "report_to": null,
31
+ "resume_adapter_file": null,
32
+ "save_every": 100,
33
+ "seed": 0,
34
+ "steps_per_eval": 100,
35
+ "steps_per_report": 10,
36
+ "test": false,
37
+ "test_batches": 500,
38
+ "train": true,
39
+ "val_batches": 25
40
+ }
adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d891778b9be35ef4b1b51a7db2646761047c594d38c0d8a6420e551a28682092
3
+ size 25179794