Add accuracy recovery adapters
Browse files- accuracy_recovery_adapters/.gitattributes +35 -0
- accuracy_recovery_adapters/README.md +26 -0
- accuracy_recovery_adapters/flux1_dev_kontext_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/hidream_i1_full_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/qwen_image_2512_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/qwen_image_2512_torchao_uint4.safetensors +3 -0
- accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/qwen_image_edit_2511_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/qwen_image_edit_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/qwen_image_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/wan22_14b_i2v_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/wan22_14b_i2v_torchao_uint4.safetensors +3 -0
- accuracy_recovery_adapters/wan22_14b_t2i_torchao_uint3.safetensors +3 -0
- accuracy_recovery_adapters/wan22_14b_t2i_torchao_uint4.safetensors +3 -0
accuracy_recovery_adapters/.gitattributes
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
| 2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
accuracy_recovery_adapters/README.md
ADDED
|
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- accuracy recovery adapter
|
| 4 |
+
- AI Toolkit
|
| 5 |
+
---
|
| 6 |
+
# Accuracy Recovery Adapters
|
| 7 |
+
|
| 8 |
+
This repo contains various accuracy recovery adapters (ARAs) that I have trained, primarialy for use with [AI Toolkit](https://github.com/ostris/ai-toolkit).
|
| 9 |
+
An ARA is a LoRA that is trained via student teacher training with the student being quantized down to a low precision and the teacher having a high precision.
|
| 10 |
+
The goal is to have a side chain LoRA, at bfloat16, that runs parallel to highly quantized layers in a network to compensate for the loss in precision that happens
|
| 11 |
+
when weights are quantized. The training is done on a per layer basis in order to match the parent output as much as possible.
|
| 12 |
+
|
| 13 |
+
While this can be used on inference, my primary goal is to make large models finetunable on consumer grade hardware. With the 3bit Qwen Image adapter, it
|
| 14 |
+
is now possible to train a LoRA on top of it, with 1 MP images, on a 24 GB GPU, such as a 3090/4090.
|
| 15 |
+
|
| 16 |
+
I have found the sweet spot, at least for [Qwen-Image](https://huggingface.co/Qwen/Qwen-Image), is 3 bit quantization with a rank 16 adapter.
|
| 17 |
+
|
| 18 |
+
More info, examples, links, training scripts, AI Toolkit example configs, and adapters to some soon.
|
| 19 |
+
|
| 20 |
+
## License
|
| 21 |
+
|
| 22 |
+
All adapters inherit the parent model license. Apache 2.0 for Apache 2.0, BFL License for BFL License, etc.
|
| 23 |
+
|
| 24 |
+
## Qwen-Image 3 bit quantization
|
| 25 |
+
|
| 26 |
+

|
accuracy_recovery_adapters/flux1_dev_kontext_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eefe8672057d111ad8dd8367a08da1cceb191b909d98dad33a9a34a8e3c5b56c
|
| 3 |
+
size 173623624
|
accuracy_recovery_adapters/hidream_i1_full_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4b89cbd7c1e22e554d64f062bbbf0bac5378a46d0c95c2b981f7b711914f3215
|
| 3 |
+
size 306532608
|
accuracy_recovery_adapters/qwen_image_2512_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d3b456a335cf60b5ef9a531a40fce5be5f03a69c2667a544794551f4669a54ac
|
| 3 |
+
size 295146200
|
accuracy_recovery_adapters/qwen_image_2512_torchao_uint4.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f50363dc04c49ef3bedf2bfa975e5a31a8dcaaa9031e99d307df7e6f516363ed
|
| 3 |
+
size 295146200
|
accuracy_recovery_adapters/qwen_image_edit_2509_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:914e1952f450c809a916fff3613542f124b0a6395c92accc782675024d066c3d
|
| 3 |
+
size 296159384
|
accuracy_recovery_adapters/qwen_image_edit_2511_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c22d6c4d22e5f80a6099c83e1a98942e27eae134891268b21a5bbee5f598afa8
|
| 3 |
+
size 296159376
|
accuracy_recovery_adapters/qwen_image_edit_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6592df2b461a4e8fca4919fbff39999015f56fec554ebda805f91af82f9f696a
|
| 3 |
+
size 296159368
|
accuracy_recovery_adapters/qwen_image_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1791fea0181da84e411c8ed8ff00c00cc5c6cdd32b1a26e2f590ece6e5fa8186
|
| 3 |
+
size 296157792
|
accuracy_recovery_adapters/wan22_14b_i2v_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1708e008e8ba36e8618b8a6effd62fccbf8431420e4341dc5b56e29ad05d92c1
|
| 3 |
+
size 311804040
|
accuracy_recovery_adapters/wan22_14b_i2v_torchao_uint4.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5c5cd2d232ff37fbd866d20eac632067f3a6bf5d4eb5b125b6c09479fd851bd
|
| 3 |
+
size 311804040
|
accuracy_recovery_adapters/wan22_14b_t2i_torchao_uint3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8b9d26ddd919e380ecfffb8d027d20f2f3ee20fbceb19592ac7c233b3d0d30b
|
| 3 |
+
size 311804032
|
accuracy_recovery_adapters/wan22_14b_t2i_torchao_uint4.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3baad095fc0b56ad88dc5f2c92b12f1a0c4229c13034a9ae9e92c2c6a7ebf74c
|
| 3 |
+
size 311804032
|