nhannt201 commited on
Commit
4475eff
Β·
verified Β·
1 Parent(s): 366d6f1

feat: upload new research

Browse files
README.md CHANGED
@@ -1,94 +1,117 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - vi
5
- - en
6
- tags:
7
- - acne
8
- - skincare
9
- - dermatology
10
- - gguf
11
- - qwen3
12
- base_model: Qwen/Qwen3-0.6B
13
- pipeline_tag: text-generation
14
- ---
15
-
16
- # Acnoryx/Airy-Lite β€” Research GGUF Bundle
17
-
18
- Research & evaluation companion to the main release. Contains sub-4-bit quantizations
19
- (<4-bit) for low-memory benchmarking on the **0.6B** model.
20
-
21
- ## Model Details
22
-
23
- | | |
24
- |-|-|
25
- | **Base model** | Qwen/Qwen3-0.6B (596M params) |
26
- | **Fine-tune** | SFT on 30,007 acne/skincare/dermatology samples |
27
- | **Training** | 4 epochs, batch=2, grad_acc=8, lr=5e-5 |
28
- | **Languages** | Vietnamese, English |
29
- | **Domain** | Acne analysis, skincare routines, scan interpretation |
30
- | **Identity** | Acnoryx AI β€” in-app dermatology assistant |
31
-
32
- ## Research Quantization Results
33
-
34
- Tested with 100 domain-specific questions Γ— 2 modes (thinking / non-thinking).
35
- All quantizations in this bundle are **sub-4-bit (<4-bit)**, ordered high-to-low bit depth.
36
-
37
- | Quant | Size | Thinking | Non-Think | Avg | Status |
38
- |-------|------|----------|-----------|-----|--------|
39
- | **Q3_K_M** | 395 MB | **77%** | **77%** | **77.0%** | ⚠️ Degraded |
40
- | **IQ3_M** | 384 MB | 3% | 7% | 5.0% | ❌ Not usable |
41
- | **Q2_K** | 331 MB | 0% | 0% | 0.0% | ❌ Not usable |
42
- | **IQ2_M** | 316 MB | 0% | 0% | 0.0% | ❌ Not usable |
43
- | **IQ2_XS** | 280 MB | 0% | 0% | 0.0% | ❌ Skipped (early-stop) |
44
- | **IQ2_XXS** | 268 MB | 0% | 0% | 0.0% | ❌ Skipped (early-stop) |
45
- | **IQ1_M** | 255 MB | 0% | 0% | 0.0% | ❌ Skipped (early-stop) |
46
- | **IQ1_S** | 247 MB | 0% | 0% | 0.0% | ❌ Skipped (early-stop) |
47
-
48
- ### imatrix
49
-
50
- IQ2/IQ1 quants were generated with importance matrix (imatrix) calibration from
51
- a 44KB domain-specific corpus. Despite imatrix, the 0.6B model fails at IQ3_M (5%)
52
- and collapses completely at 2-bit and below.
53
-
54
- ## Full Quantization Map (Release + Research)
55
-
56
- Combined view across all quantizations for the 0.6B model, ordered by bit depth (high β†’ low):
57
-
58
- | Quant | Size | Thinking | Non-Think | Avg | Bundle |
59
- |-------|------|----------|-----------|-----|--------|
60
- | F16 | 1439 MB | 95% | 90% | 92.5% | Release |
61
- | Q8_0 | 768 MB | 91% | 90% | 90.5% | Release |
62
- | Q5_K_M | 526 MB | 94% | 92% | 93.0% | Release |
63
- | Q4_K_M | 462 MB | 86% | 84% | 85.0% | Release |
64
- | Q4_0 | 447 MB | 88% | 84% | 86.0% | Release |
65
- | IQ4_NL | 448 MB | 90% | 90% | 90.0% | Release |
66
- | IQ4_XS | 431 MB | 84% | 91% | 87.5% | Release |
67
- | **Q3_K_M** | **395 MB** | **77%** | **77%** | **77.0%** | **Research** |
68
- | **IQ3_M** | **384 MB** | **3%** | **7%** | **5.0%** | **Research** |
69
- | **Q2_K** | **331 MB** | **0%** | **0%** | **0.0%** | **Research** |
70
- | **IQ2_M** | **316 MB** | **0%** | **0%** | **0.0%** | **Research** |
71
- | **IQ2_XS** | **280 MB** | **0%** | **0%** | **0.0%** | **Research** |
72
- | **IQ2_XXS** | **268 MB** | **0%** | **0%** | **0.0%** | **Research** |
73
- | **IQ1_M** | **255 MB** | **0%** | **0%** | **0.0%** | **Research** |
74
- | **IQ1_S** | **247 MB** | **0%** | **0%** | **0.0%** | **Research** |
75
-
76
- ### Key findings
77
-
78
- - **Release floor (4-bit):** All release quants score β‰₯84% β€” production-ready
79
- - **Research ceiling (3-bit):** Q3_K_M (77%) is usable but noticeably degraded
80
- - **Hard cliff:** IQ3_M collapses to 5%, and everything below 3-bit hits 0%
81
- - imatrix calibration does not rescue the 0.6B model below 3-bit
82
- - For usable sub-4-bit results, switch to the 0.8B model (Acnoryx/Airy)
83
-
84
- ## Usage
85
-
86
- ```bash
87
- # llama.cpp β€” Q3_K_M is the only viable research quant
88
- ./llama-cli -m acnoryx-0.6b-q3_k_m.gguf -cnv -p "Xin chΓ o"
89
- ```
90
-
91
- ## Related
92
-
93
- - **Release bundle:** Production quantizations (F16 β†’ IQ4_XS, β‰₯4-bit)
94
- - **0.8B research:** [Acnoryx/Airy](https://huggingface.co/Acnoryx/Airy) β€” larger model with better low-bit resilience
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - vi
5
+ - en
6
+ base_model: Qwen/Qwen3-0.6B
7
+ tags:
8
+ - dermatology
9
+ - skincare
10
+ - acne
11
+ - gguf
12
+ - llama-cpp
13
+ - qwen3
14
+ - quantization-research
15
+ - low-bit
16
+ - acnoryx
17
+ model-index:
18
+ - name: Acnoryx-0.6B
19
+ results:
20
+ - task:
21
+ type: text-generation
22
+ metrics:
23
+ - name: Q3_K_M Pass Rate (118 questions)
24
+ type: custom
25
+ value: 90.7
26
+ - name: IQ3_M Pass Rate (118 questions)
27
+ type: custom
28
+ value: 24.6
29
+ ---
30
+
31
+ # Acnoryx 0.6B β€” Research GGUF (Low-Bit Quantization)
32
+
33
+ > **⚠️ NOT for production use** β€” These are aggressive low-bit quantizations for research purposes only.
34
+
35
+ # Acnoryx 0.6B β€” Research GGUF
36
+
37
+ **Acnoryx AI** is a dermatology-focused language model fine-tuned for the [Acnoryx AI - Acne Assistant](https://play.google.com/store/apps/details?id=com.fivecanh.acnoryx). It provides skincare guidance, acne analysis, and scan interpretation in **Vietnamese** and **English**.
38
+
39
+
40
+
41
+ | | |
42
+ |---|---|
43
+ | **Base model** | `Qwen/Qwen3-0.6B` |
44
+ | **Method** | SFT (Supervised Fine-Tuning) via Unsloth LoRA |
45
+ | **Languages** | Vietnamese, English |
46
+ | **Purpose** | Low-bit quantization quality research |
47
+
48
+ ## Evaluation Results β€” 118-Question Strict Scoring
49
+
50
+ Same **118-question multi-criteria strict scoring** as release models.
51
+ All failures manually verified.
52
+
53
+ | Quantization | Size | Passed | Pass Rate | Status |
54
+ |---|---|---|---|---|
55
+ | **Q3_K_M** | 395 MB | 107/118 | **90.7%** | ⚠️ Usable with caveats |
56
+ | **IQ3_M** | 384 MB | 29/118 | **24.6%** | ❌ Fail β€” garbled output |
57
+ | Q2_K | 332 MB | β€” | β€” | β›” Skipped (IQ3_M < 50%) |
58
+ | IQ2_M | 317 MB | β€” | β€” | β›” Skipped |
59
+ | IQ2_XS | 276 MB | β€” | β€” | β›” Skipped |
60
+ | IQ2_XXS | 268 MB | β€” | β€” | β›” Skipped |
61
+ | IQ1_S | 248 MB | β€” | β€” | β›” Skipped |
62
+ | IQ1_M | 5.7 MB | β€” | β€” | ⚠️ Corrupted export |
63
+
64
+ ### Category Breakdown β€” Q3_K_M (90.7%)
65
+
66
+ | Category | Tests | Passed | Pass Rate |
67
+ |---|---|---|---|
68
+ | Identity (EN/VI) | 12 | 11 | 92% |
69
+ | Acne Types & Definitions | 20 | 19 | 95% |
70
+ | Acne Causes & Triggers | 10 | 10 | 100% |
71
+ | Skincare Ingredients | 10 | 10 | 100% |
72
+ | Skincare Routines | 8 | 6 | 75% |
73
+ | Scan Analysis | 12 | 12 | 100% |
74
+ | Boundary / Refusal | 22 | 18 | 82% |
75
+ | Format Checks (think tags) | 4 | 4 | 100% |
76
+ | Out-of-Distribution (OOD) | 20 | 17 | 85% |
77
+
78
+ ### Key Findings
79
+
80
+ - **Q3_K_M** (90.7%) is still usable but shows degradation in routines (75%) and boundary (82%) categories compared to release quants (97.5% for Q4_0)
81
+ - **IQ3_M** (24.6%) is catastrophic β€” produces garbled/broken text, CJK leakage, and nonsensical responses in the majority of tests
82
+ - Quality **cliff** occurs between Q3_K_M and IQ3_M β€” a 66-point drop
83
+ - Testing stopped at IQ3_M since pass rate fell below 50%, making lower quants pointless
84
+
85
+ ### Comparison with Release Models
86
+
87
+ | Quantization | Tier | Size | Pass Rate |
88
+ |---|---|---|---|
89
+ | Q4_0 | Release | 448 MB | **97.5%** |
90
+ | IQ4_XS | Release | 431 MB | **96.6%** |
91
+ | Q8_0 | Release | 768 MB | **95.8%** |
92
+ | Q5_K_M | Release | 526 MB | **94.9%** |
93
+ | F16 | Release | 1.5 GB | **93.2%** |
94
+ | IQ4_NL | Release | 449 MB | **93.2%** |
95
+ | Q4_K_M | Release | 462 MB | **91.5%** |
96
+ | **Q3_K_M** | **Research** | **395 MB** | **90.7%** |
97
+ | **IQ3_M** | **Research** | **384 MB** | **24.6%** |
98
+
99
+ ## Provided Files
100
+
101
+ | File | Size | Description |
102
+ |---|---|---|
103
+ | `acnoryx-0.6b-q3_k_m.gguf` | 395 MB | Best research quant (90.7%) |
104
+ | `acnoryx-0.6b-iq3_m.gguf` | 384 MB | Garbled below usable (24.6%) |
105
+ | `acnoryx-0.6b-q2_k.gguf` | 332 MB | Not tested β€” expected worse |
106
+ | `acnoryx-0.6b-iq2_m.gguf` | 317 MB | Not tested |
107
+ | `acnoryx-0.6b-iq2_xs.gguf` | 276 MB | Not tested |
108
+ | `acnoryx-0.6b-iq2_xxs.gguf` | 268 MB | Not tested |
109
+ | `acnoryx-0.6b-iq1_s.gguf` | 248 MB | Not tested |
110
+ | `acnoryx-0.6b-iq1_m.gguf` | 5.7 MB | ⚠️ Corrupted export |
111
+
112
+
113
+ | | |
114
+ |---|---|
115
+ | **Data** | 29,726 samples across 94 cleaned JSONL files |
116
+ | **Method** | Unsloth SFT with LoRA (r=16, alpha=16) |
117
+ | **Checkpoint** | checkpoint-9795 |
gguf/acnoryx-0.6b-iq1_m.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc40a6f8d3c53a22c76888a65acdccda20ba3c3e197e261cfbfaced56fc192e5
3
- size 267099392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83daa62489a1711fc36258a9fc8270dcf95178e7eb8480a3ccdaacbf78140627
3
+ size 5946666
gguf/acnoryx-0.6b-iq1_s.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a809139da4c4440dcda77c47a61b8a4a33a78faf02b344ff2227145103973e4
3
- size 259063040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1bfbaabf8564a03036ea5b9b196808f728d38322ee6e4114d280ba14b2fcf9d
3
+ size 259063584
gguf/acnoryx-0.6b-iq2_m.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bba771db4041b41f5c57cc8a92ac1255cea250f2f8c37b38addf775c852c4566
3
- size 331757824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85951e6ab9bc5ff0e4743248cd52903436abbe9cbc8dee8676062f4ead3eea26
3
+ size 331758368
gguf/acnoryx-0.6b-iq2_xs.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76a49927b10c0143d5d657172601794a2840de2d303bb4362eec14ab335ecee6
3
- size 293043456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81e4c9af16c007afe5f2bc84f6f1030bf01d9eaf7864be5667e58b5b76b031c9
3
+ size 288447264
gguf/acnoryx-0.6b-iq2_xxs.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a34558921d04859ba080bf55cb8837cbf195969b6ae55e3e551b6b5411a7c8d
3
- size 280493312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:108be9bc479a7fc456de9936861c22c4d4c4593cbf8d57a735e9ef03428eecad
3
+ size 280493856
gguf/acnoryx-0.6b-iq3_m.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1555a760d3095b865abc92ed27cdbfb5684442a2c79ec1c372958bdbf7faf454
3
- size 402875360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d21108ff45214322026034972e1b705761e361c5901177f68abe71d29587c12
3
+ size 402875872
gguf/acnoryx-0.6b-q2_k.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3735fc651e0e213794bdcbd59d96afa72a981cc0591c81ab228e88de7680a0ae
3
- size 347285472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa0a9f41d0334f737d5e4076898ed45f8577f1ed76b1621798f504f2c16eeace
3
+ size 347285984
gguf/acnoryx-0.6b-q3_k_m.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e45157f9acd6e2741522bac5bd07eb5795fb995bd412d75ce6e33b1a46528ad
3
- size 413975520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29ed30950e948de092bebb25228babf5ba9e6b755e7700c7a92a3b582cbe4927
3
+ size 413976032