nhannt201 commited on
Commit
8f415ee
·
verified ·
1 Parent(s): 4eadd75

Update Airy research model card (0.8b)

Browse files
Files changed (1) hide show
  1. README.md +87 -23
README.md CHANGED
@@ -1,31 +1,95 @@
1
- # Acnoryx AI Research Bundle
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
 
3
- ## Overview
4
 
5
- - Base model: Qwen/Qwen3.5-0.8B
6
- - Model size: 0.8b
7
- - Research quantizations: Q3_K_M, IQ3_M, Q2_K, IQ2_M, IQ2_XS, IQ2_XXS, IQ1_M, IQ1_S
8
- - Purpose: evaluate quality vs. size trade-offs below the production threshold
9
 
10
- ## Notes
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
- - IQ1/IQ2 formats require an importance matrix (imatrix).
13
- - These files are more experimental than the release bundle.
14
- - Production-facing use should prefer the release bundle.
15
- - If prompting in Vietnamese, write with full accents for best consistency.
 
 
16
 
17
- ## Evaluation snapshot
18
 
19
- - Dataset: curated `release_eval_v1-58` benchmark with 58 bilingual cases aligned to the training system style.
20
- - Protocol: quants were evaluated from higher quality to lower quality, in both `thinking` and `non-thinking` modes.
21
- - Early-stop rule: stop the downward sweep when the average of the two modes drops below 50%.
22
- - Stop point reached at `acnoryx-0.8b-iq2_m`: 27.6% thinking, 34.5% non-thinking, 31.1% average.
23
- - Because of that stop rule, lower research quants such as `Q2_K`, `IQ2_XS`, `IQ2_XXS`, `IQ1_M`, and `IQ1_S` were not executed in the combined all-model release gate.
 
 
24
 
25
- ## Tested research results
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
 
27
- | Quant | Size | Thinking | Non-thinking | Avg | Interpretation |
28
- |---|---:|---:|---:|---:|---|
29
- | Q3_K_M | 445 MB | 81.0% | 79.3% | 80.2% | Smallest clearly usable research quant |
30
- | IQ3_M | 433 MB | 70.7% | 70.7% | 70.7% | Experimental but still above minimum viability |
31
- | IQ2_M | 360 MB | 27.6% | 34.5% | 31.1% | Below threshold, triggered early stop |
 
1
+ ---
2
+ language:
3
+ - en
4
+ - vi
5
+ license: other
6
+ library_name: gguf
7
+ pipeline_tag: text-generation
8
+ base_model:
9
+ - Qwen/Qwen3.5-0.8B
10
+ tags:
11
+ - Airy
12
+ - dermatology
13
+ - skincare
14
+ - acne
15
+ - gguf
16
+ - quantization
17
+ - research
18
+ - bilingual
19
+ ---
20
 
21
+ # Airy
22
 
23
+ Airy is the public research branch of the Acnoryx dermatology model family. It contains the lower-bit GGUF experiments derived from Airy-Core-0.8B, with emphasis on size reduction, edge deployment trade-offs, and empirical failure tracking rather than production safety.
 
 
 
24
 
25
+ ## Scope
26
+
27
+ - Public repo: `Acnoryx/Airy`
28
+ - Production counterpart: `Acnoryx/Airy-Core-0.8B`
29
+ - Base model: `Qwen/Qwen3.5-0.8B`
30
+ - Languages: Vietnamese and English
31
+ - Runtime format: GGUF for llama.cpp-style runtimes
32
+ - Goal: identify the smallest research quant that still remains meaningfully useful
33
+
34
+ ## Intended use
35
+
36
+ - Quantization research
37
+ - Edge-device experiments
38
+ - Quality-vs-size benchmarking
39
+ - Comparative analysis against the release branch
40
+
41
+ ## Not intended use
42
+
43
+ - Default production deployment
44
+ - Unsupervised medical advice
45
+ - General-purpose assistant behavior outside skincare and dermatology
46
+
47
+ ## Evaluation protocol
48
 
49
+ - Benchmark: `release_eval_v1-58`
50
+ - Coverage: 58 bilingual cases across identity, knowledge, refusal, scan interpretation, language, format, hallucination, subtitle, rude-user handling, and medical-defer behavior
51
+ - Modes: `thinking` and `non-thinking`
52
+ - Sweep order: higher quality to lower quality
53
+ - Early-stop rule: stop when the average of the 2 modes falls below 50%
54
+ - Result: the sweep stopped at `Airy-0.8b-IQ2_M` with 31.1% average, so lower public research quants were not considered viable in the combined release-gate run
55
 
56
+ ## Research vs release comparison
57
 
58
+ | Branch | Representative quant | Size | Thinking | Non-thinking | Avg | Interpretation |
59
+ |---|---|---:|---:|---:|---:|---|
60
+ | Release | Airy-Core-0.8b-Q4_K_M | 505 MB | 93.1% | 82.8% | 87.9% | Best practical release balance |
61
+ | Release | Airy-Core-0.8b-Q4_0 | 478 MB | 87.9% | 82.8% | 85.4% | Smallest release quant still solid |
62
+ | Research | Airy-0.8b-Q3_K_M | 445 MB | 81.0% | 79.3% | 80.2% | Strongest research quant tested |
63
+ | Research | Airy-0.8b-IQ3_M | 433 MB | 70.7% | 70.7% | 70.7% | Usable only for experimental work |
64
+ | Research | Airy-0.8b-IQ2_M | 360 MB | 27.6% | 34.5% | 31.1% | Collapse point, triggered early stop |
65
 
66
+ ## Detailed research results
67
+
68
+ | Public model file | Local source file | Size | Thinking | Non-thinking | Avg | Notes |
69
+ |---|---|---:|---:|---:|---:|---|
70
+ | Airy-0.8b-Q3_K_M.gguf | acnoryx-0.8b-q3_k_m.gguf | 445 MB | 81.0% | 79.3% | 80.2% | Smallest clearly useful research quant |
71
+ | Airy-0.8b-IQ3_M.gguf | acnoryx-0.8b-iq3_m.gguf | 433 MB | 70.7% | 70.7% | 70.7% | Experimental floor before major degradation |
72
+ | Airy-0.8b-IQ2_M.gguf | acnoryx-0.8b-iq2_m.gguf | 360 MB | 27.6% | 34.5% | 31.1% | Below threshold, not recommended |
73
+ | Airy-0.8b-Q2_K.gguf | acnoryx-0.8b-q2_k.gguf | 403 MB | not run | not run | not run | Skipped after early stop |
74
+ | Airy-0.8b-IQ2_XS.gguf | acnoryx-0.8b-iq2_xs.gguf | 347 MB | not run | not run | not run | Skipped after early stop |
75
+ | Airy-0.8b-IQ2_XXS.gguf | acnoryx-0.8b-iq2_xxs.gguf | 336 MB | not run | not run | not run | Skipped after early stop |
76
+ | Airy-0.8b-IQ1_M.gguf | acnoryx-0.8b-iq1_m.gguf | 323 MB | not run | not run | not run | Skipped after early stop |
77
+ | Airy-0.8b-IQ1_S.gguf | acnoryx-0.8b-iq1_s.gguf | 315 MB | not run | not run | not run | Skipped after early stop |
78
+
79
+ ## Interpretation
80
+
81
+ - `Airy-0.8b-Q3_K_M` is the best public research checkpoint when file size matters more than absolute quality.
82
+ - `Airy-0.8b-IQ3_M` is still measurable but already loses substantial domain fidelity.
83
+ - `Airy-0.8b-IQ2_M` marks the quality collapse point for this benchmark.
84
+ - If you need stable app-facing behavior, use the release branch instead of this repo.
85
+
86
+ ## Public naming
87
+
88
+ Public Hugging Face files in this repository are published with the `Airy-...` prefix for clarity. The local workspace may still keep the original `acnoryx-...` filenames for build compatibility.
89
+
90
+ ## Notes
91
 
92
+ - IQ1 and IQ2 formats require an importance matrix.
93
+ - These quants are more experimental than the release branch.
94
+ - For Vietnamese prompts, use full accents for the most consistent behavior.
95
+ - Outputs remain reference-only and do not replace dermatologist care.