splats commited on
Commit
f049db6
·
verified ·
1 Parent(s): f32fbf3

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -0
README.md ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Community Quantization Requests
2
+
3
+ This space is for requesting oQe builds. These quants utilize multi-stage calibration and Hessian-based error compensation to maintain logic stability, specifically tuned for Apple Silicon performance.
4
+
5
+ ### How to Request
6
+ Open a new Discussion for requests. To ensure a valid build, please include:
7
+
8
+ 1. Model Link: URL to the official Hugging Face repository. Note that I only process builds starting from original BF16 or FP16 source weights.
9
+ 2. Quantization Format: Specify if you need BF16 or FP16 quants.
10
+ * FP16 is generally recommended for M1/M2 series to utilize AMX units for faster prefill.
11
+ * BF16 is recommended for M3/M4 series with native support.
12
+ 3. Preferred Tiers: Specify the target bitrate (e.g., oQ5e, oQ4e) based on your available Unified Memory.
13
+
14
+ ### Guidelines
15
+ * Hardware: Builds are processed on a 192GB M2 Ultra. Models up to 70B parameters (standard dense) are supported. Anything significantly larger (100B+ or large MoE architectures) will exceed memory limits when loading source weights for calibration.
16
+ * Selection Criteria: Priority is given to base models and official instruct tunes. Experimental merges or low-epoch fine-tunes are generally excluded unless there is significant community interest.
17
+ * The Process: Every oQe build undergoes a 600-sample calibration pass. These are not one-pass streaming quants.
18
+
19
+ ### Technical Spec
20
+ All fulfilled requests are processed using the oMLX Enhanced Quantization process:
21
+ * Sensitivity Mapping: Calibration pass measures precision requirements per layer to prevent output drift.
22
+ * Hessian-Based Tuning: GPTQ-Hessian error compensation is applied during weight rounding.
23
+ * Precision Anchoring: Native BF16 for routing gates and FP16 for attention heads to maximize Apple Silicon AMX throughput.
24
+ * Logic Floor: The lm_head and critical early blocks are locked at 8-bit to ensure core reasoning stability.