| --- |
| language: |
| - en |
| --- |
| |
| # Community Quantization Requests |
|
|
| This space is for requesting oQe builds. These quants utilize multi-stage calibration and Hessian-based error compensation to maintain logic stability, specifically tuned for Apple Silicon performance. |
|
|
| ### How to Request |
| Open a new Discussion for requests. To ensure a valid build, please include: |
|
|
| 1. Model Link: URL to the official Hugging Face repository. Note that I only process builds starting from original BF16 or FP16 source weights. |
| 2. Quantization Format: Specify if you need BF16 or FP16 quants. |
| * FP16 is generally recommended for M1/M2 series to utilize AMX units for faster prefill. |
| * BF16 is recommended for M3/M4 series with native support. |
| 3. Preferred Tiers: Specify the target bitrate (e.g., oQ5e, oQ4e) based on your available Unified Memory. |
|
|
| ### Guidelines |
| * Hardware: Builds are processed on a 192GB M2 Ultra. Models up to 70B parameters (standard dense) are supported. Anything significantly larger (100B+ or large MoE architectures) will exceed memory limits when loading source weights for calibration. |
| * Selection Criteria: Priority is given to base models and official instruct tunes. Experimental merges or low-epoch fine-tunes are generally excluded unless there is significant community interest. |
| * The Process: Every oQe build undergoes a 600-sample calibration pass. These are not one-pass streaming quants. |
|
|
| ### Technical Spec |
| All fulfilled requests are processed using the oMLX Enhanced Quantization process: |
| * Sensitivity Mapping: Calibration pass measures precision requirements per layer to prevent output drift. |
| * Hessian-Based Tuning: GPTQ-Hessian error compensation is applied during weight rounding. |
| * Precision Anchoring: Native BF16 for routing gates and FP16 for attention heads to maximize Apple Silicon AMX throughput. |
| * Logic Floor: The lm_head and critical early blocks are locked at 8-bit to ensure core reasoning stability. |