YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
Model Card: GenomeOcean-100M (FP8)
Generated: 2026-05-09T18:03:44-0700
Architecture
| Parameter | Value |
|---|---|
| Architecture | MistralForCausalLM |
| Model Type | mistral |
| Vocab Size | 4096 |
| Hidden Size | 768 |
| Num Hidden Layers | 12 |
| Num Attention Heads | 8 |
| Intermediate Size | 3072 |
| Max Position Embeddings | 32768 |
| RoPE Theta | 1000000.0 |
Quantization Method
- Format: FP8 (E4M3) per-channel weight-only quantization
- Scale DType: float32 per-channel scales
- Method: Post-training quantization (PTQ) with per-channel E4M3 weights
Perplexity Results
| Metric | Value |
|---|---|
| Original PPL (BF16) | 40804.858 |
| Quantized PPL (FP8) | 41000.839 |
| PPL Difference | 195.981 |
| PPL Difference (%) | 0.48% |
Quality Assessment: Excellent - negligible quality loss
Weight Fidelity
| Metric | Value |
|---|---|
| Mean Cosine Similarity | 0.999581 |
| Min Cosine Similarity | 0.999509 |
| Mean Relative L2 Error | 0.026446 |
| Max Relative L2 Error | 0.026671 |
| Layers Compared | 85 |
Compression
| Metric | Value |
|---|---|
| Original Size | 0.2394 GB |
| Quantized Size | 0.1171 GB |
| Compression Ratio | 48.91% |
| Space Saved | 0.12 GB |
Summary
The GenomeOcean-100M model was quantized from BF16 to FP8. Perplexity changed by 0.48% (original: 40804.858, quantized: 41000.839). Mean weight cosine similarity is 0.9996. Compression ratio is 48.91% (saved 0.12 GB).
- Downloads last month
- 15
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support