| | --- |
| | tags: |
| | - fp8 |
| | - vllm |
| | license: llama3 |
| | license_link: https://llama.meta.com/llama3/license/ |
| | language: |
| | - en |
| | --- |
| | |
| |
|
| |
|
| | # Meta-Llama-3-8B-Instruct-FP8 |
| |
|
| | ## Model Overview |
| | - **Model Architecture:** Meta-Llama-3 |
| | - **Input:** Text |
| | - **Output:** Text |
| | - **Model Optimizations:** |
| | - **Weight quantization:** FP8 |
| | - **Activation quantization:** FP8 |
| | - **KV cache quantization:** FP8 |
| | - **Intended Use Cases:** Intended for commercial and research use in English. Similarly to [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct), this models is intended for assistant-like chat. |
| | - **Out-of-scope:** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. |
| | - **Release Date:** 6/8/2024 |
| | - **Version:** 1.0 |
| | - **License(s):** [Llama3](https://llama.meta.com/llama3/license/) |
| | - **Model Developers:** Neural Magic |
| |
|
| | Quantized version of [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct). |
| |
|
| |
|
| | ``` |
| | lm_eval --model vllm --model_args pretrained=nm-testing/Meta-Llama-3-8B-Instruct-FP8-K-V,kv_cache_dtype=fp8,add_bos_token=True --tasks gsm8k --num_fewshot 5 --batch_size auto |
| | |
| | vllm (pretrained=nm-testing/Meta-Llama-3-8B-Instruct-FP8-K-V,kv_cache_dtype=fp8,add_bos_token=True), gen_kwargs: (None), limit: None, num_fewshot: 5, batch_size: auto |
| | |Tasks|Version| Filter |n-shot| Metric | |Value | |Stderr| |
| | |-----|------:|----------------|-----:|-----------|---|-----:|---|-----:| |
| | |gsm8k| 3|flexible-extract| 5|exact_match|↑ |0.7748|± |0.0115| |
| | | | |strict-match | 5|exact_match|↑ |0.7763|± |0.0115| |
| | ``` |