|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- speculators |
|
|
- eagle |
|
|
- speculative-decoding |
|
|
--- |
|
|
|
|
|
# Eagle Llama 3.1 8B Instruct |
|
|
|
|
|
This is a converted Eagle speculator checkpoint for Llama 3.1 8B Instruct, compatible with the [speculators](https://github.com/neuralmagic/speculators) library. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
- **Original checkpoint**: [yuhuili/EAGLE-LLaMA3.1-Instruct-8B](https://huggingface.co/yuhuili/EAGLE-LLaMA3.1-Instruct-8B) |
|
|
- **Base model**: [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) |
|
|
- **Architecture**: EAGLE (standard, without layernorms) |
|
|
- **Size**: 481MB |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from speculators.models.eagle import EagleSpeculator |
|
|
from transformers import AutoModelForCausalLM |
|
|
|
|
|
# Load the Eagle speculator |
|
|
eagle_model = EagleSpeculator.from_pretrained("nm-testing/eagle-llama3.1-8b-instruct") |
|
|
|
|
|
# Attach verifier model |
|
|
verifier = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3.1-8B-Instruct") |
|
|
eagle_model.attach_verifier(verifier) |
|
|
|
|
|
# Use for speculative decoding |
|
|
# ... your speculative decoding code here ... |
|
|
``` |
|
|
|
|
|
## Conversion Details |
|
|
|
|
|
This checkpoint was converted using: |
|
|
```bash |
|
|
speculators convert --eagle \ |
|
|
yuhuili/EAGLE-LLaMA3.1-Instruct-8B \ |
|
|
./eagle-standard-converted \ |
|
|
meta-llama/Meta-Llama-3.1-8B-Instruct |
|
|
``` |
|
|
|
|
|
## License |
|
|
|
|
|
Apache 2.0 |