|
|
--- |
|
|
language: |
|
|
- en |
|
|
- zh |
|
|
license: apache-2.0 |
|
|
pipeline_tag: text-generation |
|
|
library_name: transformers |
|
|
--- |
|
|
|
|
|
# BlockFFN-3B-SFT-EAGLE |
|
|
|
|
|
This is the 3B BlockFFN model used in the paper *BlockFFN: Towards End-Side Acceleration-Friendly Mixture-of-Experts with Chunk-Level Activation Sparsity* for acceleration tests. |
|
|
|
|
|
**BlockFFN** introduces a novel Mixture-of-Experts (MoE) architecture designed for efficient inference, particularly on end-side devices. It aims to achieve high token-level and chunk-level sparsity, making it acceleration-friendly and compatible with techniques like speculative decoding. This model is based on the [paper](https://arxiv.org/pdf/2507.08771). |
|
|
|
|
|
For the full codebase and more details, visit the official [GitHub repository](https://github.com/thunlp/BlockFFN). |
|
|
|
|
|
### Usage |
|
|
|
|
|
You can easily load and use this model with the Hugging Face `transformers` library: |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline |
|
|
import torch |
|
|
|
|
|
model_name = "SparseLLM/BlockFFN-3B-SFT-EAGLE" |
|
|
|
|
|
# Load model and tokenizer |
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model_name, |
|
|
torch_dtype=torch.bfloat16, # or torch.float16 if bfloat16 is not supported |
|
|
device_map="auto", |
|
|
trust_remote_code=True, |
|
|
) |
|
|
|
|
|
# Create a text generation pipeline |
|
|
pipe = pipeline( |
|
|
"text-generation", |
|
|
model=model, |
|
|
tokenizer=tokenizer, |
|
|
) |
|
|
|
|
|
# Generate text |
|
|
prompt = "The quick brown fox jumps over the lazy" |
|
|
output = pipe(prompt, max_new_tokens=50, do_sample=True, temperature=0.7) |
|
|
|
|
|
print(output[0]['generated_text']) |
|
|
``` |
|
|
|
|
|
### Citation |
|
|
|
|
|
If you find our work useful for your research, please kindly cite our paper as follows: |
|
|
|
|
|
``` |
|
|
@article{song2025blockffn, |
|
|
title={{BlockFFN}: Towards End-Side Acceleration-Friendly Mixture-of-Experts with Chunk-Level Activation Sparsity}, |
|
|
author={Chenyang Song and Weilin Zhao and Xu Han and Chaojun Xiao and Yingfa Chen and Yuxuan Li and Zhiyuan Liu and Maosong Sun}, |
|
|
journal={arXiv preprint arXiv:2507.08771}, |
|
|
year={2025}, |
|
|
url={https://arxiv.org/pdf/2507.08771}, |
|
|
} |