Bonsai-8B-TRNQ
Bonsai-8B-TRNQ is a Trillim-packaged, requantized build of prism-ml/Bonsai-8B-unpacked, intended for efficient text generation with the Trillim inference engine.
Overview
- Based on Prism ML’s Bonsai-8B release
- Requantized and packaged for Trillim runtime
- Includes tokenizer, chat template, and Trillim-specific runtime artifacts
- Distributed under the upstream Apache 2.0 license with attribution preserved
Model Details
| Item | Value |
|---|---|
| Architecture | Qwen3-8B dense |
| Parameters | 8B class |
| Source model | prism-ml/Bonsai-8B-unpacked |
| Packaging | Trillim requantized bundle |
| License | Apache 2.0 |
Usage
pip install trillim
trillim pull Trillim/Bonsai-8B-TRNQ
trillim chat Trillim/Bonsai-8B-TRNQ
This launches an interactive CLI chat session.
Repository Contents
| File | Description |
|---|---|
qmodel.tensors |
Quantized weights in Trillim format |
rope.cache |
Precomputed RoPE cache for runtime |
config.json |
Model configuration |
generation_config.json |
Generation defaults |
trillim_config.json |
Trillim runtime metadata |
tokenizer.json |
Tokenizer data |
tokenizer_config.json |
Tokenizer configuration |
special_tokens_map.json |
Special token mapping |
chat_template.jinja |
Chat template |
vocab.json, merges.txt, added_tokens.json |
Tokenizer assets |
Provenance
This repository is derived from:
prism-ml/Bonsai-8B-unpacked- Qwen3-8B by Alibaba Cloud
Changes made by Trillim:
- Requantized from the upstream release
- Repackaged for the Trillim inference engine
This repository is not affiliated with or endorsed by Prism ML or Alibaba Cloud.
License
Released under the Apache 2.0 License, consistent with the upstream model license.
See:
LICENSENOTICE.txt
- Downloads last month
- 96
Model tree for Trillim/Bonsai-8B-TRNQ
Base model
prism-ml/Bonsai-8B-unpacked