Instructions to use Xenova/tiny-random-mistral with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers.js
How to use Xenova/tiny-random-mistral with Transformers.js:
// npm i @huggingface/transformers import { pipeline } from '@huggingface/transformers'; // Allocate pipeline const pipe = await pipeline('text-generation', 'Xenova/tiny-random-mistral');
Add/update the quantized ONNX model files and README.md for Transformers.js v3
#2
by whitphx - opened
Applied Quantizations
β
Based on decoder_model_merged.onnx with slimming
The base model decoder_model_merged.onnx has been renamed to model.onnx.
β³ β
fp16: model_fp16.onnx (added)
β³ β
int8: model_int8.onnx (added)
β³ β
uint8: model_uint8.onnx (added)
β³ β
q4: model_q4.onnx (added)
β³ β
q4f16: model_q4f16.onnx (added)
β³ β
bnb4: model_bnb4.onnx (added)
Xenova changed pull request status to merged