Open4bits / Schematron-3B-GGUF

This repository provides the Schematron-3B model converted to GGUF format, published by Open4bits to enable efficient local inference with reduced memory usage and broad CPU compatibility.

The underlying base model is meta-llama/Llama-3.2-3B-Instruct, fine-tuned by Inference-Net. This repository contains a quantized GGUF conversion of the fine-tuned model weights produced by Open4bits.

The model is designed for instruction-based text generation tasks and is suitable for resource-constrained and local deployments.


Model Overview

Schematron-3B is an instruction-tuned language model built on the LLaMA 3.2-3B architecture. After fine-tuning by Inference-Net for enhanced instruction following and generation quality, the model has been quantized and released in GGUF format to support efficient CPU-friendly inference.


Model Details

  • Base Model: meta-llama/Llama-3.2-3B-Instruct
  • Fine-Tuned By: Inference-Net
  • Parameters: ~3 billion
  • Format: GGUF (quantized)
  • Task: Instruction-based text generation
  • Weight tying: Preserved
  • Compatibility: GGUF-compatible inference engines and CPU environments

This quantized release is designed to balance performance and resource efficiency while maintaining strong instruction following capabilities.


Intended Use

This model is intended for:

  • Instruction-guided text generation
  • Local and CPU-based inference workflows
  • Research, prototyping, and experimentation
  • Self-hosted or offline AI systems

Limitations

  • Reduced generation quality compared to larger or full-precision variants
  • Performance depends on prompt design and inference parameters
  • Not fine-tuned for highly specialized or domain-specific tasks

License

This model follows the original LLaMA 3.2 licensing terms as defined by Meta AI. Users must comply with the licensing conditions of the base model and the fine-tuning provider.


Support

If you find this model valuable, please consider supporting the project. Your support helps Open4bits continue releasing and maintaining high-quality quantized models for the community.

Downloads last month
258
GGUF
Model size
3B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Open4bits/Schematron-3B-gguf

Quantized
(9)
this model

Collection including Open4bits/Schematron-3B-gguf