fmasterpro27 commited on
Commit
d2fc335
·
verified ·
1 Parent(s): 0dc5117

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +64 -1
README.md CHANGED
@@ -1,4 +1,67 @@
1
  ---
2
  base_model:
3
  - inference-net/Schematron-3B
4
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  base_model:
3
  - inference-net/Schematron-3B
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - open4bits
7
+ ---
8
+ # Open4bits / Schematron-3B-GGUF
9
+
10
+ This repository provides the **Schematron-3B model converted to GGUF format**, published by Open4bits to enable efficient local inference with reduced memory usage and broad CPU compatibility.
11
+
12
+ The underlying base model is **meta-llama/Llama-3.2-3B-Instruct**, fine-tuned by Inference-Net. This repository contains a quantized GGUF conversion of the fine-tuned model weights produced by Open4bits.
13
+
14
+ The model is designed for instruction-based text generation tasks and is suitable for resource-constrained and local deployments.
15
+
16
+ ---
17
+
18
+ ## Model Overview
19
+
20
+ Schematron-3B is an instruction-tuned language model built on the **LLaMA 3.2-3B architecture**. After fine-tuning by Inference-Net for enhanced instruction following and generation quality, the model has been quantized and released in GGUF format to support efficient CPU-friendly inference.
21
+
22
+ ---
23
+
24
+ ## Model Details
25
+
26
+ * **Base Model:** meta-llama/Llama-3.2-3B-Instruct
27
+ * **Fine-Tuned By:** Inference-Net
28
+ * **Parameters:** ~3 billion
29
+ * **Format:** GGUF (quantized)
30
+ * **Task:** Instruction-based text generation
31
+ * **Weight tying:** Preserved
32
+ * **Compatibility:** GGUF-compatible inference engines and CPU environments
33
+
34
+ This quantized release is designed to balance performance and resource efficiency while maintaining strong instruction following capabilities.
35
+
36
+ ---
37
+
38
+ ## Intended Use
39
+
40
+ This model is intended for:
41
+
42
+ * Instruction-guided text generation
43
+ * Local and CPU-based inference workflows
44
+ * Research, prototyping, and experimentation
45
+ * Self-hosted or offline AI systems
46
+
47
+ ---
48
+
49
+ ## Limitations
50
+
51
+ * Reduced generation quality compared to larger or full-precision variants
52
+ * Performance depends on prompt design and inference parameters
53
+ * Not fine-tuned for highly specialized or domain-specific tasks
54
+
55
+ ---
56
+
57
+ ## License
58
+
59
+ This model follows the **original LLaMA 3.2 licensing terms** as defined by Meta AI.
60
+ Users must comply with the licensing conditions of the base model and the fine-tuning provider.
61
+
62
+ ---
63
+
64
+ ## Support
65
+
66
+ If you find this model valuable, please consider supporting the project.
67
+ Your support helps Open4bits continue releasing and maintaining high-quality quantized models for the community.