|
|
--- |
|
|
language: |
|
|
- en |
|
|
- de |
|
|
- fr |
|
|
- it |
|
|
- pt |
|
|
- hi |
|
|
- es |
|
|
- th |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- facebook |
|
|
- meta |
|
|
- pytorch |
|
|
- llama |
|
|
- llama-3 |
|
|
- open4bits |
|
|
license: llama3.2 |
|
|
base_model: |
|
|
- meta-llama/Llama-3.2-1B |
|
|
--- |
|
|
# Open4bits / llama3.2-1b-gguf |
|
|
|
|
|
This repository provides the **LLaMA 3.2-1B model converted to GGUF format**, published by Open4bits to enable highly efficient local inference with reduced memory usage and broad CPU compatibility. |
|
|
|
|
|
The underlying LLaMA 3.2 model and architecture are **owned by Meta AI**. This repository contains only a quantized GGUF conversion of the original model weights. |
|
|
|
|
|
The model is designed for fast, lightweight text generation and instruction-following tasks and is well suited for resource-constrained environments. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Overview |
|
|
|
|
|
LLaMA (Large Language Model Meta AI) is a family of transformer-based language models developed by Meta AI. |
|
|
This release uses the **3.2 variant with 1 billion parameters**, striking a balance between performance and efficiency. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Details |
|
|
|
|
|
* **Architecture:** LLaMA 3.2 |
|
|
* **Parameters:** ~1 billion |
|
|
* **Format:** GGUF (quantized) |
|
|
* **Task:** Text generation, instruction following |
|
|
* **Weight tying:** Preserved |
|
|
* **Compatibility:** GGUF-compatible inference runtimes (CPU-focused) |
|
|
|
|
|
Compared to larger LLaMA variants, this model offers significantly faster inference with lower memory requirements, with proportionally reduced capacity for complex reasoning. |
|
|
|
|
|
--- |
|
|
|
|
|
## Intended Use |
|
|
|
|
|
This model is intended for: |
|
|
|
|
|
* Local text generation and chat applications |
|
|
* CPU-based or low-resource deployments |
|
|
* Research, experimentation, and prototyping |
|
|
* Offline or self-hosted AI systems |
|
|
|
|
|
--- |
|
|
|
|
|
## Limitations |
|
|
|
|
|
* Lower generation quality compared to larger LLaMA 3.2 models |
|
|
* Output quality depends on prompt design and decoding settings |
|
|
* Not fine-tuned for domain-specific or high-precision tasks |
|
|
|
|
|
--- |
|
|
|
|
|
## License |
|
|
|
|
|
This model is released under the **original LLaMA 3.2 license terms** as defined by Meta AI. |
|
|
Users must comply with the licensing conditions of the base LLaMA 3.2-1B model. |
|
|
|
|
|
--- |
|
|
|
|
|
## Support |
|
|
|
|
|
If you find this model useful, please consider supporting the project. |
|
|
Your support helps Open4bits continue releasing and maintaining high-quality open models for the community. |
|
|
|