Commit ·
31fcc9b
1
Parent(s): f54a5f3
Add professional model card and metadata
Browse files- README.md +63 -5
- TM-1B-Q80.gguf +3 -0
README.md
CHANGED
|
@@ -1,14 +1,72 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
-
datasets:
|
| 4 |
-
- tiiuae/falcon-refinedweb
|
| 5 |
-
- HuggingFaceTB/cosmopedia
|
| 6 |
language:
|
| 7 |
- en
|
|
|
|
| 8 |
pipeline_tag: text-generation
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
tags:
|
|
|
|
| 10 |
- gguf
|
| 11 |
- quantized
|
| 12 |
-
-
|
| 13 |
- llama-cpp
|
| 14 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
+
base_model: []
|
| 6 |
pipeline_tag: text-generation
|
| 7 |
+
datasets:
|
| 8 |
+
- HuggingFaceTB/cosmopedia
|
| 9 |
+
- tiiuae/falcon-refinedweb
|
| 10 |
+
library_name: gguf
|
| 11 |
tags:
|
| 12 |
+
- text-generation
|
| 13 |
- gguf
|
| 14 |
- quantized
|
| 15 |
+
- 1b
|
| 16 |
- llama-cpp
|
| 17 |
+
---
|
| 18 |
+
|
| 19 |
+
# PT1S-1B-Q8.gguf
|
| 20 |
+
|
| 21 |
+
This model is a 1-billion parameter text generation model trained on a high-quality mixture of synthetic and web-crawled data. It is optimized for efficiency and performance in a small footprint.
|
| 22 |
+
|
| 23 |
+
## Model Details
|
| 24 |
+
|
| 25 |
+
- **Model Type:** Text Generation
|
| 26 |
+
- **Parameters:** 1B
|
| 27 |
+
- **Quantization:** Q8_0 (8-bit quantization for high precision with reduced memory)
|
| 28 |
+
- **Training Data:**
|
| 29 |
+
- [HuggingFaceTB/cosmopedia](https://huggingface.co/datasets/HuggingFaceTB/cosmopedia)
|
| 30 |
+
- [tiiuae/falcon-refinedweb](https://huggingface.co/datasets/tiiuae/falcon-refinedweb)
|
| 31 |
+
- **Language(s):** English
|
| 32 |
+
- **License:** Apache 2.0
|
| 33 |
+
|
| 34 |
+
## Training Information
|
| 35 |
+
|
| 36 |
+
The model was trained on a curated blend of:
|
| 37 |
+
1. **Cosmopedia**: A large-scale synthetic dataset designed to provide high-quality educational content across various domains.
|
| 38 |
+
2. **Falcon RefinedWeb**: A massive, filtered web dataset that provides broad world knowledge and linguistic diversity.
|
| 39 |
+
|
| 40 |
+
This combination allows the model to have both structured knowledge from synthetic sources and a natural "web-aware" conversational style.
|
| 41 |
+
|
| 42 |
+
## Usage
|
| 43 |
+
|
| 44 |
+
### llama.cpp
|
| 45 |
+
|
| 46 |
+
You can use this model with [llama.cpp](https://github.com/ggerganov/llama.cpp) by running:
|
| 47 |
+
|
| 48 |
+
```bash
|
| 49 |
+
./main -m PT1S-1B-Q8.gguf -p "Once upon a time," -n 128
|
| 50 |
+
```
|
| 51 |
+
|
| 52 |
+
### Python (via llama-cpp-python)
|
| 53 |
+
|
| 54 |
+
```python
|
| 55 |
+
from llama_cpp import Llama
|
| 56 |
+
|
| 57 |
+
llm = Llama(model_path="./PT1S-1B-Q8.gguf")
|
| 58 |
+
output = llm("Q: What is the importance of cosmopedia dataset? A:", max_tokens=100)
|
| 59 |
+
print(output)
|
| 60 |
+
```
|
| 61 |
+
|
| 62 |
+
## Intended Use
|
| 63 |
+
|
| 64 |
+
This model is ideal for:
|
| 65 |
+
- Lightweight text generation tasks.
|
| 66 |
+
- Educational applications.
|
| 67 |
+
- On-device inference where memory is limited.
|
| 68 |
+
- Research into small language models (SLMs).
|
| 69 |
+
|
| 70 |
+
## Limitations and Bias
|
| 71 |
+
|
| 72 |
+
While trained on filtered data, small models may still exhibit biases or generate incorrect information (hallucinations). Users should always verify the output of the model for critical applications.
|
TM-1B-Q80.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:432f310a77f4650a88d0fd59ecdd7cebed8d684bafea53cbff0473542964f0c3
|
| 3 |
+
size 1321083008
|