Text Generation
GGUF
English
quantized
1b
llama-cpp
imatrix
conversational
MrDevCoder01 commited on
Commit
31fcc9b
·
1 Parent(s): f54a5f3

Add professional model card and metadata

Browse files
Files changed (2) hide show
  1. README.md +63 -5
  2. TM-1B-Q80.gguf +3 -0
README.md CHANGED
@@ -1,14 +1,72 @@
1
  ---
2
  license: apache-2.0
3
- datasets:
4
- - tiiuae/falcon-refinedweb
5
- - HuggingFaceTB/cosmopedia
6
  language:
7
  - en
 
8
  pipeline_tag: text-generation
 
 
 
 
9
  tags:
 
10
  - gguf
11
  - quantized
12
- - 1B
13
  - llama-cpp
14
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
 
 
 
3
  language:
4
  - en
5
+ base_model: []
6
  pipeline_tag: text-generation
7
+ datasets:
8
+ - HuggingFaceTB/cosmopedia
9
+ - tiiuae/falcon-refinedweb
10
+ library_name: gguf
11
  tags:
12
+ - text-generation
13
  - gguf
14
  - quantized
15
+ - 1b
16
  - llama-cpp
17
+ ---
18
+
19
+ # PT1S-1B-Q8.gguf
20
+
21
+ This model is a 1-billion parameter text generation model trained on a high-quality mixture of synthetic and web-crawled data. It is optimized for efficiency and performance in a small footprint.
22
+
23
+ ## Model Details
24
+
25
+ - **Model Type:** Text Generation
26
+ - **Parameters:** 1B
27
+ - **Quantization:** Q8_0 (8-bit quantization for high precision with reduced memory)
28
+ - **Training Data:**
29
+ - [HuggingFaceTB/cosmopedia](https://huggingface.co/datasets/HuggingFaceTB/cosmopedia)
30
+ - [tiiuae/falcon-refinedweb](https://huggingface.co/datasets/tiiuae/falcon-refinedweb)
31
+ - **Language(s):** English
32
+ - **License:** Apache 2.0
33
+
34
+ ## Training Information
35
+
36
+ The model was trained on a curated blend of:
37
+ 1. **Cosmopedia**: A large-scale synthetic dataset designed to provide high-quality educational content across various domains.
38
+ 2. **Falcon RefinedWeb**: A massive, filtered web dataset that provides broad world knowledge and linguistic diversity.
39
+
40
+ This combination allows the model to have both structured knowledge from synthetic sources and a natural "web-aware" conversational style.
41
+
42
+ ## Usage
43
+
44
+ ### llama.cpp
45
+
46
+ You can use this model with [llama.cpp](https://github.com/ggerganov/llama.cpp) by running:
47
+
48
+ ```bash
49
+ ./main -m PT1S-1B-Q8.gguf -p "Once upon a time," -n 128
50
+ ```
51
+
52
+ ### Python (via llama-cpp-python)
53
+
54
+ ```python
55
+ from llama_cpp import Llama
56
+
57
+ llm = Llama(model_path="./PT1S-1B-Q8.gguf")
58
+ output = llm("Q: What is the importance of cosmopedia dataset? A:", max_tokens=100)
59
+ print(output)
60
+ ```
61
+
62
+ ## Intended Use
63
+
64
+ This model is ideal for:
65
+ - Lightweight text generation tasks.
66
+ - Educational applications.
67
+ - On-device inference where memory is limited.
68
+ - Research into small language models (SLMs).
69
+
70
+ ## Limitations and Bias
71
+
72
+ While trained on filtered data, small models may still exhibit biases or generate incorrect information (hallucinations). Users should always verify the output of the model for critical applications.
TM-1B-Q80.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:432f310a77f4650a88d0fd59ecdd7cebed8d684bafea53cbff0473542964f0c3
3
+ size 1321083008