File size: 1,998 Bytes
abdf333
 
81dcf54
 
 
 
 
 
550cc0e
abdf333
 
5ca3198
 
a91b842
672c426
a91b842
672c426
6def784
a91b842
 
 
 
6def784
a91b842
 
6def784
22b9a0d
 
 
 
 
 
 
 
 
04b8b68
a91b842
 
6def784
a91b842
 
 
 
 
 
 
5c1cb7f
 
a91b842
 
5c1cb7f
 
 
a91b842
 
 
 
 
 
5c1cb7f
 
 
 
 
 
 
 
b426132
caac39b
5c1cb7f
 
 
 
 
b426132
5c1cb7f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
---
license: mit
language:
- en
pipeline_tag: text-generation
library_name: transformers
tags:
- text-generation-inference
new_version: GoofyLM/N2-Nemo
---

![banner by CroissantWhyNot](banner.png)

*Banner by [Croissant](https://huggingface.co/CroissantWhyNot)*

# N1 - A Chain-of-Thought Language Model

N1 is a small, experimental Chain-of-Thought (COT) model based on the LLaMA architecture, developed by GoofyLM.

## Model Details

- **Architecture**: LLaMA-based
- **Parameter Count**: 135M
- **Training Data**: Closed-source dataset
- **Special Features**: Chain-of-Thought reasoning capabilities
- **Note**: The model often shows "schizophrenia"
- **Note**: You may need to add this Jinja to the model:
```jinja
{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
You are a helpful AI assistant named N1, trained by GoofyLM<|im_end|>
' }}{% endif %}{{'<|im_start|>' + message['role'] + '
' + message['content'] + '<|im_end|>' + '
'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
' }}{% endif %}
```
- ONNX available at [onnx-community/N1-ONNX](https://huggingface.co/onnx-community/N1-ONNX).
## Intended Use

This model is designed for text generation tasks with a focus on reasoning through problems step-by-step (using its Chain-of-Thought).

## Limitations

- Small parameter size may limit reasoning capabilities
- May produce unstable or inconsistent outputs
- Not suitable for production use without further testing

---

## Usage

The model can be loaded using the following: 

### Transformers:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("GoofyLM/N1")
tokenizer = AutoTokenizer.from_pretrained("GoofyLM/N1")
```

### llama-cpp-python:

```python
from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="GoofyLM/N1-quant",
	filename="N1_Q8_0.gguf",
)
```

### Ollama:
```python
ollama run hf.co/GoofyLM/N1-quant:Q8_0
```