File size: 1,633 Bytes
a3b87af
be61277
a3b87af
 
 
 
 
1642df8
829bf5f
be61277
 
5f7dba7
be61277
 
 
 
 
5f7dba7
be61277
 
 
 
5f7dba7
be61277
 
5f7dba7
9b7aa7a
 
 
 
 
 
 
 
 
be61277
 
 
5f7dba7
be61277
 
 
 
 
 
 
 
 
 
 
5f7dba7
be61277
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
---
license: mit
language:
- en
pipeline_tag: text-generation
tags:
- gguf
base_model:
- GoofyLM/N1
---

![banner by CroissantWhyNot](banner.png)

*Banner by [Croissant](https://huggingface.co/CroissantWhyNot)*

# N1 - A Chain-of-Thought Language Model

N1 is a small, experimental Chain-of-Thought (COT) model based on the LLaMA architecture, developed by GoofyLM.

## Model Details

- **Architecture**: LLaMA-based
- **Parameter Count**: 135M
- **Training Data**: Closed-source dataset
- **Special Features**: Chain-of-Thought reasoning capabilities
- **Note**: The model often shows "schizophrenia"
- **Note**: You may need to add this Jinja to the model:
```jinja
{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system
You are a helpful AI assistant named N1, trained by GoofyLM<|im_end|>
' }}{% endif %}{{'<|im_start|>' + message['role'] + '
' + message['content'] + '<|im_end|>' + '
'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
' }}{% endif %}
```

## Intended Use

This model is designed for text generation tasks with a focus on reasoning through problems step-by-step (using its Chain-of-Thought).

## Limitations

- Small parameter size may limit reasoning capabilities
- May produce unstable or inconsistent outputs
- Not suitable for production use without further testing

---

## Usage

The model can be loaded using the following: 

### llama-cpp-python:

```python
from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="GoofyLM/N1",
	filename="N1_Q8_0.gguf",
)
```

### Ollama:
```python
ollama run hf.co/GoofyLM/N1:Q4_K_M
```