Update README.md
Browse files
README.md
CHANGED
|
@@ -109,7 +109,7 @@ for chat in prompts:
|
|
| 109 |
print(f"Response:\n{response}\n{'-'*80}\n")
|
| 110 |
```
|
| 111 |
|
| 112 |
-
|
| 113 |
1. **System Prompts:** Each prompt defines a different role or persona for the AI to adopt.
|
| 114 |
2. **User Prompts:** These specify the context or task for the assistant, ranging from teaching to storytelling or career advice.
|
| 115 |
3. **Looping Through Prompts:** Each prompt is processed in a loop to showcase the model's versatility.
|
|
@@ -125,3 +125,93 @@ You can expand the list of prompts to explore a variety of scenarios and respons
|
|
| 125 |
# **Technical Details**
|
| 126 |
|
| 127 |
Triangulum 10B employs a state-of-the-art autoregressive architecture inspired by LLaMA. The optimized transformer framework ensures both efficiency and scalability, making it suitable for a variety of use cases.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 109 |
print(f"Response:\n{response}\n{'-'*80}\n")
|
| 110 |
```
|
| 111 |
|
| 112 |
+
# **Key Adjustments**
|
| 113 |
1. **System Prompts:** Each prompt defines a different role or persona for the AI to adopt.
|
| 114 |
2. **User Prompts:** These specify the context or task for the assistant, ranging from teaching to storytelling or career advice.
|
| 115 |
3. **Looping Through Prompts:** Each prompt is processed in a loop to showcase the model's versatility.
|
|
|
|
| 125 |
# **Technical Details**
|
| 126 |
|
| 127 |
Triangulum 10B employs a state-of-the-art autoregressive architecture inspired by LLaMA. The optimized transformer framework ensures both efficiency and scalability, making it suitable for a variety of use cases.
|
| 128 |
+
|
| 129 |
+
# **How to Run Triangulum 10B on Ollama Locally**
|
| 130 |
+
|
| 131 |
+
```markdown
|
| 132 |
+
# How to Run Ollama Locally
|
| 133 |
+
|
| 134 |
+
This guide demonstrates the power of using open-source LLMs locally, showcasing examples with different open-source models for various use cases. By the end, you'll be equipped to run any future open-source LLM models with ease.
|
| 135 |
+
|
| 136 |
+
---
|
| 137 |
+
|
| 138 |
+
## Example 1: How to Run the Triangulum-10B Model
|
| 139 |
+
|
| 140 |
+
The **Triangulum-10B** model is an open-source LLM known for its capabilities across text-based tasks. We'll interact with it similarly to ChatGPT, but run it locally with support for quants.
|
| 141 |
+
|
| 142 |
+
### Step 1: Download the Model
|
| 143 |
+
|
| 144 |
+
First, download the **Triangulum-10B-F16.gguf** model using the following command:
|
| 145 |
+
|
| 146 |
+
```bash
|
| 147 |
+
ollama run triangulum-10b-f16.gguf
|
| 148 |
+
```
|
| 149 |
+
|
| 150 |
+
### Step 2: Model Initialization and Download
|
| 151 |
+
|
| 152 |
+
Upon running the command, Ollama will initialize and download the model files. You should see output similar to the following:
|
| 153 |
+
|
| 154 |
+
```plaintext
|
| 155 |
+
pulling manifest
|
| 156 |
+
pulling 8934d96d3f08... 100% ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ 3.8 GB
|
| 157 |
+
pulling 8c17c2ebb0ea... 100% ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ 7.0 KB
|
| 158 |
+
pulling 7c23fb36d801... 100% ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ 4.8 KB
|
| 159 |
+
pulling 2e0493f67d0c... 100% ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ 59 B
|
| 160 |
+
pulling fa304d675061... 100% ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ 91 B
|
| 161 |
+
pulling 42ba7f8a01dd... 100% ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ 557 B
|
| 162 |
+
verifying sha256 digest
|
| 163 |
+
writing manifest
|
| 164 |
+
removing any unused layers
|
| 165 |
+
success
|
| 166 |
+
>>> Send a message (/? for help)
|
| 167 |
+
```
|
| 168 |
+
|
| 169 |
+
### Step 3: Interact with the Model
|
| 170 |
+
|
| 171 |
+
Once the model is ready, you can send a prompt. For example, let's ask:
|
| 172 |
+
|
| 173 |
+
```plaintext
|
| 174 |
+
>>> What can you do for me?
|
| 175 |
+
```
|
| 176 |
+
|
| 177 |
+
If you asked the same question, you should receive a response like this:
|
| 178 |
+
|
| 179 |
+
```plaintext
|
| 180 |
+
As a responsible AI language model, I am here to assist you with any questions or tasks you may have. Here are some examples of things I can help with:
|
| 181 |
+
|
| 182 |
+
1. Answering questions: I can provide information on a wide range of topics, from science and technology to history and culture.
|
| 183 |
+
2. Generating ideas: I can help you brainstorm ideas for creative projects, or provide suggestions for solving problems.
|
| 184 |
+
3. Writing assistance: I can help you with writing tasks such as proofreading, editing, and suggesting alternative words or phrases.
|
| 185 |
+
4. Translation: I can translate text from one language to another.
|
| 186 |
+
5. Summarizing content: I can summarize long pieces of text, such as articles or documents, into shorter, more digestible versions.
|
| 187 |
+
6. Creativity: I can help you generate creative ideas for stories, poems, or other forms of writing.
|
| 188 |
+
7. Language learning: I can assist you in learning a new language by providing grammar explanations, vocabulary lists, and practice exercises.
|
| 189 |
+
8. Chatting: I'm here to chat with you and provide a response to any question or topic you'd like to discuss.
|
| 190 |
+
|
| 191 |
+
Please let me know if there is anything specific you would like me to help you with.
|
| 192 |
+
```
|
| 193 |
+
|
| 194 |
+
### Step 4: Exit the Program
|
| 195 |
+
|
| 196 |
+
To exit the program, simply type:
|
| 197 |
+
|
| 198 |
+
```plaintext
|
| 199 |
+
/exit
|
| 200 |
+
```
|
| 201 |
+
## Example 2: Running Multi-Modal Models (Future Use)
|
| 202 |
+
|
| 203 |
+
Ollama supports running multi-modal models where you can send images and ask questions based on them. This section will be updated as more models become available.
|
| 204 |
+
|
| 205 |
+
## Notes on Using Quantized Models
|
| 206 |
+
|
| 207 |
+
Quantized models like **triangulum-10b-f16.gguf** are optimized for performance on resource-constrained hardware, making it accessible for local inference.
|
| 208 |
+
|
| 209 |
+
1. Ensure your system has sufficient VRAM or CPU resources.
|
| 210 |
+
2. Use the `.gguf` model format for compatibility with Ollama.
|
| 211 |
+
|
| 212 |
+
## Conclusion
|
| 213 |
+
|
| 214 |
+
Running the **Triangulum-10B** model with Ollama provides a robust way to leverage open-source LLMs locally for diverse use cases. By following these steps, you can explore the capabilities of other open-source models in the future.
|
| 215 |
+
|
| 216 |
+
Happy experimenting!
|
| 217 |
+
```
|