Update README.md
Browse files
README.md
CHANGED
|
@@ -68,7 +68,7 @@ If the documents small like 10-20 Pages, its better you copy the whole text insi
|
|
| 68 |
<br>
|
| 69 |
...
|
| 70 |
<br>
|
| 71 |
-
Nevertheless, the main model is also important
|
| 72 |
Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.<br>
|
| 73 |
<br>
|
| 74 |
<b>Important -> The Systemprompt (an example):</b><br>
|
|
@@ -79,12 +79,12 @@ The context of the entire article should not be given too much weight.
|
|
| 79 |
Answer the user's question!
|
| 80 |
After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
|
| 81 |
(change it for your needs, this example works well when I consult a book about a person and a term related to them)
|
| 82 |
-
<br>
|
| 83 |
usual models like (works):<br>
|
| 84 |
llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
|
| 85 |
(llama3 or phi3.5 are not working well) <br>
|
| 86 |
|
| 87 |
-
btw. Jinja templates very new ... the usual templates with
|
| 88 |
|
| 89 |
|
| 90 |
...
|
|
|
|
| 68 |
<br>
|
| 69 |
...
|
| 70 |
<br>
|
| 71 |
+
Nevertheless, the main model is also <b>important</b>! especially to deal with the context length and I don't mean just the theoretical number you can set.
|
| 72 |
Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.<br>
|
| 73 |
<br>
|
| 74 |
<b>Important -> The Systemprompt (an example):</b><br>
|
|
|
|
| 79 |
Answer the user's question!
|
| 80 |
After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
|
| 81 |
(change it for your needs, this example works well when I consult a book about a person and a term related to them)
|
| 82 |
+
<br><br>
|
| 83 |
usual models like (works):<br>
|
| 84 |
llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
|
| 85 |
(llama3 or phi3.5 are not working well) <br>
|
| 86 |
|
| 87 |
+
btw. <b>Jinja</b> templates very new ... the usual templates with new/usual models are fine, but merged models have a lot of optimization potential (but dont ask me iam not a coder)<br>
|
| 88 |
|
| 89 |
|
| 90 |
...
|