Update README.md
Browse files
README.md
CHANGED
|
@@ -68,6 +68,9 @@ If the documents small like 10-20 Pages, its better you copy the whole text insi
|
|
| 68 |
<br>
|
| 69 |
...
|
| 70 |
<br>
|
|
|
|
|
|
|
|
|
|
| 71 |
<b>Important -> The Systemprompt (an example):</b><br>
|
| 72 |
You are a helpful assistant who provides an overview of ... under the aspects of ... .
|
| 73 |
You use attached excerpts from the collection to generate your answers!
|
|
@@ -77,6 +80,10 @@ Answer the user's question!
|
|
| 77 |
After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
|
| 78 |
(change it for your needs, this example works well when I consult a book about a person and a term related to them)
|
| 79 |
<br>
|
|
|
|
|
|
|
|
|
|
|
|
|
| 80 |
...
|
| 81 |
<br>
|
| 82 |
<br>
|
|
|
|
| 68 |
<br>
|
| 69 |
...
|
| 70 |
<br>
|
| 71 |
+
Nevertheless, the main model is also important! especially to deal with the context length and I don't mean just the theoretical number you can set.
|
| 72 |
+
Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.
|
| 73 |
+
<br>
|
| 74 |
<b>Important -> The Systemprompt (an example):</b><br>
|
| 75 |
You are a helpful assistant who provides an overview of ... under the aspects of ... .
|
| 76 |
You use attached excerpts from the collection to generate your answers!
|
|
|
|
| 80 |
After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
|
| 81 |
(change it for your needs, this example works well when I consult a book about a person and a term related to them)
|
| 82 |
<br>
|
| 83 |
+
usual models like (works):
|
| 84 |
+
llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
|
| 85 |
+
(llama3 or phi3.5 are not working well)
|
| 86 |
+
|
| 87 |
...
|
| 88 |
<br>
|
| 89 |
<br>
|