Update README.md
Browse files
README.md
CHANGED
|
@@ -36,12 +36,12 @@ give me a ❤️, if you like ;)<br>
|
|
| 36 |
<li>Ger-RAG-BGE-M3 (german)</li>
|
| 37 |
<li>german-roberta</li>
|
| 38 |
</ul>
|
| 39 |
-
Working well, all other its up to you!
|
| 40 |
<br>
|
| 41 |
|
| 42 |
-
<b>Short hints for using:</b><br>
|
| 43 |
-
Set your (Max Tokens)context-lenght 16000t main-model, set your embedder-model (Max Embedding Chunk Length) 1024t,set (Max Context Snippets)
|
| 44 |
-
but in ALLM its cutting all in 1024 character parts, so aprox two times or bit more ~
|
| 45 |
<br>
|
| 46 |
|
| 47 |
-> Ok what that mean!<br>
|
|
@@ -57,23 +57,26 @@ You can play and set for your needs, eg 8-snippets a 2048t, or 28-snippets a 512
|
|
| 57 |
...
|
| 58 |
<br>
|
| 59 |
<b>How embedding and search works:</b><br>
|
| 60 |
-
You have a txt/pdf file maybe 90000words(~300pages). You ask the model lets say "what is described in chapter called XYZ in relation to person ZYX".
|
| 61 |
Now it searches for keywords or similar semantic terms in the document. if it has found them, lets say word and meaning around “XYZ and ZYX” ,
|
| 62 |
now a piece of text 1024token around this word “XYZ/ZYX” is cut out at this point.
|
| 63 |
-
This text snippet is then used for your answer. If, for example, the word “XYZ” occurs 100 times in one file, not all 100 are found
|
|
|
|
| 64 |
<br>
|
| 65 |
-
if you expect multible search results in your docs try
|
| 66 |
<br>
|
| 67 |
if you use snipets-size ~1024t you receive more content, if you use ~256t you receive more facts.
|
| 68 |
<br>
|
| 69 |
A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.
|
| 70 |
<br>
|
|
|
|
|
|
|
| 71 |
If the documents small like 10-20 Pages, its better you copy the whole text inside the prompt.
|
| 72 |
<br>
|
| 73 |
...
|
| 74 |
<br>
|
| 75 |
Nevertheless, the main model is also <b>important</b>! especially to deal with the context length and I don't mean just the theoretical number you can set.
|
| 76 |
-
Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.<br>
|
| 77 |
<br>
|
| 78 |
<b>Important -> The Systemprompt (an example):</b><br>
|
| 79 |
You are a helpful assistant who provides an overview of ... under the aspects of ... .
|
|
@@ -82,7 +85,7 @@ Weight each individual excerpt in order, with the most important excerpts at the
|
|
| 82 |
The context of the entire article should not be given too much weight.
|
| 83 |
Answer the user's question!
|
| 84 |
After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
|
| 85 |
-
<i>(change it for your needs, this example works well when I consult a book about a person and a term related to them)</i><br>
|
| 86 |
or:<br>
|
| 87 |
You are an imaginative storyteller who crafts compelling narratives with depth, creativity, and coherence.
|
| 88 |
Your goal is to develop rich, engaging stories that captivate readers, staying true to the themes, tone, and style appropriate for the given prompt.
|
|
@@ -91,17 +94,19 @@ When generating stories, ensure the coherence in characters, setting, and plot p
|
|
| 91 |
or:<br>
|
| 92 |
You are are a warm and engaging companion who loves to talk about cooking, recipes and the joy of food.
|
| 93 |
Your aim is to share delicious recipes, cooking tips and the stories behind different cultures in a personal, welcoming and knowledgeable way.<br>
|
|
|
|
|
|
|
| 94 |
<br><br>
|
| 95 |
usual models like (works):<br>
|
| 96 |
llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
|
| 97 |
(llama3 or phi3.5 are not working well) <br>
|
| 98 |
|
| 99 |
-
btw. <b>Jinja</b> templates very new ... the usual templates with
|
| 100 |
-
|
| 101 |
|
| 102 |
...
|
| 103 |
<br>
|
| 104 |
<br>
|
|
|
|
| 105 |
on discord (sevenof9)
|
| 106 |
...
|
| 107 |
<br>
|
|
|
|
| 36 |
<li>Ger-RAG-BGE-M3 (german)</li>
|
| 37 |
<li>german-roberta</li>
|
| 38 |
</ul>
|
| 39 |
+
Working well, all other its up to you! (jina and qwen based not yet supported)
|
| 40 |
<br>
|
| 41 |
|
| 42 |
+
<b>Short hints for using (Example for a large context with many expected hits):</b><br>
|
| 43 |
+
Set your (Max Tokens)context-lenght 16000t main-model, set your embedder-model (Max Embedding Chunk Length) 1024t,set (Max Context Snippets) 14,
|
| 44 |
+
but in ALLM its cutting all in 1024 character parts, so aprox two times or bit more ~20!
|
| 45 |
<br>
|
| 46 |
|
| 47 |
-> Ok what that mean!<br>
|
|
|
|
| 57 |
...
|
| 58 |
<br>
|
| 59 |
<b>How embedding and search works:</b><br>
|
| 60 |
+
You have a txt/pdf file maybe 90000words(~300pages) a book. You ask the model lets say "what is described in chapter called XYZ in relation to person ZYX".
|
| 61 |
Now it searches for keywords or similar semantic terms in the document. if it has found them, lets say word and meaning around “XYZ and ZYX” ,
|
| 62 |
now a piece of text 1024token around this word “XYZ/ZYX” is cut out at this point.
|
| 63 |
+
This text snippet is then used for your answer. If, for example, the word “XYZ” occurs 100 times in one file, not all 100 are found.
|
| 64 |
+
If only one snippet corresponds to your question all other snippets can negatively influence your answer because they do not fit the topic (usually 4 to 32 snippet are fine)
|
| 65 |
<br>
|
| 66 |
+
if you expect multible search results in your docs try 16-snippets or more, if you expect only 2 than dont use more!
|
| 67 |
<br>
|
| 68 |
if you use snipets-size ~1024t you receive more content, if you use ~256t you receive more facts.
|
| 69 |
<br>
|
| 70 |
A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.
|
| 71 |
<br>
|
| 72 |
+
If a book has a table of contents or a bibliography, I would delete these pages as they often contain relevant search terms but do not help answer the user's question.
|
| 73 |
+
<br>
|
| 74 |
If the documents small like 10-20 Pages, its better you copy the whole text inside the prompt.
|
| 75 |
<br>
|
| 76 |
...
|
| 77 |
<br>
|
| 78 |
Nevertheless, the main model is also <b>important</b>! especially to deal with the context length and I don't mean just the theoretical number you can set.
|
| 79 |
+
Some models can handle 128k or 1M tokens, but even with 16k input the response with the same snippets as input is worse than with other well developed models.<br>
|
| 80 |
<br>
|
| 81 |
<b>Important -> The Systemprompt (an example):</b><br>
|
| 82 |
You are a helpful assistant who provides an overview of ... under the aspects of ... .
|
|
|
|
| 85 |
The context of the entire article should not be given too much weight.
|
| 86 |
Answer the user's question!
|
| 87 |
After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
|
| 88 |
+
<i>(change it for your needs, this example works well when I consult a book about a person and a term related to them, the explanation was just a test for myself)</i><br>
|
| 89 |
or:<br>
|
| 90 |
You are an imaginative storyteller who crafts compelling narratives with depth, creativity, and coherence.
|
| 91 |
Your goal is to develop rich, engaging stories that captivate readers, staying true to the themes, tone, and style appropriate for the given prompt.
|
|
|
|
| 94 |
or:<br>
|
| 95 |
You are are a warm and engaging companion who loves to talk about cooking, recipes and the joy of food.
|
| 96 |
Your aim is to share delicious recipes, cooking tips and the stories behind different cultures in a personal, welcoming and knowledgeable way.<br>
|
| 97 |
+
<br>
|
| 98 |
+
The system prompt is weighted with a certain amount of influence around your question. You can easily test it once without or with a nonsensical system prompt.
|
| 99 |
<br><br>
|
| 100 |
usual models like (works):<br>
|
| 101 |
llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
|
| 102 |
(llama3 or phi3.5 are not working well) <br>
|
| 103 |
|
| 104 |
+
btw. <b>Jinja</b> templates very new ... the usual templates with usual models are fine, but merged models have a lot of optimization potential (but dont ask me iam not a coder)<br>
|
|
|
|
| 105 |
|
| 106 |
...
|
| 107 |
<br>
|
| 108 |
<br>
|
| 109 |
+
...
|
| 110 |
on discord (sevenof9)
|
| 111 |
...
|
| 112 |
<br>
|