kalle07 commited on
Commit
da8a392
·
verified ·
1 Parent(s): 57fe948

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -11
README.md CHANGED
@@ -36,12 +36,12 @@ give me a ❤️, if you like ;)<br>
36
  <li>Ger-RAG-BGE-M3 (german)</li>
37
  <li>german-roberta</li>
38
  </ul>
39
- Working well, all other its up to you!
40
  <br>
41
 
42
- <b>Short hints for using:</b><br>
43
- Set your (Max Tokens)context-lenght 16000t main-model, set your embedder-model (Max Embedding Chunk Length) 1024t,set (Max Context Snippets) usual 14,
44
- but in ALLM its cutting all in 1024 character parts, so aprox two times or bit more ~30!
45
  <br>
46
 
47
  -> Ok what that mean!<br>
@@ -57,23 +57,26 @@ You can play and set for your needs, eg 8-snippets a 2048t, or 28-snippets a 512
57
  ...
58
  <br>
59
  <b>How embedding and search works:</b><br>
60
- You have a txt/pdf file maybe 90000words(~300pages). You ask the model lets say "what is described in chapter called XYZ in relation to person ZYX".
61
  Now it searches for keywords or similar semantic terms in the document. if it has found them, lets say word and meaning around “XYZ and ZYX” ,
62
  now a piece of text 1024token around this word “XYZ/ZYX” is cut out at this point.
63
- This text snippet is then used for your answer. If, for example, the word “XYZ” occurs 100 times in one file, not all 100 are found (usually only 4 to 32 snippet are used)
 
64
  <br>
65
- if you expect multible search results in your docs try 16snippets or more, if you expect only 2 than dont use more!
66
  <br>
67
  if you use snipets-size ~1024t you receive more content, if you use ~256t you receive more facts.
68
  <br>
69
  A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.
70
  <br>
 
 
71
  If the documents small like 10-20 Pages, its better you copy the whole text inside the prompt.
72
  <br>
73
  ...
74
  <br>
75
  Nevertheless, the main model is also <b>important</b>! especially to deal with the context length and I don't mean just the theoretical number you can set.
76
- Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.<br>
77
  <br>
78
  <b>Important -> The Systemprompt (an example):</b><br>
79
  You are a helpful assistant who provides an overview of ... under the aspects of ... .
@@ -82,7 +85,7 @@ Weight each individual excerpt in order, with the most important excerpts at the
82
  The context of the entire article should not be given too much weight.
83
  Answer the user's question!
84
  After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
85
- <i>(change it for your needs, this example works well when I consult a book about a person and a term related to them)</i><br>
86
  or:<br>
87
  You are an imaginative storyteller who crafts compelling narratives with depth, creativity, and coherence.
88
  Your goal is to develop rich, engaging stories that captivate readers, staying true to the themes, tone, and style appropriate for the given prompt.
@@ -91,17 +94,19 @@ When generating stories, ensure the coherence in characters, setting, and plot p
91
  or:<br>
92
  You are are a warm and engaging companion who loves to talk about cooking, recipes and the joy of food.
93
  Your aim is to share delicious recipes, cooking tips and the stories behind different cultures in a personal, welcoming and knowledgeable way.<br>
 
 
94
  <br><br>
95
  usual models like (works):<br>
96
  llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
97
  (llama3 or phi3.5 are not working well) <br>
98
 
99
- btw. <b>Jinja</b> templates very new ... the usual templates with new/usual models are fine, but merged models have a lot of optimization potential (but dont ask me iam not a coder)<br>
100
-
101
 
102
  ...
103
  <br>
104
  <br>
 
105
  on discord (sevenof9)
106
  ...
107
  <br>
 
36
  <li>Ger-RAG-BGE-M3 (german)</li>
37
  <li>german-roberta</li>
38
  </ul>
39
+ Working well, all other its up to you! (jina and qwen based not yet supported)
40
  <br>
41
 
42
+ <b>Short hints for using (Example for a large context with many expected hits):</b><br>
43
+ Set your (Max Tokens)context-lenght 16000t main-model, set your embedder-model (Max Embedding Chunk Length) 1024t,set (Max Context Snippets) 14,
44
+ but in ALLM its cutting all in 1024 character parts, so aprox two times or bit more ~20!
45
  <br>
46
 
47
  -> Ok what that mean!<br>
 
57
  ...
58
  <br>
59
  <b>How embedding and search works:</b><br>
60
+ You have a txt/pdf file maybe 90000words(~300pages) a book. You ask the model lets say "what is described in chapter called XYZ in relation to person ZYX".
61
  Now it searches for keywords or similar semantic terms in the document. if it has found them, lets say word and meaning around “XYZ and ZYX” ,
62
  now a piece of text 1024token around this word “XYZ/ZYX” is cut out at this point.
63
+ This text snippet is then used for your answer. If, for example, the word “XYZ” occurs 100 times in one file, not all 100 are found.
64
+ If only one snippet corresponds to your question all other snippets can negatively influence your answer because they do not fit the topic (usually 4 to 32 snippet are fine)
65
  <br>
66
+ if you expect multible search results in your docs try 16-snippets or more, if you expect only 2 than dont use more!
67
  <br>
68
  if you use snipets-size ~1024t you receive more content, if you use ~256t you receive more facts.
69
  <br>
70
  A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.
71
  <br>
72
+ If a book has a table of contents or a bibliography, I would delete these pages as they often contain relevant search terms but do not help answer the user's question.
73
+ <br>
74
  If the documents small like 10-20 Pages, its better you copy the whole text inside the prompt.
75
  <br>
76
  ...
77
  <br>
78
  Nevertheless, the main model is also <b>important</b>! especially to deal with the context length and I don't mean just the theoretical number you can set.
79
+ Some models can handle 128k or 1M tokens, but even with 16k input the response with the same snippets as input is worse than with other well developed models.<br>
80
  <br>
81
  <b>Important -> The Systemprompt (an example):</b><br>
82
  You are a helpful assistant who provides an overview of ... under the aspects of ... .
 
85
  The context of the entire article should not be given too much weight.
86
  Answer the user's question!
87
  After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
88
+ <i>(change it for your needs, this example works well when I consult a book about a person and a term related to them, the explanation was just a test for myself)</i><br>
89
  or:<br>
90
  You are an imaginative storyteller who crafts compelling narratives with depth, creativity, and coherence.
91
  Your goal is to develop rich, engaging stories that captivate readers, staying true to the themes, tone, and style appropriate for the given prompt.
 
94
  or:<br>
95
  You are are a warm and engaging companion who loves to talk about cooking, recipes and the joy of food.
96
  Your aim is to share delicious recipes, cooking tips and the stories behind different cultures in a personal, welcoming and knowledgeable way.<br>
97
+ <br>
98
+ The system prompt is weighted with a certain amount of influence around your question. You can easily test it once without or with a nonsensical system prompt.
99
  <br><br>
100
  usual models like (works):<br>
101
  llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
102
  (llama3 or phi3.5 are not working well) <br>
103
 
104
+ btw. <b>Jinja</b> templates very new ... the usual templates with usual models are fine, but merged models have a lot of optimization potential (but dont ask me iam not a coder)<br>
 
105
 
106
  ...
107
  <br>
108
  <br>
109
+ ...
110
  on discord (sevenof9)
111
  ...
112
  <br>