kalle07 commited on
Commit
ff38d49
·
verified ·
1 Parent(s): ba0e352

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -68,7 +68,7 @@ If the documents small like 10-20 Pages, its better you copy the whole text insi
68
  <br>
69
  ...
70
  <br>
71
- Nevertheless, the main model is also important! especially to deal with the context length and I don't mean just the theoretical number you can set.
72
  Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.<br>
73
  <br>
74
  <b>Important -> The Systemprompt (an example):</b><br>
@@ -79,12 +79,12 @@ The context of the entire article should not be given too much weight.
79
  Answer the user's question!
80
  After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
81
  (change it for your needs, this example works well when I consult a book about a person and a term related to them)
82
- <br>
83
  usual models like (works):<br>
84
  llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
85
  (llama3 or phi3.5 are not working well) <br>
86
 
87
- btw. Jinja templates very new ... the usual templates with these new models are fine, but merged models have a lot of optimization potential (but dont ask me iam not a coder)<br>
88
 
89
 
90
  ...
 
68
  <br>
69
  ...
70
  <br>
71
+ Nevertheless, the main model is also <b>important</b>! especially to deal with the context length and I don't mean just the theoretical number you can set.
72
  Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.<br>
73
  <br>
74
  <b>Important -> The Systemprompt (an example):</b><br>
 
79
  Answer the user's question!
80
  After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
81
  (change it for your needs, this example works well when I consult a book about a person and a term related to them)
82
+ <br><br>
83
  usual models like (works):<br>
84
  llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
85
  (llama3 or phi3.5 are not working well) <br>
86
 
87
+ btw. <b>Jinja</b> templates very new ... the usual templates with new/usual models are fine, but merged models have a lot of optimization potential (but dont ask me iam not a coder)<br>
88
 
89
 
90
  ...