kalle07 commited on
Commit
225a69e
·
verified ·
1 Parent(s): 28b99a8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -0
README.md CHANGED
@@ -68,6 +68,9 @@ If the documents small like 10-20 Pages, its better you copy the whole text insi
68
  <br>
69
  ...
70
  <br>
 
 
 
71
  <b>Important -> The Systemprompt (an example):</b><br>
72
  You are a helpful assistant who provides an overview of ... under the aspects of ... .
73
  You use attached excerpts from the collection to generate your answers!
@@ -77,6 +80,10 @@ Answer the user's question!
77
  After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
78
  (change it for your needs, this example works well when I consult a book about a person and a term related to them)
79
  <br>
 
 
 
 
80
  ...
81
  <br>
82
  <br>
 
68
  <br>
69
  ...
70
  <br>
71
+ Nevertheless, the main model is also important! especially to deal with the context length and I don't mean just the theoretical number you can set.
72
+ Some models can handle 128k tokens, but even with 16k input the response with the same snippets as input is worse than with other models.
73
+ <br>
74
  <b>Important -> The Systemprompt (an example):</b><br>
75
  You are a helpful assistant who provides an overview of ... under the aspects of ... .
76
  You use attached excerpts from the collection to generate your answers!
 
80
  After your answer, briefly explain why you included excerpts (1 to X) in your response and justify briefly if you considered some of them unimportant!<br>
81
  (change it for your needs, this example works well when I consult a book about a person and a term related to them)
82
  <br>
83
+ usual models like (works):
84
+ llama3.1, llama3.2, qwen2.5, deepseek-r1-distill, SauerkrautLM-Nemo(german) ... <br>
85
+ (llama3 or phi3.5 are not working well)
86
+
87
  ...
88
  <br>
89
  <br>