kalle07 commited on
Commit
1c41c3d
·
verified ·
1 Parent(s): 5d4c508

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -28,6 +28,7 @@ architecture:
28
  <b> the setup for local documents described below is allmost the same, GPT4All has only one model (nomic)</b><br>
29
 
30
  (sometimes the results are more truthful if the “chat with document only” option is used)<br>
 
31
  <b>&#x21e8;</b> give me a ❤️, if you like ;)<br>
32
  <br>
33
  <b>My short impression:</b>
@@ -49,9 +50,10 @@ but in ALLM its cutting all in 1024 character parts, so aprox two times or bit m
49
  <br>
50
 
51
  -> Ok what that mean!<br>
 
52
  You can receive 14-snippets a 1024t (14336t) from your document ~10000words and 1600t left for the answer ~1000words (2 pages)
53
  <br>
54
- You can play and set for your needs, eg 8-snippets a 2048t, or 28-snippets a 512t ...
55
  <ul style="line-height: 1;">
56
  <li>8000t (~6000words) ~0.8GB VRAM usage</li>
57
  <li>16000t (~12000words) ~1.5GB VRAM usage</li>
@@ -80,7 +82,7 @@ This text snippet is then used for your answer. <br>
80
 
81
  <li>If you expect multible search results in your docs try 16-snippets or more, if you expect only 2 than dont use more!</li>
82
 
83
- <li>If you use snipets-size ~1024t you receive more content, if you use ~256t you receive more facts.</li>
84
 
85
  <li>A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.</li>
86
 
 
28
  <b> the setup for local documents described below is allmost the same, GPT4All has only one model (nomic)</b><br>
29
 
30
  (sometimes the results are more truthful if the “chat with document only” option is used)<br>
31
+ BTW embedder is only a part of a good RAG<br>
32
  <b>&#x21e8;</b> give me a ❤️, if you like ;)<br>
33
  <br>
34
  <b>My short impression:</b>
 
50
  <br>
51
 
52
  -> Ok what that mean!<br>
53
+ Your document will be embedd in x times 1024t Chunks(snippets),<br>
54
  You can receive 14-snippets a 1024t (14336t) from your document ~10000words and 1600t left for the answer ~1000words (2 pages)
55
  <br>
56
+ You can play and set for your needs, eg 8-snippets a 2048t, or 28-snippets a 512t ... (every time you change the chunk-length the document must be embedd again)
57
  <ul style="line-height: 1;">
58
  <li>8000t (~6000words) ~0.8GB VRAM usage</li>
59
  <li>16000t (~12000words) ~1.5GB VRAM usage</li>
 
82
 
83
  <li>If you expect multible search results in your docs try 16-snippets or more, if you expect only 2 than dont use more!</li>
84
 
85
+ <li>If you use chunk-length ~1024t you receive more content, if you use ~256t you receive more facts.</li>
86
 
87
  <li>A question for "summary of the document" is most time not useful, if the document has an introduction or summaries its searching there if you have luck.</li>
88