kalle07 commited on
Commit
806cfb4
·
verified ·
1 Parent(s): a2a30ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -1
README.md CHANGED
@@ -10,4 +10,27 @@ tags:
10
  - moedels
11
  - GGUF
12
 
13
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  - moedels
11
  - GGUF
12
 
13
+ ---
14
+
15
+
16
+ All models tested with ALLM(AnythingLLM) with LM as server
17
+ they work more or less
18
+
19
+ my short impression:
20
+ - nomic-embed-text
21
+ - mxbai-embed-large
22
+ - mug-b-1.6
23
+ working well, all other its up to you!
24
+
25
+ short hints for using:
26
+ set your (Max Tokens)context-lenght 16000t main-model, set your embedder-model (Max Embedding Chunk Length) 1024t,set (Max Context Snippets) 14
27
+ ok what that mean!
28
+ you can receive 14snippets a 1024t (14336t) from your document ~10000words and 1600t left for the answer ~1000words
29
+ you can play and set for your needs, eg 8snippets 2048t or 28snippets 512t ...
30
+ 16000t ~1GB VRAM usage
31
+
32
+
33
+
34
+
35
+
36
+ (ALL Licenses and terms of use go to original author)