Update README.md
Browse files
README.md
CHANGED
|
@@ -45,17 +45,15 @@ to see all files<br>
|
|
| 45 |
<b> the setup for local documents described below is allmost the same, GPT4All has only one model (nomic), and koboldcpp and JAN(Menlo) is not build in right now but in development</b><br>
|
| 46 |
|
| 47 |
(sometimes the results are more truthful if the “chat with document only” option is used)<br>
|
| 48 |
-
BTW the embedder-model is only a part of a good RAG (Retrieval-Augmented Generation)<br>
|
| 49 |
<b>⇨</b> give me a ❤️, if you like ;)<br>
|
| 50 |
<br>
|
| 51 |
<b>My short impression:</b>
|
| 52 |
<ul style="line-height: 1.05;">
|
| 53 |
-
<li>nomic-embed-text (up to
|
| 54 |
-
<li>mxbai-embed-large</li>
|
| 55 |
<li>mug-b-1.6</li>
|
| 56 |
<li>snowflake-arctic-embed-l-v2.0 (up to 8192t context length)</li>
|
| 57 |
-
<li>ger-RAG-bge-M3-merg-snowf-artic-hessian-AI (german, up to 8192t context length)</li>
|
| 58 |
-
<li>german-roberta</li>
|
| 59 |
<li>bge-m3 (up to 8192t context length)</li>
|
| 60 |
</ul>
|
| 61 |
Working well, all other its up to you! Some models are very similar! (jina and qwen based you can add manual to LM-Studio, set model "gear wheel" below "overide domain type")<br>
|
|
@@ -64,6 +62,10 @@ Further tests have shown that the following models are suitable for complex task
|
|
| 64 |
<ul style="line-height: 1.05;">
|
| 65 |
<li>GTE large</li>
|
| 66 |
<li>cross-en-de-es-roberta</li>
|
|
|
|
|
|
|
|
|
|
|
|
|
| 67 |
</ul>
|
| 68 |
There are two embedder to find toxic content (toxic-prompt-roberta and minilmv2-toxic-jigsaw), dont know how good it works, and from ibm it give a whole LLM model (granite-guardian).
|
| 69 |
<br>
|
|
|
|
| 45 |
<b> the setup for local documents described below is allmost the same, GPT4All has only one model (nomic), and koboldcpp and JAN(Menlo) is not build in right now but in development</b><br>
|
| 46 |
|
| 47 |
(sometimes the results are more truthful if the “chat with document only” option is used)<br>
|
| 48 |
+
BTW the embedder-model is only a part of a good RAG (Retrieval-Augmented Generation), 512t are ~2000 characters most cases enough.<br>
|
| 49 |
<b>⇨</b> give me a ❤️, if you like ;)<br>
|
| 50 |
<br>
|
| 51 |
<b>My short impression:</b>
|
| 52 |
<ul style="line-height: 1.05;">
|
| 53 |
+
<li>nomic-embed-text-v2-moe (up to 512t context length)</li>
|
| 54 |
+
<li>mxbai-embed-large (small and fast model)</li>
|
| 55 |
<li>mug-b-1.6</li>
|
| 56 |
<li>snowflake-arctic-embed-l-v2.0 (up to 8192t context length)</li>
|
|
|
|
|
|
|
| 57 |
<li>bge-m3 (up to 8192t context length)</li>
|
| 58 |
</ul>
|
| 59 |
Working well, all other its up to you! Some models are very similar! (jina and qwen based you can add manual to LM-Studio, set model "gear wheel" below "overide domain type")<br>
|
|
|
|
| 62 |
<ul style="line-height: 1.05;">
|
| 63 |
<li>GTE large</li>
|
| 64 |
<li>cross-en-de-es-roberta</li>
|
| 65 |
+
<li>ger-RAG-bge-M3-merg-snowf-artic-hessian-AI (very good for german, up to 8192t context length)</li>
|
| 66 |
+
<li>German-RAG-BGE-M3-TRIPLES-HESSIAN-AI (very good for german, up to 8192t context length)</li>
|
| 67 |
+
<li>bge-m3 (good for german, up to 8192t context length)</li>
|
| 68 |
+
<li>jina-embeddings-v3 (good for german, up to 8192t context length)</li>
|
| 69 |
</ul>
|
| 70 |
There are two embedder to find toxic content (toxic-prompt-roberta and minilmv2-toxic-jigsaw), dont know how good it works, and from ibm it give a whole LLM model (granite-guardian).
|
| 71 |
<br>
|