Update README.md
Browse files
README.md
CHANGED
|
@@ -10,6 +10,8 @@ tags:
|
|
| 10 |
# "[We] are joined by the bonds of love. And you cannot track that, not with a thousand bloodhounds, and you cannot break it, not with a thousand swords."
|
| 11 |
[GGUF FILES HERE](https://huggingface.co/Kquant03/Buttercup-4x7B-GGUF)
|
| 12 |
|
|
|
|
|
|
|
| 13 |
A frankenMoE not only using far better methodology and fundamental understanding of SMoE, but completely focused around intellectual roleplay. It may have a bit of a redundancy issue (I have actually been playing with it while GGUF uploads on q8_k and it has nice variety). However, just in case, to battle this, try to keep things fresh with the model by either introducing new concepts often, or through [drμgs](https://github.com/EGjoni/DRUGS). (no not that kind)
|
| 14 |
|
| 15 |
The config looks like this...(detailed version is in the files and versions):
|
|
@@ -19,7 +21,6 @@ The config looks like this...(detailed version is in the files and versions):
|
|
| 19 |
- [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) - expert #3
|
| 20 |
- [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) - expert #4
|
| 21 |
|
| 22 |
-
[Join our Discord!](https://discord.gg/CAfWPV82)
|
| 23 |
# "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
|
| 24 |
### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
|
| 25 |
|
|
|
|
| 10 |
# "[We] are joined by the bonds of love. And you cannot track that, not with a thousand bloodhounds, and you cannot break it, not with a thousand swords."
|
| 11 |
[GGUF FILES HERE](https://huggingface.co/Kquant03/Buttercup-4x7B-GGUF)
|
| 12 |
|
| 13 |
+
[Join our Discord!](https://discord.gg/CAfWPV82)
|
| 14 |
+
|
| 15 |
A frankenMoE not only using far better methodology and fundamental understanding of SMoE, but completely focused around intellectual roleplay. It may have a bit of a redundancy issue (I have actually been playing with it while GGUF uploads on q8_k and it has nice variety). However, just in case, to battle this, try to keep things fresh with the model by either introducing new concepts often, or through [drμgs](https://github.com/EGjoni/DRUGS). (no not that kind)
|
| 16 |
|
| 17 |
The config looks like this...(detailed version is in the files and versions):
|
|
|
|
| 21 |
- [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) - expert #3
|
| 22 |
- [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) - expert #4
|
| 23 |
|
|
|
|
| 24 |
# "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
|
| 25 |
### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
|
| 26 |
|