Update README.md
Browse files
README.md
CHANGED
|
@@ -10,16 +10,19 @@ tags:
|
|
| 10 |

|
| 11 |
|
| 12 |
# ConvexAI/Solutus-3x7B
|
| 13 |
-
[Join our Discord!](https://discord.gg/XFChPkGd)
|
| 14 |
|
| 15 |
A model to test how MoE will route without square expansion.
|
| 16 |
|
|
|
|
|
|
|
| 17 |
The config looks like this...(detailed version is in the files and versions):
|
| 18 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - base
|
| 19 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #1
|
| 20 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - expert #2
|
| 21 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #3
|
| 22 |
|
|
|
|
|
|
|
| 23 |
# "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
|
| 24 |
### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
|
| 25 |
|
|
|
|
| 10 |

|
| 11 |
|
| 12 |
# ConvexAI/Solutus-3x7B
|
|
|
|
| 13 |
|
| 14 |
A model to test how MoE will route without square expansion.
|
| 15 |
|
| 16 |
+
[Join our Discord!](https://discord.gg/rJXGjmxqzS)
|
| 17 |
+
|
| 18 |
The config looks like this...(detailed version is in the files and versions):
|
| 19 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - base
|
| 20 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #1
|
| 21 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - expert #2
|
| 22 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #3
|
| 23 |
|
| 24 |
+
|
| 25 |
+
|
| 26 |
# "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
|
| 27 |
### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
|
| 28 |
|