Update README.md
Browse files
README.md
CHANGED
|
@@ -17,7 +17,7 @@ pipeline_tag: text-generation
|
|
| 17 |
|
| 18 |

|
| 19 |
|
| 20 |
-
# Muse-12B
|
| 21 |
|
| 22 |
Quantized NVFP4 weights of the [Muse-12B](https://huggingface.co/LatitudeGames/Muse-12B) model, for use with nVidia Blackwell GPUs.
|
| 23 |
|
|
@@ -45,7 +45,7 @@ dataset_utils.SUPPORTED_DATASET_CONFIG["distilled-roleplay"] = {
|
|
| 45 |
|
| 46 |
## Inference
|
| 47 |
|
| 48 |
-
Tested on a RTX 5060 Ti 16GB with TensorRT-LLM, vLLM, and
|
| 49 |
|
| 50 |
Recommended generation settings (a mix of what it says on the Muse-12B model card and the [AI Dungeon Model Guide](https://help.aidungeon.com/ai-models-and-their-differences)):
|
| 51 |
- Temperature: 1.0
|
|
|
|
| 17 |
|
| 18 |

|
| 19 |
|
| 20 |
+
# Muse-12B-NVFP4
|
| 21 |
|
| 22 |
Quantized NVFP4 weights of the [Muse-12B](https://huggingface.co/LatitudeGames/Muse-12B) model, for use with nVidia Blackwell GPUs.
|
| 23 |
|
|
|
|
| 45 |
|
| 46 |
## Inference
|
| 47 |
|
| 48 |
+
Tested on a RTX 5060 Ti 16GB with TensorRT-LLM, vLLM, SGLang, and Aphrodite Engine.
|
| 49 |
|
| 50 |
Recommended generation settings (a mix of what it says on the Muse-12B model card and the [AI Dungeon Model Guide](https://help.aidungeon.com/ai-models-and-their-differences)):
|
| 51 |
- Temperature: 1.0
|