First model version
Browse files
README.md
CHANGED
|
@@ -1,13 +1,3 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
tags:
|
| 4 |
-
+ dllm
|
| 5 |
-
+ diffusion
|
| 6 |
-
+ llm
|
| 7 |
-
+ text_generation
|
| 8 |
-
library_name: transformers
|
| 9 |
-
---
|
| 10 |
-
|
| 11 |
# LLaDA2-mini-preview
|
| 12 |
**LLaDA2-mini-preview** is a diffusion language model featuring a 16BA1B Mixture-of-Experts (MoE) architecture. As an enhanced, instruction-tuned iteration of the LLaDA series, it is optimized for practical applications.
|
| 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
# LLaDA2-mini-preview
|
| 2 |
**LLaDA2-mini-preview** is a diffusion language model featuring a 16BA1B Mixture-of-Experts (MoE) architecture. As an enhanced, instruction-tuned iteration of the LLaDA series, it is optimized for practical applications.
|
| 3 |
|