Update README.md
Browse files
README.md
CHANGED
|
@@ -63,7 +63,7 @@ This model is a significant step forward for the Korean LLM community, demonstra
|
|
| 63 |
We followed the methodology outlined in the Fast-dLLM-v2 approach (as seen in the model: Efficient-Large-Model/Fast_dLLM_v2_7B [https://huggingface.co/Efficient-Large-Model/Fast_dLLM_v2_7B]).
|
| 64 |
|
| 65 |
Continual Pre-training from Tri-7B:
|
| 66 |
-
Trida-7B-Preview was continually pre-trained starting from our
|
| 67 |
|
| 68 |
|
| 69 |
## 🚀 Quickstart
|
|
|
|
| 63 |
We followed the methodology outlined in the Fast-dLLM-v2 approach (as seen in the model: Efficient-Large-Model/Fast_dLLM_v2_7B [https://huggingface.co/Efficient-Large-Model/Fast_dLLM_v2_7B]).
|
| 64 |
|
| 65 |
Continual Pre-training from Tri-7B:
|
| 66 |
+
Trida-7B-Preview was continually pre-trained starting from our autoregressive model, trillionlabs/Tri-7B. This process was executed using a Block Diffusion training paradigm to transition the efficient base model into a highly capable generative model.
|
| 67 |
|
| 68 |
|
| 69 |
## 🚀 Quickstart
|