tl-hyungguk commited on
Commit
e2ab06a
·
verified ·
1 Parent(s): 179691f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -63,7 +63,7 @@ This model is a significant step forward for the Korean LLM community, demonstra
63
  We followed the methodology outlined in the Fast-dLLM-v2 approach (as seen in the model: Efficient-Large-Model/Fast_dLLM_v2_7B [https://huggingface.co/Efficient-Large-Model/Fast_dLLM_v2_7B]).
64
 
65
  Continual Pre-training from Tri-7B:
66
- Trida-7B-Preview was continually pre-trained starting from our proprietary model, trillionlabs/Tri-7B. This process was executed using a Block Diffusion training paradigm to transition the efficient base model into a highly capable generative model.
67
 
68
 
69
  ## 🚀 Quickstart
 
63
  We followed the methodology outlined in the Fast-dLLM-v2 approach (as seen in the model: Efficient-Large-Model/Fast_dLLM_v2_7B [https://huggingface.co/Efficient-Large-Model/Fast_dLLM_v2_7B]).
64
 
65
  Continual Pre-training from Tri-7B:
66
+ Trida-7B-Preview was continually pre-trained starting from our autoregressive model, trillionlabs/Tri-7B. This process was executed using a Block Diffusion training paradigm to transition the efficient base model into a highly capable generative model.
67
 
68
 
69
  ## 🚀 Quickstart