aifeifei798 commited on
Commit
7c3bbd5
·
verified ·
1 Parent(s): 9c26760

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -52,8 +52,8 @@ This model represents a fundamental shift in how we approach LLM fine-tuning. In
52
 
53
  ---
54
 
55
- > *"While denoising objectives exist in pre-training (e.g., BART, T5), applying **heavy stochastic token shuffling (70%)** strictly during the **Instruction Fine-Tuning (SFT)** phase for Causal LLMs to decouple logic from syntax is, to the best of our knowledge, a novel approach introduced by aifeifei798, Gemini."*
56
- >
57
  > *(虽然去噪任务在预训练中存在,但在因果语言模型的 SFT 阶段使用高强度的随机词序打乱(70%)来剥离逻辑与句法,据我们所知,这是由 aifeifei798, Gemini 首创的方法。)*
58
 
59
  ---
 
52
 
53
  ---
54
 
55
+ > "While denoising objectives exist in pre-training (e.g., BART, T5), applying **heavy stochastic token shuffling (70%)** strictly during the **Instruction Fine-Tuning (SFT)** phase for Causal LLMs to decouple logic from syntax is, to the best of our knowledge, a novel approach introduced by **aifeifei798** and **Gemini**."
56
+ >
57
  > *(虽然去噪任务在预训练中存在,但在因果语言模型的 SFT 阶段使用高强度的随机词序打乱(70%)来剥离逻辑与句法,据我们所知,这是由 aifeifei798, Gemini 首创的方法。)*
58
 
59
  ---