C10X commited on
Commit
5c16f1a
·
verified ·
1 Parent(s): 919416f

Update README.md

Browse files

fix: Resolve LaTeX rendering issue
- Replace $6\times$ with 6× in README.md

Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -36,7 +36,7 @@ We open-source the KDA kernel in [FLA](https://github.com/fla-org/flash-linear-a
36
  - **Kimi Delta Attention (KDA):** A linear attention mechanism that refines the gated delta rule with finegrained gating.
37
  - **Hybrid Architecture:** A 3:1 KDA-to-global MLA ratio reduces memory usage while maintaining or surpassing the quality of full attention.
38
  - **Superior Performance:** Outperforms full attention in a variety of tasks, including long-context and RL-style benchmarks on 1.4T token training runs with fair comparisons.
39
- - **High Throughput:** Achieves up to $6\times$ faster decoding and significantly reduces time per output token (TPOT).
40
 
41
  <div align="center">
42
  <img width="60%" src="figures/arch.png">
 
36
  - **Kimi Delta Attention (KDA):** A linear attention mechanism that refines the gated delta rule with finegrained gating.
37
  - **Hybrid Architecture:** A 3:1 KDA-to-global MLA ratio reduces memory usage while maintaining or surpassing the quality of full attention.
38
  - **Superior Performance:** Outperforms full attention in a variety of tasks, including long-context and RL-style benchmarks on 1.4T token training runs with fair comparisons.
39
+ - **High Throughput:** Achieves up to 6&times; faster decoding and significantly reduces time per output token (TPOT).
40
 
41
  <div align="center">
42
  <img width="60%" src="figures/arch.png">