AscendKernelGen commited on
Commit
8ea5972
·
verified ·
1 Parent(s): 85ed22e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -15,12 +15,12 @@ KernelGen-LM-32B is a state-of-the-art domain-adaptive large language model spec
15
  The Ascend KernelGen Technical Report is published at https://arxiv.org/abs/2601.07160.
16
 
17
  **Other artifacts:**
18
- * The **Ascend KernelGen Technical Report** is published at https://arxiv.org/abs/2601.07160.
19
  * The **NPUKernelBench** evaluation framework is published at https://git.openi.org.cn/PCL-Benchmark/NPUKernelBench.
20
 
21
  ## Introduction
22
 
23
- Our framework, **Ascend KernelGen (AKGen)**, bridges the gap between general-purpose code generation and hardware-specific programming through a closed-loop system of data construction, training, and evaluation. Key innovations include:
24
 
25
  * **Ascend-CoT Dataset:** A high-quality, domain-specific dataset incorporating **Chain-of-Thought (CoT)** reasoning. It combines documentation-based reasoning, code-centric reasoning derived from real-world kernel implementations, and general reasoning chains to capture the structured logic required for low-level NPU programming.
26
  * **Domain-Adaptive Post-Training:** A two-stage optimization process that yields **KernelGen-LM**. We first employ **Supervised Fine-Tuning (SFT)** with error-derived supervision (correcting API misuse and numerical errors). This is followed by **Reinforcement Learning (RL)** using Direct Preference Optimization (DPO), driven by execution-based correctness and performance signals.
@@ -29,7 +29,7 @@ Our framework, **Ascend KernelGen (AKGen)**, bridges the gap between general-pur
29
 
30
  ## Citation
31
  @article{cao2026ascendkernelgen,
32
- title={Ascend KernelGen: A Systematic Study of LLM-Based Kernel Generation for Neural Processing Units},
33
  author={Xinzi Cao and Jianyang Zhai and Pengfei Li and Zhiheng Hu and Cen Yan and Bingxu Mu and Guanghuan Fang and Bin She and Jiayu Li and Yihan Su and Dongyang Tao and Xiansong Huang and Fan Xu and Feidiao Yang and Yao Lu and Chang-Dong Wang and Yutong Lu and Weicheng Xue and Bin Zhou and Yonghong Tian},
34
  journal={arXiv preprint arXiv:2601.07160},
35
  year={2026},
 
15
  The Ascend KernelGen Technical Report is published at https://arxiv.org/abs/2601.07160.
16
 
17
  **Other artifacts:**
18
+ * The **AscendKernelGen Technical Report** is published at https://arxiv.org/abs/2601.07160.
19
  * The **NPUKernelBench** evaluation framework is published at https://git.openi.org.cn/PCL-Benchmark/NPUKernelBench.
20
 
21
  ## Introduction
22
 
23
+ Our framework, **AscendKernelGen (AKGen)**, bridges the gap between general-purpose code generation and hardware-specific programming through a closed-loop system of data construction, training, and evaluation. Key innovations include:
24
 
25
  * **Ascend-CoT Dataset:** A high-quality, domain-specific dataset incorporating **Chain-of-Thought (CoT)** reasoning. It combines documentation-based reasoning, code-centric reasoning derived from real-world kernel implementations, and general reasoning chains to capture the structured logic required for low-level NPU programming.
26
  * **Domain-Adaptive Post-Training:** A two-stage optimization process that yields **KernelGen-LM**. We first employ **Supervised Fine-Tuning (SFT)** with error-derived supervision (correcting API misuse and numerical errors). This is followed by **Reinforcement Learning (RL)** using Direct Preference Optimization (DPO), driven by execution-based correctness and performance signals.
 
29
 
30
  ## Citation
31
  @article{cao2026ascendkernelgen,
32
+ title={AscendKernelGen: A Systematic Study of LLM-Based Kernel Generation for Neural Processing Units},
33
  author={Xinzi Cao and Jianyang Zhai and Pengfei Li and Zhiheng Hu and Cen Yan and Bingxu Mu and Guanghuan Fang and Bin She and Jiayu Li and Yihan Su and Dongyang Tao and Xiansong Huang and Fan Xu and Feidiao Yang and Yao Lu and Chang-Dong Wang and Yutong Lu and Weicheng Xue and Bin Zhou and Yonghong Tian},
34
  journal={arXiv preprint arXiv:2601.07160},
35
  year={2026},