redponike commited on
Commit
0f0fff9
·
verified ·
1 Parent(s): 9b236a4

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - Kwaipilot/KAT-Dev-72B-Exp
5
+ pipeline_tag: text-generation
6
+ library_name: transformers
7
+ ---
8
+
9
+ GGUF quants of [Kwaipilot/KAT-Dev-72B-Exp](https://huggingface.co/Kwaipilot/KAT-Dev-72B-Exp)
10
+
11
+ Using llama.cpp b6730 (commit [e60f01d941bc5b7fae62dd57fee4cec76ec0ea6e](https://github.com/ggml-org/llama.cpp/commit/e60f01d941bc5b7fae62dd57fee4cec76ec0ea6e))
12
+
13
+ The importance matrix was generated with calibration_datav3.txt.
14
+
15
+ All quants were generated/calibrated with the imatrix, including the K quants.
16
+
17
+ ---
18
+
19
+ <div align="center">
20
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/61ee40a269351366e29972ad/KIYEa1c_WJEWPpeS0L_k1.png" width="100%" alt="Kwaipilot" />
21
+ </div>
22
+
23
+ <hr>
24
+
25
+
26
+ # News
27
+
28
+ 🔥 We’re thrilled to announce the release of **KAT-Dev-72B-Exp**, our latest and most powerful model yet!
29
+
30
+ 🔥 You can now try our **strongest** proprietary coder model **KAT-Coder** directly on the [**StreamLake**](https://www.streamlake.ai/product/kat-coder) platform **for free**.
31
+
32
+ # Highlights
33
+ **KAT-Dev-72B-Exp** is an open-source 72B-parameter model for software engineering tasks.
34
+
35
+ On SWE-Bench Verified, **KAT-Dev-72B-Exp** achieves **74.6%** accuracy ⚡ — **when evaluated strictly with the SWE-agent scaffold**.
36
+
37
+ **KAT-Dev-72B-Exp** is the experimental reinforcement-learning version of the KAT-Coder model. Through this open-source release, we aim to reveal the technical innovations behind KAT-Coder’s large-scale RL to developers and researchers.
38
+
39
+
40
+ ![Kim 2025-10-10 165138](https://cdn-uploads.huggingface.co/production/uploads/61ee40a269351366e29972ad/-1nx5HYc-wTjUFNbf-GfO.png)
41
+
42
+ # Introduction
43
+
44
+ We rewrote the attention kernel and redesigned the training engine for shared prefix trajectories to achieve highly efficient RL training, especially for scaffolds leveraging context management.
45
+
46
+ Furthermore, to prevent exploration collapse observed in RL training, we reshaped advantage distribution based on pass rates: amplifying the advantage scale of highly exploratory groups while reducing that of low-exploration ones.
47
+
48
+
49
+ # SWE agent Evaluation Parameters
50
+
51
+ ```
52
+ temperature: 0.6
53
+ max_turns: 150
54
+ history_processors.n: 100
55
+ ```
56
+
57
+ For full settings please refer to inference.yaml