keypa commited on
Commit
a66e3b7
·
verified ·
1 Parent(s): bbcf8f0

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -0
README.md ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ task_categories:
4
+ - text-generation
5
+ - feature-extraction
6
+ language:
7
+ - en
8
+ tags:
9
+ - gpt-oss
10
+ - mla
11
+ - calibration
12
+ - fineweb-edu
13
+ size_categories:
14
+ - 10K<n<100K
15
+ ---
16
+
17
+ # GPT-OSS Calibration Data (SWA/Full Splits)
18
+
19
+ This dataset contains high-quality calibration artifacts for converting **GPT-OSS-20B** (and larger MoE models) from GQA to **Multi-Head Latent Attention (MLA)**.
20
+
21
+ It was generated using **GTE-Large-v1.5** (8192 context window) on an L40S GPU to ensure long-range dependencies are preserved, addressing the "V_latent compression" quality gap.
22
+
23
+ ## Files
24
+
25
+ | File | Description | Target Layers |
26
+ | :--- | :--- | :--- |
27
+ | **`calib_swa_golden.jsonl`** | Dense local coherence data (< 2048 tokens). | Sliding Window Attention Layers |
28
+ | **`calib_full_golden.jsonl`** | Long-range dependency outliers (> 4096 tokens). | Full Attention Layers |
29
+ | **`calib_mixed_golden.jsonl`** | A diverse mix of both (General purpose). | Fallback / Baseline |
30
+
31
+ ## Methodology
32
+ 1. **Source:** [FineWeb-Edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) (SOTA reasoning/academic text).
33
+ 2. **Embeddings:** `Alibaba-NLP/gte-large-en-v1.5` (8k context) used to encode documents.
34
+ 3. **Filtering:**
35
+ * **Clustering:** MiniBatchKMeans to find representative centroids.
36
+ * **Hard Negatives:** Top 5% outliers (highest distance from centroids) selected to maximize gradient signal during distillation.
37
+ * **Splitting:** Data strictly stratified by length to match GPT-OSS SWA/Full layer requirements.
38
+
39
+ ## Usage
40
+ Intended for Knowledge Distillation (Teacher -> Student) to repair Perplexity degradation during MLA conversion.