Expand DISC Mathematical Foundations with formal axioms, update portfolio stats (49 models, 22,598 downloads)
Browse files
README.md
CHANGED
|
@@ -226,13 +226,21 @@ inputs = tokenizer(prompt, return_tensors="pt")
|
|
| 226 |
outputs = model.generate(**inputs, max_new_tokens=256)
|
| 227 |
```
|
| 228 |
|
| 229 |
-
## Mathematical
|
| 230 |
|
| 231 |
-
The metric attention mechanism is grounded in
|
| 232 |
|
| 233 |
-
|
| 234 |
|
| 235 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 236 |
|
| 237 |
## Lineage
|
| 238 |
|
|
@@ -291,7 +299,7 @@ HuggingFace: [reaperdoesntknow](https://huggingface.co/reaperdoesntknow)
|
|
| 291 |
| [Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT-GGUF](https://huggingface.co/reaperdoesntknow/Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT-GGUF) | 203 |
|
| 292 |
| [Qwen3-1.7B-Coder-Distilled-SFT-GGUF](https://huggingface.co/reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT-GGUF) | 194 |
|
| 293 |
|
| 294 |
-
**Total Portfolio:
|
| 295 |
|
| 296 |
|
| 297 |
*Last updated: 2026-03-28 12:55 UTC*
|
|
@@ -313,4 +321,4 @@ Full methodology: [Structure Over Scale (DOI: 10.57967/hf/8165)](https://doi.org
|
|
| 313 |
<!-- CIX-CROSSLINK-END -->
|
| 314 |
|
| 315 |
---
|
| 316 |
-
<sub>Part of the [reaperdoesntknow research portfolio](https://huggingface.co/reaperdoesntknow) —
|
|
|
|
| 226 |
outputs = model.generate(**inputs, max_new_tokens=256)
|
| 227 |
```
|
| 228 |
|
| 229 |
+
## Mathematical Foundations: Discrepancy Calculus (DISC)
|
| 230 |
|
| 231 |
+
The metric attention mechanism is grounded in Discrepancy Calculus — a measure-theoretic framework for singularity analysis developed by the author. The core operator:
|
| 232 |
|
| 233 |
+
$$Df(x) := \limsup_{r \downarrow 0} \sup_{0 < d(x,y) < r} \frac{|f(y) - f(x)|}{d(x,y)}$$
|
| 234 |
|
| 235 |
+
On metric-measure spaces (Axiom 11.1), this replaces the classical derivative with a scale-free local Lipschitz seminorm. The triangle inequality regularizer enforces that the learned attention geometry satisfies $d(a,c) \leq d(a,b) + d(b,c)$ across sampled triples — ensuring the distance function is a proper metric, not merely a similarity function.
|
| 236 |
+
|
| 237 |
+
The **ball pruning mechanism** (learnable per-head origins and radii) creates adaptive sparse attention patterns that emerge from **gap geometry** (Axiom 11.4): the gap set $\Delta(E) = \{x : \theta^*_E(x) > \theta_{E*}(x)\}$ defines where attention should be sparse. Tokens outside the ball are in the gap — structurally irrelevant to the current computation.
|
| 238 |
+
|
| 239 |
+
**BlackHoleRoPE** extends standard rotary position encoding with learned phase perturbations from a Fourier basis, maintaining the unitary property on Q/K while adding bounded amplitude modulation on V. The boundedness condition is the DISC **discrepancy energy** criterion: $E_{\text{disc}} < \infty$ ensures Lyapunov stability.
|
| 240 |
+
|
| 241 |
+
**L2-star discrepancy** for thermodynamic governance measures training dynamics as a discrepancy field — gradient magnitude distributions analyzed through the same measure-theoretic lens that DISC applies to function spaces.
|
| 242 |
+
|
| 243 |
+
Full theory: *"On the Formal Analysis of Discrepancy Calculus"* (Colca, 2026; Convergent Intelligence LLC: Research Division). Full methodology: [Structure Over Scale (DOI: 10.57967/hf/8165)](https://doi.org/10.57967/hf/8165).
|
| 244 |
|
| 245 |
## Lineage
|
| 246 |
|
|
|
|
| 299 |
| [Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT-GGUF](https://huggingface.co/reaperdoesntknow/Qwen3-0.6B-Distilled-30B-A3B-Thinking-SFT-GGUF) | 203 |
|
| 300 |
| [Qwen3-1.7B-Coder-Distilled-SFT-GGUF](https://huggingface.co/reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT-GGUF) | 194 |
|
| 301 |
|
| 302 |
+
**Total Portfolio: 49 models, 22,598 total downloads**
|
| 303 |
|
| 304 |
|
| 305 |
*Last updated: 2026-03-28 12:55 UTC*
|
|
|
|
| 321 |
<!-- CIX-CROSSLINK-END -->
|
| 322 |
|
| 323 |
---
|
| 324 |
+
<sub>Part of the [reaperdoesntknow research portfolio](https://huggingface.co/reaperdoesntknow) — 49 models, 22,598 total downloads | Last refreshed: 2026-03-30 12:10 UTC</sub>
|