kgrabko commited on
Commit
ef962ee
·
verified ·
1 Parent(s): 2b35bcf

Update claims_dense.md

Browse files
Files changed (1) hide show
  1. claims_dense.md +45 -45
claims_dense.md CHANGED
@@ -1,46 +1,46 @@
1
- # Patent Claims: JiRack Dense Transformer Architecture (V.1.2)
2
-
3
- **Inventor:** Konstantin Vladimirovich Grabko
4
- **Entity:** CMS Manhattan
5
- **Priority Date:** December 21, 2025
6
- **Technology Class:** Large Language Model (LLM) Optimization / Neural Compute Kernels
7
-
8
- ## I. PRIMARY ARCHITECTURAL CLAIMS
9
-
10
- ### Claim 1: SwiGLU-Attention (SWA) Fused Computational Block
11
- A method for executing a transformer layer in a deep neural network, characterized by the parallelization and unified dispatch of a Multi-Head Attention (MHA) mechanism and a Gated Linear Unit (SwiGLU) Feed-Forward Network (FFN), comprising:
12
-
13
- - An integrated input bus that loads hidden states into local GPU shared memory (SRAM) for simultaneous projection by both the attention and FFN sub-layers.
14
- - A fused residual accumulation path where $x_{out} = x + \text{Attn}(x) + \text{FFN}(x)$ is computed within a single kernel boundary to minimize Global Memory (VRAM) round-trips.
15
-
16
- ### Claim 2: Buffered Routing Embedding (BRE) System
17
- A hardware-level memory management system for High Bandwidth Memory (HBM) optimized for large-scale embedding tables, comprising:
18
-
19
- - A look-ahead predictive buffer that identifies high-probability token sequences.
20
- - A dynamic routing mechanism that pages relevant embedding vectors from standard VRAM into a specialized high-speed "High-Bandwidth Partition" (L3 Cache or Shared Memory) prior to the activation of the first transformer block.
21
-
22
- ## II. SCALING & ATTENTION CLAIMS
23
-
24
- ### Claim 3: Dynamic GQA-Dense Scaling Logic
25
- An attention mechanism for models exceeding 7 billion parameters, utilizing a specific Grouped-Query Attention (GQA) ratio (specifically 5:1 for 13B configurations and 8:1 for 70B+ configurations) implemented in a dense, non-quantized format to maintain high-precision gradient flow while reducing KV-cache memory footprint.
26
-
27
- ### Claim 4: Sliding Window Memory Efficiency (SWME)
28
- A method as described in Claim 3, further comprising a specialized WINDOW_SIZE cache logic that restricts the KV-cache to a fixed-length rolling buffer (e.g., 1024 tokens) during the inference phase, ensuring constant-time memory complexity for the Attention sub-layer.
29
-
30
- ## III. HARDWARE OPTIMIZATION CLAIMS
31
-
32
- ### Claim 5: ROCm-Specific Matrix Core Optimization
33
- A computational kernel optimized for AMD ROCm/HIP environments that utilizes specialized primitives to map the fused SWA block (Claim 1) directly onto hardware matrix cores, specifically targeting the reduction of instruction-level overhead in the RoPE (Rotary Positional Embedding) application.
34
-
35
- ### Claim 6: Gradient Stability in Ultra-Deep Dense Networks
36
- An architectural arrangement of $N$ layers ($N \ge 40$) where the specific mathematical interaction between the RMSNorm layers and the fused SWA blocks enables training without auxiliary skip-connections or intermediate normalization, preserving 16-bit floating-point (BF16) precision throughout the stack.
37
-
38
- ## IV. INTELLECTUAL PROPERTY PROTECTIONS
39
-
40
- ### Claim 7: Digital Authorship Proof
41
- A non-functional, immutable data buffer registered within the model's weight distribution (proof_of_authorship) containing the cryptographically or string-encoded signature of the inventor, used to verify the origin of the architecture in derivative or fine-tuned versions of the weights.
42
-
43
- ---
44
-
45
- **Legal Notice**
46
  Any implementation of the logic described in Claims 1 through 7 without a valid license from Konstantin Vladimirovich Grabko constitutes an infringement of proprietary technology. Filing for patent rights based on the disclosed SWA or BRE logic by third parties is strictly prohibited under the terms of the JiRack Commercial License V.1.2.
 
1
+ # Patent Claims: JiRack Dense Transformer Architecture (V.1.2) (7B Edition)
2
+
3
+ **Inventor:** Konstantin Vladimirovich Grabko
4
+ **Entity:** CMS Manhattan
5
+ **Priority Date:** December 22, 2025
6
+ **Technology Class:** Large Language Model (LLM) Optimization / Neural Compute Kernels
7
+
8
+ ## I. PRIMARY ARCHITECTURAL CLAIMS
9
+
10
+ ### Claim 1: SwiGLU-Attention (SWA) Fused Computational Block
11
+ A method for executing a transformer layer in a deep neural network, characterized by the parallelization and unified dispatch of a Multi-Head Attention (MHA) mechanism and a Gated Linear Unit (SwiGLU) Feed-Forward Network (FFN), comprising:
12
+
13
+ - An integrated input bus that loads hidden states into local GPU shared memory (SRAM) for simultaneous projection by both the attention and FFN sub-layers.
14
+ - A fused residual accumulation path where $x_{out} = x + \text{Attn}(x) + \text{FFN}(x)$ is computed within a single kernel boundary to minimize Global Memory (VRAM) round-trips.
15
+
16
+ ### Claim 2: Buffered Routing Embedding (BRE) System
17
+ A hardware-level memory management system for High Bandwidth Memory (HBM) optimized for large-scale embedding tables, comprising:
18
+
19
+ - A look-ahead predictive buffer that identifies high-probability token sequences.
20
+ - A dynamic routing mechanism that pages relevant embedding vectors from standard VRAM into a specialized high-speed "High-Bandwidth Partition" (L3 Cache or Shared Memory) prior to the activation of the first transformer block.
21
+
22
+ ## II. SCALING & ATTENTION CLAIMS
23
+
24
+ ### Claim 3: Dynamic GQA-Dense Scaling Logic
25
+ An attention mechanism for models exceeding 7 billion parameters, utilizing a specific Grouped-Query Attention (GQA) ratio (specifically 5:1 for 13B configurations and 8:1 for 70B+ configurations) implemented in a dense, non-quantized format to maintain high-precision gradient flow while reducing KV-cache memory footprint.
26
+
27
+ ### Claim 4: Sliding Window Memory Efficiency (SWME)
28
+ A method as described in Claim 3, further comprising a specialized WINDOW_SIZE cache logic that restricts the KV-cache to a fixed-length rolling buffer (e.g., 1024 tokens) during the inference phase, ensuring constant-time memory complexity for the Attention sub-layer.
29
+
30
+ ## III. HARDWARE OPTIMIZATION CLAIMS
31
+
32
+ ### Claim 5: ROCm-Specific Matrix Core Optimization
33
+ A computational kernel optimized for AMD ROCm/HIP environments that utilizes specialized primitives to map the fused SWA block (Claim 1) directly onto hardware matrix cores, specifically targeting the reduction of instruction-level overhead in the RoPE (Rotary Positional Embedding) application.
34
+
35
+ ### Claim 6: Gradient Stability in Ultra-Deep Dense Networks
36
+ An architectural arrangement of $N$ layers ($N \ge 40$) where the specific mathematical interaction between the RMSNorm layers and the fused SWA blocks enables training without auxiliary skip-connections or intermediate normalization, preserving 16-bit floating-point (BF16) precision throughout the stack.
37
+
38
+ ## IV. INTELLECTUAL PROPERTY PROTECTIONS
39
+
40
+ ### Claim 7: Digital Authorship Proof
41
+ A non-functional, immutable data buffer registered within the model's weight distribution (proof_of_authorship) containing the cryptographically or string-encoded signature of the inventor, used to verify the origin of the architecture in derivative or fine-tuned versions of the weights.
42
+
43
+ ---
44
+
45
+ **Legal Notice**
46
  Any implementation of the logic described in Claims 1 through 7 without a valid license from Konstantin Vladimirovich Grabko constitutes an infringement of proprietary technology. Filing for patent rights based on the disclosed SWA or BRE logic by third parties is strictly prohibited under the terms of the JiRack Commercial License V.1.2.