Madwand1 commited on
Commit
df098b7
Β·
verified Β·
1 Parent(s): d36730c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +212 -5
README.md CHANGED
@@ -1,5 +1,212 @@
1
- ---
2
- license: other
3
- license_name: codynamics-commercial
4
- license_link: https://www.codynamicslab.com/license
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: codynamics-commercial
4
+ license_link: https://www.codynamicslab.com/license
5
+ language:
6
+ - en
7
+ tags:
8
+ - document-question-answering
9
+ - text-generation
10
+ - long-context
11
+ - information-retrieval
12
+ - enterprise-ai
13
+ - latch
14
+ - multi-document-reasoning
15
+ base_model:
16
+ - Qwen/Qwen2.5-14B-Instruct
17
+ pipeline_tag: text-generation
18
+ library_name: vllm
19
+ ---
20
+
21
+ # LATCH β€” Qwen 2.5 14B
22
+
23
+ **CoDynamics Lab Corporation** | [Website](https://www.codynamicslab.com) | [πŸ›’ Buy Self-Hosted License β€” $79](https://codynamicslab.gumroad.com/l/latch-qwen14b) | [Request Gated Access](#request-access) | [Contact](mailto:mike@codynamicslab.com)
24
+
25
+ > ⚠️ **This is a gated repository.** Model weights are available via two paths β€” see [Deployment Options](#deployment-options) below.
26
+
27
+ ---
28
+
29
+ ## What Is LATCH
30
+
31
+ **LATCH** is a proprietary inference layer built on top of `Qwen/Qwen2.5-14B-Instruct` that eliminates the long-context performance penalty for document-heavy workloads.
32
+
33
+ Standard LLMs re-process every document from scratch on every query. LATCH removes this cost entirely β€” documents are prepared once and subsequent queries run at dramatically reduced latency regardless of document length or count.
34
+
35
+ **This is not RAG. This is not prompt compression.** It is a fundamentally different approach to long-context inference that operates at the model level.
36
+
37
+ Architectural details are proprietary.
38
+
39
+ ---
40
+
41
+ ## Performance Results
42
+
43
+ All benchmarks run on **NVIDIA A100 80GB** with vLLM serving infrastructure.
44
+
45
+ ### Speed
46
+
47
+ | Metric | Baseline (Qwen 2.5 14B) | LATCH | Improvement |
48
+ |---|---|---|---|
49
+ | **Time-To-First-Token (cold)** | 23.1s | **0.11s** | **210Γ— faster** |
50
+ | **TTFT Speedup (avg, customer pack)** | 4.47s | 0.11s | **42.9Γ—** |
51
+ | **End-to-End Query Speedup** | 6.55s | 2.02s | **5.2Γ—** |
52
+ | **Cache Reload Time** | 23.1s | **0.0016s** | **246Γ— faster** |
53
+
54
+ ### Quality β€” Customer Document Pack
55
+
56
+ | Benchmark Category | Baseline | LATCH | Delta |
57
+ |---|---|---|---|
58
+ | Cross-Document Comparison | 41.5% | **49.4%** | +7.9pp |
59
+ | Cross-Document Format | 40.5% | **68.8%** | +28.3pp |
60
+ | Cross-Document Retrieval | 40.4% | **48.1%** | +7.7pp |
61
+ | Selective Retrieval | 35.2% | **47.2%** | +12.0pp |
62
+ | **Overall Mean token-F1** | **39.4%** | **53.4%** | **+14.0pp** |
63
+
64
+ ### Benchmark Gates
65
+
66
+ | Gate | Result |
67
+ |---|---|
68
+ | Single-Document Gate | 11/12 βœ… |
69
+ | Multi-Document Gate | 11/12 βœ… |
70
+ | 256K Memory Sweep | Passing |
71
+
72
+ > **Multi-doc pass rate: 91.7%** β€” the highest of any model family in the current LATCH portfolio.
73
+
74
+ ---
75
+
76
+ ## How It Works
77
+
78
+ LATCH intercepts the standard inference path and replaces the costly per-query document processing step with a persistent representation that is prepared once and reused across all subsequent queries against the same document set.
79
+
80
+ The result is a response that begins in under 120 milliseconds β€” before the user has practically finished pressing Enter β€” regardless of how many documents are in the corpus.
81
+
82
+ The underlying method is proprietary and patent pending. CoDynamics Lab does not publish architectural details.
83
+
84
+ ---
85
+
86
+ ## Hardware Requirements
87
+
88
+ | Component | Minimum | Recommended |
89
+ |---|---|---|
90
+ | GPU | NVIDIA A100 40GB | NVIDIA A100 80GB |
91
+ | VRAM | ~30 GB | 80 GB |
92
+ | CPU RAM | 64 GB | 128 GB |
93
+ | Storage | 50 GB | 100 GB |
94
+ | Inference Runtime | vLLM | vLLM β‰₯ 0.4 |
95
+
96
+ > LATCH reduces peak VRAM consumption by approximately **50%** versus standard Qwen 2.5 14B serving, enabling more concurrent instances per node.
97
+
98
+ ---
99
+
100
+ ## Deployment Options
101
+
102
+ ### πŸ”’ Option 1: Self-Hosted License β€” $79
103
+
104
+ Run LATCH on your own A100 or H100. Your documents never leave your infrastructure.
105
+
106
+ **[Buy now at codynamicslab.gumroad.com](https://codynamicslab.gumroad.com/l/latch-qwen14b)**
107
+
108
+ Upon purchase you receive:
109
+ - Private registry pull token for the LATCH Docker image
110
+ - License key (validated at container startup)
111
+ - One-line deployment command
112
+ - Access to future runtime updates
113
+
114
+ ```bash
115
+ LICENSE_KEY=xxxx-xxxx docker compose pull && docker compose up -d
116
+ ```
117
+
118
+ Compatible with standard OpenAI-format API clients.
119
+
120
+ ---
121
+
122
+ ### ☁️ Option 2: Managed Hosted Instance β€” Coming Soon
123
+
124
+ Spin up a LATCH-ready GPU instance directly from CoDynamics Lab. No infrastructure setup required.
125
+
126
+ - Pay by the hour β€” billed by wall-clock second
127
+ - Includes batch JSON query interface
128
+ - Upload documents, submit a structured prompt list, export results with full telemetry
129
+ - Every session outputs side-by-side cost savings vs. standard Qwen baseline
130
+
131
+ **[Join the waitlist](mailto:mike@codynamicslab.com?subject=LATCH%20Managed%20Instance%20Waitlist)**
132
+
133
+ ---
134
+
135
+ ### πŸ”‘ Option 3: Gated Repository Access (Research / Enterprise)
136
+
137
+ Request direct access for evaluation, research, or enterprise licensing discussions.
138
+
139
+ ---
140
+
141
+ ## Intended Use
142
+
143
+ **Primary use cases:**
144
+ - M&A and private equity due diligence (multi-document data room analysis)
145
+ - Legal document review and cross-contract comparison
146
+ - Compliance and regulatory document monitoring
147
+ - Financial research and filing analysis
148
+ - Any high-volume, repeated-query workload against a fixed document corpus
149
+
150
+ **Out of scope:**
151
+ - Real-time web search or retrieval-augmented generation
152
+ - General-purpose conversational AI without a document corpus
153
+ - Consumer applications
154
+
155
+ ---
156
+
157
+ ## Limitations & Known Weaknesses
158
+
159
+ - **Short-context standard QA:** LATCH is optimized for long-context, multi-document workloads. It does not improve performance on standard short-context QA benchmarks.
160
+ - **Document pre-preparation required:** Documents must be prepared before querying. This is a one-time cost per document set that is fully amortized across subsequent queries.
161
+ - **Cross-document retrieval is the weakest benchmark slice:** Document-selection tasks with heavy distractors are the most challenging workload category.
162
+
163
+ ---
164
+
165
+ ## Request Access
166
+
167
+ **Three ways to get started:**
168
+
169
+ | Path | Best for | Action |
170
+ |---|---|---|
171
+ | **Self-hosted license** | Teams with their own A100/H100 who need full data privacy | [Buy on Gumroad β€” $79](https://codynamicslab.gumroad.com/l/latch-qwen14b) |
172
+ | **Managed hosted instance** | Teams who want zero infrastructure setup | [Join waitlist](mailto:mike@codynamicslab.com?subject=LATCH%20Managed%20Instance%20Waitlist) |
173
+ | **Gated repo access** | Research, enterprise evaluation, volume licensing | Click Request Access above |
174
+
175
+ For gated access requests:
176
+ 1. Click the **Request Access** button above
177
+ 2. Briefly describe your use case and organization
178
+ 3. Our team will review and respond within 2 business days
179
+
180
+ πŸ“§ [mike@codynamicslab.com](mailto:mike@codynamicslab.com)
181
+ 🌐 [www.codynamicslab.com](https://www.codynamicslab.com)
182
+
183
+ ---
184
+
185
+ ## License
186
+
187
+ This model is released under the **CoDynamics Commercial License**.
188
+ - Purchase includes a single-instance deployment license
189
+ - Commercial or production use beyond the licensed instance requires a separate agreement
190
+ - Redistribution of model weights is strictly prohibited
191
+
192
+ See [LICENSE](https://www.codynamicslab.com/license) for full terms.
193
+
194
+ ---
195
+
196
+ ## Citation
197
+
198
+ If you cite LATCH benchmark results in research, please use:
199
+
200
+ ```bibtex
201
+ @misc{codynamics2026latch,
202
+ title = {LATCH: Proprietary Long-Context Inference Layer},
203
+ author = {CoDynamics Lab Corporation},
204
+ year = {2026},
205
+ howpublished = {\url{https://huggingface.co/CoDynamicsLab/LATCH-Qwen2.5-14B}},
206
+ note = {Patent Pending. Architectural details proprietary.}
207
+ }
208
+ ```
209
+
210
+ ---
211
+
212
+ *CoDynamics Lab Corporation β€” Eliminating the Long-Context Tax.*