Update README.md
Browse files
README.md
CHANGED
|
@@ -144,7 +144,7 @@ for doc, score in zip(documents, scores):
|
|
| 144 |
```
|
| 145 |
|
| 146 |
### Offline Mode (Cached Doc State)
|
| 147 |
-
For scenarios where documents are static but queries change (e.g., Search Engines, RAG), you can **pre-compute and cache the document states**. This reduces query-time latency from
|
| 148 |
|
| 149 |
#### Workflow
|
| 150 |
|
|
@@ -218,7 +218,7 @@ for doc, score in zip(documents, scores):
|
|
| 218 |
| Feature | 1. Embedding (Cosine) | 2. Online Reranking | 3. Offline Reranking |
|
| 219 |
| :--- | :--- | :--- | :--- |
|
| 220 |
| **Accuracy** | Good | **Best** | **Best** (Identical to Online) |
|
| 221 |
-
| **Latency** | Extremely Fast | Slow (
|
| 222 |
| **Input** | Query & Doc separate | `Instruct + Doc + Query` | `Query` (on top of cached Doc) |
|
| 223 |
| **Storage** | Low (Vector only) | None | High (Stores Hidden States) |
|
| 224 |
| **Best For** | Initial Retrieval (Top-k) | Reranking few candidates | Reranking many candidates |
|
|
|
|
| 144 |
```
|
| 145 |
|
| 146 |
### Offline Mode (Cached Doc State)
|
| 147 |
+
For scenarios where documents are static but queries change (e.g., Search Engines, RAG), you can **pre-compute and cache the document states**. This reduces query-time latency from O(L_doc + L_query) to just O(L_query).
|
| 148 |
|
| 149 |
#### Workflow
|
| 150 |
|
|
|
|
| 218 |
| Feature | 1. Embedding (Cosine) | 2. Online Reranking | 3. Offline Reranking |
|
| 219 |
| :--- | :--- | :--- | :--- |
|
| 220 |
| **Accuracy** | Good | **Best** | **Best** (Identical to Online) |
|
| 221 |
+
| **Latency** | Extremely Fast | Slow O(L_doc + L_query) | Fast O(L_query) only |
|
| 222 |
| **Input** | Query & Doc separate | `Instruct + Doc + Query` | `Query` (on top of cached Doc) |
|
| 223 |
| **Storage** | Low (Vector only) | None | High (Stores Hidden States) |
|
| 224 |
| **Best For** | Initial Retrieval (Top-k) | Reranking few candidates | Reranking many candidates |
|