Update index.html
Browse files- index.html +59 -13
index.html
CHANGED
|
@@ -208,6 +208,25 @@
|
|
| 208 |
font-size: 0.9em;
|
| 209 |
}
|
| 210 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 211 |
.category-badge {
|
| 212 |
display: inline-block;
|
| 213 |
padding: 4px 12px;
|
|
@@ -321,7 +340,9 @@
|
|
| 321 |
indexing: "Preprocessing β Fixed Chunking β Simple Embedding β Vector DB Storage",
|
| 322 |
inference: "User Query β Embedding β Vector DB Lookup (Top-K) β Concatenation β LLM Generate",
|
| 323 |
benefits: "Establishes the knowledge retrieval baseline; Simple and cheap to implement.",
|
| 324 |
-
challenges: "Context loss due to rigid chunking; High hallucination risk; Poor handling of complex/multi-step queries."
|
|
|
|
|
|
|
| 325 |
},
|
| 326 |
{
|
| 327 |
type: "Self RAG",
|
|
@@ -330,7 +351,9 @@
|
|
| 330 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 331 |
inference: "User Query β LLM Generates a thought β Retrieval β LLM Generates/Evaluates Retrieved Passages β LLM Decides if Answer is Ready β Final LLM Generate",
|
| 332 |
benefits: "Reduces hallucinations by self-critique/verification; Filters out poor quality retrieved passages.",
|
| 333 |
-
challenges: "Increases inference latency (multiple LLM calls per query); Requires careful tuning of reflection/critique prompt."
|
|
|
|
|
|
|
| 334 |
},
|
| 335 |
{
|
| 336 |
type: "Modular RAG",
|
|
@@ -339,7 +362,9 @@
|
|
| 339 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 340 |
inference: "User Query β Router/Module Selection β Selected Module Executes β LLM Generate",
|
| 341 |
benefits: "Improves flexibility and component reusability; Enables optimal module selection for specific tasks.",
|
| 342 |
-
challenges: "Requires complex routing/planning logic; Overhead of training/managing multiple specialized components."
|
|
|
|
|
|
|
| 343 |
},
|
| 344 |
{
|
| 345 |
type: "Graph RAG",
|
|
@@ -348,7 +373,9 @@
|
|
| 348 |
indexing: "Preprocessing β Entity/Relation Extraction β Store in Knowledge Graph (KG) & Vector DB",
|
| 349 |
inference: "User Query β Embedding/KG Query β Simultaneous Retrieval (Vector + KG Path) β Concatenation β LLM Generate",
|
| 350 |
benefits: "Resolves complex, multi-hop queries by leveraging factual relationships; Improves interpretability and fact consistency.",
|
| 351 |
-
challenges: "High indexing complexity (KG construction); Expensive maintenance for rapidly changing data; Retrieval latency can be high."
|
|
|
|
|
|
|
| 352 |
},
|
| 353 |
{
|
| 354 |
type: "MultiModal RAG",
|
|
@@ -357,7 +384,9 @@
|
|
| 357 |
indexing: "Preprocessing β Multi-Modal Embedding (e.g., CLIP) β Stores representations of all modalities in Vector DB",
|
| 358 |
inference: "User Query (Text or Image) β Multi-Modal Embedding β Vector DB Lookup (Retrieves related text, image, metadata) β LLM Generate",
|
| 359 |
benefits: "Unlocks knowledge stored in non-text data (images, charts, tables); Provides a richer context.",
|
| 360 |
-
challenges: "Requires specialized multimodal embeddings/models; Indexing is computationally expensive; Difficult to combine disparate modalities coherently."
|
|
|
|
|
|
|
| 361 |
},
|
| 362 |
{
|
| 363 |
type: "Recursive RAG",
|
|
@@ -366,7 +395,9 @@
|
|
| 366 |
indexing: "Preprocessing β Chunking & Summarization β Embeddings of both chunks & summaries β Vector DB Storage",
|
| 367 |
inference: "User Query β Retrieval β LLM evaluates initial result β Recursive Query β Retrieve Specific Chunks β LLM Generate",
|
| 368 |
benefits: "Summarizes context or decomposes queries recursively; Handles high-level questions that require abstract understanding.",
|
| 369 |
-
challenges: "Risk of information loss during aggressive summarization; Chain of thought adds significant latency."
|
|
|
|
|
|
|
| 370 |
},
|
| 371 |
{
|
| 372 |
type: "Cache RAG",
|
|
@@ -375,7 +406,9 @@
|
|
| 375 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 376 |
inference: "User Query β Cache Lookup β If Hit: Return Cached Answer β If Miss: Standard Retrieval β LLM Generate β Cache Store",
|
| 377 |
benefits: "Dramatically improves latency and reduces LLM cost for repeated or highly similar queries.",
|
| 378 |
-
challenges: "Complex cache invalidation logic; Requires robust query similarity and hashing functions."
|
|
|
|
|
|
|
| 379 |
},
|
| 380 |
{
|
| 381 |
type: "Corrective RAG",
|
|
@@ -384,7 +417,9 @@
|
|
| 384 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 385 |
inference: "User Query β Standard Retrieval β Retrieved Docs Evaluated β Corrective Action β LLM Generate",
|
| 386 |
benefits: "Detects and corrects poor quality retrieval/generation post-hoc; Increases overall trustworthiness.",
|
| 387 |
-
challenges: "High latency due to iterative correction loops; Requires training a dedicated evaluation model."
|
|
|
|
|
|
|
| 388 |
},
|
| 389 |
{
|
| 390 |
type: "Multi-Hop RAG",
|
|
@@ -393,7 +428,9 @@
|
|
| 393 |
indexing: "Preprocessing β Chunking/Entity Extraction β Embedding β Structured Storage (Vector DB + Optional KG)",
|
| 394 |
inference: "User Query β Query Decomposition β Hop 1 Retrieval β Iterative Reasoning β Hop 2 Retrieval β Final Evidence Aggregation β LLM Generate",
|
| 395 |
benefits: "Solves questions requiring reasoning across multiple independent documents or retrieval steps.",
|
| 396 |
-
challenges: "Prone to error propagation (if one hop fails); Significantly higher latency; Requires generation of accurate intermediate queries."
|
|
|
|
|
|
|
| 397 |
},
|
| 398 |
{
|
| 399 |
type: "Agentic RAG",
|
|
@@ -402,7 +439,9 @@
|
|
| 402 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 403 |
inference: "User Query β Agent Planning/Tool Selection β Agent Executes RAG Retrieval β Agent Reflects/Synthesizes β Final LLM Generate",
|
| 404 |
benefits: "Handles complex, goal-oriented tasks via dynamic planning, tool use, and state tracking.",
|
| 405 |
-
challenges: "Highest development/orchestration complexity; Slowest inference due to planning/execution loops; Failure in planning leads to catastrophic task failure."
|
|
|
|
|
|
|
| 406 |
},
|
| 407 |
{
|
| 408 |
type: "Adaptive RAG",
|
|
@@ -411,7 +450,9 @@
|
|
| 411 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage β Train Query Complexity Classifier",
|
| 412 |
inference: "User Query β Query Classification (Router) β Adaptive Decision β Retrieval Execution β LLM Generate",
|
| 413 |
benefits: "Optimizes pipeline complexity and cost based on query assessment.",
|
| 414 |
-
challenges: "Requires training a robust query classifier/router; Misclassification can lead to poor quality results."
|
|
|
|
|
|
|
| 415 |
},
|
| 416 |
{
|
| 417 |
type: "Hierarchical RAG",
|
|
@@ -420,7 +461,9 @@
|
|
| 420 |
indexing: "Preprocessing β Hierarchical Chunking (Multiple levels) β Multiple Embeddings (for each level) β Vector DB Storage",
|
| 421 |
inference: "User Query β Embedding β Multi-Level Retrieval β Concatenation β LLM Generate",
|
| 422 |
benefits: "Solves the 'needle-in-a-haystack' problem for very long documents; Efficiently prunes non-relevant sections.",
|
| 423 |
-
challenges: "Complex, multi-level chunking and indexing structure; Requires multiple retrieval passes."
|
|
|
|
|
|
|
| 424 |
},
|
| 425 |
{
|
| 426 |
type: "Speculative RAG",
|
|
@@ -429,7 +472,9 @@
|
|
| 429 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 430 |
inference: "User Query β Standard Retrieval β Drafting LLM Generates Tokens β Verifier LLM Checks Drafted Tokens Against Context β LLM Generate",
|
| 431 |
benefits: "Significantly reduces token generation latency and LLM inference cost.",
|
| 432 |
-
challenges: "Does not inherently improve semantic quality or hallucination rate; Requires careful balance between drafting and verifier models."
|
|
|
|
|
|
|
| 433 |
}
|
| 434 |
];
|
| 435 |
|
|
@@ -483,6 +528,7 @@
|
|
| 483 |
<div class="timeline-content">
|
| 484 |
<div class="timeline-type">${rag.type}</div>
|
| 485 |
<span class="category-badge ${getCategoryBadgeClass(rag.category)}">${rag.category}</span>
|
|
|
|
| 486 |
</div>
|
| 487 |
`;
|
| 488 |
yearTimeline.appendChild(item);
|
|
|
|
| 208 |
font-size: 0.9em;
|
| 209 |
}
|
| 210 |
|
| 211 |
+
.timeline-reference {
|
| 212 |
+
color: #888;
|
| 213 |
+
font-size: 0.85em;
|
| 214 |
+
font-style: italic;
|
| 215 |
+
margin-top: 5px;
|
| 216 |
+
line-height: 1.4;
|
| 217 |
+
}
|
| 218 |
+
|
| 219 |
+
.timeline-reference a {
|
| 220 |
+
color: #667eea;
|
| 221 |
+
text-decoration: none;
|
| 222 |
+
transition: color 0.3s ease;
|
| 223 |
+
}
|
| 224 |
+
|
| 225 |
+
.timeline-reference a:hover {
|
| 226 |
+
color: #764ba2;
|
| 227 |
+
text-decoration: underline;
|
| 228 |
+
}
|
| 229 |
+
|
| 230 |
.category-badge {
|
| 231 |
display: inline-block;
|
| 232 |
padding: 4px 12px;
|
|
|
|
| 340 |
indexing: "Preprocessing β Fixed Chunking β Simple Embedding β Vector DB Storage",
|
| 341 |
inference: "User Query β Embedding β Vector DB Lookup (Top-K) β Concatenation β LLM Generate",
|
| 342 |
benefits: "Establishes the knowledge retrieval baseline; Simple and cheap to implement.",
|
| 343 |
+
challenges: "Context loss due to rigid chunking; High hallucination risk; Poor handling of complex/multi-step queries.",
|
| 344 |
+
references: "Lewis, P., et al. (2020). Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. NeurIPS 2020.",
|
| 345 |
+
paperUrl: "https://arxiv.org/abs/2005.11401"
|
| 346 |
},
|
| 347 |
{
|
| 348 |
type: "Self RAG",
|
|
|
|
| 351 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 352 |
inference: "User Query β LLM Generates a thought β Retrieval β LLM Generates/Evaluates Retrieved Passages β LLM Decides if Answer is Ready β Final LLM Generate",
|
| 353 |
benefits: "Reduces hallucinations by self-critique/verification; Filters out poor quality retrieved passages.",
|
| 354 |
+
challenges: "Increases inference latency (multiple LLM calls per query); Requires careful tuning of reflection/critique prompt.",
|
| 355 |
+
references: "Asai, A., et al. (2023). Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection. arXiv:2310.11511",
|
| 356 |
+
paperUrl: "https://arxiv.org/abs/2310.11511"
|
| 357 |
},
|
| 358 |
{
|
| 359 |
type: "Modular RAG",
|
|
|
|
| 362 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 363 |
inference: "User Query β Router/Module Selection β Selected Module Executes β LLM Generate",
|
| 364 |
benefits: "Improves flexibility and component reusability; Enables optimal module selection for specific tasks.",
|
| 365 |
+
challenges: "Requires complex routing/planning logic; Overhead of training/managing multiple specialized components.",
|
| 366 |
+
references: "Gao, Y., et al. (2024). Modular RAG: Transforming RAG Systems into LEGO-like Reconfigurable Frameworks. arXiv:2407.21059",
|
| 367 |
+
paperUrl: "https://arxiv.org/abs/2407.21059"
|
| 368 |
},
|
| 369 |
{
|
| 370 |
type: "Graph RAG",
|
|
|
|
| 373 |
indexing: "Preprocessing β Entity/Relation Extraction β Store in Knowledge Graph (KG) & Vector DB",
|
| 374 |
inference: "User Query β Embedding/KG Query β Simultaneous Retrieval (Vector + KG Path) β Concatenation β LLM Generate",
|
| 375 |
benefits: "Resolves complex, multi-hop queries by leveraging factual relationships; Improves interpretability and fact consistency.",
|
| 376 |
+
challenges: "High indexing complexity (KG construction); Expensive maintenance for rapidly changing data; Retrieval latency can be high.",
|
| 377 |
+
references: "Barry, M., et al. (2025). GraphRAG: Leveraging Graph-Based Efficiency to Minimize Hallucinations in LLM-Driven RAG. GenAIK Workshop.",
|
| 378 |
+
paperUrl: "https://aclanthology.org/2025.genaik-1.6/"
|
| 379 |
},
|
| 380 |
{
|
| 381 |
type: "MultiModal RAG",
|
|
|
|
| 384 |
indexing: "Preprocessing β Multi-Modal Embedding (e.g., CLIP) β Stores representations of all modalities in Vector DB",
|
| 385 |
inference: "User Query (Text or Image) β Multi-Modal Embedding β Vector DB Lookup (Retrieves related text, image, metadata) β LLM Generate",
|
| 386 |
benefits: "Unlocks knowledge stored in non-text data (images, charts, tables); Provides a richer context.",
|
| 387 |
+
challenges: "Requires specialized multimodal embeddings/models; Indexing is computationally expensive; Difficult to combine disparate modalities coherently.",
|
| 388 |
+
references: "Gao, Y., et al. (2024). Retrieval-Augmented Multimodal Language Modeling. CVPR 2024.",
|
| 389 |
+
paperUrl: "https://arxiv.org/abs/2211.12561"
|
| 390 |
},
|
| 391 |
{
|
| 392 |
type: "Recursive RAG",
|
|
|
|
| 395 |
indexing: "Preprocessing β Chunking & Summarization β Embeddings of both chunks & summaries β Vector DB Storage",
|
| 396 |
inference: "User Query β Retrieval β LLM evaluates initial result β Recursive Query β Retrieve Specific Chunks β LLM Generate",
|
| 397 |
benefits: "Summarizes context or decomposes queries recursively; Handles high-level questions that require abstract understanding.",
|
| 398 |
+
challenges: "Risk of information loss during aggressive summarization; Chain of thought adds significant latency.",
|
| 399 |
+
references: "Liu, Y., et al. (2024). RAG-GPT: Retrieval-Augmented Generation for Open-Domain Question Answering. IJCNN 2024.",
|
| 400 |
+
paperUrl: "https://arxiv.org/abs/2405.10627"
|
| 401 |
},
|
| 402 |
{
|
| 403 |
type: "Cache RAG",
|
|
|
|
| 406 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 407 |
inference: "User Query β Cache Lookup β If Hit: Return Cached Answer β If Miss: Standard Retrieval β LLM Generate β Cache Store",
|
| 408 |
benefits: "Dramatically improves latency and reduces LLM cost for repeated or highly similar queries.",
|
| 409 |
+
challenges: "Complex cache invalidation logic; Requires robust query similarity and hashing functions.",
|
| 410 |
+
references: "Jin, C., et al. (2024). RAGCache: Efficient Knowledge Caching for Retrieval-Augmented Generation. ACM TOCS.",
|
| 411 |
+
paperUrl: "https://arxiv.org/abs/2404.12457"
|
| 412 |
},
|
| 413 |
{
|
| 414 |
type: "Corrective RAG",
|
|
|
|
| 417 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 418 |
inference: "User Query β Standard Retrieval β Retrieved Docs Evaluated β Corrective Action β LLM Generate",
|
| 419 |
benefits: "Detects and corrects poor quality retrieval/generation post-hoc; Increases overall trustworthiness.",
|
| 420 |
+
challenges: "High latency due to iterative correction loops; Requires training a dedicated evaluation model.",
|
| 421 |
+
references: "Yan, S.-Q., et al. (2024). Corrective Retrieval Augmented Generation. arXiv:2401.15884",
|
| 422 |
+
paperUrl: "https://arxiv.org/abs/2401.15884"
|
| 423 |
},
|
| 424 |
{
|
| 425 |
type: "Multi-Hop RAG",
|
|
|
|
| 428 |
indexing: "Preprocessing β Chunking/Entity Extraction β Embedding β Structured Storage (Vector DB + Optional KG)",
|
| 429 |
inference: "User Query β Query Decomposition β Hop 1 Retrieval β Iterative Reasoning β Hop 2 Retrieval β Final Evidence Aggregation β LLM Generate",
|
| 430 |
benefits: "Solves questions requiring reasoning across multiple independent documents or retrieval steps.",
|
| 431 |
+
challenges: "Prone to error propagation (if one hop fails); Significantly higher latency; Requires generation of accurate intermediate queries.",
|
| 432 |
+
references: "Tang, B., & Yang, Y. (2024). MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries. COLM 2024.",
|
| 433 |
+
paperUrl: "https://arxiv.org/abs/2401.15391"
|
| 434 |
},
|
| 435 |
{
|
| 436 |
type: "Agentic RAG",
|
|
|
|
| 439 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 440 |
inference: "User Query β Agent Planning/Tool Selection β Agent Executes RAG Retrieval β Agent Reflects/Synthesizes β Final LLM Generate",
|
| 441 |
benefits: "Handles complex, goal-oriented tasks via dynamic planning, tool use, and state tracking.",
|
| 442 |
+
challenges: "Highest development/orchestration complexity; Slowest inference due to planning/execution loops; Failure in planning leads to catastrophic task failure.",
|
| 443 |
+
references: "Singh, A., et al. (2025). Agentic Retrieval-Augmented Generation: A Survey on Agentic RAG.",
|
| 444 |
+
paperUrl: "https://arxiv.org/abs/2412.09550"
|
| 445 |
},
|
| 446 |
{
|
| 447 |
type: "Adaptive RAG",
|
|
|
|
| 450 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage β Train Query Complexity Classifier",
|
| 451 |
inference: "User Query β Query Classification (Router) β Adaptive Decision β Retrieval Execution β LLM Generate",
|
| 452 |
benefits: "Optimizes pipeline complexity and cost based on query assessment.",
|
| 453 |
+
challenges: "Requires training a robust query classifier/router; Misclassification can lead to poor quality results.",
|
| 454 |
+
references: "Jeong, S., et al. (2024). Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity. NAACL 2024.",
|
| 455 |
+
paperUrl: "https://arxiv.org/abs/2403.14403"
|
| 456 |
},
|
| 457 |
{
|
| 458 |
type: "Hierarchical RAG",
|
|
|
|
| 461 |
indexing: "Preprocessing β Hierarchical Chunking (Multiple levels) β Multiple Embeddings (for each level) β Vector DB Storage",
|
| 462 |
inference: "User Query β Embedding β Multi-Level Retrieval β Concatenation β LLM Generate",
|
| 463 |
benefits: "Solves the 'needle-in-a-haystack' problem for very long documents; Efficiently prunes non-relevant sections.",
|
| 464 |
+
challenges: "Complex, multi-level chunking and indexing structure; Requires multiple retrieval passes.",
|
| 465 |
+
references: "Huang, H., et al. (2025). Retrieval-Augmented Generation with Hierarchical Knowledge (HiRAG). EMNLP 2025.",
|
| 466 |
+
paperUrl: "https://aclanthology.org/2025.emnlp-main.1/"
|
| 467 |
},
|
| 468 |
{
|
| 469 |
type: "Speculative RAG",
|
|
|
|
| 472 |
indexing: "Preprocessing β Standard Chunking β Embedding β Vector DB Storage",
|
| 473 |
inference: "User Query β Standard Retrieval β Drafting LLM Generates Tokens β Verifier LLM Checks Drafted Tokens Against Context β LLM Generate",
|
| 474 |
benefits: "Significantly reduces token generation latency and LLM inference cost.",
|
| 475 |
+
challenges: "Does not inherently improve semantic quality or hallucination rate; Requires careful balance between drafting and verifier models.",
|
| 476 |
+
references: "Wang, Z., et al. (2025). Speculative RAG: Enhancing Retrieval Augmented Generation through Drafting. ICLR 2025.",
|
| 477 |
+
paperUrl: "https://arxiv.org/abs/2407.08223"
|
| 478 |
}
|
| 479 |
];
|
| 480 |
|
|
|
|
| 528 |
<div class="timeline-content">
|
| 529 |
<div class="timeline-type">${rag.type}</div>
|
| 530 |
<span class="category-badge ${getCategoryBadgeClass(rag.category)}">${rag.category}</span>
|
| 531 |
+
<div class="timeline-reference">π <a href="${rag.paperUrl}" target="_blank">${rag.references}</a></div>
|
| 532 |
</div>
|
| 533 |
`;
|
| 534 |
yearTimeline.appendChild(item);
|