Soha85 commited on
Commit
90e2d2b
Β·
verified Β·
1 Parent(s): 6f657ba

Update index.html

Browse files
Files changed (1) hide show
  1. index.html +59 -13
index.html CHANGED
@@ -208,6 +208,25 @@
208
  font-size: 0.9em;
209
  }
210
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
211
  .category-badge {
212
  display: inline-block;
213
  padding: 4px 12px;
@@ -321,7 +340,9 @@
321
  indexing: "Preprocessing β†’ Fixed Chunking β†’ Simple Embedding β†’ Vector DB Storage",
322
  inference: "User Query β†’ Embedding β†’ Vector DB Lookup (Top-K) β†’ Concatenation β†’ LLM Generate",
323
  benefits: "Establishes the knowledge retrieval baseline; Simple and cheap to implement.",
324
- challenges: "Context loss due to rigid chunking; High hallucination risk; Poor handling of complex/multi-step queries."
 
 
325
  },
326
  {
327
  type: "Self RAG",
@@ -330,7 +351,9 @@
330
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
331
  inference: "User Query β†’ LLM Generates a thought β†’ Retrieval β†’ LLM Generates/Evaluates Retrieved Passages β†’ LLM Decides if Answer is Ready β†’ Final LLM Generate",
332
  benefits: "Reduces hallucinations by self-critique/verification; Filters out poor quality retrieved passages.",
333
- challenges: "Increases inference latency (multiple LLM calls per query); Requires careful tuning of reflection/critique prompt."
 
 
334
  },
335
  {
336
  type: "Modular RAG",
@@ -339,7 +362,9 @@
339
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
340
  inference: "User Query β†’ Router/Module Selection β†’ Selected Module Executes β†’ LLM Generate",
341
  benefits: "Improves flexibility and component reusability; Enables optimal module selection for specific tasks.",
342
- challenges: "Requires complex routing/planning logic; Overhead of training/managing multiple specialized components."
 
 
343
  },
344
  {
345
  type: "Graph RAG",
@@ -348,7 +373,9 @@
348
  indexing: "Preprocessing β†’ Entity/Relation Extraction β†’ Store in Knowledge Graph (KG) & Vector DB",
349
  inference: "User Query β†’ Embedding/KG Query β†’ Simultaneous Retrieval (Vector + KG Path) β†’ Concatenation β†’ LLM Generate",
350
  benefits: "Resolves complex, multi-hop queries by leveraging factual relationships; Improves interpretability and fact consistency.",
351
- challenges: "High indexing complexity (KG construction); Expensive maintenance for rapidly changing data; Retrieval latency can be high."
 
 
352
  },
353
  {
354
  type: "MultiModal RAG",
@@ -357,7 +384,9 @@
357
  indexing: "Preprocessing β†’ Multi-Modal Embedding (e.g., CLIP) β†’ Stores representations of all modalities in Vector DB",
358
  inference: "User Query (Text or Image) β†’ Multi-Modal Embedding β†’ Vector DB Lookup (Retrieves related text, image, metadata) β†’ LLM Generate",
359
  benefits: "Unlocks knowledge stored in non-text data (images, charts, tables); Provides a richer context.",
360
- challenges: "Requires specialized multimodal embeddings/models; Indexing is computationally expensive; Difficult to combine disparate modalities coherently."
 
 
361
  },
362
  {
363
  type: "Recursive RAG",
@@ -366,7 +395,9 @@
366
  indexing: "Preprocessing β†’ Chunking & Summarization β†’ Embeddings of both chunks & summaries β†’ Vector DB Storage",
367
  inference: "User Query β†’ Retrieval β†’ LLM evaluates initial result β†’ Recursive Query β†’ Retrieve Specific Chunks β†’ LLM Generate",
368
  benefits: "Summarizes context or decomposes queries recursively; Handles high-level questions that require abstract understanding.",
369
- challenges: "Risk of information loss during aggressive summarization; Chain of thought adds significant latency."
 
 
370
  },
371
  {
372
  type: "Cache RAG",
@@ -375,7 +406,9 @@
375
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
376
  inference: "User Query β†’ Cache Lookup β†’ If Hit: Return Cached Answer β†’ If Miss: Standard Retrieval β†’ LLM Generate β†’ Cache Store",
377
  benefits: "Dramatically improves latency and reduces LLM cost for repeated or highly similar queries.",
378
- challenges: "Complex cache invalidation logic; Requires robust query similarity and hashing functions."
 
 
379
  },
380
  {
381
  type: "Corrective RAG",
@@ -384,7 +417,9 @@
384
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
385
  inference: "User Query β†’ Standard Retrieval β†’ Retrieved Docs Evaluated β†’ Corrective Action β†’ LLM Generate",
386
  benefits: "Detects and corrects poor quality retrieval/generation post-hoc; Increases overall trustworthiness.",
387
- challenges: "High latency due to iterative correction loops; Requires training a dedicated evaluation model."
 
 
388
  },
389
  {
390
  type: "Multi-Hop RAG",
@@ -393,7 +428,9 @@
393
  indexing: "Preprocessing β†’ Chunking/Entity Extraction β†’ Embedding β†’ Structured Storage (Vector DB + Optional KG)",
394
  inference: "User Query β†’ Query Decomposition β†’ Hop 1 Retrieval β†’ Iterative Reasoning β†’ Hop 2 Retrieval β†’ Final Evidence Aggregation β†’ LLM Generate",
395
  benefits: "Solves questions requiring reasoning across multiple independent documents or retrieval steps.",
396
- challenges: "Prone to error propagation (if one hop fails); Significantly higher latency; Requires generation of accurate intermediate queries."
 
 
397
  },
398
  {
399
  type: "Agentic RAG",
@@ -402,7 +439,9 @@
402
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
403
  inference: "User Query β†’ Agent Planning/Tool Selection β†’ Agent Executes RAG Retrieval β†’ Agent Reflects/Synthesizes β†’ Final LLM Generate",
404
  benefits: "Handles complex, goal-oriented tasks via dynamic planning, tool use, and state tracking.",
405
- challenges: "Highest development/orchestration complexity; Slowest inference due to planning/execution loops; Failure in planning leads to catastrophic task failure."
 
 
406
  },
407
  {
408
  type: "Adaptive RAG",
@@ -411,7 +450,9 @@
411
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage β†’ Train Query Complexity Classifier",
412
  inference: "User Query β†’ Query Classification (Router) β†’ Adaptive Decision β†’ Retrieval Execution β†’ LLM Generate",
413
  benefits: "Optimizes pipeline complexity and cost based on query assessment.",
414
- challenges: "Requires training a robust query classifier/router; Misclassification can lead to poor quality results."
 
 
415
  },
416
  {
417
  type: "Hierarchical RAG",
@@ -420,7 +461,9 @@
420
  indexing: "Preprocessing β†’ Hierarchical Chunking (Multiple levels) β†’ Multiple Embeddings (for each level) β†’ Vector DB Storage",
421
  inference: "User Query β†’ Embedding β†’ Multi-Level Retrieval β†’ Concatenation β†’ LLM Generate",
422
  benefits: "Solves the 'needle-in-a-haystack' problem for very long documents; Efficiently prunes non-relevant sections.",
423
- challenges: "Complex, multi-level chunking and indexing structure; Requires multiple retrieval passes."
 
 
424
  },
425
  {
426
  type: "Speculative RAG",
@@ -429,7 +472,9 @@
429
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
430
  inference: "User Query β†’ Standard Retrieval β†’ Drafting LLM Generates Tokens β†’ Verifier LLM Checks Drafted Tokens Against Context β†’ LLM Generate",
431
  benefits: "Significantly reduces token generation latency and LLM inference cost.",
432
- challenges: "Does not inherently improve semantic quality or hallucination rate; Requires careful balance between drafting and verifier models."
 
 
433
  }
434
  ];
435
 
@@ -483,6 +528,7 @@
483
  <div class="timeline-content">
484
  <div class="timeline-type">${rag.type}</div>
485
  <span class="category-badge ${getCategoryBadgeClass(rag.category)}">${rag.category}</span>
 
486
  </div>
487
  `;
488
  yearTimeline.appendChild(item);
 
208
  font-size: 0.9em;
209
  }
210
 
211
+ .timeline-reference {
212
+ color: #888;
213
+ font-size: 0.85em;
214
+ font-style: italic;
215
+ margin-top: 5px;
216
+ line-height: 1.4;
217
+ }
218
+
219
+ .timeline-reference a {
220
+ color: #667eea;
221
+ text-decoration: none;
222
+ transition: color 0.3s ease;
223
+ }
224
+
225
+ .timeline-reference a:hover {
226
+ color: #764ba2;
227
+ text-decoration: underline;
228
+ }
229
+
230
  .category-badge {
231
  display: inline-block;
232
  padding: 4px 12px;
 
340
  indexing: "Preprocessing β†’ Fixed Chunking β†’ Simple Embedding β†’ Vector DB Storage",
341
  inference: "User Query β†’ Embedding β†’ Vector DB Lookup (Top-K) β†’ Concatenation β†’ LLM Generate",
342
  benefits: "Establishes the knowledge retrieval baseline; Simple and cheap to implement.",
343
+ challenges: "Context loss due to rigid chunking; High hallucination risk; Poor handling of complex/multi-step queries.",
344
+ references: "Lewis, P., et al. (2020). Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. NeurIPS 2020.",
345
+ paperUrl: "https://arxiv.org/abs/2005.11401"
346
  },
347
  {
348
  type: "Self RAG",
 
351
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
352
  inference: "User Query β†’ LLM Generates a thought β†’ Retrieval β†’ LLM Generates/Evaluates Retrieved Passages β†’ LLM Decides if Answer is Ready β†’ Final LLM Generate",
353
  benefits: "Reduces hallucinations by self-critique/verification; Filters out poor quality retrieved passages.",
354
+ challenges: "Increases inference latency (multiple LLM calls per query); Requires careful tuning of reflection/critique prompt.",
355
+ references: "Asai, A., et al. (2023). Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection. arXiv:2310.11511",
356
+ paperUrl: "https://arxiv.org/abs/2310.11511"
357
  },
358
  {
359
  type: "Modular RAG",
 
362
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
363
  inference: "User Query β†’ Router/Module Selection β†’ Selected Module Executes β†’ LLM Generate",
364
  benefits: "Improves flexibility and component reusability; Enables optimal module selection for specific tasks.",
365
+ challenges: "Requires complex routing/planning logic; Overhead of training/managing multiple specialized components.",
366
+ references: "Gao, Y., et al. (2024). Modular RAG: Transforming RAG Systems into LEGO-like Reconfigurable Frameworks. arXiv:2407.21059",
367
+ paperUrl: "https://arxiv.org/abs/2407.21059"
368
  },
369
  {
370
  type: "Graph RAG",
 
373
  indexing: "Preprocessing β†’ Entity/Relation Extraction β†’ Store in Knowledge Graph (KG) & Vector DB",
374
  inference: "User Query β†’ Embedding/KG Query β†’ Simultaneous Retrieval (Vector + KG Path) β†’ Concatenation β†’ LLM Generate",
375
  benefits: "Resolves complex, multi-hop queries by leveraging factual relationships; Improves interpretability and fact consistency.",
376
+ challenges: "High indexing complexity (KG construction); Expensive maintenance for rapidly changing data; Retrieval latency can be high.",
377
+ references: "Barry, M., et al. (2025). GraphRAG: Leveraging Graph-Based Efficiency to Minimize Hallucinations in LLM-Driven RAG. GenAIK Workshop.",
378
+ paperUrl: "https://aclanthology.org/2025.genaik-1.6/"
379
  },
380
  {
381
  type: "MultiModal RAG",
 
384
  indexing: "Preprocessing β†’ Multi-Modal Embedding (e.g., CLIP) β†’ Stores representations of all modalities in Vector DB",
385
  inference: "User Query (Text or Image) β†’ Multi-Modal Embedding β†’ Vector DB Lookup (Retrieves related text, image, metadata) β†’ LLM Generate",
386
  benefits: "Unlocks knowledge stored in non-text data (images, charts, tables); Provides a richer context.",
387
+ challenges: "Requires specialized multimodal embeddings/models; Indexing is computationally expensive; Difficult to combine disparate modalities coherently.",
388
+ references: "Gao, Y., et al. (2024). Retrieval-Augmented Multimodal Language Modeling. CVPR 2024.",
389
+ paperUrl: "https://arxiv.org/abs/2211.12561"
390
  },
391
  {
392
  type: "Recursive RAG",
 
395
  indexing: "Preprocessing β†’ Chunking & Summarization β†’ Embeddings of both chunks & summaries β†’ Vector DB Storage",
396
  inference: "User Query β†’ Retrieval β†’ LLM evaluates initial result β†’ Recursive Query β†’ Retrieve Specific Chunks β†’ LLM Generate",
397
  benefits: "Summarizes context or decomposes queries recursively; Handles high-level questions that require abstract understanding.",
398
+ challenges: "Risk of information loss during aggressive summarization; Chain of thought adds significant latency.",
399
+ references: "Liu, Y., et al. (2024). RAG-GPT: Retrieval-Augmented Generation for Open-Domain Question Answering. IJCNN 2024.",
400
+ paperUrl: "https://arxiv.org/abs/2405.10627"
401
  },
402
  {
403
  type: "Cache RAG",
 
406
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
407
  inference: "User Query β†’ Cache Lookup β†’ If Hit: Return Cached Answer β†’ If Miss: Standard Retrieval β†’ LLM Generate β†’ Cache Store",
408
  benefits: "Dramatically improves latency and reduces LLM cost for repeated or highly similar queries.",
409
+ challenges: "Complex cache invalidation logic; Requires robust query similarity and hashing functions.",
410
+ references: "Jin, C., et al. (2024). RAGCache: Efficient Knowledge Caching for Retrieval-Augmented Generation. ACM TOCS.",
411
+ paperUrl: "https://arxiv.org/abs/2404.12457"
412
  },
413
  {
414
  type: "Corrective RAG",
 
417
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
418
  inference: "User Query β†’ Standard Retrieval β†’ Retrieved Docs Evaluated β†’ Corrective Action β†’ LLM Generate",
419
  benefits: "Detects and corrects poor quality retrieval/generation post-hoc; Increases overall trustworthiness.",
420
+ challenges: "High latency due to iterative correction loops; Requires training a dedicated evaluation model.",
421
+ references: "Yan, S.-Q., et al. (2024). Corrective Retrieval Augmented Generation. arXiv:2401.15884",
422
+ paperUrl: "https://arxiv.org/abs/2401.15884"
423
  },
424
  {
425
  type: "Multi-Hop RAG",
 
428
  indexing: "Preprocessing β†’ Chunking/Entity Extraction β†’ Embedding β†’ Structured Storage (Vector DB + Optional KG)",
429
  inference: "User Query β†’ Query Decomposition β†’ Hop 1 Retrieval β†’ Iterative Reasoning β†’ Hop 2 Retrieval β†’ Final Evidence Aggregation β†’ LLM Generate",
430
  benefits: "Solves questions requiring reasoning across multiple independent documents or retrieval steps.",
431
+ challenges: "Prone to error propagation (if one hop fails); Significantly higher latency; Requires generation of accurate intermediate queries.",
432
+ references: "Tang, B., & Yang, Y. (2024). MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries. COLM 2024.",
433
+ paperUrl: "https://arxiv.org/abs/2401.15391"
434
  },
435
  {
436
  type: "Agentic RAG",
 
439
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
440
  inference: "User Query β†’ Agent Planning/Tool Selection β†’ Agent Executes RAG Retrieval β†’ Agent Reflects/Synthesizes β†’ Final LLM Generate",
441
  benefits: "Handles complex, goal-oriented tasks via dynamic planning, tool use, and state tracking.",
442
+ challenges: "Highest development/orchestration complexity; Slowest inference due to planning/execution loops; Failure in planning leads to catastrophic task failure.",
443
+ references: "Singh, A., et al. (2025). Agentic Retrieval-Augmented Generation: A Survey on Agentic RAG.",
444
+ paperUrl: "https://arxiv.org/abs/2412.09550"
445
  },
446
  {
447
  type: "Adaptive RAG",
 
450
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage β†’ Train Query Complexity Classifier",
451
  inference: "User Query β†’ Query Classification (Router) β†’ Adaptive Decision β†’ Retrieval Execution β†’ LLM Generate",
452
  benefits: "Optimizes pipeline complexity and cost based on query assessment.",
453
+ challenges: "Requires training a robust query classifier/router; Misclassification can lead to poor quality results.",
454
+ references: "Jeong, S., et al. (2024). Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity. NAACL 2024.",
455
+ paperUrl: "https://arxiv.org/abs/2403.14403"
456
  },
457
  {
458
  type: "Hierarchical RAG",
 
461
  indexing: "Preprocessing β†’ Hierarchical Chunking (Multiple levels) β†’ Multiple Embeddings (for each level) β†’ Vector DB Storage",
462
  inference: "User Query β†’ Embedding β†’ Multi-Level Retrieval β†’ Concatenation β†’ LLM Generate",
463
  benefits: "Solves the 'needle-in-a-haystack' problem for very long documents; Efficiently prunes non-relevant sections.",
464
+ challenges: "Complex, multi-level chunking and indexing structure; Requires multiple retrieval passes.",
465
+ references: "Huang, H., et al. (2025). Retrieval-Augmented Generation with Hierarchical Knowledge (HiRAG). EMNLP 2025.",
466
+ paperUrl: "https://aclanthology.org/2025.emnlp-main.1/"
467
  },
468
  {
469
  type: "Speculative RAG",
 
472
  indexing: "Preprocessing β†’ Standard Chunking β†’ Embedding β†’ Vector DB Storage",
473
  inference: "User Query β†’ Standard Retrieval β†’ Drafting LLM Generates Tokens β†’ Verifier LLM Checks Drafted Tokens Against Context β†’ LLM Generate",
474
  benefits: "Significantly reduces token generation latency and LLM inference cost.",
475
+ challenges: "Does not inherently improve semantic quality or hallucination rate; Requires careful balance between drafting and verifier models.",
476
+ references: "Wang, Z., et al. (2025). Speculative RAG: Enhancing Retrieval Augmented Generation through Drafting. ICLR 2025.",
477
+ paperUrl: "https://arxiv.org/abs/2407.08223"
478
  }
479
  ];
480
 
 
528
  <div class="timeline-content">
529
  <div class="timeline-type">${rag.type}</div>
530
  <span class="category-badge ${getCategoryBadgeClass(rag.category)}">${rag.category}</span>
531
+ <div class="timeline-reference">πŸ“„ <a href="${rag.paperUrl}" target="_blank">${rag.references}</a></div>
532
  </div>
533
  `;
534
  yearTimeline.appendChild(item);