Spaces:
Running
Running
RAG Documents
This folder contains documents used by the AI chatbot for Retrieval-Augmented Generation (RAG).
Supported File Types
.txt- Plain text files.md- Markdown files
How to Add Documents
- Place your documentation files in this folder
- Delete
embeddings_cache.jsonif it exists (to force re-indexing) - The chatbot will automatically index new documents on startup
- Documents are chunked and embedded for semantic search
Document Inventory
Category 1: General LLM/Transformer Knowledge
what_is_an_llm.md- Neural networks, language models, next-token predictiontransformer_architecture.md- Layers, encoder/decoder, residual streamtokenization_explained.md- Subword tokenization, BPE, token IDsembeddings_explained.md- Lookup tables, vector spaces, positional encodingsattention_mechanism.md- Q/K/V, multi-head attention, intuitive explanationsmlp_layers_explained.md- Feed-forward networks, knowledge storage, expand-compressoutput_and_prediction.md- Logits, softmax, temperature, greedy vs. samplingkey_terminology.md- Extended glossary of ML/transformer terms
Category 2: Dashboard Components
dashboard_overview.md- Layout tour, navigation, typical workflowpipeline_stages.md- What each of the 5 pipeline stages showsablation_panel_guide.md- How to use the ablation experiment panelattribution_panel_guide.md- How to use the token attribution panelbeam_search_and_generation.md- Beam search, generation controlshead_categories_explained.md- Previous-Token, Positional, BoW, Syntactic, Othermodel_selector_guide.md- Choosing models, auto-detection, generation settings
Category 3: Model-Specific Documentation
gpt2_overview.md- GPT-2 architecture, why it's a good starter, variantsgpt_neo_overview.md- GPT-Neo architecture, local attention, comparison with GPT-2pythia_overview.md- Pythia architecture, RoPE, parallel attn+MLP, interpretability focusopt_overview.md- OPT architecture, ReLU activation, comparison with GPT-2qwen_overview.md- Qwen2.5 (LLaMA-like) architecture, RMSNorm, SiLU, GQA
Category 4: Guided Experiments (Step-by-Step)
experiment_first_analysis.md- Your first analysis with GPT-2experiment_exploring_attention.md- Reading attention patterns and head categoriesexperiment_first_ablation.md- Removing a head and observing the effectexperiment_token_attribution.md- Measuring token influence with gradientsexperiment_comparing_heads.md- Systematic comparison across head categoriesexperiment_beam_search.md- Exploring alternative generation paths
Category 5: Interpretation, Troubleshooting, and Research
interpreting_ablation_results.md- How to read ablation probability changesinterpreting_attribution_scores.md- Understanding attribution score valuesinterpreting_attention_maps.md- Reading BertViz patterns visuallytroubleshooting_and_faq.md- Common issues and frequently asked questionsrecommended_starting_points.md- Best models, prompts, and experiment ordermechanistic_interpretability_intro.md- Mech interp research context
Notes
- Large files will be automatically chunked (~500 tokens per chunk)
- Embeddings are cached in
embeddings_cache.jsonfor faster subsequent loads - Delete
embeddings_cache.jsonto force re-indexing of all documents