Spaces:
Sleeping
Sleeping
metadata
title: RAG Capstone Project
emoji: π€
colorFrom: blue
colorTo: purple
sdk: docker
app_file: streamlit_app.py
pinned: false
license: mit
π€ RAG Capstone Project
A comprehensive Retrieval-Augmented Generation (RAG) system with TRACE evaluation metrics for medical/clinical domains.
Features
- π Multiple RAG Bench Datasets: CovidQA, CUAD, FinQA, HotpotQA, PubMedQA, and more
- π§© Chunking Strategies: Dense, Sparse, Hybrid, Re-ranking, Row-based, Entity-based
- π€ Medical Embedding Models:
- sentence-transformers/all-mpnet-base-v2
- emilyalsentzer/Bio_ClinicalBERT
- microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract
- πΎ ChromaDB Vector Storage: Persistent vector storage with efficient retrieval
- βοΈ Groq LLM Integration: Cloud-based inference with rate limiting
- π TRACE Evaluation Metrics:
- Utilization: How well the system uses retrieved documents
- Relevance: Relevance of retrieved documents to the query
- Adherence: How well the response adheres to the retrieved context
- Completeness: How complete the response is
- π¬ Chat Interface: Streamlit-based interactive chat with history
Usage
- Enter your Groq API Key in the sidebar
- Select a dataset from RAG Bench
- Choose a chunking strategy (dense, sparse, hybrid, re-ranking)
- Select an embedding model for document vectorization
- Choose an LLM model for response generation
- Click "Load Data & Create Collection" to initialize
- Start chatting in the chat interface
- View retrieved documents and evaluation metrics
- Run TRACE evaluation on test data
Environment Variables
Set these in your Hugging Face Space secrets:
GROQ_API_KEY: Your Groq API key (required)GROQ_API_KEYS: Comma-separated list of API keys for rotation (optional)
License
MIT License