Spaces:
Sleeping
Sleeping
| import os | |
| from pathlib import Path | |
| from dotenv import load_dotenv | |
| load_dotenv() | |
| # Base Paths | |
| BASE_DIR = Path(__file__).resolve().parent.parent | |
| DATA_DIR = BASE_DIR / "data" | |
| SRC_DIR = BASE_DIR / "src" | |
| # Data Paths | |
| PDF_PATH = DATA_DIR / "source.pdf" # We will rename the input PDF to this | |
| VECTORSTORE_PATH = DATA_DIR / "faiss_index" | |
| # RAG Parameters | |
| CHUNK_SIZE = 1000 | |
| CHUNK_OVERLAP = 200 | |
| EMBEDDING_MODEL_NAME = "sentence-transformers/all-MiniLM-L6-v2" | |
| # LLM Parameters (Hugging Face free Inference API) | |
| # Default router model should exist on the router. Override via HF_MODEL_ID env var or UI input. | |
| # Meta Llama 3 8B Instruct is widely available on the HF router as of Nov 2024. | |
| HF_MODEL_ID = os.getenv("HF_MODEL_ID", "meta-llama/Meta-Llama-3-8B-Instruct") | |
| HF_API_TOKEN = os.getenv("HUGGINGFACEHUB_API_TOKEN", "") # Optional for many free endpoints | |
| LOCAL_MODEL_ID = os.getenv("LOCAL_MODEL_ID", "distilgpt2") | |
| TEMPERATURE = float(os.getenv("HF_TEMPERATURE", "0.3")) | |