Spaces:
Build error
Build error
| # Ollama LLM server | |
| OLLAMA_BASE_URL=http://localhost:11434 | |
| MODEL_NAME="llama3.2" | |
| # STT (faster-whisper model size: tiny, base, small, medium, large-v3) | |
| STT_MODEL=base | |
| # TTS (edge-tts voice name, e.g. en-US-AriaNeural, en-GB-SoniaNeural) | |
| TTS_VOICE=en-US-AriaNeural | |
| # Local vision model (only used with --local-vision CLI flag) | |
| LOCAL_VISION_MODEL=HuggingFaceTB/SmolVLM2-2.2B-Instruct | |
| # Cache for local VLM (only used with --local-vision CLI flag) | |
| HF_HOME=./cache | |
| # Hugging Face token for accessing datasets/models | |
| HF_TOKEN= | |
| # Profile selection (defaults to "default" when unset) | |
| REACHY_MINI_CUSTOM_PROFILE="example" | |
| # Optional external profile/tool directories | |
| # REACHY_MINI_EXTERNAL_PROFILES_DIRECTORY=external_content/external_profiles | |
| # REACHY_MINI_EXTERNAL_TOOLS_DIRECTORY=external_content/external_tools | |
| # Optional: discover and auto-load all tools found in REACHY_MINI_EXTERNAL_TOOLS_DIRECTORY, | |
| # even if they are not listed in the selected profile's tools.txt. | |
| # This is convenient for downloaded tools used with built-in/default profiles. | |
| # AUTOLOAD_EXTERNAL_TOOLS=1 |