File size: 506 Bytes
b269c5d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 |
# Model configuration
MODEL_REPO=lmstudio-community/gemma-3n-E4B-it-text-GGUF
MODEL_FILENAME=gemma-3n-E4B-it-Q8_0.gguf
MODEL_PATH=./models/gemma-3n-E4B-it-Q8_0.gguf
HUGGINGFACE_TOKEN=
# Model parameters - optimized for Docker containers
N_CTX=4096
N_GPU_LAYERS=0
N_THREADS=4
N_BATCH=512
USE_MLOCK=false
USE_MMAP=true
F16_KV=true
SEED=42
# Server settings
HOST=0.0.0.0
GRADIO_PORT=7860
API_PORT=8000
# Generation settings
MAX_NEW_TOKENS=256
TEMPERATURE=0.1
# File upload settings
MAX_FILE_SIZE=10485760
|