Update Dockerfile to use app_v4 with background model loading 7ae4b71 MiniMax Agent commited on Jan 1
Add v4 with background model loading - prevents timeout by loading model after server starts 44ffe48 MiniMax Agent commited on Jan 1
Add complete local Ollama setup with OpenELM - includes setup script, API server, test scripts, and documentation 41831f1 MiniMax Agent commited on Jan 1
Fix tokenizer issues: add sentencepiece dependency and robust loading strategy d36a46f MiniMax Agent commited on Jan 1
Implement lazy loading - model loads on first request to avoid startup timeouts 2aeb5c7 MiniMax Agent commited on Jan 1
Fix OpenELM tokenizer loading - use LlamaTokenizer as fallback 3daef91 MiniMax Agent commited on Jan 1