Spaces:
Paused
Paused
| FROM ubuntu:22.04 | |
| # Set environment variables | |
| ENV DEBIAN_FRONTEND=noninteractive | |
| ENV HOME=/root | |
| # CUDA paths for when running on GPU (RunPod) | |
| ENV CUDA_HOME=/usr/local/cuda | |
| ENV PATH=${CUDA_HOME}/bin:${PATH} | |
| ENV LD_LIBRARY_PATH=${CUDA_HOME}/lib64:${LD_LIBRARY_PATH} | |
| # Install dependencies, code-server, and Ollama | |
| RUN apt-get update && \ | |
| apt-get install -y curl wget gpg apt-transport-https git software-properties-common && \ | |
| # Add Python 3.12 | |
| add-apt-repository ppa:deadsnakes/ppa && \ | |
| apt-get update && \ | |
| apt-get install -y python3.12 python3.12-venv python3.12-dev && \ | |
| # Make Python 3.12 the default | |
| update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.12 1 && \ | |
| update-alternatives --set python3 /usr/bin/python3.12 && \ | |
| # Install pip for Python 3.12 | |
| curl -sS https://bootstrap.pypa.io/get-pip.py | python3.12 && \ | |
| # Upgrade pip | |
| python3 -m pip install --upgrade pip setuptools wheel && \ | |
| # Install Node.js 22.x from NodeSource | |
| curl -fsSL https://deb.nodesource.com/setup_22.x | bash - && \ | |
| apt-get install -y nodejs && \ | |
| # Install code-server | |
| curl -fsSL https://code-server.dev/install.sh | sh && \ | |
| apt-get clean && \ | |
| rm -rf /var/lib/apt/lists/* | |
| # Install global npm packages | |
| RUN npm install -g @anthropic-ai/claude-code @anthropic-ai/dxt | |
| # Create a directory for the workspace | |
| RUN mkdir -p /workspace | |
| # Copy requirements files (if they exist) | |
| COPY requirements*.txt /workspace/ | |
| # Create a requirements-cpu.txt without vllm for build time | |
| RUN if [ -f /workspace/requirements.txt ]; then \ | |
| grep -v "vllm" /workspace/requirements.txt | grep -v "xformers==.*+.*" > /workspace/requirements-cpu.txt || true; \ | |
| echo "xformers==0.0.32" >> /workspace/requirements-cpu.txt; \ | |
| pip3 install --no-cache-dir -r /workspace/requirements-cpu.txt || true; \ | |
| fi | |
| # Create configuration directory for code-server and Ollama | |
| RUN mkdir -p /root/.config/code-server /root/.ollama | |
| # Configure code-server to run on port 8443 | |
| RUN echo "bind-addr: 0.0.0.0:8443\nauth: none\ncert: false" > /root/.config/code-server/config.yaml | |
| # Install Ollama after code-server is set up | |
| RUN curl -fsSL https://ollama.com/install.sh | sh || true | |
| # Install some useful VS Code extensions | |
| RUN code-server --install-extension ms-python.python && \ | |
| code-server --install-extension ritwickdey.LiveServer && \ | |
| code-server --install-extension ms-toolsai.jupyter | |
| # Create a startup script that handles vllm installation on GPU | |
| RUN echo '#!/bin/bash\n\ | |
| # Check if running on GPU and install vllm if needed\n\ | |
| if [ -d "/usr/local/cuda" ] && ! python3 -c "import vllm" 2>/dev/null; then\n\ | |
| echo "GPU detected, installing vllm for Blackwell support..."\n\ | |
| pip3 install git+https://github.com/vllm-project/vllm.git@04e1642e3 || echo "vllm installation failed, continuing..."\n\ | |
| fi\n\ | |
| \n\ | |
| # Start Ollama in the background\n\ | |
| /usr/local/bin/ollama serve &\n\ | |
| \n\ | |
| # Give Ollama a moment to start\n\ | |
| sleep 2\n\ | |
| \n\ | |
| # Start code-server in the foreground\n\ | |
| exec code-server --disable-telemetry --bind-addr 0.0.0.0:8443 /workspace\n\ | |
| ' > /start.sh && \ | |
| chmod +x /start.sh | |
| # Expose ports for both services | |
| EXPOSE 8443 11434 | |
| # Set the workspace as working directory | |
| WORKDIR /workspace | |
| # Start both services | |
| CMD ["/start.sh"] |