Spaces:
Running
Running
FROM ubuntu:22.04 | |
# Set environment variables | |
ENV DEBIAN_FRONTEND=noninteractive | |
ENV HOME=/root | |
# CUDA paths for when running on GPU (RunPod) | |
ENV CUDA_HOME=/usr/local/cuda | |
ENV PATH=${CUDA_HOME}/bin:${PATH} | |
ENV LD_LIBRARY_PATH=${CUDA_HOME}/lib64:${LD_LIBRARY_PATH} | |
# Install dependencies, code-server, and Ollama | |
RUN apt-get update && \ | |
apt-get install -y curl wget gpg apt-transport-https git software-properties-common && \ | |
# Add Python 3.12 | |
add-apt-repository ppa:deadsnakes/ppa && \ | |
apt-get update && \ | |
apt-get install -y python3.12 python3.12-venv python3.12-dev && \ | |
# Make Python 3.12 the default | |
update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.12 1 && \ | |
update-alternatives --set python3 /usr/bin/python3.12 && \ | |
# Install pip for Python 3.12 | |
curl -sS https://bootstrap.pypa.io/get-pip.py | python3.12 && \ | |
# Upgrade pip | |
python3 -m pip install --upgrade pip setuptools wheel && \ | |
# Install Node.js 22.x from NodeSource | |
curl -fsSL https://deb.nodesource.com/setup_22.x | bash - && \ | |
apt-get install -y nodejs && \ | |
# Install code-server | |
curl -fsSL https://code-server.dev/install.sh | sh && \ | |
apt-get clean && \ | |
rm -rf /var/lib/apt/lists/* | |
# Install global npm packages | |
RUN npm install -g @anthropic-ai/claude-code @anthropic-ai/dxt | |
# Create a directory for the workspace | |
RUN mkdir -p /workspace | |
# Copy requirements files (if they exist) | |
COPY requirements*.txt /workspace/ | |
# Create a requirements-cpu.txt without vllm for build time | |
RUN if [ -f /workspace/requirements.txt ]; then \ | |
grep -v "vllm" /workspace/requirements.txt | grep -v "xformers==.*+.*" > /workspace/requirements-cpu.txt || true; \ | |
echo "xformers==0.0.32" >> /workspace/requirements-cpu.txt; \ | |
pip3 install --no-cache-dir -r /workspace/requirements-cpu.txt || true; \ | |
fi | |
# Create configuration directory for code-server and Ollama | |
RUN mkdir -p /root/.config/code-server /root/.ollama | |
# Configure code-server to run on port 8443 | |
RUN echo "bind-addr: 0.0.0.0:8443\nauth: none\ncert: false" > /root/.config/code-server/config.yaml | |
# Install Ollama after code-server is set up | |
RUN curl -fsSL https://ollama.com/install.sh | sh || true | |
# Install some useful VS Code extensions | |
RUN code-server --install-extension ms-python.python && \ | |
code-server --install-extension ritwickdey.LiveServer && \ | |
code-server --install-extension ms-toolsai.jupyter | |
# Create a startup script that handles vllm installation on GPU | |
RUN echo '#!/bin/bash\n\ | |
# Check if running on GPU and install vllm if needed\n\ | |
if [ -d "/usr/local/cuda" ] && ! python3 -c "import vllm" 2>/dev/null; then\n\ | |
echo "GPU detected, installing vllm for Blackwell support..."\n\ | |
pip3 install git+https://github.com/vllm-project/vllm.git@04e1642e3 || echo "vllm installation failed, continuing..."\n\ | |
fi\n\ | |
\n\ | |
# Start Ollama in the background\n\ | |
/usr/local/bin/ollama serve &\n\ | |
\n\ | |
# Give Ollama a moment to start\n\ | |
sleep 2\n\ | |
\n\ | |
# Start code-server in the foreground\n\ | |
exec code-server --disable-telemetry --bind-addr 0.0.0.0:8443 /workspace\n\ | |
' > /start.sh && \ | |
chmod +x /start.sh | |
# Expose ports for both services | |
EXPOSE 8443 11434 | |
# Set the workspace as working directory | |
WORKDIR /workspace | |
# Start both services | |
CMD ["/start.sh"] |