Files
tlg da35e94b16 fix: add triton kernels for MXFP4, fix GGUF KV cache quantization
- Add 'kernels' package to Dockerfile for native MXFP4 execution
  (fixes gpt-oss-20b OOM: 15.2GB→13.5GB)
- Reduce GGUF n_ctx from 8192 to 4096 and quantize KV cache to Q8_0
  to reduce VRAM usage
- Use GGML_TYPE_Q8_0 constant instead of string for type_k/type_v

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-04-05 22:49:16 +02:00

67 lines
2.3 KiB
Docker

# --- Build stage: compile llama-cpp-python with CUDA ---
FROM docker.io/pytorch/pytorch:2.11.0-cuda12.8-cudnn9-devel AS builder
RUN pip install --no-cache-dir --break-system-packages --upgrade pip setuptools wheel
RUN CMAKE_ARGS="-DGGML_CUDA=on" pip install --no-cache-dir --break-system-packages \
"llama-cpp-python>=0.3.0"
# --- Runtime stage ---
FROM docker.io/pytorch/pytorch:2.11.0-cuda12.8-cudnn9-runtime
# System dependencies for audio processing
RUN apt-get update && apt-get install -y --no-install-recommends \
libsndfile1 \
ffmpeg \
&& rm -rf /var/lib/apt/lists/*
# Upgrade pip/setuptools for Python 3.12 compatibility
RUN pip install --no-cache-dir --break-system-packages --upgrade \
pip setuptools wheel
# Install deps that don't conflict with pre-installed torch stack
RUN pip install --no-cache-dir --break-system-packages \
"fastapi>=0.115.0" \
"uvicorn[standard]>=0.34.0" \
"python-multipart>=0.0.18" \
"soundfile>=0.12.0" \
"sentencepiece>=0.2.0" \
"protobuf>=5.0.0"
# Install transformers + accelerate + kernels (MXFP4/FP8 triton kernels)
RUN pip install --no-cache-dir --break-system-packages --no-build-isolation \
"transformers>=5.4.0" \
"accelerate>=1.0.0" \
"kernels"
# Install chatterbox-tts WITHOUT its dependencies (it would downgrade
# torch from 2.11 to 2.6 and pull gradio, librosa, etc.)
# Then install only the runtime deps chatterbox actually needs.
RUN pip install --no-cache-dir --break-system-packages --no-deps \
"chatterbox-tts>=0.1.0"
RUN pip install --no-cache-dir --break-system-packages --no-build-isolation \
"conformer>=0.3.2" \
"einops>=0.8.0" \
"omegaconf>=2.3.0" \
"scipy>=1.17.0" \
"diffusers>=0.29.0" \
"resemble-perth>=1.0.0" \
"s3tokenizer>=0.3.0" \
"librosa>=0.10.0" \
"diskcache>=5.6.0"
# Copy llama-cpp-python from builder
COPY --from=builder /usr/local/lib/python3.12/dist-packages/llama_cpp /usr/local/lib/python3.12/dist-packages/llama_cpp
COPY --from=builder /usr/local/lib/python3.12/dist-packages/llama_cpp_python* /usr/local/lib/python3.12/dist-packages/
# Copy application code
COPY llmux/ /app/llmux/
WORKDIR /app
# Avoid CUDA memory fragmentation when swapping models
ENV PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True
# Run the server
EXPOSE 8081
CMD ["uvicorn", "llmux.main:app", "--host", "0.0.0.0", "--port", "8081"]