Spaces:
Paused
Paused
Update Dockerfile
Browse files- Dockerfile +6 -7
Dockerfile
CHANGED
@@ -22,6 +22,7 @@ RUN wget https://developer.download.nvidia.com/compute/cuda/repos/debian12/x86_6
|
|
22 |
ENV PATH=/usr/local/cuda/bin:${PATH}
|
23 |
ENV LD_LIBRARY_PATH=/usr/local/cuda/lib64:${LD_LIBRARY_PATH}
|
24 |
ENV CUDAToolkit_ROOT=/usr/local/cuda
|
|
|
25 |
|
26 |
RUN useradd -m -u 1000 user
|
27 |
USER user
|
@@ -32,16 +33,14 @@ WORKDIR /app
|
|
32 |
|
33 |
COPY --chown=user . /app
|
34 |
|
35 |
-
#
|
36 |
-
ENV CMAKE_ARGS="-DGGML_CUDA=on -DCMAKE_CUDA_ARCHITECTURES=86"
|
37 |
-
|
38 |
RUN pip install --upgrade pip
|
39 |
|
40 |
-
#
|
41 |
-
RUN pip install --no-cache-dir llama-cpp-python
|
42 |
|
43 |
-
#
|
44 |
-
RUN pip install --no-cache-dir --
|
45 |
|
46 |
RUN pip install --no-cache-dir -r requirements.txt
|
47 |
|
|
|
22 |
ENV PATH=/usr/local/cuda/bin:${PATH}
|
23 |
ENV LD_LIBRARY_PATH=/usr/local/cuda/lib64:${LD_LIBRARY_PATH}
|
24 |
ENV CUDAToolkit_ROOT=/usr/local/cuda
|
25 |
+
ENV CMAKE_ARGS="-DGGML_CUDA=on -DCMAKE_CUDA_ARCHITECTURES=86"
|
26 |
|
27 |
RUN useradd -m -u 1000 user
|
28 |
USER user
|
|
|
33 |
|
34 |
COPY --chown=user . /app
|
35 |
|
36 |
+
# Wichtig: Isolation deaktivieren für llama-cpp-python Build
|
|
|
|
|
37 |
RUN pip install --upgrade pip
|
38 |
|
39 |
+
# Optional: zuerst llama-cpp-python bauen (cachebar)
|
40 |
+
RUN pip install --no-cache-dir --no-build-isolation llama-cpp-python
|
41 |
|
42 |
+
# Danach: outetts (zieht llama-cpp-python nicht erneut)
|
43 |
+
RUN pip install --no-cache-dir --no-build-isolation outetts
|
44 |
|
45 |
RUN pip install --no-cache-dir -r requirements.txt
|
46 |
|