Spaces:
Sleeping
Sleeping
| FROM debian:bookworm-slim | |
| ARG DEBIAN_FRONTEND=noninteractive | |
| RUN apt-get update && apt-get install -y --no-install-recommends \ | |
| git build-essential cmake curl ca-certificates pkg-config libcurl4-openssl-dev \ | |
| && rm -rf /var/lib/apt/lists/* | |
| # build llama.cpp (HTTP server in tools/server) | |
| WORKDIR /app | |
| RUN git clone --depth 1 https://github.com/ggml-org/llama.cpp.git \ | |
| && cd llama.cpp && mkdir -p build && cd build \ | |
| && cmake -DCMAKE_BUILD_TYPE=Release \ | |
| -DGGML_NATIVE=ON \ | |
| -DLLAMA_BUILD_EXAMPLES=ON \ | |
| -DLLAMA_BUILD_SERVER=ON \ | |
| -DLLAMA_BUILD_TESTS=OFF \ | |
| .. \ | |
| && cmake --build . --target llama-server -j | |
| # fetch LiquidAI LFM2-350M GGUF | |
| RUN mkdir -p /models && \ | |
| curl -fL --retry 5 --retry-delay 2 -o /models/model.gguf \ | |
| "https://huggingface.co/LiquidAI/LFM2-350M-GGUF/resolve/main/LFM2-350M-Q4_K_M.gguf?download=true" | |
| EXPOSE 7860 | |
| CMD ["/app/llama.cpp/build/bin/llama-server","-m","/models/model.gguf","-c","2048","-ngl","0","-t","4","--host","0.0.0.0","--port","7860"] | |