Spaces:
Running on Zero
Running on Zero
| services: | |
| lightdiffusion: | |
| build: | |
| context: . | |
| dockerfile: Dockerfile | |
| args: | |
| # Specify target GPU architectures for CUDA extension builds | |
| # 8.0: A100, 8.6: RTX 30xx, 8.9: RTX 40xx, 9.0: H100, 12.0: RTX 50xx (Blackwell) | |
| # Customize based on your GPU: TORCH_CUDA_ARCH_LIST: "12.0" for RTX 50xx only | |
| TORCH_CUDA_ARCH_LIST: "8.0;8.6;8.9;9.0;12.0" | |
| INSTALL_STABLE_FAST: "0" | |
| INSTALL_OLLAMA: "0" | |
| INSTALL_SAGEATTENTION: "0" | |
| INSTALL_SPARGEATTN: "0" | |
| ports: | |
| - "7860:7860" # FastAPI backend serving the built React UI | |
| volumes: | |
| # Mount output directory to persist generated images | |
| - ./output:/app/output | |
| # Mount checkpoints directory for model files | |
| - ./include/checkpoints:/app/include/checkpoints | |
| # Mount other model directories | |
| - ./include/loras:/app/include/loras | |
| - ./include/embeddings:/app/include/embeddings | |
| - ./include/ESRGAN:/app/include/ESRGAN | |
| - ./include/yolos:/app/include/yolos | |
| environment: | |
| - PORT=7860 | |
| - CUDA_VISIBLE_DEVICES=0 | |
| - CUDA_HOME=/usr/local/cuda | |
| - PROMPT_ENHANCER_MODEL=qwen3:0.6b | |
| deploy: | |
| resources: | |
| reservations: | |
| devices: | |
| - driver: nvidia | |
| count: 1 | |
| capabilities: [ gpu ] | |
| restart: unless-stopped | |
| stdin_open: true | |
| tty: true | |