## Docker Compose (Recommended) TODO: just reference the existing compose file in the repo === "CUDA" ```yaml # https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html services: faster-whisper-server: image: fedirz/faster-whisper-server:latest-cuda name: faster-whisper-server restart: unless-stopped ports: - 8000:8000 volumes: - hf-hub-cache:/home/ubuntu/.cache/huggingface/hub deploy: resources: reservations: devices: - capabilities: ["gpu"] volumes: hf-hub-cache: ``` === "CUDA (with CDI feature enabled)" ```yaml # https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html services: faster-whisper-server: image: fedirz/faster-whisper-server:latest-cuda name: faster-whisper-server restart: unless-stopped ports: - 8000:8000 volumes: - hf-hub-cache:/home/ubuntu/.cache/huggingface/hub deploy: resources: reservations: # https://docs.docker.com/reference/cli/dockerd/#enable-cdi-devices # https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/cdi-support.html devices: - driver: cdi device_ids: - nvidia.com/gpu=all volumes: hf-hub-cache: ``` === "CPU" ```yaml services: faster-whisper-server: image: fedirz/faster-whisper-server:latest-cpu name: faster-whisper-server restart: unless-stopped ports: - 8000:8000 volumes: - hf-hub-cache:/home/ubuntu/.cache/huggingface/hub volumes: hf-hub-cache: ``` ## Docker === "CUDA" ```bash docker run --rm --detach --publish 8000:8000 --name faster-whisper-server --volume hf-hub-cache:/home/ubuntu/.cache/huggingface/hub --gpus=all fedirz/faster-whisper-server:latest-cuda ``` === "CUDA (with CDI feature enabled)" ```bash docker run --rm --detach --publish 8000:8000 --name faster-whisper-server --volume hf-hub-cache:/home/ubuntu/.cache/huggingface/hub --device=nvidia.com/gpu=all fedirz/faster-whisper-server:latest-cuda ``` === "CPU" ```bash docker run --rm --detach --publish 8000:8000 --name faster-whisper-server --volume hf-hub-cache:/home/ubuntu/.cache/huggingface/hub fedirz/faster-whisper-server:latest-cpu ``` ## Kubernetes WARNING: it was written few months ago and may be outdated. Please refer to this [blog post](https://substratus.ai/blog/deploying-faster-whisper-on-k8s) ## Python (requires Python 3.12+) ```bash git clone https://github.com/fedirz/faster-whisper-server.git cd faster-whisper-server uv venv sourve .venv/bin/activate uv sync --all-extras uvicorn --factory --host 0.0.0.0 faster_whisper_server.main:create_app ```