Spaces:
Configuration error
Configuration error
Fedir Zadniprovskyi
commited on
Commit
·
608e57c
1
Parent(s):
760dfe6
chore: update docker tag to latest
Browse files- README.md +2 -2
- compose.yaml +2 -2
- examples/live-audio/script.sh +2 -2
- examples/youtube/script.sh +2 -2
README.md
CHANGED
@@ -20,9 +20,9 @@ See [OpenAI API reference](https://platform.openai.com/docs/api-reference/audio)
|
|
20 |
## Quick Start
|
21 |
Using Docker
|
22 |
```bash
|
23 |
-
docker run --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:
|
24 |
# or
|
25 |
-
docker run --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:
|
26 |
```
|
27 |
Using Docker Compose
|
28 |
```bash
|
|
|
20 |
## Quick Start
|
21 |
Using Docker
|
22 |
```bash
|
23 |
+
docker run --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:latest-cuda
|
24 |
# or
|
25 |
+
docker run --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface fedirz/faster-whisper-server:latest-cpu
|
26 |
```
|
27 |
Using Docker Compose
|
28 |
```bash
|
compose.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
# NOTE: arm images haven't been tested
|
2 |
services:
|
3 |
faster-whisper-server-cuda:
|
4 |
-
image: fedirz/faster-whisper-server:
|
5 |
build:
|
6 |
dockerfile: Dockerfile.cuda
|
7 |
context: .
|
@@ -23,7 +23,7 @@ services:
|
|
23 |
devices:
|
24 |
- capabilities: ["gpu"]
|
25 |
faster-whisper-server-cpu:
|
26 |
-
image: fedirz/faster-whisper-server:
|
27 |
build:
|
28 |
dockerfile: Dockerfile.cpu
|
29 |
context: .
|
|
|
1 |
# NOTE: arm images haven't been tested
|
2 |
services:
|
3 |
faster-whisper-server-cuda:
|
4 |
+
image: fedirz/faster-whisper-server:latest-cuda
|
5 |
build:
|
6 |
dockerfile: Dockerfile.cuda
|
7 |
context: .
|
|
|
23 |
devices:
|
24 |
- capabilities: ["gpu"]
|
25 |
faster-whisper-server-cpu:
|
26 |
+
image: fedirz/faster-whisper-server:latest-cpu
|
27 |
build:
|
28 |
dockerfile: Dockerfile.cpu
|
29 |
context: .
|
examples/live-audio/script.sh
CHANGED
@@ -10,9 +10,9 @@ set -e
|
|
10 |
export WHISPER_MODEL=Systran/faster-distil-whisper-large-v3 # or Systran/faster-whisper-tiny.en if you are running on a CPU for a faster inference.
|
11 |
|
12 |
# Ensure you have `faster-whisper-server` running. If this is your first time running it expect to wait up-to a minute for the model to be downloaded and loaded into memory. You can run `curl localhost:8000/health` to check if the server is ready or watch the logs with `docker logs -f <container_id>`.
|
13 |
-
docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:
|
14 |
# or you can run it on a CPU
|
15 |
-
# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:
|
16 |
|
17 |
# `pv` is used to limit the rate at which the audio is streamed to the server. Audio is being streamed at a rate of 32kb/s(16000 sample rate * 16-bit sample / 8 bits per byte = 32000 bytes per second). This emulutes live audio input from a microphone: `ffmpeg -loglevel quiet -f alsa -i default -ac 1 -ar 16000 -f s16le`
|
18 |
# shellcheck disable=SC2002
|
|
|
10 |
export WHISPER_MODEL=Systran/faster-distil-whisper-large-v3 # or Systran/faster-whisper-tiny.en if you are running on a CPU for a faster inference.
|
11 |
|
12 |
# Ensure you have `faster-whisper-server` running. If this is your first time running it expect to wait up-to a minute for the model to be downloaded and loaded into memory. You can run `curl localhost:8000/health` to check if the server is ready or watch the logs with `docker logs -f <container_id>`.
|
13 |
+
docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cuda
|
14 |
# or you can run it on a CPU
|
15 |
+
# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cpu
|
16 |
|
17 |
# `pv` is used to limit the rate at which the audio is streamed to the server. Audio is being streamed at a rate of 32kb/s(16000 sample rate * 16-bit sample / 8 bits per byte = 32000 bytes per second). This emulutes live audio input from a microphone: `ffmpeg -loglevel quiet -f alsa -i default -ac 1 -ar 16000 -f s16le`
|
18 |
# shellcheck disable=SC2002
|
examples/youtube/script.sh
CHANGED
@@ -6,9 +6,9 @@ set -e
|
|
6 |
export WHISPER_MODEL=Systran/faster-distil-whisper-large-v3 # or Systran/faster-whisper-tiny.en if you are running on a CPU for a faster inference.
|
7 |
|
8 |
# Ensure you have `faster-whisper-server` running. If this is your first time running it expect to wait up-to a minute for the model to be downloaded and loaded into memory. You can run `curl localhost:8000/health` to check if the server is ready or watch the logs with `docker logs -f <container_id>`.
|
9 |
-
docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:
|
10 |
# or you can run it on a CPU
|
11 |
-
# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:
|
12 |
|
13 |
# Download the audio from a YouTube video. In this example I'm downloading "The Evolution of the Operating System" by Asionometry YouTube channel. I highly checking this channel out, the guy produces very high content. If you don't have `youtube-dl`, you'll have to install it. https://github.com/ytdl-org/youtube-dl
|
14 |
youtube-dl --extract-audio --audio-format mp3 -o the-evolution-of-the-operating-system.mp3 'https://www.youtube.com/watch?v=1lG7lFLXBIs'
|
|
|
6 |
export WHISPER_MODEL=Systran/faster-distil-whisper-large-v3 # or Systran/faster-whisper-tiny.en if you are running on a CPU for a faster inference.
|
7 |
|
8 |
# Ensure you have `faster-whisper-server` running. If this is your first time running it expect to wait up-to a minute for the model to be downloaded and loaded into memory. You can run `curl localhost:8000/health` to check if the server is ready or watch the logs with `docker logs -f <container_id>`.
|
9 |
+
docker run --detach --gpus=all --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cuda
|
10 |
# or you can run it on a CPU
|
11 |
+
# docker run --detach --publish 8000:8000 --volume ~/.cache/huggingface:/root/.cache/huggingface --env WHISPER_MODEL=$WHISPER_MODEL fedirz/faster-whisper-server:latest-cpu
|
12 |
|
13 |
# Download the audio from a YouTube video. In this example I'm downloading "The Evolution of the Operating System" by Asionometry YouTube channel. I highly checking this channel out, the guy produces very high content. If you don't have `youtube-dl`, you'll have to install it. https://github.com/ytdl-org/youtube-dl
|
14 |
youtube-dl --extract-audio --audio-format mp3 -o the-evolution-of-the-operating-system.mp3 'https://www.youtube.com/watch?v=1lG7lFLXBIs'
|