Skip to content

Latest commit

 

History

History
90 lines (56 loc) · 3.44 KB

README.md

File metadata and controls

90 lines (56 loc) · 3.44 KB

whisperbox-transcribe

HTTP wrapper around openai/whisper.

Overview

This project wraps OpenAI's whisper speech-to-text models with a HTTP API.

The API design draws inspiration from the rev.ai async speech-to-text API. Transcription jobs are submitted by making a HTTP POST request to the service. Once the job is accepted, an ID is returned, which can be later utilized to retrieve the transcription results. These results are stored in an internal database until they are retrieved and can optionally be deleted afterwards.

It is assumed that the service is used by exactly one consumer, so a pre-shared API key is used as authentication method. OpenAPI documentation for the service is available at <service_url>/docs.

Deploy

0. Choose model & instance size Whisper offers a range of models in [different sizes](https://github.com/openai/whisper#available-models-and-languages). The model size affects factors such as accuracy, resource usage, and transcription speed. Smaller models are generally faster and consume fewer resources, but they may be less accurate, especially when working with non-English languages or translation tasks.

Whisper supports inference on both CPU and GPU, and this project includes slightly modified Docker Compose configurations to enable both options. CPU inference is slower but usually more cost-effective for hosting purposes. CPU inference performance typically scales well with the CPU speed.

When selecting an instance for your application, it's important to consider the disk size. Media files need to be downloaded before they can be transcribed, so the disk must have sufficient free space to accommodate them.

As a starting point, the "small" model can run on a 4GB Digital Ocean droplet with, achieving approximately a 1-2x speed-up over to the original audio length when transcribing.

1. Prepare host environment

This project is intended to be run via docker compose. In order to get started, install docker engine. Then, clone this repository to the machine.

Note
If you want to use a GPU, uncomment the sections tagged <GPU SUPPORT> in docker-compose.prod.yml.

2. Configure service

  1. Create an .env file from .env.example and configure it. Refer to comments for available envs and their usage.

3. Run service

Run make run to start the server. To launch at system startup, wrap it in a systemd launch service.

Develop

docker compose is required for local development.

It is recommended to setup a virtual environment for python tooling. To install dependencies in your virtual env, run pip install -e .[tooling,web,worker].

Copy .env.dev to .env to configure the service.

Start

make dev

Builds and starts the docker containers.

# Bindings
http://localhost:5555                        => Celery dashboard
http://localhost:15672                       => RabbitMQ dashboard
http://whisperbox-transcribe.localhost       => API
http://whisperbox-transcribe.localhost/docs  => API docs
./whisperbox-transcribe.sqlite               => Database

Clean

This removes all containers and attached volumes.

make clean

Test

make test

Lint

make lint

Format

make fmt