Skip to content
Snippets Groups Projects
Commit b2089ae3 authored by Tanay Upadhyaya's avatar Tanay Upadhyaya
Browse files

Download whisper model and convert for faster_whisper while building Dockerfile

parent 9963d632
No related branches found
No related tags found
No related merge requests found
......@@ -34,4 +34,9 @@ COPY . .
RUN poetry install
RUN $POETRY_VENV/bin/pip install torch==1.13.0+cu117 -f https://download.pytorch.org/whl/torch
ENV ASR_MODEL="base"
# TODO: Skip based on ENV variable
RUN ./faster_whisper_model_conversion.sh ${ASR_MODEL}
CMD gunicorn --bind 0.0.0.0:9000 --workers 1 --timeout 0 app.webservice:app -k uvicorn.workers.UvicornWorker
......@@ -70,6 +70,11 @@ Starting the Webservice:
poetry run gunicorn --bind 0.0.0.0:9000 --workers 1 --timeout 0 app.webservice:app -k uvicorn.workers.UvicornWorker
```
With docker compose:
```sh
docker-compose up --build
```
## Quick start
After running the docker image interactive Swagger API documentation is available at [localhost:9000/docs](http://localhost:9000/docs)
......
......@@ -137,11 +137,10 @@ def run_asr(
with model_lock:
model = get_modal(faster)
if faster:
# TODO: options_dict
segments = []
text = ""
i = 0
segment_generator, info = model.transcribe(audio, beam_size=5)
segment_generator, info = model.transcribe(audio, beam_size=5, **options_dict)
for segment in segment_generator:
segments.append(segment)
text = text + segment.text
......
......@@ -13,12 +13,18 @@ services:
count: 1
capabilities: [gpu]
environment:
- ASR_MODEL=large
- ASR_MODEL=base
ports:
- 9000:9000
volumes:
- ./app:/app/app
- ~/.cache/poetry:/root/.cache/poetry
- ~/.cache/whisper:/root/.cache/whisper
- ~/.cache/faster_whisper:/root/.cache/faster_whisper
- cache-pip:/root/.cache/pip
- cache-poetry:/root/.cache/poetry
- cache-whisper:/root/.cache/whisper
- cache-faster-whisper:/root/.cache/faster_whisper
volumes:
cache-pip:
cache-poetry:
cache-whisper:
cache-faster-whisper:
#!/bin/bash
ct2-transformers-converter --model openai/whisper-"$ASR_MODEL" --output_dir /root/.cache/faster_whisper --quantization float16
#!/bin/bash
# TODO: Add step to build setup based on ENV variable
ct2-transformers-converter --model openai/whisper-large-v2 --output_dir /root/.cache/faster_whisper --quantization float16
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment