Skip to content
Snippets Groups Projects
Commit 7e842996 authored by Ahmet Öner's avatar Ahmet Öner
Browse files

Updated model conversion method to use Hugging Face downloader

parent ef85c04a
No related branches found
No related tags found
No related merge requests found
...@@ -4,6 +4,10 @@ Changelog ...@@ -4,6 +4,10 @@ Changelog
Unreleased Unreleased
---------- ----------
### Updated
- Updated model conversion method (for Faster Whisper) to use Hugging Face downloader
[1.1.1] (2023-05-29) [1.1.1] (2023-05-29)
-------------------- --------------------
......
...@@ -7,16 +7,15 @@ import torch ...@@ -7,16 +7,15 @@ import torch
import whisper import whisper
from faster_whisper import WhisperModel from faster_whisper import WhisperModel
from .utils import model_converter, ResultWriter, WriteTXT, WriteSRT, WriteVTT, WriteTSV, WriteJSON from .utils import ResultWriter, WriteTXT, WriteSRT, WriteVTT, WriteTSV, WriteJSON
model_name = os.getenv("ASR_MODEL", "base") model_name = os.getenv("ASR_MODEL", "base")
model_path = os.path.join("/root/.cache/faster_whisper", model_name) model_path = os.getenv("ASR_MODEL_PATH", "/root/.cache/whisper")
model_converter(model_name, model_path)
if torch.cuda.is_available(): if torch.cuda.is_available():
model = WhisperModel(model_path, device="cuda", compute_type="float32") model = WhisperModel(model_size_or_path=model_name, device="cuda", compute_type="float32", download_root=model_path)
else: else:
model = WhisperModel(model_path, device="cpu", compute_type="int8") model = WhisperModel(model_size_or_path=model_name, device="cpu", compute_type="int8", download_root=model_path)
model_lock = Lock() model_lock = Lock()
......
...@@ -2,16 +2,6 @@ import json ...@@ -2,16 +2,6 @@ import json
import os import os
from typing import TextIO from typing import TextIO
from ctranslate2.converters.transformers import TransformersConverter
def model_converter(model, model_output):
converter = TransformersConverter("openai/whisper-" + model)
try:
converter.convert(model_output, None, "float16", False)
except Exception as e:
print(e)
def format_timestamp(seconds: float, always_include_hours: bool = False, decimal_marker: str = '.'): def format_timestamp(seconds: float, always_include_hours: bool = False, decimal_marker: str = '.'):
assert seconds >= 0, "non-negative timestamp expected" assert seconds >= 0, "non-negative timestamp expected"
......
...@@ -1590,4 +1590,4 @@ files = [ ...@@ -1590,4 +1590,4 @@ files = [
[metadata] [metadata]
lock-version = "2.0" lock-version = "2.0"
python-versions = "^3.10" python-versions = "^3.10"
content-hash = "0f767e3a851b5f487a902468cb556217c791a841703c4a0c3e6edfde34cb576f" content-hash = "a0d7846f244ddef103453b6830120a9704f56d12b924a90aa904b61574b84ade"
...@@ -22,7 +22,6 @@ unidecode = "^1.3.4" ...@@ -22,7 +22,6 @@ unidecode = "^1.3.4"
uvicorn = { extras = ["standard"], version = "^0.18.2" } uvicorn = { extras = ["standard"], version = "^0.18.2" }
gunicorn = "^20.1.0" gunicorn = "^20.1.0"
tqdm = "^4.64.1" tqdm = "^4.64.1"
transformers = "^4.22.1"
python-multipart = "^0.0.5" python-multipart = "^0.0.5"
ffmpeg-python = "^0.2.0" ffmpeg-python = "^0.2.0"
fastapi = "^0.95.1" fastapi = "^0.95.1"
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please to comment