Catch OOM and run whisper on cpu automatically. #117

Merged
mrq merged 1 commits from :vram into master 2023-03-12 05:09:54 +07:00
1 changed files with 7 additions and 1 deletions

@ -2065,7 +2065,13 @@ def load_whisper_model(language=None, model_name=None, progress=None):
if args.whisper_backend == "openai/whisper":
import whisper
whisper_model = whisper.load_model(model_name)
try:
#is it possible for model to fit on vram but go oom later on while executing on data?
whisper_model = whisper.load_model(model_name)
except:
print("Out of VRAM memory.")
print(f"Falling back to loading Whisper on CPU.")
whisper_model = whisper.load_model(model_name, device="cpu")
elif args.whisper_backend == "lightmare/whispercpp":
from whispercpp import Whisper
if not language: