forked from mrq/tortoise-tts
Optimizing graphics card memory
During inference it does not store gradients, which take up most of the video memory
This commit is contained in:
parent
ea8c825ee0
commit
cc38333249
|
@ -225,6 +225,7 @@ class TextToSpeech:
|
||||||
properties.
|
properties.
|
||||||
:param voice_samples: List of 2 or more ~10 second reference clips, which should be torch tensors containing 22.05kHz waveform data.
|
:param voice_samples: List of 2 or more ~10 second reference clips, which should be torch tensors containing 22.05kHz waveform data.
|
||||||
"""
|
"""
|
||||||
|
with torch.no_grad():
|
||||||
voice_samples = [v.to('cuda') for v in voice_samples]
|
voice_samples = [v.to('cuda') for v in voice_samples]
|
||||||
|
|
||||||
auto_conds = []
|
auto_conds = []
|
||||||
|
|
Loading…
Reference in New Issue
Block a user