Collection of utilities aimed to voice clone through AI
Go to file
2023-03-07 03:01:02 +00:00
bin
config
dlas@6eb7ebf847 oops v2 2023-03-05 14:19:41 +00:00
models added option to set worker size in training config generator (because the default is overkill), for whisper transcriptions, load a specialized language model if it exists (for now, only english), output transcription to web UI when done transcribing 2023-03-05 05:17:19 +00:00
results
src brain worms and headaches 2023-03-07 03:01:02 +00:00
tortoise-tts@e2db36af60 moved (actually not working) setting to use BigVGAN to a dropdown to select between vocoders (for when slotting in future ones), and ability to load a new vocoder while TTS is loaded 2023-03-07 02:45:22 +00:00
training
voices
.gitignore experimental multi-gpu training (Linux only, because I can't into batch files) 2023-03-03 04:37:18 +00:00
.gitmodules I forgot submodules existed 2023-02-19 21:41:51 +00:00
LICENSE
notebook.ipynb oops 2023-02-24 16:23:30 +00:00
README.md huge success 2023-02-23 06:24:54 +00:00
requirements.txt added api_name for generation, added whisperx backend, relocated use whispercpp option to whisper backend list 2023-03-06 05:21:33 +00:00
setup-cuda.bat forgot to reorder the dependency install because whisperx needs to be installed before DLAS 2023-03-06 16:43:17 +00:00
setup-cuda.sh forgot to reorder the dependency install because whisperx needs to be installed before DLAS 2023-03-06 16:43:17 +00:00
setup-directml.bat forgot to reorder the dependency install because whisperx needs to be installed before DLAS 2023-03-06 16:43:17 +00:00
setup-rocm-bnb.sh added setup script for bitsandbytes-rocm (soon: multi-gpu testing, because I am finally making use of my mispurchased second 6800XT) 2023-03-03 02:58:34 +00:00
setup-rocm.sh forgot to reorder the dependency install because whisperx needs to be installed before DLAS 2023-03-06 16:43:17 +00:00
start.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
start.sh I swear I committed forwarding arguments from the start scripts 2023-02-19 15:01:16 +00:00
train.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
train.sh use torchrun instead for multigpu 2023-03-04 20:53:00 +00:00
update-force.bat one more 2023-03-06 16:47:34 +00:00
update-force.sh one more 2023-03-06 16:47:34 +00:00
update.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
update.sh added api_name for generation, added whisperx backend, relocated use whispercpp option to whisper backend list 2023-03-06 05:21:33 +00:00

AI Voice Cloning

This repo/rentry aims to serve as both a foolproof guide for setting up AI voice cloning tools for legitimate, local use on Windows/Linux, as well as a stepping stone for anons that genuinely want to play around with TorToiSe.

Similar to my own findings for Stable Diffusion image generation, this rentry may appear a little disheveled as I note my new findings with TorToiSe. Please keep this in mind if the guide seems to shift a bit or sound confusing.

>Ugh... why bother when I can just abuse 11.AI?

You're more than welcome to, but TorToiSe is shaping up to be a very promising tool, especially with finetuning now on the horizon.

This is not endorsed by neonbjb. I do not expect this to run into any ethical issues, as it seems (like me), this is mostly for making funny haha vidya characters say funny lines.

Documentation

Please consult the wiki for the documentation, including how to install, prepare voices for, and use the software.

Bug Reporting

If you run into any problems, please refer to the issues you may encounter wiki page first. Please don't hesitate to submit an issue.

Changelogs

Below will be a rather-loose changelogss, as I don't think I have a way to chronicle them outside of commit messages:

2023.02.22

  • greatly reduced VRAM consumption through the use of TimDettmers/bitsandbytes
  • cleaned up section of code that handled parsing output from training script
  • added button to reconnect to the training script's output (sometimes skips a line to update, but it's better than nothing)
  • actually update submodules from the update script (somehow forgot to pass --remote)

Before 2023.02.22

Refer to commit logs.