Collection of utilities aimed to voice clone through AI
Go to file
2023-03-04 20:53:00 +00:00
bin
config
dlas@71cc43e65c added setup script for bitsandbytes-rocm (soon: multi-gpu testing, because I am finally making use of my mispurchased second 6800XT) 2023-03-03 02:58:34 +00:00
models renamed mega batch factor to an actual real term: gradient accumulation factor, fixed halting training not actually killing the training process and freeing up resources, some logic cleanup for gradient accumulation (so many brain worms and wrong assumptions from testing on low batch sizes) (read the training section in the wiki for more details) 2023-03-04 15:55:06 +00:00
results
src use torchrun instead for multigpu 2023-03-04 20:53:00 +00:00
tortoise-tts@06bdf72b89 oops 2023-03-03 18:51:33 +00:00
training
voices
.gitignore experimental multi-gpu training (Linux only, because I can't into batch files) 2023-03-03 04:37:18 +00:00
.gitmodules
LICENSE
notebook.ipynb oops 2023-02-24 16:23:30 +00:00
README.md
requirements.txt
setup-cuda.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
setup-cuda.sh added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
setup-directml.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
setup-rocm-bnb.sh added setup script for bitsandbytes-rocm (soon: multi-gpu testing, because I am finally making use of my mispurchased second 6800XT) 2023-03-03 02:58:34 +00:00
setup-rocm.sh added setup script for bitsandbytes-rocm (soon: multi-gpu testing, because I am finally making use of my mispurchased second 6800XT) 2023-03-03 02:58:34 +00:00
start.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
start.sh
train.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
train.sh use torchrun instead for multigpu 2023-03-04 20:53:00 +00:00
update-force.bat added setup script for bitsandbytes-rocm (soon: multi-gpu testing, because I am finally making use of my mispurchased second 6800XT) 2023-03-03 02:58:34 +00:00
update-force.sh added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
update.bat added button to just load a training set's loss information, added installing broncotc/bitsandbytes-rocm when running setup-rocm.sh 2023-03-02 01:35:12 +00:00
update.sh Added optional whispercpp update functionality 2023-03-03 18:34:49 +00:00

AI Voice Cloning

This repo/rentry aims to serve as both a foolproof guide for setting up AI voice cloning tools for legitimate, local use on Windows/Linux, as well as a stepping stone for anons that genuinely want to play around with TorToiSe.

Similar to my own findings for Stable Diffusion image generation, this rentry may appear a little disheveled as I note my new findings with TorToiSe. Please keep this in mind if the guide seems to shift a bit or sound confusing.

>Ugh... why bother when I can just abuse 11.AI?

You're more than welcome to, but TorToiSe is shaping up to be a very promising tool, especially with finetuning now on the horizon.

This is not endorsed by neonbjb. I do not expect this to run into any ethical issues, as it seems (like me), this is mostly for making funny haha vidya characters say funny lines.

Documentation

Please consult the wiki for the documentation, including how to install, prepare voices for, and use the software.

Bug Reporting

If you run into any problems, please refer to the issues you may encounter wiki page first. Please don't hesitate to submit an issue.

Changelogs

Below will be a rather-loose changelogss, as I don't think I have a way to chronicle them outside of commit messages:

2023.02.22

  • greatly reduced VRAM consumption through the use of TimDettmers/bitsandbytes
  • cleaned up section of code that handled parsing output from training script
  • added button to reconnect to the training script's output (sometimes skips a line to update, but it's better than nothing)
  • actually update submodules from the update script (somehow forgot to pass --remote)

Before 2023.02.22

Refer to commit logs.