Commit Graph

110 Commits

Author SHA1 Message Date
mrq
b4ca260de9 added flag to enable/disable voicefixer using CUDA because I'll OOM on my 2060, changed from naively subdividing eavenly (2,4,8,16 pieces) to just incrementing by 1 (1,2,3,4) when trying to subdivide within constraints of the max chunk size for computing voice latents 2023-02-14 16:47:34 +00:00
mrq
b16eb99538 history tab doesn't naively reuse the voice dir instead for results, experimental "divide total sound size until it fits under requests max chunk size" doesn't have a +1 to mess things up (need to re-evaluate how I want to calculate sizes of bests fits eventually) 2023-02-14 16:23:04 +00:00
mrq
2427c98333 Implemented kv_cache "fix" (from 1f3c1b5f4a); guess I should find out why it's crashing DirectML backend 2023-02-13 13:48:31 +00:00
mrq
4ced0296a2 DirectML: fixed redaction/aligner by forcing it to stay on CPU 2023-02-12 20:52:04 +00:00
mrq
b85c9921d7 added button to recalculate voice latents, added experimental switch for computing voice latents 2023-02-12 18:11:40 +00:00
mrq
2210b49cb6 fixed regression with computing conditional latencies outside of the CPU 2023-02-12 17:44:39 +00:00
mrq
a2d95fe208 fixed silently crashing from enabling kv_cache-ing if using the DirectML backend, throw an error when reading a generated audio file that does not have any embedded metadata in it, cleaned up the blocks of code that would DMA/transfer tensors/models between GPU and CPU 2023-02-12 14:46:21 +00:00
mrq
5f1c032312 fixed regression where the auto_conds do not move to the GPU and causes a problem during CVVP compare pass 2023-02-11 20:34:12 +00:00
mrq
c5337a6b51 Added integration for "voicefixer", fixed issue where candidates>1 and lines>1 only outputs the last combined candidate, numbered step for each generation in progress, output time per generation step 2023-02-11 15:02:11 +00:00
mrq
8641cc9906 revamped result formatting, added "kludgy" stop button 2023-02-10 22:12:37 +00:00
mrq
7471bc209c Moved voices out of the tortoise folder because it kept being processed for setup.py 2023-02-10 20:11:56 +00:00
mrq
2bce24b9dd Cleanup 2023-02-10 19:55:33 +00:00
mrq
39b81318f2 Added new options: "Output Sample Rate", "Output Volume", and documentation 2023-02-10 03:02:09 +00:00
mrq
77b39e59ac oops 2023-02-09 22:17:57 +00:00
mrq
3621e16ef9 Added 'Only Load Models Locally' setting 2023-02-09 22:06:55 +00:00
mrq
d7443dfa06 Added option: listen path 2023-02-09 20:42:38 +00:00
mrq
38ee19cd57 I didn't have to suck off a wizard for DirectML support (courtesy of https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/7600 for leading the way) 2023-02-09 05:05:21 +00:00
mrq
a37546ad99 owari da... 2023-02-09 01:53:25 +00:00
mrq
6255c98006 beginning to add DirectML support 2023-02-08 23:03:52 +00:00
mrq
6ebdde58f0 (finally) added the CVVP model weigh slider, latents export more data too for weighing against CVVP 2023-02-07 20:55:56 +00:00
mrq
793515772a un-hardcoded input output sampling rates (changing them "works" but leads to wrong audio, naturally) 2023-02-07 18:34:29 +00:00
mrq
5f934c5feb (maybe) fixed an issue with using prompt redactions (emotions) on CPU causing a crash, because for some reason the wav2vec_alignment assumed CUDA was always available 2023-02-07 07:51:05 -06:00
mrq
d6b5d67f79 forgot to auto compute batch size again if set to 0 2023-02-06 23:14:17 -06:00
mrq
be6fab9dcb added setting to adjust autoregressive sample batch size 2023-02-06 22:31:06 +00:00
mrq
b441a84615 added flag (--cond-latent-max-chunk-size) that should restrict the maximum chunk size when chunking for calculating conditional latents, to avoid OOMing on VRAM 2023-02-06 05:10:07 +00:00
mrq
a1f3b6a4da fixed up the computing conditional latents 2023-02-06 03:44:34 +00:00
mrq
945136330c Forgot to rename the cached latents to the new filename 2023-02-05 23:51:52 +00:00
mrq
5bf21fdbe1 modified how conditional latents are computed (before, it just happened to only bother reading the first 102400/24000=4.26 seconds per audio input, now it will chunk it all to compute latents) 2023-02-05 23:25:41 +00:00
mrq
f66754b557 oops 2023-02-05 20:10:40 +00:00
mrq
1c582b5dc8 added button to refresh voice list, enabling KV caching for a bonerific speed increase (credit to https://github.com/152334H/tortoise-tts-fast/) 2023-02-05 17:59:13 +00:00
mrq
8831522de9 New tunable: pause size/breathing room (governs pause at the end of clips) 2023-02-05 14:45:51 +00:00
mrq
bf32efe503 Added multi-line parsing 2023-02-05 06:17:51 +00:00
mrq
84a9758ab9 Set transformer and model folder to local './models/' instead of for the user profile, because I'm sick of more bloat polluting my C:\ 2023-02-05 04:18:35 +00:00
mrq
ed33e34fcc Added choices to choose between diffusion samplers (p, ddim) 2023-02-05 01:28:31 +00:00
mrq
5c876b81f3 Added small optimization with caching latents, dropped Anaconda for just a py3.9 + pip + venv setup, added helper install scripts for such, cleaned up app.py, added flag '--low-vram' to disable minor optimizations 2023-02-04 01:50:57 +00:00
mrq
8f20afc18f Reverted slight improvement patch, as it's just enough to OOM on GPUs with low VRAM 2023-02-03 21:45:06 +00:00
mrq
e8d4a4f89c Added progress for transforming to audio, changed number inputs to sliders instead 2023-02-03 04:56:30 +00:00
mrq
ea751d7b6c forgot to copy the alleged slight performance improvement patch, added detailed progress information with passing gr.Progress, save a little more info with output 2023-02-03 04:20:01 +00:00
mrq
74f447e5d0 QoL fixes 2023-02-02 21:13:28 +00:00
James Betker
5dc3e269b3
Merge pull request #233 from kianmeng/fix-typos
Fix typos
2023-01-17 18:24:24 -07:00
원빈 정
b3d67dcc6b Add reference of univnet implementation 2023-01-06 15:57:02 +09:00
Kian-Meng Ang
551fe655ff Fix typos
Found via `codespell -S *.json -L splitted,nd,ser,broadcat`
2023-01-06 11:04:36 +08:00
James Betker
958c6d2f73 Get rid of checkpointing
It isn't needed in inference.
2022-06-15 22:09:15 -06:00
Johan Nordberg
de7c5ddec3 Typofix 2022-06-11 21:19:07 +09:00
Johan Nordberg
b876a6b32c Allow running on CPU 2022-06-11 20:03:14 +09:00
Marcus Llewellyn
700978b777 Fixed silly lack of EOF blank line, indentation 2022-06-06 15:13:29 -05:00
Marcus Llewellyn
2477a4f763 read.py combines all candidates
If candidates where greater than 1 on in read.py, only the fist candidate clips would be combined. This adds a bit of code to make a combined file for every candidate.
2022-06-04 17:47:29 -05:00
James Betker
68c1580f94
Merge pull request #74 from jnordberg/improved-cli
Add CLI tool
2022-05-28 21:33:53 -06:00
Johan Nordberg
d8f98c07b4 Remove some assumptions about working directory
This allows cli tool to run when not standing in repository dir
2022-05-29 01:10:19 +00:00
Johan Nordberg
9f6ae0f0b3 Add tortoise_cli.py 2022-05-28 05:25:23 +00:00