Commit Graph

  • dcc5c140e6 fixes mrq 2023-02-15 15:33:08 +0000
  • 729b292515 oops x2 mrq 2023-02-15 05:57:42 +0000
  • 5bf98de301 oops mrq 2023-02-15 05:55:01 +0000
  • 3e8365fdec voicefixed files do not overwrite, as my autism wants to hear the difference between them, incrementing file format fixed for real mrq 2023-02-15 05:49:28 +0000
  • ea1bc770aa added option: force cpu for conditioning latents, for when you want low chunk counts but your GPU keeps OOMing because fuck fragmentation mrq 2023-02-15 05:01:40 +0000
  • b721e395b5 modified conversion scripts to not give a shit about bitrate and formats since torchaudio.load handles all of that anyways, and it all gets resampled anyways mrq 2023-02-15 04:44:14 +0000
  • 2e777e8a67 done away with kludgy shit code, just have the user decide how many chunks to slice concat'd samples to (since it actually does improve vocie replicability) mrq 2023-02-15 04:39:31 +0000
  • 314feaeea1 added reset generation settings to default button, revamped utilities tab to double as plain jane voice importer (and runs through voicefixer despite it not really doing anything if your voice samples are already of decent quality anyways), ditched load_wav_to_torch or whatever it was called because it literally exists as torchaudio.load, sample voice is now a combined waveform of all your samples and will always return even if using a latents file mrq 2023-02-14 21:20:04 +0000
  • 0bc2c1f540 updates chunk size to the chunked tensor length, just in case mrq 2023-02-14 17:13:34 +0000
  • 48275899e8 added flag to enable/disable voicefixer using CUDA because I'll OOM on my 2060, changed from naively subdividing eavenly (2,4,8,16 pieces) to just incrementing by 1 (1,2,3,4) when trying to subdivide within constraints of the max chunk size for computing voice latents mrq 2023-02-14 16:47:34 +0000
  • b648186691 history tab doesn't naively reuse the voice dir instead for results, experimental "divide total sound size until it fits under requests max chunk size" doesn't have a +1 to mess things up (need to re-evaluate how I want to calculate sizes of bests fits eventually) mrq 2023-02-14 16:23:04 +0000
  • 47f4b5bf81 voicefixer uses CUDA if exposed mrq 2023-02-13 15:30:49 +0000
  • 8250a79b23 Implemented kv_cache "fix" (from 1f3c1b5f4a); guess I should find out why it's crashing DirectML backend mrq 2023-02-13 13:48:31 +0000
  • 80eeef01fb Merge pull request 'Download from Gradio' (#31) from Armored1065/tortoise-tts:main into main mrq 2023-02-13 13:30:09 +0000
  • 446d643d62 Merge pull request 'Update 'README.md'' (#1) from armored1065-patch-1 into main Armored1065 2023-02-13 06:21:37 +0000
  • 8c96aa02c5 Merge pull request 'Update 'README.md'' (#1) from armored1065-patch-1 into main Armored1065 2023-02-13 06:21:37 +0000
  • 99f901baa9 Update 'README.md' Armored1065 2023-02-13 06:19:42 +0000
  • d458e932be Update 'README.md' Armored1065 2023-02-13 06:19:42 +0000
  • 37d25573ac added random voice option back because I forgot I accidentally removed it mrq 2023-02-13 04:57:06 +0000
  • f92e432c8d added random voice option back because I forgot I accidentally removed it mrq 2023-02-13 04:57:06 +0000
  • a84aaa4f96 Fixed out of order settings causing other settings to flipflop mrq 2023-02-13 03:43:08 +0000
  • a2bac3fb2c Fixed out of order settings causing other settings to flipflop mrq 2023-02-13 03:43:08 +0000
  • 4ced0296a2 DirectML: fixed redaction/aligner by forcing it to stay on CPU mrq 2023-02-12 20:52:04 +0000
  • 5b5e32338c DirectML: fixed redaction/aligner by forcing it to stay on CPU mrq 2023-02-12 20:52:04 +0000
  • 409dec98d5 fixed voicefixing not working as intended, load TTS before Gradio in the webui due to how long it takes to initialize tortoise (instead of just having a block to preload it) mrq 2023-02-12 20:05:59 +0000
  • 824ad38cca fixed voicefixing not working as intended, load TTS before Gradio in the webui due to how long it takes to initialize tortoise (instead of just having a block to preload it) mrq 2023-02-12 20:05:59 +0000
  • b85c9921d7 added button to recalculate voice latents, added experimental switch for computing voice latents mrq 2023-02-12 18:11:40 +0000
  • 4d01bbd429 added button to recalculate voice latents, added experimental switch for computing voice latents mrq 2023-02-12 18:11:40 +0000
  • 2210b49cb6 fixed regression with computing conditional latencies outside of the CPU mrq 2023-02-12 17:44:39 +0000
  • 88529fda43 fixed regression with computing conditional latencies outside of the CPU mrq 2023-02-12 17:44:39 +0000
  • a2d95fe208 fixed silently crashing from enabling kv_cache-ing if using the DirectML backend, throw an error when reading a generated audio file that does not have any embedded metadata in it, cleaned up the blocks of code that would DMA/transfer tensors/models between GPU and CPU mrq 2023-02-12 14:46:21 +0000
  • 65f74692a0 fixed silently crashing from enabling kv_cache-ing if using the DirectML backend, throw an error when reading a generated audio file that does not have any embedded metadata in it, cleaned up the blocks of code that would DMA/transfer tensors/models between GPU and CPU mrq 2023-02-12 14:46:21 +0000
  • 25e70dce1a instll python3.9, wrapped try/catch when parsing args.listen in case you somehow manage to insert garbage into that field and fuck up your config, removed a very redudnant setup.py install call since that only is required if you're just going to install it for using outside of the tortoise-tts folder mrq 2023-02-12 04:35:21 +0000
  • 94757f5b41 instll python3.9, wrapped try/catch when parsing args.listen in case you somehow manage to insert garbage into that field and fuck up your config, removed a very redudnant setup.py install call since that only is required if you're just going to install it for using outside of the tortoise-tts folder mrq 2023-02-12 04:35:21 +0000
  • 6328466852 cleanup loop, save files while generating a batch in the event it crashes midway through mrq 2023-02-12 01:15:22 +0000
  • ddd0c4ccf8 cleanup loop, save files while generating a batch in the event it crashes midway through mrq 2023-02-12 01:15:22 +0000
  • 5f1c032312 fixed regression where the auto_conds do not move to the GPU and causes a problem during CVVP compare pass mrq 2023-02-11 20:34:12 +0000
  • 1b55730e67 fixed regression where the auto_conds do not move to the GPU and causes a problem during CVVP compare pass mrq 2023-02-11 20:34:12 +0000
  • 2f86565969 Merge pull request 'Only directories in the voice list' (#20) from lightmare/tortoise-tts:only_dirs_in_voice_list into main mrq 2023-02-11 20:14:36 +0000
  • 3d69274a46 Merge pull request 'Only directories in the voice list' (#20) from lightmare/tortoise-tts:only_dirs_in_voice_list into main mrq 2023-02-11 20:14:36 +0000
  • 192a510ee1 Only directories in the voice list lightmare 2023-02-11 18:26:51 +0000
  • 13b60db29c Only directories in the voice list lightmare 2023-02-11 18:26:51 +0000
  • 84316d8f80 Moved experimental settings to main tab, hidden under a check box mrq 2023-02-11 17:21:08 +0000
  • 3a8ce5a110 Moved experimental settings to main tab, hidden under a check box mrq 2023-02-11 17:21:08 +0000
  • 50073e635f sloppily guarantee stop/reloading TTS actually works mrq 2023-02-11 17:01:40 +0000
  • 126f1a0afe sloppily guarantee stop/reloading TTS actually works mrq 2023-02-11 17:01:40 +0000
  • 4b3b0ead1a Added candidate selection for outputs, hide output elements (except for the main one) to only show one progress bar mrq 2023-02-11 16:34:47 +0000
  • 6d06bcce05 Added candidate selection for outputs, hide output elements (except for the main one) to only show one progress bar mrq 2023-02-11 16:34:47 +0000
  • c5337a6b51 Added integration for "voicefixer", fixed issue where candidates>1 and lines>1 only outputs the last combined candidate, numbered step for each generation in progress, output time per generation step mrq 2023-02-11 15:02:11 +0000
  • a7330164ab Added integration for "voicefixer", fixed issue where candidates>1 and lines>1 only outputs the last combined candidate, numbered step for each generation in progress, output time per generation step mrq 2023-02-11 15:02:11 +0000
  • fa743e2e9b store generation time per generation rather than per entire request mrq 2023-02-11 13:00:39 +0000
  • 841754602e store generation time per generation rather than per entire request mrq 2023-02-11 13:00:39 +0000
  • ffb269e579 fixed using old output dir because of my autism with prefixing everything with "./" broke it, fixed incrementing filenames mrq 2023-02-11 12:39:16 +0000
  • 44eba62dc8 fixed using old output dir because of my autism with prefixing everything with "./" broke it, fixed incrementing filenames mrq 2023-02-11 12:39:16 +0000
  • 9bf1ea5b0a History tab (3/10 it works) mrq 2023-02-11 01:45:25 +0000
  • 58e2b22b0e History tab (3/10 it works) mrq 2023-02-11 01:45:25 +0000
  • 340a89f883 Numbering predicates on input_#.json files instead of "number of wavs" mrq 2023-02-10 22:51:56 +0000
  • c924ebd034 Numbering predicates on input_#.json files instead of "number of wavs" mrq 2023-02-10 22:51:56 +0000
  • 8641cc9906 revamped result formatting, added "kludgy" stop button mrq 2023-02-10 22:12:37 +0000
  • 4f903159ee revamped result formatting, added "kludgy" stop button mrq 2023-02-10 22:12:37 +0000
  • 8f789d17b9 Slight notebook adjust mrq 2023-02-10 20:22:12 +0000
  • 9e0fbff545 Slight notebook adjust mrq 2023-02-10 20:22:12 +0000
  • 7471bc209c Moved voices out of the tortoise folder because it kept being processed for setup.py mrq 2023-02-10 20:11:56 +0000
  • 52a9ed7858 Moved voices out of the tortoise folder because it kept being processed for setup.py mrq 2023-02-10 20:11:56 +0000
  • 2bce24b9dd Cleanup mrq 2023-02-10 19:55:33 +0000
  • 8b83c9083d Cleanup mrq 2023-02-10 19:55:33 +0000
  • 811539b20a Added the remaining input settings mrq 2023-02-10 16:47:57 +0000
  • a09eff5d9c Added the remaining input settings mrq 2023-02-10 16:47:57 +0000
  • f5ed5499a0 Added a link to the colab notebook mrq 2023-02-10 16:26:13 +0000
  • 7baf9e3f79 Added a link to the colab notebook mrq 2023-02-10 16:26:13 +0000
  • 07c54ad361 Colab notebook (part II) mrq 2023-02-10 16:12:11 +0000
  • 5b852da720 Colab notebook (part II) mrq 2023-02-10 16:12:11 +0000
  • 939c89f16e Colab notebook (part 1) mrq 2023-02-10 15:58:56 +0000
  • 3d6ac3afaa Colab notebook (part 1) mrq 2023-02-10 15:58:56 +0000
  • 39b81318f2 Added new options: "Output Sample Rate", "Output Volume", and documentation mrq 2023-02-10 03:02:09 +0000
  • efa556b793 Added new options: "Output Sample Rate", "Output Volume", and documentation mrq 2023-02-10 03:02:09 +0000
  • 77b39e59ac oops mrq 2023-02-09 22:17:57 +0000
  • 57af25c6c0 oops mrq 2023-02-09 22:17:57 +0000
  • 3621e16ef9 Added 'Only Load Models Locally' setting mrq 2023-02-09 22:06:55 +0000
  • 504db0d1ac Added 'Only Load Models Locally' setting mrq 2023-02-09 22:06:55 +0000
  • dccedc3f66 Added and documented mrq 2023-02-09 21:07:51 +0000
  • 460f5d6e32 Added and documented mrq 2023-02-09 21:07:51 +0000
  • 8c30cd1aa4 Oops mrq 2023-02-09 20:49:22 +0000
  • 145298b766 Oops mrq 2023-02-09 20:49:22 +0000
  • d7443dfa06 Added option: listen path mrq 2023-02-09 20:42:38 +0000
  • 729be135ef Added option: listen path mrq 2023-02-09 20:42:38 +0000
  • 38ee19cd57 I didn't have to suck off a wizard for DirectML support (courtesy of https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/7600 for leading the way) mrq 2023-02-09 05:05:21 +0000
  • 3f8302a680 I didn't have to suck off a wizard for DirectML support (courtesy of https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/7600 for leading the way) mrq 2023-02-09 05:05:21 +0000
  • 716e227953 oops mrq 2023-02-09 02:39:08 +0000
  • 50b4e2c458 oops mrq 2023-02-09 02:39:08 +0000
  • a37546ad99 owari da... mrq 2023-02-09 01:53:25 +0000
  • b23d6b4b4c owari da... mrq 2023-02-09 01:53:25 +0000
  • 6255c98006 beginning to add DirectML support mrq 2023-02-08 23:03:52 +0000
  • 494f3c84a1 beginning to add DirectML support mrq 2023-02-08 23:03:52 +0000
  • d9a9fa6a82 Added two flags/settings: embed output settings, slimmer computed voice latents mrq 2023-02-08 14:14:28 +0000
  • 81e4d261b7 Added two flags/settings: embed output settings, slimmer computed voice latents mrq 2023-02-08 14:14:28 +0000
  • f03b6b8d97 disable telemetry/what-have-you if not requesting a public Gradio URL mrq 2023-02-07 21:44:16 +0000
  • 94eab20529 disable telemetry/what-have-you if not requesting a public Gradio URL mrq 2023-02-07 21:44:16 +0000
  • 479f30c808 Merge pull request 'Added convert.sh' (#8) from lightmare/tortoise-tts:convert_sh into main mrq 2023-02-07 21:09:00 +0000
  • 0bf4fefd42 Merge pull request 'Added convert.sh' (#8) from lightmare/tortoise-tts:convert_sh into main mrq 2023-02-07 21:09:00 +0000