Commit Graph

306 Commits

Author SHA1 Message Date
James Betker
c61cd64bc9 network updates 2022-06-08 09:26:59 -06:00
James Betker
5a54d7db11 unet with ar prior 2022-06-07 17:52:36 -06:00
James Betker
a14274c845 fix memory issue 2022-06-03 11:20:09 -06:00
James Betker
c0db85bf4f music quantizer 2022-05-31 21:06:54 -06:00
James Betker
96da10415e facepalm 2022-05-30 16:48:18 -06:00
James Betker
479c3195f3 make wandb histograms work again 2022-05-30 16:44:39 -06:00
James Betker
f7d237a50a train quantizer with diffusion 2022-05-30 16:25:33 -06:00
James Betker
2e72fddaeb td_tts_2 2022-05-29 22:22:14 -06:00
James Betker
3db862dd32 adf update 2022-05-27 09:25:53 -06:00
James Betker
48aab2babe ressurect ctc code gen with some cool new ideas 2022-05-24 14:02:33 -06:00
James Betker
5d13d38119 allow opt states to be reset 2022-05-23 10:54:37 -06:00
James Betker
57d6f6d366 Big rework of flat_diffusion
Back to the drawing board, boys. Time to waste some resources catching bugs....
2022-05-22 08:09:33 -06:00
James Betker
e9fb2ead9a m2v stuff 2022-05-20 11:01:17 -06:00
James Betker
519151d83f m2v 2022-05-17 15:37:59 -06:00
James Betker
eb64d18075 Fix phoneme tokenizer 2022-05-13 17:56:26 -06:00
James Betker
1177c35dec music fid updates 2022-05-08 18:49:39 -06:00
James Betker
58ed27d7a8 new gap_filler 2022-05-07 12:44:23 -06:00
James Betker
1609101a42 musical gap filler 2022-05-05 16:47:08 -06:00
James Betker
47662b9ec5 some random crap 2022-05-04 20:29:23 -06:00
James Betker
c42c53e75a Add a trainable network for converting a normal distribution into a latent space 2022-05-02 09:47:30 -06:00
James Betker
e402089556 abstractify 2022-05-02 00:11:26 -06:00
James Betker
b712d3b72b break out get_conditioning_latent from unified_voice 2022-05-01 23:04:44 -06:00
James Betker
f02b01bd9d reverse univnet classifier 2022-04-20 21:37:55 -06:00
James Betker
419f4d37bd gen2 music 2022-04-19 23:38:37 -06:00
James Betker
48cb6a5abd misc 2022-04-16 20:28:04 -06:00
James Betker
efe12cb816 Update clvp to add masking probabilities in conditioning and to support code inputs 2022-04-15 09:11:23 -06:00
James Betker
f2c172291f fix audio_diffusion_fid for autoregressive latent inputs 2022-04-11 12:08:15 -06:00
James Betker
3f8d7955ef unified_voice with rotary embeddings 2022-04-07 20:11:14 -06:00
James Betker
e6387c7613 Fix eval logic to not run immediately 2022-04-07 11:29:57 -06:00
James Betker
3d916e7687 Fix evaluation when using multiple batch sizes 2022-04-05 07:51:09 -06:00
James Betker
572d137589 track iteration rate 2022-04-04 12:33:25 -06:00
James Betker
4cdb0169d0 update training data encountered when using force_start_step 2022-04-04 12:25:00 -06:00
James Betker
2b6ff09225 autoregressive_codegen v1 2022-04-02 15:07:39 -06:00
James Betker
2a29a71c37 attempt to force meaningful codes by adding a surrogate loss 2022-03-26 08:31:40 -06:00
James Betker
57da6d0ddf more simplifications 2022-03-22 11:46:03 -06:00
James Betker
bf08519d71 fixes 2022-03-17 10:53:39 -06:00
James Betker
54202aa099 fix mel normalization 2022-03-16 09:26:55 -06:00
James Betker
d553808d24 misc 2022-03-08 15:52:16 -07:00
James Betker
d1dc8dbb35 Support tts9 2022-03-05 20:14:36 -07:00
James Betker
e1052a5e32 Move log consensus to train for efficiency 2022-03-04 13:41:32 -07:00
James Betker
ce6dfdf255 Distributed "fixes" 2022-03-04 12:46:41 -07:00
James Betker
70fa780edb Add mechanism to export grad norms 2022-03-01 20:19:52 -07:00
James Betker
db0c3340ac Implement guidance-free diffusion in eval
And a few other fixes
2022-03-01 11:49:36 -07:00
James Betker
ac920798bb misc 2022-02-27 14:49:11 -07:00
James Betker
896ac029ae allow continuation of samples encountered 2022-02-21 19:12:50 -07:00
James Betker
79e8f36d30 Convert CLIP models into new folder 2022-02-15 20:53:07 -07:00
James Betker
2bdb515068 A few mods to make wav2vec2 trainable with DDP on DLAS 2022-02-15 06:28:54 -07:00
James Betker
29534180b2 w2v fine tuner 2022-02-12 20:00:59 -07:00
James Betker
4abc094b47 fix train bug 2022-02-11 11:18:15 -07:00
James Betker
a930f2576e Begin a migration to specifying training rate on megasamples instead of arbitrary "steps"
This should help me greatly in tuning models.  It's also necessary now that batch size isn't really
respected; we simply step once the gradient direction becomes unstable.
2022-02-09 17:25:05 -07:00