Commit Graph

14 Commits

Author SHA1 Message Date
James Betker
1b4d9567f3 tfd13 for multi-resolution superscaling 2022-07-18 16:36:22 -06:00
James Betker
15831b2576 some stuff 2022-07-13 21:26:25 -06:00
James Betker
28d95e3141 gptmusic work 2022-06-16 15:09:47 -06:00
James Betker
c61cd64bc9 network updates 2022-06-08 09:26:59 -06:00
James Betker
49568ee16f some updates 2022-06-06 09:13:47 -06:00
James Betker
1f521d6a1d add reconstruction loss to m2v 2022-05-23 09:28:41 -06:00
James Betker
47662b9ec5 some random crap 2022-05-04 20:29:23 -06:00
James Betker
d186414566 More spring cleaning 2022-03-16 12:04:00 -06:00
James Betker
8ada52ccdc Update LR layers to checkpoint better 2022-01-22 08:22:57 -07:00
James Betker
f2a31702b5 Clean stuff up, move more things into arch_util 2021-10-20 21:19:25 -06:00
James Betker
a6f0f854b9 Fix codes when inferring from dvae 2021-10-17 22:51:17 -06:00
James Betker
398185e109 More work on wave-diffusion 2021-07-27 05:36:17 -06:00
James Betker
4328c2f713 Change default ReLU slope to .2 BREAKS COMPATIBILITY
This conforms my ConvGnLelu implementation with the generally accepted negative_slope=.2. I have no idea where I got .1. This will break backwards compatibility with some older models but will likely improve their performance when freshly trained. I did some auditing to find what these models might be, and I am not actively using any of them, so probably OK.
2020-12-19 08:28:03 -07:00
James Betker
5640e4efe4 More refactoring 2020-12-18 09:18:34 -07:00