DL-Art-School/codes
James Betker 009a1e8404 Add a new diffusion_vocoder that should be trainable faster
This new one has a "cheating" top layer, that does not feed down into the unet encoder,
but does consume the outputs of the unet. This cheater only operates on half of the input,
while the rest of the unet operates on the full input. This limits the dimensionality of this last
layer, on the assumption that these last layers consume by far the most computation and memory,
but do not require the full input context.

Losses are only computed on half of the aggregate input.
2022-01-11 17:26:07 -07:00
..
.idea
data Fix dataset 2022-01-06 15:24:37 -07:00
models Add a new diffusion_vocoder that should be trainable faster 2022-01-11 17:26:07 -07:00
scripts misc updates 2022-01-11 16:25:40 -07:00
trainer More zero_grad fixes 2022-01-08 20:31:19 -07:00
utils
multi_modal_train.py
process_video.py
requirements.txt misc updates 2022-01-11 16:25:40 -07:00
test.py
train.py misc updates 2022-01-11 16:25:40 -07:00
use_discriminator_as_filter.py