DL-Art-School/codes/models/archs
James Betker 8ab595e427 Add FlatProcessorNet
After doing some thinking and reading on the subject, it occurred to me that
I was treating the generator like a discriminator by focusing the network
complexity at the feature levels. It makes far more sense to process each conv
level equally for the generator, hence the FlatProcessorNet in this commit. This
network borrows some of the residual pass-through logic from RRDB which makes
the gradient path exceptionally short for pretty much all model parameters and
can be trained in O1 optimization mode without overflows again.
2020-04-28 11:49:21 -06:00
..
dcn add deform_conv_cuda_kernel.cu 2019-08-27 17:49:12 +08:00
__init__.py mmsr 2019-08-23 21:42:47 +08:00
arch_util.py Add FlatProcessorNet 2020-04-28 11:49:21 -06:00
discriminator_vgg_arch.py Implement downsample GAN 2020-04-24 00:00:46 -06:00
DUF_arch.py mmsr 2019-08-23 21:42:47 +08:00
EDVR_arch.py mmsr 2019-08-23 21:42:47 +08:00
FlatProcessorNet_arch.py Add FlatProcessorNet 2020-04-28 11:49:21 -06:00
HighToLowResNet.py Retool HighToLowResNet 2020-04-26 01:13:42 -06:00
RRDBNet_arch.py Support variant input sizes and scales 2020-04-22 00:39:55 -06:00
SRResNet_arch.py mmsr 2019-08-23 21:42:47 +08:00
TOF_arch.py mmsr 2019-08-23 21:42:47 +08:00