DL-Art-School/codes/models/archs
James Betker 8ab595e427 Add FlatProcessorNet
After doing some thinking and reading on the subject, it occurred to me that
I was treating the generator like a discriminator by focusing the network
complexity at the feature levels. It makes far more sense to process each conv
level equally for the generator, hence the FlatProcessorNet in this commit. This
network borrows some of the residual pass-through logic from RRDB which makes
the gradient path exceptionally short for pretty much all model parameters and
can be trained in O1 optimization mode without overflows again.
2020-04-28 11:49:21 -06:00
..
dcn
__init__.py
arch_util.py Add FlatProcessorNet 2020-04-28 11:49:21 -06:00
discriminator_vgg_arch.py
DUF_arch.py
EDVR_arch.py
FlatProcessorNet_arch.py Add FlatProcessorNet 2020-04-28 11:49:21 -06:00
HighToLowResNet.py
RRDBNet_arch.py
SRResNet_arch.py
TOF_arch.py