forked from mrq/DL-Art-School
Re-enable batch norm on switch processing blocks
Found out that batch norm is causing the switches to init really poorly - not using a significant number of transforms. Might be a great time to re-consider using the attention norm, but for now just re-enable it.
This commit is contained in:
parent
4001db1ede
commit
42a10b34ce
|
@ -59,7 +59,7 @@ class HalvingProcessingBlock(nn.Module):
|
||||||
def __init__(self, filters):
|
def __init__(self, filters):
|
||||||
super(HalvingProcessingBlock, self).__init__()
|
super(HalvingProcessingBlock, self).__init__()
|
||||||
self.bnconv1 = ConvBnLelu(filters, filters * 2, stride=2, bn=False)
|
self.bnconv1 = ConvBnLelu(filters, filters * 2, stride=2, bn=False)
|
||||||
self.bnconv2 = ConvBnLelu(filters * 2, filters * 2, bn=False)
|
self.bnconv2 = ConvBnLelu(filters * 2, filters * 2, bn=True)
|
||||||
|
|
||||||
def forward(self, x):
|
def forward(self, x):
|
||||||
x = self.bnconv1(x)
|
x = self.bnconv1(x)
|
||||||
|
@ -72,7 +72,7 @@ def create_sequential_growing_processing_block(filters_init, filter_growth, num_
|
||||||
convs = []
|
convs = []
|
||||||
current_filters = filters_init
|
current_filters = filters_init
|
||||||
for i in range(num_convs):
|
for i in range(num_convs):
|
||||||
convs.append(ConvBnLelu(current_filters, current_filters + filter_growth, bn=False))
|
convs.append(ConvBnLelu(current_filters, current_filters + filter_growth, bn=True))
|
||||||
current_filters += filter_growth
|
current_filters += filter_growth
|
||||||
return nn.Sequential(*convs), current_filters
|
return nn.Sequential(*convs), current_filters
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user