fix flat
This commit is contained in:
parent
e47a759ed8
commit
5405ce4363
|
@ -71,7 +71,7 @@ class DiffusionTtsFlat(nn.Module):
|
||||||
attn_dropout=dropout,
|
attn_dropout=dropout,
|
||||||
use_rmsnorm=True,
|
use_rmsnorm=True,
|
||||||
ff_glu=True,
|
ff_glu=True,
|
||||||
rotary_emb_dim=True,
|
rotary_pos_emb=True,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
@ -91,7 +91,7 @@ class DiffusionTtsFlat(nn.Module):
|
||||||
attn_dropout=dropout,
|
attn_dropout=dropout,
|
||||||
use_rmsnorm=True,
|
use_rmsnorm=True,
|
||||||
ff_glu=True,
|
ff_glu=True,
|
||||||
rotary_emb_dim=True,
|
rotary_pos_emb=True,
|
||||||
)
|
)
|
||||||
))
|
))
|
||||||
self.conditioning_conv = nn.Conv1d(model_channels*2, model_channels, 1)
|
self.conditioning_conv = nn.Conv1d(model_channels*2, model_channels, 1)
|
||||||
|
@ -110,7 +110,7 @@ class DiffusionTtsFlat(nn.Module):
|
||||||
attn_dropout=dropout,
|
attn_dropout=dropout,
|
||||||
use_rmsnorm=True,
|
use_rmsnorm=True,
|
||||||
ff_glu=True,
|
ff_glu=True,
|
||||||
rotary_emb_dim=True,
|
rotary_pos_emb=True,
|
||||||
layerdrop_percent=0,
|
layerdrop_percent=0,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
@ -130,9 +130,10 @@ class DiffusionTtsFlat(nn.Module):
|
||||||
attn_dropout=dropout,
|
attn_dropout=dropout,
|
||||||
use_rmsnorm=True,
|
use_rmsnorm=True,
|
||||||
ff_glu=True,
|
ff_glu=True,
|
||||||
rotary_emb_dim=True,
|
rotary_pos_emb=True,
|
||||||
layerdrop_percent=layer_drop,
|
layerdrop_percent=layer_drop,
|
||||||
zero_init_branch_output=True,
|
zero_init_branch_output=True,
|
||||||
|
sandwich_coef=4,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
self.layers.transformer.norm = nn.Identity() # We don't want the final norm for the main encoder.
|
self.layers.transformer.norm = nn.Identity() # We don't want the final norm for the main encoder.
|
||||||
|
|
Loading…
Reference in New Issue
Block a user