2022-06-05 07:31:37 +00:00
import itertools
2022-05-30 15:09:42 +00:00
import torch
import torch . nn as nn
import torch . nn . functional as F
2022-06-05 15:35:43 +00:00
from models . arch_util import ResBlock
2022-06-05 04:23:08 +00:00
from models . audio . music . music_quantizer2 import MusicQuantizer2
2022-05-30 15:09:42 +00:00
from models . diffusion . nn import timestep_embedding , normalization , zero_module , conv_nd , linear
from models . diffusion . unet_diffusion import TimestepBlock
from models . lucidrains . x_transformers import Encoder , Attention , FeedForward , RMSScaleShiftNorm , RotaryEmbedding
from trainer . networks import register_model
from utils . util import checkpoint , print_network
def is_latent ( t ) :
return t . dtype == torch . float
def is_sequence ( t ) :
return t . dtype == torch . long
class MultiGroupEmbedding ( nn . Module ) :
def __init__ ( self , tokens , groups , dim ) :
super ( ) . __init__ ( )
self . m = nn . ModuleList ( [ nn . Embedding ( tokens , dim / / groups ) for _ in range ( groups ) ] )
def forward ( self , x ) :
h = [ embedding ( x [ : , : , i ] ) for i , embedding in enumerate ( self . m ) ]
return torch . cat ( h , dim = - 1 )
class TimestepRotaryEmbedSequential ( nn . Sequential , TimestepBlock ) :
def forward ( self , x , emb , rotary_emb ) :
for layer in self :
if isinstance ( layer , TimestepBlock ) :
x = layer ( x , emb , rotary_emb )
else :
x = layer ( x , rotary_emb )
return x
class DietAttentionBlock ( TimestepBlock ) :
def __init__ ( self , in_dim , dim , heads , dropout ) :
super ( ) . __init__ ( )
2022-06-05 04:23:08 +00:00
self . rms_scale_norm = RMSScaleShiftNorm ( in_dim )
2022-05-30 15:09:42 +00:00
self . proj = nn . Linear ( in_dim , dim )
self . attn = Attention ( dim , heads = heads , causal = False , dropout = dropout )
2022-06-05 16:06:17 +00:00
self . ff = FeedForward ( dim , in_dim , mult = 1 , dropout = dropout , zero_init_output = True )
2022-05-30 15:09:42 +00:00
def forward ( self , x , timestep_emb , rotary_emb ) :
2022-06-05 04:23:08 +00:00
h = self . rms_scale_norm ( x , norm_scale_shift_inp = timestep_emb )
h = self . proj ( h )
2022-06-05 16:06:17 +00:00
h , _ , _ , _ = checkpoint ( self . attn , h , None , None , None , None , None , rotary_emb )
2022-05-30 15:09:42 +00:00
h = checkpoint ( self . ff , h )
return h + x
class TransformerDiffusion ( nn . Module ) :
"""
A diffusion model composed entirely of stacks of transformer layers . Why would you do it any other way ?
"""
def __init__ (
self ,
prenet_channels = 256 ,
2022-06-05 04:23:08 +00:00
prenet_layers = 3 ,
2022-05-30 15:09:42 +00:00
model_channels = 512 ,
block_channels = 256 ,
num_layers = 8 ,
in_channels = 256 ,
rotary_emb_dim = 32 ,
input_vec_dim = 512 ,
out_channels = 512 , # mean and variance
dropout = 0 ,
use_fp16 = False ,
2022-06-08 15:26:59 +00:00
ar_prior = False ,
2022-05-30 15:09:42 +00:00
# Parameters for regularization.
unconditioned_percentage = .1 , # This implements a mechanism similar to what is used in classifier-free training.
) :
super ( ) . __init__ ( )
self . in_channels = in_channels
self . model_channels = model_channels
self . prenet_channels = prenet_channels
self . out_channels = out_channels
self . dropout = dropout
self . unconditioned_percentage = unconditioned_percentage
self . enable_fp16 = use_fp16
self . inp_block = conv_nd ( 1 , in_channels , prenet_channels , 3 , 1 , 1 )
self . time_embed = nn . Sequential (
linear ( prenet_channels , prenet_channels ) ,
nn . SiLU ( ) ,
2022-06-05 05:09:54 +00:00
linear ( prenet_channels , model_channels ) ,
2022-05-30 15:09:42 +00:00
)
prenet_heads = prenet_channels / / 64
2022-06-08 15:26:59 +00:00
self . ar_prior = ar_prior
if ar_prior :
self . ar_input = nn . Linear ( input_vec_dim , prenet_channels )
self . ar_prior_intg = Encoder (
2022-05-30 15:09:42 +00:00
dim = prenet_channels ,
2022-06-05 04:23:08 +00:00
depth = prenet_layers ,
2022-05-30 15:09:42 +00:00
heads = prenet_heads ,
ff_dropout = dropout ,
attn_dropout = dropout ,
use_rmsnorm = True ,
ff_glu = True ,
rotary_pos_emb = True ,
zero_init_branch_output = True ,
ff_mult = 1 ,
)
2022-06-08 15:26:59 +00:00
else :
self . input_converter = nn . Linear ( input_vec_dim , prenet_channels )
self . code_converter = Encoder (
dim = prenet_channels ,
depth = prenet_layers ,
heads = prenet_heads ,
ff_dropout = dropout ,
attn_dropout = dropout ,
use_rmsnorm = True ,
ff_glu = True ,
rotary_pos_emb = True ,
zero_init_branch_output = True ,
ff_mult = 1 ,
)
2022-05-30 15:09:42 +00:00
self . unconditioned_embedding = nn . Parameter ( torch . randn ( 1 , 1 , prenet_channels ) )
self . rotary_embeddings = RotaryEmbedding ( rotary_emb_dim )
self . intg = nn . Linear ( prenet_channels * 2 , model_channels )
self . layers = TimestepRotaryEmbedSequential ( * [ DietAttentionBlock ( model_channels , block_channels , block_channels / / 64 , dropout ) for _ in range ( num_layers ) ] )
self . out = nn . Sequential (
normalization ( model_channels ) ,
nn . SiLU ( ) ,
zero_module ( conv_nd ( 1 , model_channels , out_channels , 3 , padding = 1 ) ) ,
)
self . debug_codes = { }
def get_grad_norm_parameter_groups ( self ) :
groups = {
' layers ' : list ( self . layers . parameters ( ) ) + list ( self . inp_block . parameters ( ) ) ,
' code_converters ' : list ( self . input_converter . parameters ( ) ) + list ( self . code_converter . parameters ( ) ) ,
' time_embed ' : list ( self . time_embed . parameters ( ) ) ,
}
return groups
2022-06-08 15:26:59 +00:00
def timestep_independent ( self , prior , expected_seq_len ) :
code_emb = self . ar_input ( prior ) if self . ar_prior else self . input_converter ( prior )
2022-06-10 15:24:41 +00:00
code_emb = self . ar_prior_intg ( code_emb ) if self . ar_prior else self . code_converter ( code_emb )
2022-05-30 15:09:42 +00:00
# Mask out the conditioning branch for whole batch elements, implementing something similar to classifier-free guidance.
if self . training and self . unconditioned_percentage > 0 :
unconditioned_batches = torch . rand ( ( code_emb . shape [ 0 ] , 1 , 1 ) ,
device = code_emb . device ) < self . unconditioned_percentage
2022-06-08 15:26:59 +00:00
code_emb = torch . where ( unconditioned_batches , self . unconditioned_embedding . repeat ( prior . shape [ 0 ] , 1 , 1 ) ,
2022-05-30 15:09:42 +00:00
code_emb )
expanded_code_emb = F . interpolate ( code_emb . permute ( 0 , 2 , 1 ) , size = expected_seq_len , mode = ' nearest ' ) . permute ( 0 , 2 , 1 )
2022-06-05 05:09:54 +00:00
return expanded_code_emb
2022-05-30 15:09:42 +00:00
2022-06-05 05:09:54 +00:00
def forward ( self , x , timesteps , codes = None , conditioning_input = None , precomputed_code_embeddings = None , conditioning_free = False ) :
2022-05-30 15:09:42 +00:00
if precomputed_code_embeddings is not None :
assert codes is None and conditioning_input is None , " Do not provide precomputed embeddings and the other parameters. It is unclear what you want me to do here. "
unused_params = [ ]
if conditioning_free :
2022-06-05 04:23:08 +00:00
code_emb = self . unconditioned_embedding . repeat ( x . shape [ 0 ] , x . shape [ - 1 ] , 1 )
2022-05-30 15:09:42 +00:00
else :
if precomputed_code_embeddings is not None :
code_emb = precomputed_code_embeddings
else :
2022-06-05 05:09:54 +00:00
code_emb = self . timestep_independent ( codes , x . shape [ - 1 ] )
2022-05-30 15:09:42 +00:00
unused_params . append ( self . unconditioned_embedding )
2022-06-05 05:09:54 +00:00
blk_emb = self . time_embed ( timestep_embedding ( timesteps , self . prenet_channels ) )
2022-05-30 15:09:42 +00:00
x = self . inp_block ( x ) . permute ( 0 , 2 , 1 )
rotary_pos_emb = self . rotary_embeddings ( x . shape [ 1 ] , x . device )
x = self . intg ( torch . cat ( [ x , code_emb ] , dim = - 1 ) )
for layer in self . layers :
x = checkpoint ( layer , x , blk_emb , rotary_pos_emb )
x = x . float ( ) . permute ( 0 , 2 , 1 )
out = self . out ( x )
# Involve probabilistic or possibly unused parameters in loss so we don't get DDP errors.
extraneous_addition = 0
for p in unused_params :
extraneous_addition = extraneous_addition + p . mean ( )
out = out + extraneous_addition * 0
return out
2022-06-05 04:23:08 +00:00
class TransformerDiffusionWithQuantizer ( nn . Module ) :
2022-06-10 16:40:56 +00:00
def __init__ ( self , freeze_quantizer_until = 20000 , quantizer_dims = [ 1024 ] , no_reconstruction = True , * * kwargs ) :
2022-06-05 04:23:08 +00:00
super ( ) . __init__ ( )
self . internal_step = 0
self . freeze_quantizer_until = freeze_quantizer_until
self . diff = TransformerDiffusion ( * * kwargs )
2022-06-10 16:40:56 +00:00
self . quantizer = MusicQuantizer2 ( inp_channels = kwargs [ ' in_channels ' ] , inner_dim = quantizer_dims ,
codevector_dim = quantizer_dims [ 0 ] ,
codebook_size = 256 , codebook_groups = 2 ,
max_gumbel_temperature = 4 , min_gumbel_temperature = .5 )
2022-06-05 04:23:08 +00:00
self . quantizer . quantizer . temperature = self . quantizer . min_gumbel_temperature
2022-06-10 16:40:56 +00:00
if no_reconstruction :
del self . quantizer . up
2022-06-05 04:23:08 +00:00
def update_for_step ( self , step , * args ) :
self . internal_step = step
qstep = max ( 0 , self . internal_step - self . freeze_quantizer_until )
self . quantizer . quantizer . temperature = max (
self . quantizer . max_gumbel_temperature * self . quantizer . gumbel_temperature_decay * * qstep ,
self . quantizer . min_gumbel_temperature ,
)
2022-06-10 15:24:41 +00:00
def forward ( self , x , timesteps , truth_mel , conditioning_input = None , disable_diversity = False , conditioning_free = False ) :
2022-06-05 04:23:08 +00:00
quant_grad_enabled = self . internal_step > self . freeze_quantizer_until
2022-06-10 16:40:56 +00:00
mse , diversity_loss , proj = self . quantizer ( truth_mel , return_decoder_latent = True )
proj = proj . permute ( 0 , 2 , 1 )
2022-06-05 04:23:08 +00:00
# Make sure this does not cause issues in DDP by explicitly using the parameters for nothing.
if not quant_grad_enabled :
2022-06-10 16:40:56 +00:00
proj = proj . detach ( )
2022-06-05 04:23:08 +00:00
unused = 0
for p in self . quantizer . parameters ( ) :
unused = unused + p . mean ( ) * 0
proj = proj + unused
2022-06-10 16:40:56 +00:00
diff = self . diff ( x , timesteps , codes = proj , conditioning_input = conditioning_input ,
conditioning_free = conditioning_free )
if mse is None :
return diff , diversity_loss
return diff , diversity_loss , mse
2022-06-05 04:23:08 +00:00
def get_debug_values ( self , step , __ ) :
if self . quantizer . total_codes > 0 :
2022-06-10 16:40:56 +00:00
return { ' histogram_quant_codes ' : self . quantizer . codes [ : self . quantizer . total_codes ] ,
2022-06-10 03:56:47 +00:00
' gumbel_temperature ' : self . quantizer . quantizer . temperature }
2022-06-05 04:23:08 +00:00
else :
return { }
2022-06-05 05:09:54 +00:00
def get_grad_norm_parameter_groups ( self ) :
groups = {
2022-06-05 07:31:37 +00:00
' attention_layers ' : list ( itertools . chain . from_iterable ( [ lyr . attn . parameters ( ) for lyr in self . diff . layers ] ) ) ,
' ff_layers ' : list ( itertools . chain . from_iterable ( [ lyr . ff . parameters ( ) for lyr in self . diff . layers ] ) ) ,
2022-06-05 05:09:54 +00:00
' quantizer_encoder ' : list ( self . quantizer . encoder . parameters ( ) ) ,
' quant_codebook ' : [ self . quantizer . quantizer . codevectors ] ,
' rotary_embeddings ' : list ( self . diff . rotary_embeddings . parameters ( ) ) ,
' out ' : list ( self . diff . out . parameters ( ) ) ,
' x_proj ' : list ( self . diff . inp_block . parameters ( ) ) ,
' layers ' : list ( self . diff . layers . parameters ( ) ) ,
' code_converters ' : list ( self . diff . input_converter . parameters ( ) ) + list ( self . diff . code_converter . parameters ( ) ) ,
' time_embed ' : list ( self . diff . time_embed . parameters ( ) ) ,
}
return groups
2022-06-05 04:23:08 +00:00
2022-06-08 15:26:59 +00:00
class TransformerDiffusionWithARPrior ( nn . Module ) :
def __init__ ( self , freeze_diff = False , * * kwargs ) :
super ( ) . __init__ ( )
self . internal_step = 0
from models . audio . music . gpt_music import GptMusicLower
self . ar = GptMusicLower ( dim = 512 , layers = 12 )
for p in self . ar . parameters ( ) :
p . DO_NOT_TRAIN = True
p . requires_grad = False
self . diff = TransformerDiffusion ( ar_prior = True , * * kwargs )
if freeze_diff :
for p in self . diff . parameters ( ) :
p . DO_NOT_TRAIN = True
p . requires_grad = False
for p in list ( self . diff . ar_prior_intg . parameters ( ) ) + list ( self . diff . ar_input . parameters ( ) ) :
del p . DO_NOT_TRAIN
p . requires_grad = True
def get_grad_norm_parameter_groups ( self ) :
groups = {
' attention_layers ' : list ( itertools . chain . from_iterable ( [ lyr . attn . parameters ( ) for lyr in self . diff . layers ] ) ) ,
' ff_layers ' : list ( itertools . chain . from_iterable ( [ lyr . ff . parameters ( ) for lyr in self . diff . layers ] ) ) ,
' rotary_embeddings ' : list ( self . diff . rotary_embeddings . parameters ( ) ) ,
' out ' : list ( self . diff . out . parameters ( ) ) ,
' x_proj ' : list ( self . diff . inp_block . parameters ( ) ) ,
' layers ' : list ( self . diff . layers . parameters ( ) ) ,
' ar_prior_intg ' : list ( self . diff . ar_prior_intg . parameters ( ) ) ,
' time_embed ' : list ( self . diff . time_embed . parameters ( ) ) ,
}
return groups
def forward ( self , x , timesteps , truth_mel , disable_diversity = False , conditioning_input = None , conditioning_free = False ) :
with torch . no_grad ( ) :
prior = self . ar ( truth_mel , conditioning_input , return_latent = True )
diff = self . diff ( x , timesteps , prior , conditioning_free = conditioning_free )
return diff
2022-05-30 15:09:42 +00:00
@register_model
2022-06-05 04:23:08 +00:00
def register_transformer_diffusion8 ( opt_net , opt ) :
2022-05-30 15:09:42 +00:00
return TransformerDiffusion ( * * opt_net [ ' kwargs ' ] )
2022-06-05 04:23:08 +00:00
@register_model
def register_transformer_diffusion8_with_quantizer ( opt_net , opt ) :
return TransformerDiffusionWithQuantizer ( * * opt_net [ ' kwargs ' ] )
2022-06-08 15:26:59 +00:00
@register_model
def register_transformer_diffusion8_with_ar_prior ( opt_net , opt ) :
return TransformerDiffusionWithARPrior ( * * opt_net [ ' kwargs ' ] )
2022-06-05 04:23:08 +00:00
2022-06-08 15:26:59 +00:00
def test_quant_model ( ) :
2022-06-10 16:40:56 +00:00
clip = torch . randn ( 2 , 100 , 401 )
2022-06-05 04:23:08 +00:00
ts = torch . LongTensor ( [ 600 , 600 ] )
2022-06-10 16:40:56 +00:00
model = TransformerDiffusionWithQuantizer ( in_channels = 100 , out_channels = 200 , quantizer_dims = [ 1024 , 768 , 512 , 384 ] ,
model_channels = 2048 , block_channels = 1024 , prenet_channels = 1024 ,
input_vec_dim = 1024 , num_layers = 16 , prenet_layers = 6 ,
no_reconstruction = False )
#model.get_grad_norm_parameter_groups()
2022-06-05 04:23:08 +00:00
2022-06-10 16:40:56 +00:00
#quant_weights = torch.load('D:\\dlas\\experiments\\train_music_quant_r4\\models\\5000_generator.pth')
2022-06-05 04:23:08 +00:00
#diff_weights = torch.load('X:\\dlas\\experiments\\train_music_diffusion_tfd5\\models\\48000_generator_ema.pth')
2022-06-10 16:40:56 +00:00
#model.quantizer.load_state_dict(quant_weights, strict=False)
2022-06-05 04:23:08 +00:00
#model.diff.load_state_dict(diff_weights)
2022-06-10 16:40:56 +00:00
#torch.save(model.state_dict(), 'sample.pth')
2022-06-05 04:23:08 +00:00
print_network ( model )
2022-06-10 16:40:56 +00:00
o = model ( clip , ts , clip )
2022-05-30 15:09:42 +00:00
2022-06-08 15:26:59 +00:00
def test_ar_model ( ) :
2022-06-10 16:40:56 +00:00
clip = torch . randn ( 2 , 256 , 401 )
2022-06-08 15:26:59 +00:00
cond = torch . randn ( 2 , 256 , 400 )
ts = torch . LongTensor ( [ 600 , 600 ] )
model = TransformerDiffusionWithARPrior ( model_channels = 2048 , block_channels = 1024 , prenet_channels = 1024 ,
2022-06-10 15:24:41 +00:00
input_vec_dim = 512 , num_layers = 16 , prenet_layers = 6 , freeze_diff = True ,
unconditioned_percentage = .4 )
2022-06-08 15:26:59 +00:00
model . get_grad_norm_parameter_groups ( )
ar_weights = torch . load ( ' D: \\ dlas \\ experiments \\ train_music_gpt \\ models \\ 44500_generator_ema.pth ' )
model . ar . load_state_dict ( ar_weights , strict = True )
diff_weights = torch . load ( ' X: \\ dlas \\ experiments \\ train_music_diffusion_tfd8 \\ models \\ 47500_generator_ema.pth ' )
pruned_diff_weights = { }
for k , v in diff_weights . items ( ) :
if k . startswith ( ' diff. ' ) :
pruned_diff_weights [ k . replace ( ' diff. ' , ' ' ) ] = v
model . diff . load_state_dict ( pruned_diff_weights , strict = False )
torch . save ( model . state_dict ( ) , ' sample.pth ' )
model ( clip , ts , cond , conditioning_input = cond )
if __name__ == ' __main__ ' :
2022-06-10 16:40:56 +00:00
test_quant_model ( )