From 9e1989be1beb521cffc443ef2ea28504cb2a5938 Mon Sep 17 00:00:00 2001 From: mrq Date: Sat, 3 Aug 2024 09:01:37 -0500 Subject: [PATCH] tweaked initial NAR pass's initial token embeddings to use a different value, or osmething --- vall_e/models/base.py | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/vall_e/models/base.py b/vall_e/models/base.py index 2907cf0..b1c0fd5 100755 --- a/vall_e/models/base.py +++ b/vall_e/models/base.py @@ -1024,11 +1024,15 @@ class Base(nn.Module): embedding = self.tones_emb( input ) elif name == "resp": if "len" in self.capabilities and quant_level == 0: + """ # fill with "stop" tokens for NAR-only model embedding = self.resps_emb( torch.full_like(input if input.dim() == 1 else input[..., 0], self.stop_token), offset = 0 ) + """ + # fill with filler tokens for NAR-only model + embedding = self.dropout_token.repeat((input.shape[0], 1)) else: # get RVQ level 0, or up to targetted RVQ level inference if self.version <= 4: