From da4f92681e5f130594f04a5a1f3812d4f761b557 Mon Sep 17 00:00:00 2001 From: mrq Date: Thu, 16 Mar 2023 04:35:12 +0000 Subject: [PATCH] oops --- models/tokenizers/ipa.json | 3 ++- modules/tortoise-tts | 2 +- src/utils.py | 2 +- 3 files changed, 4 insertions(+), 3 deletions(-) diff --git a/models/tokenizers/ipa.json b/models/tokenizers/ipa.json index 4a8d78d..498c860 100755 --- a/models/tokenizers/ipa.json +++ b/models/tokenizers/ipa.json @@ -105,7 +105,8 @@ "ʊ": 56, "ʌ": 57, "ʒ": 58, - "θ": 59 + "θ": 59, + "ː": 60 }, "merges": [ diff --git a/modules/tortoise-tts b/modules/tortoise-tts index 9961869..1f674a4 160000 --- a/modules/tortoise-tts +++ b/modules/tortoise-tts @@ -1 +1 @@ -Subproject commit 99618694db4cd7b77e68b62753bb8e2418ac0d55 +Subproject commit 1f674a468f4202ac47feb8fb3587dc5837f2af2b diff --git a/src/utils.py b/src/utils.py index f5fcf5a..1cded68 100755 --- a/src/utils.py +++ b/src/utils.py @@ -1777,7 +1777,7 @@ def tokenize_text( text ): load_tts() encoded = tts.tokenizer.encode(text) - decoded = tts.tokenizer.tokenizer.decode(encoded, skip_special_tokens=False) + decoded = tts.tokenizer.tokenizer.decode(encoded, skip_special_tokens=False).replace(" ", "") return "\n".join([ str(encoded), decoded ])