diff --git a/tortoise/read.py b/tortoise/read.py index 75da75f..05b6658 100644 --- a/tortoise/read.py +++ b/tortoise/read.py @@ -81,13 +81,13 @@ if __name__ == '__main__': dbg_state = (seed, texts, voice_samples, conditioning_latents) torch.save(dbg_state, f'debug_states/read_debug_{selected_voice}.pth') - # Combine each candidate's audio clips. - if args.candidates > 1: - audio_clips = [] - for candidate in range(args.candidates): - for line in range(len(texts)): - wav_file = os.path.join(voice_outpath, str(line), f"{candidate}.wav") - audio_clips.append(load_audio(wav_file, 24000)) - audio_clips = torch.cat(audio_clips, dim=-1) - torchaudio.save(os.path.join(voice_outpath, f"combined_{candidate:02d}.wav"), audio_clips, 24000) - audio_clips = [] \ No newline at end of file + # Combine each candidate's audio clips. + if args.candidates > 1: + audio_clips = [] + for candidate in range(args.candidates): + for line in range(len(texts)): + wav_file = os.path.join(voice_outpath, str(line), f"{candidate}.wav") + audio_clips.append(load_audio(wav_file, 24000)) + audio_clips = torch.cat(audio_clips, dim=-1) + torchaudio.save(os.path.join(voice_outpath, f"combined_{candidate:02d}.wav"), audio_clips, 24000) + audio_clips = []