1
1
forked from mrq/tortoise-tts
Commit Graph

13 Commits

Author SHA1 Message Date
Johan Nordberg
491fe7f6d3 Remove some assumptions about working directory
This allows cli tool to run when not standing in repository dir
2022-05-29 01:10:19 +00:00
Johan Nordberg
a641d8f29b Add tortoise_cli.py 2022-05-28 05:25:23 +00:00
Johan Nordberg
b4fa8c86b9 Allow passing additional voice directories when loading voices 2022-05-19 21:02:11 +09:00
Danila Berezin
ef5fb5f5fc Fix bug in load_voices in audio.py
The read.py script did not work with pth latents, so I fix bug in audio.py. It seems that in the elif statement, instead of voice, voices should be clip, clips. And torch stack doesn't work with tuples, so I had to split this operation.
2022-05-17 18:34:54 +03:00
James Betker
f5ebd14d09 Add error message 2022-05-12 20:15:40 -06:00
James Betker
e18428166d v2.2 2022-05-06 00:11:10 -06:00
James Betker
e23e6f6696 Use librosa for loading mp3s 2022-05-03 20:44:31 -06:00
James Betker
00e84bbd86 fix paths 2022-05-02 20:56:28 -06:00
James Betker
5663e98904 misc fixes 2022-05-02 18:00:57 -06:00
James Betker
ccf16f978e more fixes 2022-05-02 16:44:47 -06:00
James Betker
ee24d3ee4b Support totally random voices (and make fixes to previous changes) 2022-05-02 15:40:03 -06:00
James Betker
01b783fc02 Add support for extracting and feeding conditioning latents directly into the model
- Adds a new script and API endpoints for doing this
- Reworks autoregressive and diffusion models so that the conditioning is computed separately (which will actually provide a mild performance boost)
- Updates README

This is untested. Need to do the following manual tests (and someday write unit tests for this behemoth before
it becomes a problem..)
1) Does get_conditioning_latents.py work?
2) Can I feed those latents back into the model by creating a new voice?
3) Can I still mix and match voices (both with conditioning latents and normal voices) with read.py?
2022-05-01 17:25:18 -06:00
James Betker
23a3d5d00b Move everything into the tortoise/ subdirectory
For eventual packaging.
2022-05-01 16:24:24 -06:00