vall-e/vall_e
2024-08-05 20:34:58 -05:00
..
emb re-adapted process_libritts.py to a 'better' way (better because it processed without needing to shuffle a bunch of things and adapt to cope or something) 2024-08-05 20:34:58 -05:00
engines fix issue with sft and shared tensors... 2024-08-04 19:56:21 -05:00
ext fixes 2024-06-04 00:07:00 -05:00
models updated process_datasets.py, added argparsing so I can mostly stop manually editing things, and some other cleanup 2024-08-05 15:59:25 -05:00
utils fix issue with sft and shared tensors... 2024-08-04 19:56:21 -05:00
__init__.py Rewrite init 2023-08-02 21:53:35 +00:00
__main__.py added option to set the causal size (how many tokens to sample per AR step), but requires the model to be trained for this (which explains why recurrent chunk sampling just doesn't work for the retnet tests, obvious in hindsight) 2024-07-30 20:53:51 -05:00
config.py implicitly load either normal pickled weights or safetensors on loading the model 2024-08-03 23:34:18 -05:00
data.py added safetensors support (with metadata) and feed whatever torch.load/torch.save into it 2024-08-03 23:15:20 -05:00
demo.py fix issue with sft and shared tensors... 2024-08-04 19:56:21 -05:00
export.py added export option to convert Llama to MixtralMoE for another dumb experiment 2024-08-04 20:25:06 -05:00
inference.py fix weird regression in handling checkpoints when backend is local, but deepspeed checkpoints are in (it was handled with LoRA loading but not real loading...) 2024-07-30 22:15:56 -05:00
plot.py ugh 2024-06-09 11:39:43 -05:00
samplers.py possible speedup for samplers that require a list of previous tokens (the DRY sampler made me realize that I should copy the tolist() thing from the rep pen sampler for everything else) 2024-07-29 20:23:26 -05:00
train.py add cap for NAR-len training, to avoid any weird cases in early training where it'll just mess up and generate long lengths 2024-08-03 21:00:32 -05:00
webui.py fix issue with sft and shared tensors... 2024-08-04 19:56:21 -05:00