vall-e/vall_e
2024-12-04 09:30:29 -06:00
..
emb fixes 2024-11-10 20:37:50 -06:00
engines m 2024-11-21 15:07:46 -06:00
models NAR-len tweaks (remasks a small amount of tokens per step, it seems to help with reducing the number of steps needed some of the time?, disable CFG for the first half to speed things up) 2024-12-04 09:30:29 -06:00
utils fixed training tqdm being stubborn 2024-11-23 09:45:23 -06:00
__init__.py
__main__.py moved stuff in the web UI around (un-experimented the max NAR-len steps because its kind of important to adjust this value for better sounding audio / quicker generated audio) 2024-11-20 20:37:33 -06:00
config.py huge oversight in the attention masking......... (i realized I have not been providing a non-causal mask to non-causal tasks) 2024-11-22 13:44:43 -06:00
data.py fixed training tqdm being stubborn 2024-11-23 09:45:23 -06:00
demo.py touch ups in docs 2024-12-02 19:10:42 -06:00
export.py cringe code to convert to LlamaForCausalLM-happy weights + tokenizer dict (still need to write logic to actually use these weights for proper inferencing) 2024-12-03 10:18:58 -06:00
inference.py added more harvard sentences to load from a text file 2024-11-21 13:18:11 -06:00
plot.py very, very naive layerskip speculative sampling (it just checks if the current layer's state is good enough) 2024-11-02 11:49:05 -05:00
samplers.py cleaned up classifier-free guidance logit processing (in order to try and cope with a bad nar-len model) 2024-11-19 10:30:05 -06:00
train.py default set cfg strength to 3.0 since the reference model is updated 2024-11-17 10:23:40 -06:00
webui.py cleanup 2024-11-21 23:08:43 -06:00