vall-e/vall_e/models/arch
2024-11-03 19:19:15 -06:00
..
attention ugh 2024-08-30 14:39:07 -05:00
mamba_vasqu
retnet_syncdoth
__init__.py layer skip training implemented (need to gut the inferencing from the repo, and to actually see if the model can benefit from this) 2024-10-30 20:05:45 -05:00
bitnet.py
llama.py Windows specific fixes (to-do: find libespeak-ng.dll automatically because it cannot be trusted to do it by default) 2024-11-03 19:19:15 -06:00
mamba.py
mixtral.py fixed attentions for MoE 2024-08-27 17:02:42 -05:00
retnet.py
transformer.py