vall-e/vall_e
2024-06-01 10:44:32 -05:00
..
emb
engines report current loss scale and adjust grad norm by loss scale (for deepspeed) 2024-06-01 10:44:32 -05:00
ext
models added model config option to set KV head count for MQA/GQA instead of MHA for llama-based models (i think its very negligible both ways on such a small model size) 2024-05-31 19:32:37 -05:00
utils ugh 2024-06-01 10:30:13 -05:00
__init__.py
__main__.py
config.py actually don't default to compute split losses, test bitnet model doesn't seem to be doing things right (despite debug printouts showing theyre roughly the same logit/loss sequences, could just be bitnet linears being not up to par on actual models) 2024-06-01 09:12:51 -05:00
data.py ugh 2024-06-01 10:30:13 -05:00
export.py
inference.py
plot.py
samplers.py
train.py nevermind it breaks training 2024-05-25 18:03:43 -05:00
webui.py