This website requires JavaScript.
Explore
Help
Register
Sign In
mrq
/
vall-e
Watch
5
Star
9
Fork
0
You've already forked vall-e
Code
Issues
8
Pull Requests
Packages
Projects
Releases
Wiki
Activity
d69a00e389
vall-e
/
vall_e
/
models
History
mrq
d69a00e389
Properly pass retention_mask for retnet-HF, attempt to fix recurrent forward for retnet (doesn't work still)
2024-04-14 13:12:50 -05:00
..
__init__.py
added FP8 support through
NVIDIA/TransformerEngine
, added RetNet_HF through
syncdoth/RetNet
(as an alternative to branch away from torchscale)
2024-04-08 20:14:51 -05:00
adaln.py
ar_nar.py
Properly pass retention_mask for retnet-HF, attempt to fix recurrent forward for retnet (doesn't work still)
2024-04-14 13:12:50 -05:00
ar.py
base.py
Properly pass retention_mask for retnet-HF, attempt to fix recurrent forward for retnet (doesn't work still)
2024-04-14 13:12:50 -05:00
nar.py
retnet_hf.py
added FP8 support through
NVIDIA/TransformerEngine
, added RetNet_HF through
syncdoth/RetNet
(as an alternative to branch away from torchscale)
2024-04-08 20:14:51 -05:00
retnet.py
added FP8 support through
NVIDIA/TransformerEngine
, added RetNet_HF through
syncdoth/RetNet
(as an alternative to branch away from torchscale)
2024-04-08 20:14:51 -05:00
transformer.py
Added cfg.bitsandbytes.replace as a less intrusive alternative to cfg.bitsandbytes.inject to replace all Linear modules in a model
2024-03-01 19:20:10 -06:00