This website requires JavaScript.
Explore
Help
Register
Sign In
mrq
/
vall-e
Watch
5
Star
9
Fork
0
You've already forked vall-e
Code
Issues
8
Pull Requests
Packages
Projects
Releases
Wiki
Activity
b6131565ad
vall-e
/
vall_e
/
models
History
mrq
0d5d545a40
crammed in DAdaptation (doesn't seem worth it) and ScheduleFree (forgot I wanted to weeks ago, seems promising), optimization wrapper cleanup, test trainer changes, etc.
2024-05-09 20:28:20 -05:00
..
__init__.py
deprecate sole AR/NAR model by only keeping the AR+NAR (the beauty of no one using this is that I can break compat as much as I want), add tone token for when I classify my dataset with tone/emotion in the future, some other things
2024-04-15 19:54:32 -05:00
adaln.py
Tweaks
2023-08-02 22:06:39 +00:00
ar_nar.py
crammed in DAdaptation (doesn't seem worth it) and ScheduleFree (forgot I wanted to weeks ago, seems promising), optimization wrapper cleanup, test trainer changes, etc.
2024-05-09 20:28:20 -05:00
base.py
crammed in DAdaptation (doesn't seem worth it) and ScheduleFree (forgot I wanted to weeks ago, seems promising), optimization wrapper cleanup, test trainer changes, etc.
2024-05-09 20:28:20 -05:00
retnet_hf.py
added FP8 support through
NVIDIA/TransformerEngine
, added RetNet_HF through
syncdoth/RetNet
(as an alternative to branch away from torchscale)
2024-04-08 20:14:51 -05:00
retnet_ts.py
backwards compat for old YAMLs with
models
, option to set flash attention 2 for Llama (and derivatives), included
syncdoth/RetNet
s torchscale retnet for shits and grins, etc.
2024-04-16 10:02:31 -05:00
retnet.py
added FP8 support through
NVIDIA/TransformerEngine
, added RetNet_HF through
syncdoth/RetNet
(as an alternative to branch away from torchscale)
2024-04-08 20:14:51 -05:00
transformer.py
Added cfg.bitsandbytes.replace as a less intrusive alternative to cfg.bitsandbytes.inject to replace all Linear modules in a model
2024-03-01 19:20:10 -06:00