forked from mrq/DL-Art-School
Disabling bitsandbytes optimization as default for now, in the off chance that it actually produces garbage (which shouldn't happen, there's no chance, if training at float16 from a model at float16 works fine, then this has to work)
This commit is contained in:
parent
918473807f
commit
58600274ac
|
@ -1,3 +1,14 @@
|
||||||
|
# (QoL improvements for) Deep Learning Art School
|
||||||
|
|
||||||
|
This fork of [neonbjb/DL-Art-School](https://github.com/neonbjb/DL-Art-School/) contains a few fixes and QoL improvements, including but not limited to:
|
||||||
|
* sanity tidying, like:
|
||||||
|
- not outputing to `./DL-Art-School/experiments/`
|
||||||
|
- the custom module loader for networks/injectors getting fixed
|
||||||
|
- BitsAndBytes integration:
|
||||||
|
+ working but output untested: Adam/AdamW
|
||||||
|
+ toggles available in `./codes/torch_indermediary/__init__.py`
|
||||||
|
|
||||||
|
---
|
||||||
# Deep Learning Art School
|
# Deep Learning Art School
|
||||||
|
|
||||||
Send your Pytorch model to art class!
|
Send your Pytorch model to art class!
|
|
@ -13,8 +13,8 @@ from torch.optim.adamw import AdamW
|
||||||
|
|
||||||
OVERRIDE_LINEAR = False
|
OVERRIDE_LINEAR = False
|
||||||
OVERRIDE_EMBEDDING = False
|
OVERRIDE_EMBEDDING = False
|
||||||
OVERRIDE_ADAM = True
|
OVERRIDE_ADAM = False # True
|
||||||
OVERRIDE_ADAMW = True
|
OVERRIDE_ADAMW = False # True
|
||||||
USE_STABLE_EMBEDDING = True
|
USE_STABLE_EMBEDDING = True
|
||||||
|
|
||||||
if OVERRIDE_LINEAR:
|
if OVERRIDE_LINEAR:
|
||||||
|
|
|
@ -258,7 +258,7 @@ class Trainer:
|
||||||
self.logger.info(message)
|
self.logger.info(message)
|
||||||
|
|
||||||
#### save models and training states
|
#### save models and training states
|
||||||
if self.current_step % opt['logger']['save_checkpoint_freq'] == 0:
|
if self.current_step > 0 and self.current_step % opt['logger']['save_checkpoint_freq'] == 0:
|
||||||
self.model.consolidate_state()
|
self.model.consolidate_state()
|
||||||
if self.rank <= 0:
|
if self.rank <= 0:
|
||||||
self.logger.info('Saving models and training states.')
|
self.logger.info('Saving models and training states.')
|
||||||
|
|
Loading…
Reference in New Issue
Block a user