Add persistent_workers options in DataLoader to make training faster by removing pauses between epochs. #42
Labels
No Label
bug
duplicate
enhancement
help wanted
insufficient info
invalid
news
not a bug
question
wontfix
No Milestone
No project
No Assignees
2 Participants
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: mrq/ai-voice-cloning#42
Loading…
Reference in New Issue
Block a user
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
The slight issue i've had in training is pretty substantial delay between epochs during training. And GPU basically doing nothing in-between.
Basically i've had the same issue with LORA training in stable diffusion using kohya sd-scripts: https://github.com/kohya-ss/sd-scripts
Then that problem got solved by adding one parameter to a simple line:
https://github.com/kohya-ss/sd-scripts/pull/140
So, by adding
to line 21 in
dlas\codes\data\__init__.py
resulting in
I tried applying the same thing here, and i think it solves the problem, accelerating the training. With large batch sizes this could potentially accelerate the training by several times. (x4 times with my test)
Might need someone to test it out on their machines.
Testing it now on a paperspace instance. It hasn't outright died yet so I suppose it's safe to push out.
Added in DLAS commit
71cc43e65c
.