forked from mrq/ai-voice-cloning
fixed user inputted LR schedule not actually getting used (oops)
This commit is contained in:
parent
6d8c2dd459
commit
6d5e1e1a80
|
@ -935,8 +935,8 @@ def calc_iterations( epochs, lines, batch_size ):
|
|||
return iterations
|
||||
|
||||
EPOCH_SCHEDULE = [ 9, 18, 25, 33 ]
|
||||
def schedule_learning_rate( iterations ):
|
||||
return [int(iterations * d) for d in EPOCH_SCHEDULE]
|
||||
def schedule_learning_rate( iterations, schedule=EPOCH_SCHEDULE ):
|
||||
return [int(iterations * d) for d in schedule]
|
||||
|
||||
def optimize_training_settings( epochs, learning_rate, text_ce_lr_weight, learning_rate_schedule, batch_size, mega_batch_factor, print_rate, save_rate, resume_path, half_p, bnb, source_model, voice ):
|
||||
name = f"{voice}-finetune"
|
||||
|
|
|
@ -307,7 +307,10 @@ def save_training_settings_proxy( epochs, learning_rate, text_ce_lr_weight, lear
|
|||
|
||||
if not learning_rate_schedule:
|
||||
learning_rate_schedule = EPOCH_SCHEDULE
|
||||
learning_rate_schedule = schedule_learning_rate( iterations / epochs )
|
||||
elif isinstance(learning_rate_schedule,str):
|
||||
learning_rate_schedule = json.loads(learning_rate_schedule)
|
||||
|
||||
learning_rate_schedule = schedule_learning_rate( iterations / epochs, learning_rate_schedule )
|
||||
|
||||
messages.append(save_training_settings(
|
||||
iterations=iterations,
|
||||
|
|
Loading…
Reference in New Issue
Block a user