100x better for subtrain/eval to be by group instead
This commit is contained in:
parent
458b95d196
commit
e3ef89f5aa
|
@ -211,7 +211,7 @@ class Dataset(_Dataset):
|
||||||
self.training = training
|
self.training = training
|
||||||
self.dataset_type = "training" if self.training else "validation"
|
self.dataset_type = "training" if self.training else "validation"
|
||||||
self.dataset = cfg.dataset.training if self.training else cfg.dataset.validation
|
self.dataset = cfg.dataset.training if self.training else cfg.dataset.validation
|
||||||
self.sampler_type = cfg.dataset.sample_type if self.dataset_type == "training" else "path"
|
self.sampler_type = cfg.dataset.sample_type if self.dataset_type == "training" else "group"
|
||||||
|
|
||||||
# to-do: do not do validation if there's nothing in the validation
|
# to-do: do not do validation if there's nothing in the validation
|
||||||
# this just makes it be happy
|
# this just makes it be happy
|
||||||
|
|
|
@ -319,6 +319,8 @@ class AR_NAR(Base):
|
||||||
def example_usage():
|
def example_usage():
|
||||||
#cfg.trainer.backend = "local"
|
#cfg.trainer.backend = "local"
|
||||||
cfg.hyperparameters.gradient_accumulation_steps = 1
|
cfg.hyperparameters.gradient_accumulation_steps = 1
|
||||||
|
if cfg.inference.audio_backend == "dac":
|
||||||
|
cfg.sample_rate = 44_000
|
||||||
|
|
||||||
from functools import partial
|
from functools import partial
|
||||||
from einops import repeat
|
from einops import repeat
|
||||||
|
@ -371,7 +373,7 @@ def example_usage():
|
||||||
'n_layers': 8, # 32
|
'n_layers': 8, # 32
|
||||||
'n_experts': 1,
|
'n_experts': 1,
|
||||||
|
|
||||||
'p_dropout': 0.0,
|
'p_dropout': 0.1,
|
||||||
|
|
||||||
'l_padding': 8 if cfg.optimizations.fp8 else 0,
|
'l_padding': 8 if cfg.optimizations.fp8 else 0,
|
||||||
|
|
||||||
|
@ -386,7 +388,7 @@ def example_usage():
|
||||||
"""
|
"""
|
||||||
|
|
||||||
model = AR_NAR(**kwargs).to(device)
|
model = AR_NAR(**kwargs).to(device)
|
||||||
steps = 1000
|
steps = 200
|
||||||
|
|
||||||
optimizer = cfg.hyperparameters.optimizer.lower() if cfg.cfg_path is not None else "prodigy"
|
optimizer = cfg.hyperparameters.optimizer.lower() if cfg.cfg_path is not None else "prodigy"
|
||||||
scheduler = cfg.hyperparameters.scheduler.lower() if cfg.cfg_path is not None else ""
|
scheduler = cfg.hyperparameters.scheduler.lower() if cfg.cfg_path is not None else ""
|
||||||
|
|
Loading…
Reference in New Issue
Block a user