James Betker
|
e1052a5e32
|
Move log consensus to train for efficiency
|
2022-03-04 13:41:32 -07:00 |
|
James Betker
|
ce6dfdf255
|
Distributed "fixes"
|
2022-03-04 12:46:41 -07:00 |
|
James Betker
|
3ff878ae85
|
Accumulate loss & grad_norm metrics from all entities within a distributed graph
|
2022-03-04 12:01:16 -07:00 |
|
James Betker
|
79e5692388
|
Fix distributed bug
|
2022-03-04 11:58:53 -07:00 |
|
James Betker
|
f87e10ffef
|
Make deterministic sampler work with distributed training & microbatches
|
2022-03-04 11:50:50 -07:00 |
|
James Betker
|
77c18b53b3
|
Cap grad booster
|
2022-03-04 10:40:24 -07:00 |
|
James Betker
|
2d1cb83c1d
|
Add a deterministic timestep sampler, with provisions to employ it every n steps
|
2022-03-04 10:40:14 -07:00 |
|
James Betker
|
f490eaeba7
|
Shuffle optimizer states back and forth between cpu memory during steps
|
2022-03-04 10:38:51 -07:00 |
|
James Betker
|
3c242403f5
|
adjust location of pre-optimizer step so I can visualize the new grad norms
|
2022-03-04 08:56:42 -07:00 |
|
James Betker
|
58019a2ce3
|
audio diffusion fid updates
|
2022-03-03 21:53:32 -07:00 |
|
James Betker
|
998c53ad4f
|
w2v_matcher mods
|
2022-03-03 21:52:51 -07:00 |
|
James Betker
|
9029e4f20c
|
Add a base-wrapper
|
2022-03-03 21:52:28 -07:00 |
|
James Betker
|
6873ad6660
|
Support functionality
|
2022-03-03 21:52:16 -07:00 |
|
James Betker
|
6af5d129ce
|
Add experimental gradient boosting into tts7
|
2022-03-03 21:51:40 -07:00 |
|
James Betker
|
7ea84f1ac3
|
asdf
|
2022-03-03 13:43:44 -07:00 |
|
James Betker
|
3cd6c7f428
|
Get rid of unused codes in vq
|
2022-03-03 13:41:38 -07:00 |
|
James Betker
|
619da9ea28
|
Get rid of discretization loss
|
2022-03-03 13:36:25 -07:00 |
|
James Betker
|
beb7c8a39d
|
asdf
|
2022-03-01 21:41:31 -07:00 |
|
James Betker
|
70fa780edb
|
Add mechanism to export grad norms
|
2022-03-01 20:19:52 -07:00 |
|
James Betker
|
d9f8f92840
|
Codified fp16
|
2022-03-01 15:46:04 -07:00 |
|
James Betker
|
45ab444c04
|
Rework minicoder to always checkpoint
|
2022-03-01 14:09:18 -07:00 |
|
James Betker
|
db0c3340ac
|
Implement guidance-free diffusion in eval
And a few other fixes
|
2022-03-01 11:49:36 -07:00 |
|
James Betker
|
2134f06516
|
Implement conditioning-free diffusion at the eval level
|
2022-02-27 15:11:42 -07:00 |
|
James Betker
|
436fe24822
|
Add conditioning-free guidance
|
2022-02-27 15:00:06 -07:00 |
|
James Betker
|
ac920798bb
|
misc
|
2022-02-27 14:49:11 -07:00 |
|
James Betker
|
ba155e4e2f
|
script for uploading models to the HF hub
|
2022-02-27 14:48:38 -07:00 |
|
James Betker
|
dbc74e96b2
|
w2v_matcher
|
2022-02-27 14:48:23 -07:00 |
|
James Betker
|
42879d7296
|
w2v_wrapper ramping dropout mode
this is an experimental feature that needs some testing
|
2022-02-27 14:47:51 -07:00 |
|
James Betker
|
c375287db9
|
Re-instate autocasting
|
2022-02-25 11:06:18 -07:00 |
|
James Betker
|
34ee32a90e
|
get rid of autocasting in tts7
|
2022-02-24 21:53:51 -07:00 |
|
James Betker
|
f458f5d8f1
|
abort early if losses reach nan too much, and save the model
|
2022-02-24 20:55:30 -07:00 |
|
James Betker
|
18dc62453f
|
Don't step if NaN losses are encountered.
|
2022-02-24 17:45:08 -07:00 |
|
James Betker
|
ea500ad42a
|
Use clustered masking in udtts7
|
2022-02-24 07:57:26 -07:00 |
|
James Betker
|
7c17c8e674
|
gurgl
|
2022-02-23 21:28:24 -07:00 |
|
James Betker
|
e6824e398f
|
Load dvae to cpu
|
2022-02-23 21:21:45 -07:00 |
|
James Betker
|
81017d9696
|
put frechet_distance on cuda
|
2022-02-23 21:21:13 -07:00 |
|
James Betker
|
9a7bbf33df
|
f
|
2022-02-23 18:03:38 -07:00 |
|
James Betker
|
68726eac74
|
.
|
2022-02-23 17:58:07 -07:00 |
|
James Betker
|
b7319ab518
|
Support vocoder type diffusion in audio_diffusion_fid
|
2022-02-23 17:25:16 -07:00 |
|
James Betker
|
58f6c9805b
|
adf
|
2022-02-22 23:12:58 -07:00 |
|
James Betker
|
03752c1cd6
|
Report NaN
|
2022-02-22 23:09:37 -07:00 |
|
James Betker
|
7201b4500c
|
default text_to_sequence cleaners
|
2022-02-21 19:14:22 -07:00 |
|
James Betker
|
ba7f54c162
|
w2v: new inference function
|
2022-02-21 19:13:03 -07:00 |
|
James Betker
|
896ac029ae
|
allow continuation of samples encountered
|
2022-02-21 19:12:50 -07:00 |
|
James Betker
|
6313a94f96
|
eval: integrate a n-gram language model into decoding
|
2022-02-21 19:12:34 -07:00 |
|
James Betker
|
af50afe222
|
pairedvoice: error out if clip is too short
|
2022-02-21 19:11:10 -07:00 |
|
James Betker
|
38802a96c8
|
remove timesteps from cond calculation
|
2022-02-21 12:32:21 -07:00 |
|
James Betker
|
668876799d
|
unet_diffusion_tts7
|
2022-02-20 15:22:38 -07:00 |
|
James Betker
|
0872e17e60
|
unified_voice mods
|
2022-02-19 20:37:35 -07:00 |
|
James Betker
|
7b12799370
|
Reformat mel_text_clip for use in eval
|
2022-02-19 20:37:26 -07:00 |
|