James Betker
|
aca2c7ab41
|
Full checkpoint-ize SSG1
|
2020-10-04 18:24:52 -06:00 |
|
James Betker
|
e3294939b0
|
Revert "SSG: offer option to use BN-based attention normalization"
Didn't work. Oh well.
This reverts commit 5cd2b37591 .
|
2020-10-03 17:54:53 -06:00 |
|
James Betker
|
5cd2b37591
|
SSG: offer option to use BN-based attention normalization
Not sure how this is going to work, lets try it.
|
2020-10-03 16:16:19 -06:00 |
|
James Betker
|
19a4075e1e
|
Allow checkpointing to be disabled in the options file
Also makes options a global variable for usage in utils.
|
2020-10-03 11:03:28 -06:00 |
|
James Betker
|
d9ae970fd9
|
SSG update
|
2020-10-01 11:27:51 -06:00 |
|
James Betker
|
f40beb5460
|
Add 'before' and 'after' defs to injections, steps and optimizers
|
2020-09-22 17:03:22 -06:00 |
|
James Betker
|
53a5657850
|
Fix SSGR
|
2020-09-20 19:07:15 -06:00 |
|
James Betker
|
fe82785ba5
|
Add some new architectures to ssg
|
2020-09-19 21:47:10 -06:00 |
|
James Betker
|
9a17ade550
|
Some convenience adjustments to ExtensibleTrainer
|
2020-09-17 21:05:32 -06:00 |
|
James Betker
|
723754c133
|
Update attention debugger outputting for SSG
|
2020-09-16 13:09:46 -06:00 |
|
James Betker
|
0918430572
|
SSG network
This branches off of SPSR. It is identical but substantially reduced
in complexity. It's intended to be my long term working arch.
|
2020-09-15 20:59:24 -06:00 |
|