DL-Art-School/codes
James Betker fba29d7dcc Move to apex distributeddataparallel and add switch all_reduce
Torch's distributed_data_parallel is missing "delay_allreduce", which is
necessary to get gradient checkpointing to work with recurrent models.
2020-10-08 11:20:05 -06:00
..
.idea
data
data_scripts
metrics
models Move to apex distributeddataparallel and add switch all_reduce 2020-10-08 11:20:05 -06:00
options
scripts
switched_conv@004dda04e3
temp
utils
process_video.py
recover_tensorboard_log.py
requirements.txt
run_scripts.sh
test.py
train2.py
train.py