diff --git a/.compute b/.compute index 71c4d0c1..c4bf43c9 100644 --- a/.compute +++ b/.compute @@ -3,5 +3,5 @@ ls ${SHARED_DIR}/data/ pip3 install https://download.pytorch.org/whl/cu100/torch-1.0.1.post2-cp36-cp36m-linux_x86_64.whl yes | apt-get install espeak python3 setup.py develop -python3 distribute.py --config_path config_cluster.json --data_path ${SHARED_DIR}/data/Blizzard/Nancy/ --restore_path ${USER_DIR}/best_model.pth.tar -# python3 distribute.py --config_path config_cluster.json --data_path ${SHARED_DIR}/data/Blizzard/Nancy/ +# python3 distribute.py --config_path config_cluster.json --data_path ${SHARED_DIR}/data/Blizzard/Nancy/ --restore_path ${USER_DIR}/best_model.pth.tar +python3 distribute.py --config_path config_cluster.json --data_path ${SHARED_DIR}/data/Blizzard/Nancy/ diff --git a/config_cluster.json b/config_cluster.json index e12111be..d707b929 100644 --- a/config_cluster.json +++ b/config_cluster.json @@ -46,7 +46,7 @@ "save_step": 1000, // Number of training steps expected to save traning stats and checkpoints. "print_step": 10, // Number of steps to log traning on console. "tb_model_param_stats": false, // true, plots param stats per layer on tensorboard. Might be memory consuming, but good for debugging. - "batch_group_size": 8, //Number of batches to shuffle after bucketing. + "batch_group_size": 4, //Number of batches to shuffle after bucketing. "run_eval": true, "test_delay_epochs": 100, //Until attention is aligned, testing only wastes computation time.