From a41ec4668bce5348bfec6a3d97cf9a0766cc1e1c Mon Sep 17 00:00:00 2001 From: lizhenyu Date: Wed, 16 Dec 2020 16:48:40 +0800 Subject: [PATCH] refine wide and deep model zoo log file --- .../script/run_parameter_server_train_cluster.sh | 6 +++--- .../script/run_parameter_server_train_distribute.sh | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_cluster.sh b/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_cluster.sh index 0015bf470d..de312f6f98 100644 --- a/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_cluster.sh +++ b/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_cluster.sh @@ -70,9 +70,9 @@ if [[ "$MS_ROLE" == "MS_WORKER" ]]; then rm -rf ${execute_path}/worker/ mkdir ${execute_path}/worker/ cd ${execute_path}/worker/ || exit - mpirun --allow-run-as-root -n $LOCAL_WORKER_NUM \ - python -s ${self_path}/../train_and_eval_parameter_server_distribute.py \ - --device_target=$DEVICE --data_path=$DATASET --epochs=$EPOCH_SIZE --parameter_server=1 \ + mpirun --allow-run-as-root -n $LOCAL_WORKER_NUM --output-filename log_output --merge-stderr-to-stdout \ + python -s ${self_path}/../train_and_eval_parameter_server_distribute.py \ + --device_target=$DEVICE --data_path=$DATASET --epochs=$EPOCH_SIZE --parameter_server=1 \ --vocab_cache_size=$VOCAB_CACHE_SIZE --dropout_flag=1 >worker.log 2>&1 & else for((i=0;i<$LOCAL_WORKER_NUM;i++)); diff --git a/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_distribute.sh b/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_distribute.sh index 10186751f3..b2cf337753 100644 --- a/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_distribute.sh +++ b/model_zoo/official/recommend/wide_and_deep/script/run_parameter_server_train_distribute.sh @@ -62,7 +62,7 @@ if [[ "X$DEVICE_TARGET" == "XGPU" ]]; then rm -rf ${execute_path}/worker/ mkdir ${execute_path}/worker/ cd ${execute_path}/worker/ || exit - mpirun --allow-run-as-root -n $RANK_SIZE \ + mpirun --allow-run-as-root -n $RANK_SIZE --output-filename log_output --merge-stderr-to-stdout \ python -s ${self_path}/../train_and_eval_parameter_server_distribute.py \ --device_target=$DEVICE_TARGET --data_path=$DATASET --epochs=$EPOCH_SIZE --parameter_server=1 \ --vocab_cache_size=$VOCAB_CACHE_SIZE --dropout_flag=1 >worker.log 2>&1 &