|
|
cd ../../../.. |
|
|
|
|
|
|
|
|
|
|
|
DEVICES=0,1,2,3,4,5,6,7, |
|
|
|
|
|
RUN_FILE='./tools/dist_test.sh' |
|
|
PORT=$(( ((RANDOM<<15)|RANDOM) % 63001 + 2000 )) |
|
|
|
|
|
|
|
|
|
|
|
DATASET='coco' |
|
|
MODEL="sapiens_1b-210e_${DATASET}-1024x768" |
|
|
JOB_NAME="test_pose_whole_$MODEL" |
|
|
TEST_BATCH_SIZE_PER_GPU=32 |
|
|
|
|
|
CHECKPOINT="/home/$USER/sapiens_host/pose/checkpoints/sapiens_1b/sapiens_1b_coco_best_coco_AP_821.pth" |
|
|
|
|
|
|
|
|
|
|
|
mode='multi-gpu' |
|
|
|
|
|
|
|
|
CONFIG_FILE=configs/sapiens_pose/${DATASET}/${MODEL}.py |
|
|
OUTPUT_DIR="Outputs/test/${DATASET}/${MODEL}/node" |
|
|
OUTPUT_DIR="$(echo "${OUTPUT_DIR}/$(date +"%m-%d-%Y_%H:%M:%S")")" |
|
|
|
|
|
export TF_CPP_MIN_LOG_LEVEL=2 |
|
|
|
|
|
|
|
|
OPTIONS="$(echo "test_dataloader.batch_size=$TEST_BATCH_SIZE_PER_GPU")" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if [ "$mode" = "debug" ]; then |
|
|
TEST_BATCH_SIZE_PER_GPU=16 |
|
|
|
|
|
OPTIONS="$(echo "test_dataloader.batch_size=${TEST_BATCH_SIZE_PER_GPU} test_dataloader.num_workers=0 test_dataloader.persistent_workers=False")" |
|
|
CUDA_VISIBLE_DEVICES=${DEVICES} python tools/test.py ${CONFIG_FILE} ${CHECKPOINT} --work-dir ${OUTPUT_DIR} --cfg-options ${OPTIONS} |
|
|
|
|
|
elif [ "$mode" = "multi-gpu" ]; then |
|
|
NUM_GPUS_STRING_LEN=${#DEVICES} |
|
|
NUM_GPUS=$((NUM_GPUS_STRING_LEN/2)) |
|
|
|
|
|
LOG_FILE="$(echo "${OUTPUT_DIR}/log.txt")" |
|
|
mkdir -p ${OUTPUT_DIR}; touch ${LOG_FILE} |
|
|
|
|
|
CUDA_VISIBLE_DEVICES=${DEVICES} PORT=${PORT} ${RUN_FILE} ${CONFIG_FILE} ${CHECKPOINT}\ |
|
|
${NUM_GPUS} \ |
|
|
--work-dir ${OUTPUT_DIR} \ |
|
|
--cfg-options ${OPTIONS} \ |
|
|
| tee ${LOG_FILE} |
|
|
|
|
|
fi |
|
|
|