|
|
#!/bin/bash |
|
|
MODELSCOPE_CACHE_DIR_IN_CONTAINER=/modelscope_cache |
|
|
CODE_DIR=$PWD |
|
|
CODE_DIR_IN_CONTAINER=/ms-swift |
|
|
echo "$USER" |
|
|
gpus='0,1 2,3' |
|
|
cpu_sets='0-15 16-31' |
|
|
cpu_sets_arr=($cpu_sets) |
|
|
is_get_file_lock=false |
|
|
CI_COMMAND=${CI_COMMAND:-bash .dev_scripts/ci_container_test.sh python tests/run.py --parallel 2 --run_config tests/run_config.yaml} |
|
|
echo "ci command: $CI_COMMAND" |
|
|
PR_CHANGED_FILES="${PR_CHANGED_FILES:-}" |
|
|
echo "PR modified files: $PR_CHANGED_FILES" |
|
|
PR_CHANGED_FILES=${PR_CHANGED_FILES//[ ]/#} |
|
|
echo "PR_CHANGED_FILES: $PR_CHANGED_FILES" |
|
|
idx=0 |
|
|
for gpu in $gpus |
|
|
do |
|
|
exec {lock_fd}>"/tmp/gpu$gpu" || exit 1 |
|
|
flock -n "$lock_fd" || { echo "WARN: gpu $gpu is in use!" >&2; idx=$((idx+1)); continue; } |
|
|
echo "get gpu lock $gpu" |
|
|
|
|
|
CONTAINER_NAME="swift-ci-$idx" |
|
|
let is_get_file_lock=true |
|
|
|
|
|
|
|
|
docker pull ${IMAGE_NAME}:${IMAGE_VERSION} |
|
|
if [ "$MODELSCOPE_SDK_DEBUG" == "True" ]; then |
|
|
echo 'debugging' |
|
|
docker run --rm --name $CONTAINER_NAME --shm-size=16gb \ |
|
|
--cpuset-cpus=${cpu_sets_arr[$idx]} \ |
|
|
--gpus='"'"device=$gpu"'"' \ |
|
|
-v $CODE_DIR:$CODE_DIR_IN_CONTAINER \ |
|
|
-v $MODELSCOPE_CACHE:$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
|
|
-v $MODELSCOPE_HOME_CACHE/$idx:/root \ |
|
|
-v /home/admin/pre-commit:/home/admin/pre-commit \ |
|
|
-e CI_TEST=True \ |
|
|
-e TEST_LEVEL=$TEST_LEVEL \ |
|
|
-e MODELSCOPE_CACHE=$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
|
|
-e MODELSCOPE_DOMAIN=$MODELSCOPE_DOMAIN \ |
|
|
-e MODELSCOPE_SDK_DEBUG=True \ |
|
|
-e HUB_DATASET_ENDPOINT=$HUB_DATASET_ENDPOINT \ |
|
|
-e TEST_ACCESS_TOKEN_CITEST=$TEST_ACCESS_TOKEN_CITEST \ |
|
|
-e TEST_ACCESS_TOKEN_SDKDEV=$TEST_ACCESS_TOKEN_SDKDEV \ |
|
|
-e TEST_LEVEL=$TEST_LEVEL \ |
|
|
-e MODELSCOPE_ENVIRONMENT='ci' \ |
|
|
-e TEST_UPLOAD_MS_TOKEN=$TEST_UPLOAD_MS_TOKEN \ |
|
|
-e MODEL_TAG_URL=$MODEL_TAG_URL \ |
|
|
-e MODELSCOPE_API_TOKEN=$MODELSCOPE_API_TOKEN \ |
|
|
-e PR_CHANGED_FILES=$PR_CHANGED_FILES \ |
|
|
--workdir=$CODE_DIR_IN_CONTAINER \ |
|
|
${IMAGE_NAME}:${IMAGE_VERSION} \ |
|
|
$CI_COMMAND |
|
|
else |
|
|
docker run --rm --name $CONTAINER_NAME --shm-size=16gb \ |
|
|
--cpuset-cpus=${cpu_sets_arr[$idx]} \ |
|
|
--gpus='"'"device=$gpu"'"' \ |
|
|
-v $CODE_DIR:$CODE_DIR_IN_CONTAINER \ |
|
|
-v $MODELSCOPE_CACHE:$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
|
|
-v $MODELSCOPE_HOME_CACHE/$idx:/root \ |
|
|
-v /home/admin/pre-commit:/home/admin/pre-commit \ |
|
|
-e CI_TEST=True \ |
|
|
-e TEST_LEVEL=$TEST_LEVEL \ |
|
|
-e MODELSCOPE_CACHE=$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
|
|
-e MODELSCOPE_DOMAIN=$MODELSCOPE_DOMAIN \ |
|
|
-e HUB_DATASET_ENDPOINT=$HUB_DATASET_ENDPOINT \ |
|
|
-e TEST_ACCESS_TOKEN_CITEST=$TEST_ACCESS_TOKEN_CITEST \ |
|
|
-e TEST_ACCESS_TOKEN_SDKDEV=$TEST_ACCESS_TOKEN_SDKDEV \ |
|
|
-e TEST_LEVEL=$TEST_LEVEL \ |
|
|
-e MODELSCOPE_ENVIRONMENT='ci' \ |
|
|
-e TEST_UPLOAD_MS_TOKEN=$TEST_UPLOAD_MS_TOKEN \ |
|
|
-e MODEL_TAG_URL=$MODEL_TAG_URL \ |
|
|
-e MODELSCOPE_API_TOKEN=$MODELSCOPE_API_TOKEN \ |
|
|
-e PR_CHANGED_FILES=$PR_CHANGED_FILES \ |
|
|
--workdir=$CODE_DIR_IN_CONTAINER \ |
|
|
${IMAGE_NAME}:${IMAGE_VERSION} \ |
|
|
$CI_COMMAND |
|
|
fi |
|
|
if [ $? -ne 0 ]; then |
|
|
echo "Running test case failed, please check the log!" |
|
|
exit -1 |
|
|
fi |
|
|
break |
|
|
done |
|
|
if [ "$is_get_file_lock" = false ] ; then |
|
|
echo 'No free GPU!' |
|
|
exit 1 |
|
|
fi |
|
|
|