|
| 1 | +#!/bin/bash |
| 2 | +MODELSCOPE_CACHE_DIR_IN_CONTAINER=/modelscope_cache |
| 3 | +CODE_DIR=$PWD |
| 4 | +CODE_DIR_IN_CONTAINER=/twinkle |
| 5 | +MODELSCOPE_SDK_DEBUG=True |
| 6 | +echo "$USER" |
| 7 | +gpus='0,1 2,3' |
| 8 | +cpu_sets='0-15 16-31' |
| 9 | +cpu_sets_arr=($cpu_sets) |
| 10 | +is_get_file_lock=false |
| 11 | +CI_COMMAND=${CI_COMMAND:-bash .dev_scripts/ci_container_test.sh python tests/run.py --parallel 2 --run_config tests/run_config.yaml} |
| 12 | +echo "ci command: $CI_COMMAND" |
| 13 | +PR_CHANGED_FILES="${PR_CHANGED_FILES:-}" |
| 14 | +echo "PR modified files: $PR_CHANGED_FILES" |
| 15 | +PR_CHANGED_FILES=${PR_CHANGED_FILES//[ ]/#} |
| 16 | +echo "PR_CHANGED_FILES: $PR_CHANGED_FILES" |
| 17 | +idx=0 |
| 18 | +for gpu in $gpus |
| 19 | +do |
| 20 | + exec {lock_fd}>"/tmp/gpu$gpu" || exit 1 |
| 21 | + flock -n "$lock_fd" || { echo "WARN: gpu $gpu is in use!" >&2; idx=$((idx+1)); continue; } |
| 22 | + echo "get gpu lock $gpu" |
| 23 | + |
| 24 | + CONTAINER_NAME="twinkle-ci-$idx" |
| 25 | + let is_get_file_lock=true |
| 26 | + |
| 27 | + # pull image if there are update |
| 28 | + docker pull ${IMAGE_NAME}:${IMAGE_VERSION} |
| 29 | + if [ "$MODELSCOPE_SDK_DEBUG" == "True" ]; then |
| 30 | + echo 'debugging' |
| 31 | + docker run --rm --name $CONTAINER_NAME --shm-size=16gb \ |
| 32 | + --cpuset-cpus=${cpu_sets_arr[$idx]} \ |
| 33 | + --gpus='"'"device=$gpu"'"' \ |
| 34 | + -v $CODE_DIR:$CODE_DIR_IN_CONTAINER \ |
| 35 | + -v $MODELSCOPE_CACHE:$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
| 36 | + -v $MODELSCOPE_HOME_CACHE/$idx:/root \ |
| 37 | + -v /home/admin/pre-commit:/home/admin/pre-commit \ |
| 38 | + -e CI_TEST=True \ |
| 39 | + -e TEST_LEVEL=$TEST_LEVEL \ |
| 40 | + -e MODELSCOPE_CACHE=$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
| 41 | + -e MODELSCOPE_DOMAIN=$MODELSCOPE_DOMAIN \ |
| 42 | + -e MODELSCOPE_SDK_DEBUG=True \ |
| 43 | + -e HUB_DATASET_ENDPOINT=$HUB_DATASET_ENDPOINT \ |
| 44 | + -e TEST_ACCESS_TOKEN_CITEST=$TEST_ACCESS_TOKEN_CITEST \ |
| 45 | + -e TEST_ACCESS_TOKEN_SDKDEV=$TEST_ACCESS_TOKEN_SDKDEV \ |
| 46 | + -e TEST_LEVEL=$TEST_LEVEL \ |
| 47 | + -e MODELSCOPE_ENVIRONMENT='ci' \ |
| 48 | + -e TEST_UPLOAD_MS_TOKEN=$TEST_UPLOAD_MS_TOKEN \ |
| 49 | + -e MODEL_TAG_URL=$MODEL_TAG_URL \ |
| 50 | + -e MODELSCOPE_API_TOKEN=$MODELSCOPE_API_TOKEN \ |
| 51 | + -e PR_CHANGED_FILES=$PR_CHANGED_FILES \ |
| 52 | + --workdir=$CODE_DIR_IN_CONTAINER \ |
| 53 | + ${IMAGE_NAME}:${IMAGE_VERSION} \ |
| 54 | + $CI_COMMAND |
| 55 | + else |
| 56 | + docker run --rm --name $CONTAINER_NAME --shm-size=16gb \ |
| 57 | + --cpuset-cpus=${cpu_sets_arr[$idx]} \ |
| 58 | + --gpus='"'"device=$gpu"'"' \ |
| 59 | + -v $CODE_DIR:$CODE_DIR_IN_CONTAINER \ |
| 60 | + -v $MODELSCOPE_CACHE:$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
| 61 | + -v $MODELSCOPE_HOME_CACHE/$idx:/root \ |
| 62 | + -v /home/admin/pre-commit:/home/admin/pre-commit \ |
| 63 | + -e CI_TEST=True \ |
| 64 | + -e TEST_LEVEL=$TEST_LEVEL \ |
| 65 | + -e MODELSCOPE_CACHE=$MODELSCOPE_CACHE_DIR_IN_CONTAINER \ |
| 66 | + -e MODELSCOPE_DOMAIN=$MODELSCOPE_DOMAIN \ |
| 67 | + -e HUB_DATASET_ENDPOINT=$HUB_DATASET_ENDPOINT \ |
| 68 | + -e TEST_ACCESS_TOKEN_CITEST=$TEST_ACCESS_TOKEN_CITEST \ |
| 69 | + -e TEST_ACCESS_TOKEN_SDKDEV=$TEST_ACCESS_TOKEN_SDKDEV \ |
| 70 | + -e TEST_LEVEL=$TEST_LEVEL \ |
| 71 | + -e MODELSCOPE_ENVIRONMENT='ci' \ |
| 72 | + -e TEST_UPLOAD_MS_TOKEN=$TEST_UPLOAD_MS_TOKEN \ |
| 73 | + -e MODEL_TAG_URL=$MODEL_TAG_URL \ |
| 74 | + -e MODELSCOPE_API_TOKEN=$MODELSCOPE_API_TOKEN \ |
| 75 | + -e PR_CHANGED_FILES=$PR_CHANGED_FILES \ |
| 76 | + --workdir=$CODE_DIR_IN_CONTAINER \ |
| 77 | + ${IMAGE_NAME}:${IMAGE_VERSION} \ |
| 78 | + $CI_COMMAND |
| 79 | + fi |
| 80 | + if [ $? -ne 0 ]; then |
| 81 | + echo "Running test case failed, please check the log!" |
| 82 | + exit -1 |
| 83 | + fi |
| 84 | + break |
| 85 | +done |
| 86 | +if [ "$is_get_file_lock" = false ] ; then |
| 87 | + echo 'No free GPU!' |
| 88 | + exit 1 |
| 89 | +fi |
0 commit comments