|
| 1 | +#!/bin/bash |
| 2 | + |
| 3 | +# SC2206 (warning): Quote to prevent word splitting/globbing, or split robustly with mapfile or read -a. |
| 4 | +# bloopDirsGenerated=() |
| 5 | +# bloopDir='spark 356' |
| 6 | +# bloopDirsGenerated+=("$bloopDir") |
| 7 | +# for item in "${bloopDirsGenerated[@]}"; do |
| 8 | +# echo $item |
| 9 | +# done |
| 10 | + |
| 11 | +# SC2124 (warning): Assigning an array to a string! Assign as array, or use * instead of @ to concatenate. |
| 12 | +# IFS=" " read -ra ARGS <<< "$(find . -name "*.yaml" -printf "%p ")" |
| 13 | +# for item in "${ARGS[@]}"; do |
| 14 | +# echo $item |
| 15 | +# done |
| 16 | +# echo "${ARGS[@]}" |
| 17 | + |
| 18 | +SKIP_TESTS="TRUE" |
| 19 | +if [[ $( echo ${SKIP_TESTS} | tr [:upper:] [:lower:] ) == "true" ]]; |
| 20 | +then |
| 21 | + echo "PYTHON INTEGRATION TESTS SKIPPED..." |
| 22 | +fi |
| 23 | + |
| 24 | +# GPU_MEM_PARALLEL=1 |
| 25 | +# CPU_CORES=2 |
| 26 | +# TMP_PARALLEL=$(( GPU_MEM_PARALLEL > CPU_CORES ? CPU_CORES : GPU_MEM_PARALLEL )) |
| 27 | +# echo $TMP_PARALLEL |
| 28 | + |
| 29 | +# TMP_PARALLEL=10 |
| 30 | +# if (( TMP_PARALLEL <= 1 )); then |
| 31 | +# TEST_PARALLEL=1 |
| 32 | +# else |
| 33 | +# TEST_PARALLEL=$TMP_PARALLEL |
| 34 | +# fi |
| 35 | +# echo $TEST_PARALLEL |
| 36 | + |
| 37 | +# TESTS="api_test login_test db_query" |
| 38 | + |
| 39 | +# read -ra RAW_TESTS <<< "${TESTS}" |
| 40 | +# for raw_test in "${RAW_TESTS[@]}"; do |
| 41 | +# echo "Running test: $raw_test" |
| 42 | +# done |
| 43 | + |
| 44 | +# sed_i() { |
| 45 | +# sed -e "$1" "$2" > "$2.tmp" && mv "$2.tmp" "$2" |
| 46 | +# } |
| 47 | +# SPARK_VERSION='spark330' |
| 48 | +# sed_i '/<java\.major\.version>/,/<spark\.version>\${spark[0-9]\+\.version}</s/<spark\.version>\${spark[0-9]\+\.version}</<spark.version>\${'$SPARK_VERSION'.version}</' \ |
| 49 | +# "pom.xml" |
| 50 | + |
| 51 | +# BASE_SPARK_VERSION="320.dev" |
| 52 | +# BUILDVER=${BASE_SPARK_VERSION//./}db |
| 53 | +# echo $BUILDVER |
| 54 | + |
| 55 | +# REQUIRED_PACKAGES=("existing-pkg" "with space" "c*") |
| 56 | +# CUDA_VER="11.0" |
| 57 | +# CUDF_VER="22.0" |
| 58 | +# REQUIRED_PACKAGES=( |
| 59 | +# "cuda-version=$CUDA_VER" |
| 60 | +# "cudf=$CUDF_VER" |
| 61 | +# "${REQUIRED_PACKAGES[@]}" |
| 62 | +# ) |
| 63 | +# for item in "${REQUIRED_PACKAGES[@]}"; do |
| 64 | +# echo $item |
| 65 | +# done |
| 66 | + |
| 67 | +# CLASSIFIERS="cuda12,cuda12-arm64" |
| 68 | +# DEPLOY_TYPES=${CLASSIFIERS//[^,]?/jar} |
| 69 | +# # DEPLOY_TYPES=$(echo $CLASSIFIERS | sed -e 's;[^,]*;jar;g') |
| 70 | +# echo $DEPLOY_TYPES |
| 71 | + |
| 72 | +# PHASE_TYPE="330 331 340 350 356" |
| 73 | +# echo "original output:" |
| 74 | +# SPARK_SHIM_VERSIONS=(`echo "$PHASE_TYPE"`) |
| 75 | +# echo "${SPARK_SHIM_VERSIONS[@]}" |
| 76 | +# echo "new output:" |
| 77 | +# IFS=" " read -ra SPARK_SHIM_VERSIONS <<< $PHASE_TYPE |
| 78 | +# echo "${SPARK_SHIM_VERSIONS[@]}" |
| 79 | +# # for item in "${SPARK_SHIM_VERSIONS[@]}"; do |
| 80 | +# # echo $item |
| 81 | +# # done |
| 82 | + |
| 83 | + |
| 84 | +# PROFILE_OPT="-PpremergeUT1" |
| 85 | +# # SPARK_SHIM_VERSIONS_STR=$(mvn -B help:evaluate -q -pl dist $PROFILE_OPT -Dexpression=included_buildvers -DforceStdout) |
| 86 | +# SPARK_SHIM_VERSIONS_STR=$(echo -n $(mvn -B help:evaluate -q -pl dist $PROFILE_OPT -Dexpression=included_buildvers -DforceStdout)) |
| 87 | +# # SPARK_SHIM_VERSIONS_STR=$(echo $SPARK_SHIM_VERSIONS_STR) |
| 88 | +# # echo "$SPARK_SHIM_VERSIONS_STR" |
| 89 | +# IFS=", " <<< $SPARK_SHIM_VERSIONS_STR read -r -a SPARK_SHIM_VERSIONS_ARR |
| 90 | +# for item in "${SPARK_SHIM_VERSIONS_ARR[@]}"; do |
| 91 | +# echo $item |
| 92 | +# done |
| 93 | + |
| 94 | +# SPARK_SHIM_VERSIONS_NOSNAPSHOTS_TAIL=(320 310 310db 330) |
| 95 | +# SPARK_SHIM_VERSIONS_PREMERGE_UT_1=(321 330) |
| 96 | +# for version in "${SPARK_SHIM_VERSIONS_NOSNAPSHOTS_TAIL[@]}" |
| 97 | +# do |
| 98 | +# echo "Spark version: $version" |
| 99 | +# # build and run unit tests on one specific version for each sub-version (e.g. 320, 330) except base version |
| 100 | +# # separate the versions to two ci stages (mvn_verify, ci_2) for balancing the duration |
| 101 | +# match=1 |
| 102 | +# for element in "${SPARK_SHIM_VERSIONS_PREMERGE_UT_1[@]}"; do |
| 103 | +# if [[ "$element" == "$version" ]]; then |
| 104 | +# match=0 |
| 105 | +# break |
| 106 | +# fi |
| 107 | +# done |
| 108 | +# if [[ $match == 0 ]]; then |
| 109 | +# echo "yes" |
| 110 | +# fi |
| 111 | +# done |
| 112 | + |
| 113 | + |
| 114 | + |
| 115 | +# HOST_MEM_PARALLEL=$(awk '/MemAvailable/ {print int($2 / (8 * 1024 * 1024))}' /proc/meminfo) |
| 116 | +# # HOST_MEM_PARALLEL=`cat /proc/meminfo | grep MemAvailable | awk '{print int($2 / (8 * 1024 * 1024))}'` |
| 117 | +# echo $HOST_MEM_PARALLEL |
| 118 | + |
| 119 | +# REPORT_CHARS=${REPORT_CHARS:="fE"} |
| 120 | +# if [[ "${TEST_TAGS}" != "" ]]; |
| 121 | +# then |
| 122 | +# TEST_TAGS="-m $TEST_TAGS" |
| 123 | +# fi |
| 124 | +# STD_INPUT_PATH="test/path" |
| 125 | +# TEST_TYPE_PARAM="" |
| 126 | +# TEST_TYPE="pre-commit" |
| 127 | +# if [[ "${TEST_TYPE}" != "" ]]; |
| 128 | +# then |
| 129 | +# TEST_TYPE_PARAM="--test_type=$TEST_TYPE" |
| 130 | +# fi |
| 131 | +# if [[ "${TEST}" != "" ]]; |
| 132 | +# then |
| 133 | +# TEST_ARGS="-k $TEST" |
| 134 | +# fi |
| 135 | +# TEST_COMMON_OPTS=(-v |
| 136 | +# -r"$REPORT_CHARS" |
| 137 | +# "$TEST_TAGS" |
| 138 | +# --std_input_path="$STD_INPUT_PATH" |
| 139 | +# --color=yes |
| 140 | +# "$TEST_TYPE_PARAM" |
| 141 | +# "$TEST_ARGS" |
| 142 | +# --junitxml=TEST-pytest-`date +%s%N`.xml |
| 143 | +# "$@") |
| 144 | +# echo "${TEST_COMMON_OPTS[@]}" |
| 145 | + |
| 146 | +# declare -a path_parts='([0]="" [1]="spark320" [2]="com" [3]="nvidia" [4]="spark" [5]="udf" [6]="Repr\$UnknownCapturedArg\$.class")' |
0 commit comments