name: benchmark description: "Run benchmarks" inputs: cloudProvider: description: "Which cloud provider to use." required: true kubeconfig: description: "The kubeconfig of the cluster to test." required: true awsOpenSearchDomain: description: "AWS OpenSearch Endpoint Domain to upload the results." required: false awsOpenSearchUsers: description: "AWS OpenSearch User to upload the results." required: false awsOpenSearchPwd: description: "AWS OpenSearch Password to upload the results." required: false runs: using: "composite" steps: - name: Setup python uses: actions/setup-python@d27e3f3d7c64b4bbf8e4abfb9b63b83e846e0435 # v4.5.0 with: python-version: "3.10" - name: Install kubestr shell: bash env: KUBESTR_VER: "0.4.37" run: | HOSTOS="$(go env GOOS)" HOSTARCH="$(go env GOARCH)" curl -fsSLO https://github.com/kastenhq/kubestr/releases/download/v${KUBESTR_VER}/kubestr_${KUBESTR_VER}_${HOSTOS}_${HOSTARCH}.tar.gz tar -xzf kubestr_${KUBESTR_VER}_${HOSTOS}_${HOSTARCH}.tar.gz install kubestr /usr/local/bin - name: Checkout k8s-bench-suite uses: actions/checkout@8f4b7f84864484a7bf31766abe9204da3cbe65b3 # v3.5.0 with: fetch-depth: 0 repository: "InfraBuilder/k8s-bench-suite" ref: 1698974913b7b18ad54cf5860838029c295c77b1 path: k8s-bench-suite - name: Run FIO benchmark without caching in Azure if: inputs.cloudProvider == 'azure' shell: bash env: KUBECONFIG: ${{ inputs.kubeconfig }} run: | cat <> $GITHUB_ENV - name: Get previous benchmark records from S3 shell: bash env: CSP: ${{ inputs.cloudProvider }} run: | mkdir -p benchmarks aws s3 cp --recursive ${S3_PATH} benchmarks --no-progress if [[ -f benchmarks/constellation-${CSP}.json ]]; then mv benchmarks/constellation-${CSP}.json benchmarks/constellation-${CSP}-previous.json else echo "::warning::Couldn't retrieve previous benchmark records from s3" fi - name: Parse results, create diagrams and post the progression summary shell: bash env: # Original result directory BENCH_RESULTS: out/ # Working directory containing the previous results as JSON and to contain the graphs BDIR: benchmarks # Paths to benchmark results as JSON of the previous run and the current run PREV_BENCH: benchmarks/constellation-${{ inputs.cloudProvider }}-previous.json CURR_BENCH: benchmarks/constellation-${{ inputs.cloudProvider }}.json CSP: ${{ inputs.cloudProvider }} run: | python .github/actions/e2e_benchmark/evaluate/parse.py export BENCHMARK_SUCCESS=true if [[ -f "$PREV_BENCH" ]]; then # Sets $BENCHMARK_SUCCESS=false if delta is bigger than defined in compare.py python .github/actions/e2e_benchmark/evaluate/compare.py >> $GITHUB_STEP_SUMMARY fi echo BENCHMARK_SUCCESS=$BENCHMARK_SUCCESS >> $GITHUB_ENV - name: Upload benchmark results to action run if: (!env.ACT) uses: actions/upload-artifact@0b7f8abb1508181956e8e162db84b466c27e18ce # v3.1.2 with: path: | benchmarks/constellation-${{ inputs.cloudProvider }}.json name: "benchmarks" - name: Upload benchmark results to opensearch if: (!env.ACT) shell: bash env: OPENSEARCH_DOMAIN: ${{ inputs.awsOpenSearchDomain }} OPENSEARCH_USER: ${{ inputs.awsOpenSearchUsers }} OPENSEARCH_PWD: ${{ inputs.awsOpenSearchPwd }} run: | curl -XPOST \ -u "${OPENSEARCH_USER}:${OPENSEARCH_PWD}" \ "${OPENSEARCH_DOMAIN}/benchmarks-${{ inputs.cloudProvider }}-$(date '+%Y')"/_doc \ --data-binary @benchmarks/constellation-${{ inputs.cloudProvider }}.json \ -H 'Content-Type: application/json' - name: Update benchmark records in S3 if: github.ref_name == 'main' shell: bash env: CSP: ${{ inputs.cloudProvider }} run: | aws s3 cp benchmarks/constellation-${CSP}.json ${S3_PATH}/constellation-${CSP}.json - name: Check performance comparison result shell: bash run: | if [[ $BENCHMARK_SUCCESS == true ]] ; then echo "Benchmark successful, all metrics in the expected range." else echo "::error::Benchmark failed, some metrics are outside of the expected range." exit 1 fi