aquavm/.github/workflows/benchmark.yml
renovate[bot] 2d87fa753f
chore(deps): update actions/setup-python action to v5 (#785)
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2024-04-18 12:41:34 +03:00

74 lines
1.8 KiB
YAML

name: Run benchmark with workflow_call
on:
workflow_call:
inputs:
ref:
description: "git ref to checkout to"
type: string
default: "master"
tag:
description: "Release tag to upload results to"
type: string
default: "null"
jobs:
benchmark:
name: Run benchmark
runs-on: ubuntu-latest
steps:
- name: Checkout AquaVM
uses: actions/checkout@v4
with:
repository: fluencelabs/aquavm
ref: ${{ inputs.ref }}
- name: Setup Rust toolchain
uses: dsherret/rust-toolchain-file@v1
- name: Setup cache
uses: Swatinem/rust-cache@v2
with:
shared-key: aquavm
save-if: false
- name: Setup marine
uses: fluencelabs/setup-marine@v1
- name: Generate benchmark data
working-directory: junk/gen-bench-data
run: ./gen_benchmark_data.sh
- name: Setup python
uses: actions/setup-python@v5
with:
python-version: "3.x"
- name: Install python module
run: pip install tools/cli/performance_metering
- name: Run benchmark
run: aquavm_performance_metering run
- name: Upload benchmark results to artifacts
uses: actions/upload-artifact@v3
with:
name: ./benches/PERFORMANCE.txt
path: lcov.info
- name: Print benchmark results to checks
run: |
{
echo '```';
cat ./benches/PERFORMANCE.txt;
echo '```';
} >> $GITHUB_STEP_SUMMARY
- name: Upload benchmark results to release
if: inputs.tag != 'null'
uses: softprops/action-gh-release@v1
with:
files: ./benches/PERFORMANCE.txt
tag_name: ${{ inputs.tag }}