| #!/bin/bash |
| |
| # This is a very simple script that runs all three benchmark suites |
| # and then runs a log parsing script to get the logs into a csv form |
| # and uploads the result to GitHub Gist for easy copy paste into Google |
| # Spreadsheet. |
| # |
| # Useful flag sets: |
| # |
| # Run models that are skipped in dynamic CI only, to update skips |
| # ./run_all.sh --training --backend aot_eager --dynamic-ci-skips-only |
| # |
| # Run CI models with dynamic shapes, training, aot_eager |
| # ./run_all.sh --training --backend aot_eager --dynamic-shapes --ci |
| # |
| # Run CI models with dynamic shapes, inference, inductor |
| # ./run_all.sh --backend inductor --dynamic-shapes --ci |
| # |
| # WARNING: this will silently clobber .csv and .log files in your CWD! |
| |
| set -x |
| |
| # Some QoL for people running this script on Meta servers |
| if getent hosts fwdproxy; then |
| export https_proxy=http://fwdproxy:8080 http_proxy=http://fwdproxy:8080 no_proxy=.fbcdn.net,.facebook.com,.thefacebook.com,.tfbnw.net,.fb.com,.fburl.com,.facebook.net,.sb.fbsbx.com,localhost |
| fi |
| |
| # Feel free to edit these, but we expect most users not to need to modify this |
| BASE_FLAGS=( --accuracy --explain --timing --print-graph-breaks ) |
| DATE="$(date)" |
| WORK="$PWD" |
| |
| cd "$(dirname "$BASH_SOURCE")"/../.. |
| |
| python benchmarks/dynamo/benchmarks.py --output "$WORK"/benchmarks.csv "${BASE_FLAGS[@]}" "$@" 2>&1 | tee "$WORK"/sweep.log |
| gh gist create -d "Sweep logs for $(git rev-parse --abbrev-ref HEAD) $* - $(git rev-parse HEAD) $DATE" "$WORK"/sweep.log | tee -a "$WORK"/sweep.log |
| python benchmarks/dynamo/parse_logs.py "$WORK"/sweep.log > "$WORK"/final.csv |
| gh gist create "$WORK"/final.csv |