This repository includes a set of benchmarks to regularly test the performance of various parts of the astropy core package. These benchmarks are run for every new commit in the core package repository and the results are published using a user-friendly web interface. The benchmarks are run using airspeed velocity on dedicated physical machines belonging to members of the Astropy developer community.
If you want to try and run the benchmarks locally you will first need to install asv
pip install asvthen clone the benchmarks repository
git clone git@github.com:astropy/astropy-benchmarks.git --single-branch
cd astropy-benchmarksNote that the --single-branch option is to avoid downloading the results
branch, which is large.
The easiest/fastest way to try out the benchmarks is to make sure you have either a stable or a developer version of astropy installed, then run
asv run --python=same --quick --show-stderr --dry-runThis will run the benchmarks against the local astropy version and will do some basic timing, but the timing will not be very accurate, because this only runs each benchmark once instead of taking an average of many runs. Nevertheless, this is the first step to make sure things are running correctly and can still give order of magnitude timings.
To run asv properly on the latest commit in the upstream astropy main, you can do
asv runThis will set up a temporary environment in which astropy will be installed, and
the benchmark functions will be run multiple times and averaged to get accurate
timings. The results from this will be saved into a results/<machine-name>
directory, which will store one file per commit (or more than one file if the
tests are set up to run for multiple python or numpy versions, which they are
not configured to do by default for Astropy).
You can specify the commit(s) to run the benchmarks for by using the same syntax
as you would for the git log command. For example, to run the benchmarks for
a single specific commit, you can do
asv run 88fbbc33^!replacing 88fbbc33 by the commit you want to test (the ^! Indicates to just run
this commit, not all commits up to that point). If
you want to run a range of commits, use
asv run 827f322b..729abcf3For the best accuracy, if you are using Linux, you can prevent core-swapping
(which can introduce some noise in the timings) by running asv with
taskset -c 0 asv ...This will ensure that the whole process runs on the same core.
You can generate a user-friendly web interface for your results locally by running
asv publish
asv previewThe asv preview command will give the URL of the local web server (e.g.
http://127.0.0.1:21331) - go to this address in your favorite browser to see
the results.
If you would like to run the benchmarks against specific version(s) of an
installed dependency (or dependencies), modify asv.conf.json to add
matrix.req with the desired version(s):
{
"version": 1,
"project": "astropy",
"project_url": "http://www.astropy.org/",
"repo": "https://github.com/astropy/astropy.git",
"install_command": [
"pip install . matplotlib scipy"
],
+ "matrix": {
+ "req": {
+ "matplotlib": [
+ "3.5.1"
+ ],
+ "numpy": [
+ "1.26.0",
+ "2.0.0rc1"
+ ]
+ }
+ },
"branches": ["main"],
"show_commit_url": "http://github.com/astropy/astropy/commit/",
"pythons": ["3.11"],
"environment_type": "virtualenv"
}To write a new benchmark, fork this repository and take a look inside the
benchmarks folder - each module in astropy that already has benchmarks has a
corresponding file or directory with files for different parts of the module.
Either add your benchmark to one of the existing files or create a new file as
needed.
A benchmark is a Python function whose name starts with time_. A function
should test as little as possible and therefore be as short/simple as possible.
Any required imports should be done outside the function. Here is an example of
a benchmark to test unit conversion:
from astropy import units as u
def time_my_benchmark():
(u.m / u.s).to(u.km / u.h)Once you have added a benchmark, you can make sure it runs by running
asv run --python=same --quick --show-stderr --dry-runAs mentioned in [Running the benchmarks locally], this will run all the benchmarks in fast mode (running each function once).
You can select just the benchmark you have written using the --bench option
asv run --python=same --quick --show-stderr --dry-run --bench time_my_benchmarkSay you have a local clone of astropy, and you would like to run benchmarks
against your local changes. To do this, edit asv.conf.json to switch the
repo entry from the upstream repository to your local repository:
{
"version": 1,
"project": "astropy",
"project_url": "http://www.astropy.org/",
- "repo": "https://github.com/astropy/astropy.git",
+ "repo": "./astropy/",
"install_command": [
"pip install . matplotlib scipy"
],
"branches": ["main"],
"show_commit_url": "http://github.com/astropy/astropy/commit/",
"pythons": ["3.11"],
"environment_type": "virtualenv"
}You will then be able to run the benchmarks against a commit in your local repository, e.g.
asv run 827f322b^!To compare two commits (e.g. the latest upstream commit and a local commit),
use asv compare:
asv compare 88fbbc33 827f322bThis will show a table with a comparison of the benchmark times for the two commits:
All benchmarks:
before after ratio
[e8f1432a] [c378f912]
+ 1.17±0ms 1.69ms 1.44 coordinates.FrameBenchmarks.time_init_array
+ 388±0.3μs 543μs 1.40 coordinates.FrameBenchmarks.time_init_nodata
+ 932±2μs 1.34ms 1.44 coordinates.FrameBenchmarks.time_init_scalar
1.08s 1.17s 1.08 coordinates.SkyCoordBenchmarks.time_icrs_to_galactic_array
+ 28.6±0.03ms 43.0ms 1.50 coordinates.SkyCoordBenchmarks.time_icrs_to_galactic_scalar
- 48.2±7ms 37.4ms 0.78 coordinates.SkyCoordBenchmarks.time_init_array
+ 2.84±0.01ms 5.22ms 1.84 coordinates.SkyCoordBenchmarks.time_init_scalar
+ 168±2ms 29.5s 175.69 coordinates.SkyCoordBenchmarks.time_iter_array
- 118±0.2ms 46.4ms 0.39 coordinates.SkyCoordBenchmarks.time_repr_array
- 876±1μs 760μs 0.87 coordinates.SkyCoordBenchmarks.time_repr_scalar
...Once you are happy with your benchmark(s), open a pull request to the astropy-benchmarks repository. You do not need to add any result files for the benchmarks - we have machines that automatically do this every night.
The main branch in this repository should not contain any results or built
website. Results should be added to the results branch, and commits to the
results branch trigger a build to the gh-pages branch.