BenchmarkTools makes performance tracking of Julia code easy by supplying a framework for writing and running groups of benchmarks as well as comparing benchmark results.
This package is used to write and run the benchmarks found in BaseBenchmarks.jl.
The CI infrastructure for automated performance testing of the Julia language is not in this package, but can be found in Nanosoldier.jl.
BenchmarkTools is a Julia Language package. To install BenchmarkTools, please open Julia's interactive session (known as REPL) and press ] key in the REPL to use the package mode, then type the following command
pkg> add BenchmarkTools
If you're just getting started, check out the manual for a thorough explanation of BenchmarkTools.
If you want to explore the BenchmarkTools API, see the reference document.
If you want a short example of a toy benchmark suite, see the sample file in this repo (benchmark/benchmarks.jl).
If you want an extensive example of a benchmark suite being used in the real world, you can look at the source code of BaseBenchmarks.jl.
If you're benchmarking on Linux, I wrote up a series of tips and tricks to help eliminate noise during performance tests.
The primary macro provided by BenchmarkTools is
julia> using BenchmarkTools # The `setup` expression is run once per sample, and is not included in the # timing results. Note that each sample can require multiple evaluations # benchmark kernel evaluations. See the BenchmarkTools manual for details. julia> @benchmark sin(x) setup=(x=rand()) BenchmarkTools.Trial: memory estimate: 0 bytes allocs estimate: 0 -------------- minimum time: 4.248 ns (0.00% GC) median time: 4.631 ns (0.00% GC) mean time: 5.502 ns (0.00% GC) maximum time: 60.995 ns (0.00% GC) -------------- samples: 10000 evals/sample: 1000
julia> @btime sin(x) setup=(x=rand()) 4.361 ns (0 allocations: 0 bytes) 0.49587200950472454
If the expression you want to benchmark depends on external variables, you should use
$ to "interpolate" them into the benchmark expression to
avoid the problems of benchmarking with globals.
Essentially, any interpolated variable
$x or expression
$(...) is "pre-computed" before benchmarking begins:
julia> A = rand(3,3); julia> @btime inv($A); # we interpolate the global variable A with $A 1.191 μs (10 allocations: 2.31 KiB) julia> @btime inv($(rand(3,3))); # interpolation: the rand(3,3) call occurs before benchmarking 1.192 μs (10 allocations: 2.31 KiB) julia> @btime inv(rand(3,3)); # the rand(3,3) call is included in the benchmark time 1.295 μs (11 allocations: 2.47 KiB)
Sometimes, interpolating variables into very simple expressions can give the compiler more information than you intended, causing it to "cheat" the benchmark by hoisting the calculation out of the benchmark code
julia> a = 1; b = 2 2 julia> @btime $a + $b 0.024 ns (0 allocations: 0 bytes) 3
As a rule of thumb, if a benchmark reports that it took less than a nanosecond to perform, this hoisting probably occured. You can avoid this by referencing and dereferencing the interpolated variables
julia> @btime $(Ref(a)) + $(Ref(b)) 1.277 ns (0 allocations: 0 bytes) 3
As described the manual, the BenchmarkTools package supports many other features, both for additional output and for more fine-grained control over the benchmarking process.
Our story begins with two packages, "Benchmarks" and "BenchmarkTrackers". The Benchmarks package implemented an execution strategy for collecting and summarizing individual benchmark results, while BenchmarkTrackers implemented a framework for organizing, running, and determining regressions of groups of benchmarks. Under the hood, BenchmarkTrackers relied on Benchmarks for actual benchmark execution.
For a while, the Benchmarks + BenchmarkTrackers system was used for automated performance testing of Julia's Base library. It soon became apparent that the system suffered from a variety of issues:
The BenchmarkTools package is a response to these issues, designed by examining user reports and the benchmark data generated by the old system. BenchmarkTools offers the following solutions to the corresponding issues above:
This package was authored primarily by Jarrett Revels (@jrevels). Additionally, I'd like to thank the following people: