This is the root directory of the "NoFib Haskell benchmark suite". It should be part of a GHC source tree, that is the 'nofib' directory should be at the same level in the tree as 'compiler' and 'libraries'. This makes sure that NoFib picks up the stage 2 compiler from the surrounding GHC source tree.
You can also clone this repository in isolation, in which case it will
pick $(which ghc)
or whatever the HC
environment variable is set to.
Additional information can also be found on NoFib's wiki page.
There's also a easy.sh
helper script, which as name implies, is
automated and easy way to run nofib
.
See the section at the end of README for its usage.
Git symlink support for Windows machines
NoFib uses a few symlinks here and there to share code between benchmarks.
Git for Windows has symlinks support for some time now, but
it may not be enabled by default.
You will notice strange make boot
failures if it's not enabled for you.
Make sure you follow the instructions in the link to enable symlink support,
possibly as simple as through git config core.symlinks true
or cloning with
git clone -c core.symlinks=true <URL>
.
Install cabal-install-2.4
or later.
Then, to run the tests, execute:
$ make clean # or git clean -fxd, it's faster
$ # Generates input files for the benchmarks and builds compilation
$ # dependencies for make (ghc -M)
$ make boot
$ # Builds the benchmarks and runs them $NoFibRuns (default: 5) times
$ make
This will put the results in the file nofib-log
. You can pass extra
options to a nofib run using the EXTRA_HC_OPTS
variable like this:
$ make clean
$ make boot
$ make EXTRA_HC_OPTS="-fllvm"
Note: to get all the results, you have to clean
and boot
between
separate nofib
runs.
To compare the results of multiple runs, save the output in a logfile
and use the program in ./nofib-analyse/nofib-analyse
, for example:
...
$ make 2>&1 | tee nofib-log-6.4.2
...
$ make 2>&1 | tee nofib-log-6.6
$ nofib-analyse nofib-log-6.4.2 nofib-log-6.6 | less
to generate a comparison of the runs in captured in nofib-log-6.4.2
and nofib-log-6.6
. When making comparisons, be careful to ensure
that the things that changed between the builds are only the things
that you wanted to change. There are lots of variables: machine,
GHC version, GCC version, C libraries, static vs. dynamic GMP library,
build options, run options, and probably lots more. To be on the safe
side, make both runs on the same unloaded machine.
Each benchmark is runnable in three different time mode
s:
fast
: 0.1-0.2snorm
: 1-2sslow
: 5-10s
You can control which mode to run by setting an additional mode
variable for
make
. The default is mode=norm
. Example for mode=fast
:
$ make clean
$ make boot mode=fast
$ make mode=fast
Note that the mode
s set in make boot
and make
need to agree. Otherwise you
will get output errors, because make boot
will generate input files for a
different mode
. A more DRY way to control the mode
would be
$ make clean
$ export mode=fast
$ make boot
$ make
As CPU architectures advance, the above running times may drift and occasionally, all benchmarks will need adjustments.
Be aware that nofib-analyse
will ignore the result if it falls below 0.2s.
This is the default of its -i
option, which is of course incompatible with
mode=fast
. In that case, you should just set -i
as appropriate, even
deactivate it with -i 0
.
The nofib-analyse
utility is compiled with BOOT_HC
compiler,
which may be different then the GHC under the benchmark.
You can control which GHC you benchmark with HC
variable
$ make clean
$ make boot HC=ghc-head
$ make HC=ghc-head 2>&1 | tee nofib-log-ghc-head
There are some options you might want to tweak; search for nofib in
../mk/config.mk
, and override settings in ../mk/build.mk
as usual.
To get instruction counts, memory reads/writes, and "cache misses", you'll need to get hold of Cachegrind, which is part of Valgrind.
You can then pass -cachegrind
as EXTRA_RUNTEST_OPTS
. Counting
instructions slows down execution by a factor of ~30. But it's
a deterministic metric, so you can combine it with NoFibRuns=1
:
$ (make EXTRA_RUNTEST_OPTS="-cachegrind" NoFibRuns=1) 2>&1 | tee nofib-log
Optionally combine this with mode=fast
, see Modes.
Some benchmarks aren't run by default and require extra packages are installed for the GHC compiler being tested. These packages include:
- stm - for smp benchmarks
If you add a benchmark try to set the problem sizes for fast/normal/slow reasonably. Modes lists the recommended brackets for each mode.
So you have a benchmark to submit but don't know in which subfolder to put it? Here's some advice on the intended semantics of each category.
These are run when you just type make
. Their semantics is explained in
the Nofib paper
(You can find a .ps online, thanks to @bgamari. Alternatively grep for
'Spectral' in docs/paper/paper.verb).
imaginary
: Mostly toy benchmarks, solving puzzles like n-queens.spectral
: Algorithmic kernels, like FFT. If you want to add a benchmark of a library, this most certainly the place to put it.real
: Actual applications, with a command-line interface and all. Because of the large dependency footprint of today's applications, these have become rather aged.shootout
: Benchmarks from the benchmarks game, formerly known as "language shootout".
Most of the benchmarks are quite old and aren't really written in way one would
write high-performance Haskell code today (e.g., use of String
, lists,
redefining own list combinators that don't take part in list fusion, rare use of
strictness annotations or unboxed data), so new benchmarks for the real
and
spectral
in brackets in particular are always welcome!
Other than the default single-threaded categories above, there are the following (SG: I'm guessing here, have never run them):
gc
: Run bymake -C gc
(though you'll probably have to edit the Makefile to your specific config). Select benchmarks fromspectral
andreal
, plus a few more (Careful, these have not been touched by #15999/!5, see the next subsection). Testdrives different GC configs, apparently.smp
: Microbenchmarks for the-threaded
runtime, measuring scheduler performance on concurrent and STM-heavy code.
Additionally, pay attention that your benchmarks are stable wrt. different GC paramerisations, so that small changes in allocation don't lead to big, unexplicable jumps in performance. See #15999 for details. Also make sure that you run the benchmark with the default GC settings, as enlarging Gen 0 or Gen 1 heaps just amplifies the problem.
As a rule of thumb on how to ensure this: Make sure that your benchmark doesn't
just build up one big data and consume it in a final step, but rather that the
working set grows and shrinks (e.g. is approximately constant) over the whole
run of the benchmark. You can ensure this by iterating your main logic $n times
(how often depends on your program, but in the ball park of 100-1000).
You can test stability by plotting productivity curves for your fast
settings
with the prod.py
script attached to #15999.
If in doubt, ask Sebastian Graf for help.
./easy.sh - easy nofib
Usage: ./easy.sh [ -m mode ] /path/to/baseline/ghc /path/to/new/ghc"
GHC paths can point to the root of the GHC repository,
if it's build with Hadrian.
Available options:
-m MODE nofib mode: fast norm slow
This script caches the results using the sha256 of ghc executable.
Remove these files, if you want to rerun the benchmark.