Provided by: hyperfine_1.14.0-2_amd64 bug

NAME

       hyperfine - hyperfine

DESCRIPTION

       hyperfine 1.14.0 A command-line benchmarking tool.

   USAGE:
              hyperfine [OPTIONS] <command>...

   ARGS:
              <command>...

              The  command  to  benchmark.  This can be the name of an executable, a command line
              like "grep -i todo" or a shell command like "sleep 0.5 && echo test". The latter is
              only  available  if  the  shell  is  not explicitly disabled via '--shell=none'. If
              multiple commands are given, hyperfine will show a  comparison  of  the  respective
              runtimes.

   OPTIONS:
       -w, --warmup <NUM>

              Perform  NUM  warmup  runs  before  the  actual benchmark. This can be used to fill
              (disk) caches for I/O-heavy programs.

       -m, --min-runs <NUM>

              Perform at least NUM runs for each command (default: 10).

       -M, --max-runs <NUM>

              Perform at most NUM runs for each command. By default, there is no limit.

       -r, --runs <NUM>

              Perform exactly NUM runs for  each  command.  If  this  option  is  not  specified,
              hyperfine automatically determines the number of runs.

       -s, --setup <CMD>

              Execute  CMD  before  each  set  of  timing runs. This is useful for compiling your
              software with the provided parameters, or to do any other work that  should  happen
              once  before  a  series  of benchmark runs, not every time as would happen with the
              --prepare option.

       -p, --prepare <CMD>

              Execute CMD before each timing run. This is useful for clearing  disk  caches,  for
              example.   The  --prepare option can be specified once for all commands or multiple
              times, once for each command. In the latter case, each preparation command will  be
              run prior to the corresponding benchmark command.

       -c, --cleanup <CMD>

              Execute  CMD  after  the  completion  of  all benchmarking runs for each individual
              command to be benchmarked. This is useful if the commands to be benchmarked produce
              artifacts that need to be cleaned up.

       -P, --parameter-scan <VAR> <MIN> <MAX>

              Perform  benchmark  runs  for each value in the range MIN..MAX. Replaces the string
              '{VAR}' in each command by the current parameter value.

       Example:
              hyperfine -P threads 1 8 'make -j {threads}'

              This performs benchmarks for 'make -j 1', 'make -j 2', ???, 'make -j 8'.

              To have the value increase following different patterns, use shell arithmetics.

              Example: hyperfine -P size 0 3 'sleep $((2**{size}))'

              This performs benchmarks with power of 2 increases: 'sleep 1',  'sleep  2',  'sleep
              4', ???  The exact syntax may vary depending on your shell and OS.

       -D, --parameter-step-size <DELTA>

              This argument requires --parameter-scan to be specified as well. Traverse the range
              MIN..MAX in steps of DELTA.

       Example:
              hyperfine -P delay 0.3 0.7 -D 0.2 'sleep {delay}'

              This performs benchmarks for 'sleep 0.3', 'sleep 0.5' and 'sleep 0.7'.

       -L, --parameter-list <VAR> <VALUES>

              Perform benchmark runs for each value in the comma-separated list VALUES.  Replaces
              the string '{VAR}' in each command by the current parameter value.

       Example:
              hyperfine -L compiler gcc,clang '{compiler} -O2 main.cpp'

              This performs benchmarks for 'gcc -O2 main.cpp' and 'clang -O2 main.cpp'.

              The  option  can  be  specified  multiple  times to run benchmarks for all possible
              parameter combinations.

       --style <TYPE>

              Set output style type (default: auto).  Set  this  to  'basic'  to  disable  output
              coloring  and  interactive elements. Set it to 'full' to enable all effects even if
              no interactive terminal was detected. Set this to 'nocolor' to keep the interactive
              output  without  any  colors.  Set  this  to 'color' to keep the colors without any
              interactive output. Set this to 'none' to disable all the output of the tool.

       -S, --shell <SHELL>

              Set the shell to use for executing benchmarked commands. This can be  the  name  or
              the path to the shell executable, or a full command line like "bash --norc". It can
              also be set to "default" to explicitly select the default shell on  this  platform.
              Finally,  this  can  also  be  set  to  "none"  to disable the shell. In this case,
              commands will be executed directly. They can still have arguments, but more complex
              things like "sleep 0.1; sleep 0.2" are not possible without a shell.

       -N

              An alias for '--shell=none'.

       -i, --ignore-failure

              Ignore non-zero exit codes of the benchmarked programs.

       -u, --time-unit <UNIT>

              Set the time unit to be used. Possible values: millisecond, second.

       --export-asciidoc <FILE>

              Export the timing summary statistics as an AsciiDoc table to the given FILE.

       --export-csv <FILE>

              Export  the  timing  summary  statistics  as CSV to the given FILE. If you need the
              timing results for each individual run, use the JSON export format.

       --export-json <FILE>

              Export the timing summary statistics and timings of individual runs as JSON to  the
              given FILE.

       --export-markdown <FILE>

              Export the timing summary statistics as a Markdown table to the given FILE.

       --export-orgmode <FILE>

              Export the timing summary statistics as a Emacs org-mode table to the given FILE.

       --show-output

              Print  the  stdout and stderr of the benchmark instead of suppressing it. This will
              increase the time it takes for benchmarks to run, so it should  only  be  used  for
              debugging purposes or when trying to benchmark output speed.

       --output <WHERE>

              Control where the output of the benchmark is redirected. <WHERE> can be:

              null:  Redirect  output  to  /dev/null  (the default). Note that some programs like
              'grep' detect when standard output is /dev/null and apply certain optimizations. To
              avoid that, consider using '--output=pipe'.

              pipe: Feed the output through a pipe before discarding it.

              inherit: Don't redirect the output at all (same as '--show-output').

              <FILE>: Write the output to the given file.

       -n, --command-name <NAME>

              Give  a  meaningful  name  to  a  command.  This can be specified multiple times if
              several commands are benchmarked.

       -h, --help

              Print help information

       -V, --version

              Print version information