View Source Benchee.Configuration (Benchee v1.3.0)
Functions to handle the configuration of Benchee, exposes init/1
function.
Summary
Types
Generated configuration struct from the user supplied configuration options.
The configuration supplied by the user as either a map or a keyword list
Functions
Returns the initial benchmark configuration for Benchee, composed of defaults and an optional custom configuration.
Types
@type t() :: %Benchee.Configuration{ after_each: Benchee.Benchmark.Hooks.hook_function() | nil, after_scenario: Benchee.Benchmark.Hooks.hook_function() | nil, assigns: map(), before_each: Benchee.Benchmark.Hooks.hook_function() | nil, before_scenario: Benchee.Benchmark.Hooks.hook_function() | nil, formatters: [ (Benchee.Suite.t() -> Benchee.Suite.t()) | module() | {module(), user_configuration()} ], input_names: [String.t()], inputs: %{required(Benchee.Suite.key()) => any()} | [{Benchee.Suite.key(), any()}] | nil, load: String.t() | [String.t()] | false, measure_function_call_overhead: boolean(), memory_time: number(), parallel: pos_integer(), percentiles: term(), pre_check: boolean(), print: map(), profile_after: boolean() | atom() | {atom(), keyword()}, reduction_time: number(), save: map() | false, time: number(), title: String.t() | nil, unit_scaling: Benchee.Conversion.Scale.scaling_strategy(), warmup: number() }
Generated configuration struct from the user supplied configuration options.
Filled in with a lot of defaults. Also notably every option is already converted to a map or struct at this point for easier handling in Benchee.
The configuration supplied by the user as either a map or a keyword list
Possible options are:
Possible options:
warmup
- the time in seconds for which a benchmarking job should be run without measuring times before "real" measurements start. This simulates a "warm" running system. Defaults to 2.time
- the time in seconds for how long each individual benchmarking job should be run for measuring the execution times (run time performance). Defaults to 5.memory_time
- the time in seconds for how long memory measurements should be conducted. Defaults to 0 (turned off).inputs
- a map from descriptive input names to some different input, your benchmarking jobs will then be run with each of these inputs. For this to work your benchmarking function gets the current input passed in as an argument into the function. Defaults tonil
, aka no input specified and functions are called without an argument.title
- this option is purely cosmetic. If you would like to add a title with some meaning to a given suite, you can do so by providing a single string here. This is only for use by formatters.formatters
- list of formatters either as a module implementing the formatter behaviour, a tuple of said module and options it should take or formatter functions. They are run when usingBenchee.run/2
or you can invoktem them throughBenchee.Formatter.output/1
. Functions need to accept one argument (which is the benchmarking suite with all data) and then use that to produce output. Used for plugins. Defaults to the builtin console formatterBenchee.Formatters.Console
. See Formatters.pre_check
- whether or not to run each job with each input - including all given before or after scenario or each hooks - before the benchmarks are measured to ensure that your code executes without error. This can save time while developing your suites. Defaults tofalse
.parallel
- each the function of each job will be executed inparallel
number processes. Ifparallel
is4
then 4 processes will be spawned that all execute the same function for the given time. When these finish/the time is up 4 new processes will be spawned for the next job/function. This gives you more data in the same time, but also puts a load on the system interfering with benchmark results. For more on the pros and cons of parallel benchmarking check the wiki. Defaults to 1 (no parallel execution).save
- specify apath
where to store the results of the current benchmarking suite, tagged with the specifiedtag
.load
- load saved suit or suits to compare your current benchmarks against. Can be a string or a list of strings or patterns.print
- a map from atoms totrue
orfalse
to configure if the output identified by the atom will be printed. All options are enabled by default (true). Options are::benchmarking
- print when Benchee starts benchmarking a new job (Benchmarking name ..) as well as when statistics are being calculated or formatting begins.:configuration
- a summary of configured benchmarking options including estimated total run time is printed before benchmarking starts:fast_warning
- warnings are displayed if functions are executed too fast leading to inaccurate measures
console
- options for the built-in console formatter::comparison
- if the comparison of the different benchmarking jobs (x times slower than) is shown (true/false). Enabled by default.extended_statistics
- display more statistics, akaminimum
,maximum
,sample_size
andmode
. Disabled by default.
percentiles
- if you are using extended statistics and want to see the results for certain percentiles of results beyond just the median. Defaults to [50, 99] to calculate the 50th and 99th percentiles.:unit_scaling
- the strategy for choosing a unit for durations and counts. May or may not be implemented by a given formatter (The console formatter implements it). When scaling a value, Benchee finds the "best fit" unit (the largest unit for which the result is at least 1). For example, 1_200_000 scales to1.2 M
, while800_000
scales to800 K
. Theunit_scaling
strategy determines how Benchee chooses the best fit unit for an entire list of values, when the individual values in the list may have different best fit units. There are four strategies, defaulting to:best
::best
- the most frequent best fit unit will be used, a tie will result in the larger unit being selected.:largest
- the largest best fit unit will be used (i.e. thousand and seconds if values are large enough).:smallest
- the smallest best fit unit will be used (i.e. millisecond and one):none
- no unit scaling will occur. Durations will be displayed in microseconds, and counts will be displayed in ones (this is equivalent to the behaviour Benchee had pre 0.5.0)
:before_scenario
/after_scenario
/before_each
/after_each
- read up on them in the hooks section in the README:measure_function_call_overhead
- Measure how long an empty function call takes and deduct this from each measure run time. This overhead should be negligible for all but the most micro benchmarks. Defaults to false.profile_after
- accepts any of the following options:- a boolean -
true
will enable profiling with the default profiler (:eprof
) andfalse
will disable profiling. Defaults tofalse
. - a profiler - either as a tuple of
{profiler, opts}
(e.g.,{:fprof, [sort: :own]}
) or just the profiler (e.g.,:fprof
), which is equivalent to{profiler, []}
. The accepted built-in profilers are:cprof
,:eprof
and:fprof
.
- a boolean -
Functions
@spec init(user_configuration()) :: Benchee.Suite.t()
Returns the initial benchmark configuration for Benchee, composed of defaults and an optional custom configuration.
Configuration times are given in seconds, but are converted to microseconds internally.
For a list of all possible options see user_configuration/0
Examples
iex> init()
%Benchee.Suite{
configuration:
%Benchee.Configuration{
parallel: 1,
time: 5_000_000_000.0,
warmup: 2_000_000_000.0,
inputs: nil,
input_names: [],
save: false,
load: false,
formatters: [Benchee.Formatters.Console],
print: %{
benchmarking: true,
fast_warning: true,
configuration: true
},
percentiles: [50, 99],
unit_scaling: :best,
assigns: %{},
before_each: nil,
after_each: nil,
before_scenario: nil,
after_scenario: nil
},
system: nil,
scenarios: []
}
iex> init(time: 1, warmup: 0.2)
%Benchee.Suite{
configuration:
%Benchee.Configuration{
parallel: 1,
time: 1_000_000_000.0,
warmup: 200_000_000.0,
inputs: nil,
input_names: [],
save: false,
load: false,
formatters: [Benchee.Formatters.Console],
print: %{
benchmarking: true,
fast_warning: true,
configuration: true
},
percentiles: [50, 99],
unit_scaling: :best,
assigns: %{},
before_each: nil,
after_each: nil,
before_scenario: nil,
after_scenario: nil
},
system: nil,
scenarios: []
}
iex> init(
...> parallel: 2,
...> time: 1,
...> warmup: 0.2,
...> formatters: [&IO.puts/1],
...> print: [fast_warning: false],
...> inputs: %{"Small" => 5, "Big" => 9999},
...> unit_scaling: :smallest)
%Benchee.Suite{
configuration:
%Benchee.Configuration{
parallel: 2,
time: 1_000_000_000.0,
warmup: 200_000_000.0,
inputs: [{"Big", 9999}, {"Small", 5}],
input_names: ["Big", "Small"],
save: false,
load: false,
formatters: [&IO.puts/1],
print: %{
benchmarking: true,
fast_warning: false,
configuration: true
},
percentiles: [50, 99],
unit_scaling: :smallest,
assigns: %{},
before_each: nil,
after_each: nil,
before_scenario: nil,
after_scenario: nil
},
system: nil,
scenarios: []
}