aboutsummaryrefslogblamecommitdiffstats
path: root/HOWTO/BENCHMARKS.md
blob: 28a590dfd2a837e53f0083e6931c44347dc76e58 (plain) (tree)



















                                                                               
                                                                             




                                                                              
                                                                          










































                                                                                

                                                                   
 
Benchmarking Erlang/OTP
=======================

The Erlang/OTP source tree contains a number of benchmarks. The same framework
is used to run these benchmarks as is used to run tests. Therefore in order to
run benchmarks you have to [release the tests][] just as you normally would.

Note that many of these benchmarks were developed to test a specific feature
under a specific setting. We strive to keep the benchmarks up-to-date, but alas
time is not an endless resource so some benchmarks will be outdated and 
irrelevant.

Running the benchmarks
----------------------

As with testing, `ts` is used to run the benchmarks. Before running any 
benchmarks you have to [install the tests][]. To get a listing of all 
benchmarks you have available call `ts:benchmarks()`.

To run all benchmarks call `ts:bench()`. This will run all benchmarks using 
the emulator which is in your `$PATH` (Note that this does not have to be the
same as from which the benchmarks were built from). All the results of the 
benchmarks are put in a folder in `$TESTROOT/test_server/` called 
`YYYY_MO_DDTHH_MI_SS`. 

Each benchmark is run multiple times and the data for all runs is collected in
the files within the benchmark folder. All benchmarks are written so that 
higher values are better. 

Writing benchmarks
------------------

Benchmarks are just normal testcases in Common Test suites. They are marked as
benchmarks by being included in the `AppName_bench.spec` which is located in
`lib/AppName/test/` for the applications which have benchmarks. Note that you
might want to add a skip clause to `AppName.spec` for the benchmarks if you do
not want them to be run in the nightly tests.

Results of benchmarks are sent using the ct_event mechanism and automatically
collected and formatted by ts. 

    ct_event:notify(
        #event{name = benchmark_data, 
               data = [{value,TPS}]}).

The application, suite and testcase associated with the value is automatically
detected. If you want to supply your own you can include `suite` andor `name`
with the data. i.e.

    ct_event:notify(
        #event{name = benchmark_data, 
               data = [{suite,"erts_bench"},
			           {name,"ets_transactions_per_sec"},
			           {value,TPS}]}).
					   
The reason for using the internal ct_event and not ct is because the benchmark 
code has to be backwards compatible with at least R14. 

The value which is reported should be as raw as possible. i.e. you should not
do any averaging of the value before reporting. The tools we use to collect the
benchmark data over time will do averages, means, stddev and more with the data.
So the more data which is sent using `ct_event` the better.

Viewing benchmarks
------------------

At the moment of writing this HOWTO the tool for viewing benchmark results is
not available as opensource. This will hopefully change in the near future.


   [release the tests]: TESTING.md#releasing-tests
   [install the tests]: TESTING.md#configuring-the-test-environment