Erlang/OTP provides a standard API for logging. The backend of this API can be used as is, or it can be customized to suite specific needs.
It consists of two parts - the logger part and the handler part. The logger part forwards log events to one or more handler(s).
Filters can be added to the logger part and to each handler. The filters decide if an event is to be forwarded or not, and they can also modify all parts of the log event.
A formatter can be set for each handler. The formatter does the final formatting of the log event, including the log message itself, and possibly a timestamp, header and other metadata.
In accordance with the Syslog protocol, RFC-5424, eight severity levels can be specified:
A log event is allowed by Logger if the integer value of
its
A handler is defined as a module exporting the following function:
log(Log, Config) -> ok
The handler callback is called after filtering on logger level and on handler level for the handler in question. The function call is done on the client process, and it is up to the handler implementation if other processes are to be involved or not.
Multiple instances of the same handler can be added. Configuration is per instance.
Filters can be set on the logger part, or on a handler. Logger filters are applied first, and if passed, the handler filters for each handler are applied. The handler callback is only called if all handler filters for the handler in question also pass.
A filter is specified as:
{fun((Log,Extra) -> Log | stop | ignore), Extra}
The configuration parameter
The
A formatter is defined as a module exporting the following function:
format(Log,Extra) -> unicode:chardata()
The formatter callback is called by each handler, and the returned string can be printed to the handler's destination (stdout, file, ...).
This is the default handler used by OTP. Multiple instances can be started, and each instance will write log events to a given destination, console or file. Filters can be used for selecting which event to send to which handler instance.
This handler behaves much like logger_std_h, except it uses
This handler is to be used for backwards compatibility
only. It is not started by default, but will be automatically
started the first time an event handler is added
with
No built-in event handlers exist.
This filter provides a way of filtering log events based on a
This filter provides a way of filtering log events based
on the log level. See
This filter matches all progress reports
from
This filter matches all events originating from a process
that has its group leader on a remote node.
See
The default formatter is
Logger can be configured either when the system starts through
Logger is best configured by using the configuration parameters
of Kernel. There are four possible configuration parameters:
logger
The application configuration parameter
Disable the default handler. This allows another application
to add its own default handler. See
Only one entry of this option is allowed.
Add a handler as if
It is allowed to have multiple entries of this option.
Add the specified
Only one entry of this option is allowed.
This option configures
It is allowed to have multiple entries of this option.
Examples:
Output logs into the file "logs/erlang.log"
[{kernel,
[{logger,
[{handler, default, logger_std_h,
#{ logger_std_h => #{ type => {file,"log/erlang.log"}}}}]}]}].
Output logs in single line format
[{kernel,
[{logger,
[{handler, default, logger_std_h,
#{ formatter => { logger_formatter,#{ single_line => true}}}}]}]}].
Add the pid to each log event
[{kernel,
[{logger,
[{handler, default, logger_std_h,
#{ formatter => { logger_formatter,
#{ template => [time," ",pid," ",msg,"\n"]}}
}}]}]}].
Use a different file for debug logging
[{kernel,
[{logger,
[{handler, default, logger_std_h,
#{ level => error,
logger_std_h => #{ type => {file, "log/erlang.log"}}}},
{handler, info, logger_std_h,
#{ level => debug,
logger_std_h => #{ type => {file, "log/debug.log"}}}}
]}]}].
Specifies the severity level to log.
Logger filters are added or removed with
See
By default, no filters exist.
Specifies what to do with an event if all filters
return
Default is
Handlers are added or removed with
See
Specifies the severity level to log.
Handler filters can be specified when adding the handler,
or added or removed later with
See
By default, no filters exist.
Specifies what to do with an event if all filters
return
Default is
See
The default module is
Note that
Logger provides backwards compatibility with the old
The old
Calls
to
To get log events on the same format as produced
by
By default, all log events originating from within OTP, except the former so called "SASL reports", look the same as before.
By SASL reports we mean supervisor reports, crash reports and progress reports.
In earlier releases, these reports were only logged when
the SASL application was running, and they were printed
trough specific event handlers
named
The destination of these log events were configured by
Due to the specific event handlers, the output format slightly differed from other log events.
As of OTP-21, the concept of SASL reports is removed, meaning that the default behaviour is as follows:
If the old behaviour is preferred, the Kernel configuation
parameter
All SASL reports have a metadata
field
See the
To use event handlers written for
error_logger:add_report_handler/1,2.
This will automatically start the
#{level=>info,
filter_default=>log,
filters=>[]}.
Notice that this handler will ignore events that do not
originate from the old
Also notice that
Log data is expected to be either a format string and arguments, a string (unicode:chardata), or a report (map or key-value list) which can be converted to a format string and arguments by the handler. A default report callback should be included in the log event's metadata, which can be used for converting the report to a format string and arguments. The handler might also do a custom conversion if the default format is not desired.
Logger does, to a certain extent, check its input data before forwarding a log event to the handlers, but it does not evaluate conversion funs or check the validity of format strings and arguments. This means that any filter or handler must be careful when formatting the data of a log event, making sure that it does not crash due to bad input data or faulty callbacks.
If a filter or handler still crashes, Logger will remove the filter or handler in question from the configuration, and then print a short error message on the console. A debug event containing the crash reason and other details is also issued, and can be seen if a handler is installed which logs on debug level.
When starting an erlang node, the default behaviour is that all
log events with level info and above are logged to the
console. In order to also log debug events, you can either
change the global log level to
First, we add an instance of
1> Config = #{level=>debug,logger_std_h=>#{type=>{file,"./debug.log"}}}. #{logger_std_h => #{type => {file,"./debug.log"}}, level => debug} 2> logger:add_handler(debug_handler,logger_std_h,Config). ok
By default, the handler receives all events
(
3> Fun = fun(#{level:=debug}=Log,_) -> Log; (_,_) -> stop end. #Fun<erl_eval.12.98642416> 4> logger:add_handler_filter(debug_handler,allow_debug,{Fun,[]}). ok
And finally, we need to make sure that Logger itself allows debug events. This can either be done by setting the global log level:
5> logger:set_logger_config(level,debug). ok
Or by allowing debug events from one or a few modules only:
6> logger:set_module_level(mymodule,debug). ok
The only requirement that a handler MUST fulfill is to export the following function:
log(logger:log(),logger:config()) ->ok
It may also implement the following callbacks:
adding_handler(logger:handler_id(),logger:config()) -> {ok,logger:config()} | {error,term()}
removing_handler(logger:handler_id(),logger:config()) -> ok
changing_config(logger:handler_id(),logger:config(),logger:config()) -> {ok,logger:config()} | {error,term()}
When
A handler can be removed by calling
When
A simple handler that prints to the console could be implemented as follows:
-module(myhandler).
-export([log/2]).
log(Log,#{formatter:={FModule,FConfig}) ->
io:put_chars(FModule:format(Log,FConfig)).
A simple handler which prints to file could be implemented like this:
-module(myhandler).
-export([adding_handler/2, removing_handler/2, log/2]).
-export([init/1, handle_call/3, handle_cast/2, terminate/2]).
adding_handler(Id,Config) ->
{ok,Fd} = file:open(File,[append,{encoding,utf8}]),
{ok,Config#{myhandler_fd=>Fd}}.
removing_handler(Id,#{myhandler_fd:=Fd}) ->
_ = file:close(Fd),
ok.
log(Log,#{myhandler_fd:=Fd,formatter:={FModule,FConfig}}) ->
io:put_chars(Fd,FModule:format(Log,FConfig)).
The above handlers do not have any overload protection, and all log events are printed directly from the client process.
For examples of overload protection, please refer to the
implementation
of
Below is a simpler example of a handler which logs through one single process.
-module(myhandler).
-export([adding_handler/2, removing_handler/2, log/2]).
-export([init/1, handle_call/3, handle_cast/2, terminate/2]).
adding_handler(Id,Config) ->
{ok,Pid} = gen_server:start(?MODULE,Config),
{ok,Config#{myhandler_pid=>Pid}}.
removing_handler(Id,#{myhandler_pid:=Pid}) ->
gen_server:stop(Pid).
log(Log,#{myhandler_pid:=Pid} = Config) ->
gen_server:cast(Pid,{log,Log,Config}).
init(#{myhandler_file:=File}) ->
{ok,Fd} = file:open(File,[append,{encoding,utf8}]),
{ok,#{file=>File,fd=>Fd}}.
handle_call(_,_,State) ->
{reply,{error,bad_request},State}.
handle_cast({log,Log,Config},#{fd:=Fd} = State) ->
do_log(Fd,Log,Config),
{noreply,State}.
terminate(Reason,#{fd:=Fd}) ->
_ = file:close(Fd),
ok.
do_log(Fd,Log,#{formatter:={FModule,FConfig}}) ->
String = FModule:format(Log,FConfig),
io:put_chars(Fd,String).
In order for the built-in handlers to survive, and stay responsive,
during periods of high load (i.e. when huge numbers of incoming
log requests must be handled), a mechanism for overload protection
has been implemented in the
The handler process keeps track of the length of its message queue and reacts in different ways depending on the current status. The purpose is to keep the handler in, or (as quickly as possible), get the handler into, a state where it can keep up with the pace of incoming log requests. The memory usage of the handler must never keep growing larger and larger, since that would eventually cause the handler to crash. Three thresholds with associated actions have been defined:
The default value of this level is
When the message queue has grown larger than this threshold, which
defaults to
Above this threshold, which defaults to
For the overload protection algorithm to work properly, it is required that:
and that:
If
During high load scenarios, the length of the handler message queue rarely grows in a linear and predictable way. Instead, whenever the handler process gets scheduled in, it can have an almost arbitrary number of messages waiting in the mailbox. It's for this reason that the overload protection mechanism is focused on acting quickly and quite drastically (such as immediately dropping or flushing messages) as soon as a large queue length is detected.
The thresholds listed above may be modified by the user if, e.g, a handler shouldn't drop or flush messages unless the message queue length grows extremely large. (The handler must be allowed to use large amounts of memory under such circumstances however). Another example of when the user might want to change the settings is if, for performance reasons, the logging processes must never get blocked by synchronous log requests, while dropping or flushing requests is perfectly acceptable (since it doesn't affect the performance of the loggers).
A configuration example:
logger:add_handler(my_standard_h, logger_std_h,
#{logger_std_h =>
#{type => {file,"./system_info.log"},
toggle_sync_qlen => 100,
drop_new_reqs_qlen => 1000,
flush_reqs_qlen => 2000}}).
A potential problem with large bursts of log requests, is that log files may get full or wrapped too quickly (in the latter case overwriting previously logged data that could be of great importance). For this reason, both built-in handlers offer the possibility to set a maximum level of how many requests to process with a certain time frame. With this burst control feature enabled, the handler will take care of bursts of log requests without choking log files, or the console, with massive amounts of printouts. These are the configuration parameters:
This is set to
This is how many requests should be processed within the
The default window is
A configuration example:
logger:add_handler(my_disk_log_h, logger_disk_log_h,
#{disk_log_opts =>
#{file => "./my_disk_log"},
logger_disk_log_h =>
#{burst_limit_size => 10,
burst_window_time => 500}}).
A handler process may grow large even if it can manage peaks of high load without crashing. The overload protection mechanism includes user configurable levels for a maximum allowed message queue length and maximum allowed memory usage. This feature is disabled by default, but can be switched on by means of the following configuration parameters:
This is set to
This is the maximum allowed queue length. If the mailbox grows larger than this, the handler process gets terminated.
This is the maximum allowed memory usage of the handler process. If the handler grows any larger, the process gets terminated.
If the handler gets terminated because of its queue length or
memory usage, it can get automatically restarted again after a
configurable delay time. The time is specified in milliseconds
and