Blackhole is an attribute-based logger with strong focus on gaining maximum performance as possible for such kind of loggers.
Attributes is the core feature of Blackhole. Technically speaking it's a key-value pairs escorting every logging record.
For example we have HTTP/1.1 server which produces access logs like:
[::] - esafronov [10/Oct/2000:13:55:36 -0700] 'GET /porn.png HTTP/1.0' 200 2326 - SUCCESS
It can be splitted into indexes or attributes:
message: SUCCESS
host: [::]
user: esafronov
timestamp: 10/Oct/2000:13:55:36 -0700
method: GET
uri: /porn.png
protocol: HTTP/1.0
status: 200
elapsed: 2326
Blackhole allows to specify any number of attributes you want, providing an ability to work with them before of while you writing them into its final destination. For example, Elasticsearch.
Despite the header-only dark past now Blackhole is developed as a shared library. Such a radical change of distributing process was chosen because of many reasons.
Mainly, header-only libraries have one big disadvantage: any code change may (or not) result in recompiling all its dependencies, otherwise having weird runtime errors with symbol loading race.
The other reason was the personal aim to reduce compile time, because it was fucking huge!
Of course, there are disadvantages, such as virtual function call cost and closed doors for inlining, but here my personal benchmark-driven development helped to avoid performance degradation.
Note, that there are some symbols, that are wrapped into experimental
namespace. These symbols don't adhere semantic versioning and, well... experimental. Use them with caution and only, where you want to try super-unstable features, which can be changed or even dropped.
Formatters in Blackhole are responsible for converting every log record passing into some byte array representation. It can be either human-readable string, JSON tree or even protobuf packed frame.
String formatter provides an ability to configure your logging output using pattern mechanics with powerful customization support.
Unlike previous Blackhole versions now string formatter uses python-like syntax for describing patterns with using {}
placeholders and format specifications inside. Moreover now you can specify timestamp specification directly inside the
general pattern or even format it as a microseconds number since epoch.
For example we have the given pattern:
[{severity:>7}] [{timestamp:{%Y-%m-%d %H:%M:%S.%f}s}] {scope}: {message}
After applying some log events we expect to receive something like this:
[ DEBUG] [2015-11-19 19:02:30.836222] accept: HTTP/1.1 GET - / - 200, 4238
[ INFO] [2015-11-19 19:02:32.106331] config: server has reload its config in 200 ms
[WARNING] [2015-11-19 19:03:12.176262] accept: HTTP/1.1 GET - /info - 404, 829
[ ERROR] [2015-11-19 19:03:12.002127] accept: HTTP/1.1 GET - /info - 503, 829
As you may notice the severity field is aligned to the right border (see that >7 spec in pattern), the timestamp is formatted using default representation with a microseconds extension and so on. Because Blackhole is all about attributes you can place and format every custom attribute you want, as we just done with scope attribute.
The Blackhole supports several predefined attributes, with convenient specifications:
Placeholder | Description |
---|---|
{severity:s} | User provided severity string representation |
{severity}, {severity:d} | Numeric severity value |
{timestamp:d} | Number of microseconds since Unix epoch |
{timestamp:{spec}s} | String representation using strftime specification in UTC |
{timestamp:{spec}l} | String representation using strftime specification in local timezone |
{timestamp}, {timestamp:s} | The same as {timestamp:{%Y-%m-%d %H:%M:%S.%f}s} |
{process:s} | Process name |
{process}, {process:d} | PID |
{thread}, {thread::x} | Thread hex id as an opaque value returned by pthread_self(3) |
{thread:s} | Thread name or unnamed |
{message} | Logging message |
{...} | All user declared attributes |
For more information please read the documentation and visit the following links:
Note, that if you need to include a brace character in the literal text, it can be escaped by doubling: {{
and }}
.
There is a special attribute placeholder - {...}
- which means to print all non-reserved attributes in a reverse order they were provided in a key-value manner separated by a comma. These kind of attributes can be configured using special syntax, similar with the timestamp attribute with an optional separator.
For example the following placeholder {...:{{name}={value}:p}{\t:x}s}
results in tab separated key-value pairs like id=42\tmethod=GET
.
For pedants there is a full placeholder grammar in EBNF:
Grammar = Ph
| OptPh
| VarPh
Ph = "{" Name "}"
OptPh = "{" Name ":" Spec? "}"
VarPh = "{...}"
| "{...:" Ext? s "}"
Ext = Pat
| Sep
| Pat Sep
| Sep Pat
Name = [a-zA-Z0-9_]
Spec = Fill? Align? Width? Type
Fill = [a character other than '{' or '}']
Align = [>^<]
Width = [1-9][0-9]*
Type = [su]
Pat = "{" PatSpec ":p}"
Sep = "{" SepLit* ":s}" ("}" SepLit* ":s}")*
SepLit = . ! (":s" | "}" | "}}" | "{" | "{{")
| LeBrace
| RiBrace
LeBrace = "{{" -> "{"
RiBrace = "}}" -> "}"
PatSpec = (AtName | AtValue | PatLit)*
AtName = "{name}"
| AtNameSpec
AtNameSpec = "{name:" AtSpec "}"
AtSpec = Align? Width? AtType
AtType = [sd]
AtValue = "{value}"
| AtValueSpec
AtValueSpec = "{value:" AtSpec "}"
PatLit = . ! ("}" | "}}" | "{" | "{{")
| LeBrace
| RiBrace
Let's describe it more precisely. Given a complex leftover placeholder, let's parse it manually to see what Blackhole see.
Given: {...:{{name}={value}:p}{\t:s}>50s}
.
Parameter | Description |
---|---|
... | Reserved placeholder name indicating for Blackhole that this is a leftover placeholder. |
: | Optional spec marker that is placed after placeholder name where you want to apply one of several extensions. There are pattern, separator, prefix, suffix and format extensions. All of them except format should be surrounded in curly braces. |
{{name}={value}:p} | Pattern extension that describes how each attribute should be formatted using typical Blackhole notation. The suffix :p, that is required for extension identification, means pattern. Inside this pattern you can write any pattern you like using two available sub-placeholders for attribute name and value, for each of them a format spec can be applied using cppformat grammar. At last a format spec can be also applied to the entire placeholder, i.e. :>50p for example. |
{\t:s} | Separator extension for configuring each key-value pair separation. Nuff said. |
{[:r} | (Not implemented yet). Prefix extension that is prepended before entire result if it is not empty. |
{]:u} | (Not implemented yet). Suffix extension that is appended after entire result if it is not empty. |
>50s | Entire result format. See cppformat rules for specification. |
JSON formatter provides an ability to format a logging record into a structured JSON tree with attribute handling features, like renaming, routing, mutating and much more.
Briefly using JSON formatter allows to build fully dynamic JSON trees for its further processing with various external tools, like logstash or rsyslog lefting it, however, in a human-readable manner.
Blackhole allows you to control of JSON tree building process using several predefined options.
Without options it will produce just a plain tree with zero level depth. For example for a log record with a severity of 3, message "fatal error, please try again" and a pair of attributes {"key": 42, "ip": "[::]"}
the result string will look like:
{
"message": "fatal error, please try again",
"severity": 3,
"timestamp": 1449859055,
"process": 12345,
"thread": 57005,
"key": 42,
"ip": "[::]"
}
Using configuration parameters for this formatter you can:
Attributes renaming acts so much transparently as it appears: it just renames the given attribute name using the specified alternative.
Attributes routing specifies a location where the listed attributes will be placed at the tree construction. Also, you can specify a default location for all attributes, which is "/" meaning root otherwise.
For example with routing {"/fields": ["message", "severity"]}
and "/" as a default pointer the mentioned JSON will look like:
{
"fields": {
"message": "fatal error, please try again",
"severity": 3
},
"timestamp": 1449859055,
"process": 12345,
"thread": 57005,
"key": 42,
"ip": "[::]"
}
Attribute renaming occurs after routing, so mapping "message" => "#message" just replaces the old name with its new alternative.
To gain maximum speed at the tree construction no filtering occurs, so this formatter by default allows duplicated keys, which means invalid JSON tree (but most of parsers are fine with it). If you are really required to deal with unique keys, you can enable unique
option, but it involves heap allocation and may slow down formatting.
Also, formatter allows to automatically append a newline character at the end of the tree, which is strangely required by some consumers, like logstash.
Note, that JSON formatter formats the tree using compact style without excess spaces, tabs etc.
For convenient formatter construction a special builder class is implemented allowing to create and configure instances of this class using streaming API. For example:
auto formatter = blackhole::formatter::json_t::builder_t()
.route("/fields", {"message", "severity", "timestamp"})
.route("/other")
.rename("message", "#message")
.rename("timestamp", "#timestamp")
.newline()
.unique()
.build();
This allows to avoid hundreds of constructors and to make a formatter creation to look eye-candy.
The full table of options:
Option | Type | Description |
---|---|---|
/route | Object of: [string] "*" |
Allows to configure nested tree mapping. Each key must satisfy JSON Pointer specification and sets new attributes location in the tree. Values must be either an array of string, meaning list of attributes that are configured with new place or an "*" literal, meaning all other attributes. |
/mapping | Object of: [string] | Simple attribute names renaming from key to value. |
/newline | bool | If true, a newline will be appended to the end of the result message. The default is false. |
/unique | bool | If true removes all backward consecutive duplicate elements from the attribute list range. For example, if there are two attributes with name "name" and values "v1" and "v2" inserted, then after filtering there will be only the last inserted, i.e. "v2". The default is false. |
/mutate/timestamp | string | Replaces the timestamp field with new value by transforming it with the given strftime pattern. |
/mutate/severity | [string] | Replaces the severity field with the string value at the current severity value. |
For example:
"formatter": {
"type": "json",
"newline": true,
"unique": true,
"mapping": {
"message": "@message",
"timestamp": "@timestamp"
},
"routing": {
"": ["message", "timestamp"],
"/fields": "*"
},
"mutate": {
"timestamp": "%Y-%m-%dT%H:%M:%S.%fZ",
"severity": ["D", "I", "W", "E"]
}
}
Sometimes we need to just drop all logging events no matter what, for example to benchmarking purposes. For these cases, there is null output (or sink), which just ignores all records.
The common configuration for this sink looks like:
"sinks": [
{
"type": "null"
}
]
Represents a console sink which is responsible for writing all incoming log events directly into the terminal using one of the selected standard outputs with an ability to optionally colorize result strings.
The sink automatically detects whether the destination stream is a TTY disabling colored output otherwise, which makes possible to redirect standard output to file without escaping codes garbage.
Note, that despite of C++ std::cout
and std::cerr
thread-safety with no undefined behavior its guarantees is insufficiently for safe working with them from multiple threads, leading to result messages intermixing. To avoid this a global mutex is used internally, which is kinda hack. Any other stdout/stderr usage outside from logger will probably results in character mixing, but no undefined behavior will be invoked.
The configuration:
"sinks": [
{
"type": "console"
}
]
Note, that currently coloring cannot be configured through dynamic factory (i.e through JSON, YAML etc.), but can be through the builder.
enum severity {
debug = 0,
info,
warn,
error
};
auto console = blackhole::builder<blackhole::sink::console_t>()
.colorize(severity::debug, blackhole::termcolor_t())
.colorize(severity::info, blackhole::termcolor_t::blue())
.colorize(severity::warn, blackhole::termcolor_t::yellow())
.colorize(severity::error, blackhole::termcolor_t::red())
.stdout()
.build();
Represents a sink that writes formatted log events to the file or files located at the specified path.
The path can contain attribute placeholders, meaning that the real destination name will be deduced at runtime using provided log record (not ready yet). No real file will be opened at construction time. All files are opened by default in append mode meaning seek to the end of stream immediately after open.
This sink supports custom flushing policies, allowing to control hardware write load. There are three implemented policies right now:
Note, that it's guaranteed that the sink always flush its buffers at destruction time. This guarantee with conjunction of thread-safe logger reassignment allows to implement common SIGHUP files reopening during log rotation.
Blackhole won't create intermediate directories, because of potential troubles with ACL. Instead an exception will be thrown, which will be anyway caught by the internal logging system notifying through stdout about it.
Note, that associated files will be opened on demand during the first write operation.
"sinks": [
{
"type": "file",
"flush": "10MB",
"path": "/var/log/blackhole.log"
}
]
Blackhole knows about the following marginal binary units:
More you can read at https://en.wikipedia.org/wiki/Binary_prefix.
The socket sinks category contains sinks that write their output to a remote destination specified by a host and port. Currently the data can be sent over either TCP or UDP.
This appender emits formatted logging events using connected TCP socket.
Option | Type | Description |
---|---|---|
host | string | Required. The name or address of the system that is listening for log events. |
port | u16 | Required. The port on the host that is listening for log events. |
Nuff said.
Option | Type | Description |
---|---|---|
priorities | [i16] | Required. Priority mapping from severity number. |
Blackhole can be configured mainly in two ways:
The first way involves using experimental yet builder. For each library component (formatter, sink, etc.) there should be appropriate builder specialization that is used to create instances of associated component in a flow-way.
For example:
// Here we are going to configure our string/console handler and to build the logger.
auto log = blackhole::experimental::partial_builder<blackhole::root_logger_t>()
// Add the blocking handler.
.handler<blackhole::handler::blocking_t>()
// Configure string formatter.
//
// Pattern syntax behaves like as usual substitution for placeholder. For example if
// the attribute named `severity` has value `2`, then pattern `{severity}` will invoke
// severity mapping function provided and the result will be `W`.
.set<blackhole::formatter::string_t>("{severity}, [{timestamp}]: {message}")
.mapping(&sevmap)
.build()
// Configure console sink to write into stdout (also stderr can be configured).
.add<blackhole::sink::console_t>()
.build()
// And build the handler. Multiple handlers can be added to a single logger, but right
// now we confine ourselves with a single handler.
.build()
// Build the logger.
.build();
The result is a std::unique_ptr<C> where C: Component
, sorry for my Rust.
This is also called static initialization, because you must know the configuration of your logging system at compile time. If this isn't suit for you there is another way.
Also called as dynamic initialization, and is the recommended way to configure the Blackhole, because it implements some kind of dependency injection through some external source, like JSON file, XML, or folly::dynamic
.
Blackhole for now implements only initialization from JSON, but it can be easily extended as a plugin, because all you need is just to implement proper interface to allow tree-like traversing through your config object.
Here there is an example how to configure the library from JSON file.
// Here we are going to build the logger using registry. The registry's responsibility is to
// track registered handlers, formatter and sinks, but for now we're not going to register
// anything else, since there are predefined types.
auto log = blackhole::registry::configured()
// Specify the concrete builder type we want to use. It may be JSON, XML, YAML or whatever
// else.
->builder<blackhole::config::json_t>(std::ifstream(argv[1]))
// Build the logger named "root".
.build("root");
The result is a std::unique_ptr<logger_t>
object.
For more information see blackhole::registry_t class and the include/blackhole/config where all magic happens. If you look for an example how to implement your own factory, please see src/config directory.
One can say that the raw logger interface is inconvenient, and this is true, unfortunately, because it must work both in simple cases, where intermediate message formatting is not required, without attributes; and in complex cases, where lazy message formatting occurs, with attributes provided, remaining at the same time as fast as possible, giving a high-performance solution.
Let's take a look on the interface:
class logger_t {
public:
virtual ~logger_t() = 0;
virtual auto log(severity_t severity, const message_t& message) -> void = 0;
virtual auto log(severity_t severity, const message_t& message, attribute_pack& pack) -> void = 0;
virtual auto log(severity_t severity, const lazy_message_t& message, attribute_pack& pack) -> void = 0;
virtual auto manager() -> scope::manager_t& = 0;
};
To avoid manually creating all these structures a special extension is provided: facade. In two words it is a thin template adapter over any given logger which extends its interface, providing methods that makes logging convenient again. We describe all these methods by abusing a random HTTP logging event of success file serve.
For simple cases, there is a thin wrapper that transforms a string into string view and passes it further.
logger.log(0, "GET /static/image.png HTTP/1.1 436 200");
Sometimes we want to provide additional attributes. In these cases, they can be passed using initializer list.
logger.log(0, "GET /static/image.png HTTP/1.1 436 200", {
{"cache", true},
{"elapsed", 435.72},
{"user-agent", "Mozilla Firefox"}
});
Often we want to format a message using predefined pattern, but with arguments obtained at runtime.
logger.log(0, "{} {} HTTP/1.1 {} {}", "GET", "/static/image.png", 436, 200);
At last, we can combine two previous examples to obtain something really useful. Note that attribute list argument must be the last.
logger.log(0, "{} {} HTTP/1.1 {} {}", "GET", "/static/image.png", 436, 200, attribute_list{
{"cache", true},
{"elapsed", 435.72},
{"user-agent", "Mozilla Firefox"}
});
To use it all you need is to create a logger, import the facade definition and wrap the logger with it. We show you an improved example:
/// This example demonstrates how to initialize Blackhole from configuration file using JSON
/// builder.
/// In this case the entire logging pipeline is initialized from file including severity mapping.
/// The logging facade is used to allow runtime formatting and attributes provisioning.
#include <fstream>
#include <iostream>
#include <blackhole/attribute.hpp>
#include <blackhole/attributes.hpp>
#include <blackhole/config/json.hpp>
#include <blackhole/extensions/facade.hpp>
#include <blackhole/extensions/writer.hpp>
#include <blackhole/registry.hpp>
#include <blackhole/root.hpp>
using namespace blackhole;
/// As always specify severity enumeration.
enum severity {
debug = 0,
info = 1,
warning = 2,
error = 3
};
auto main(int argc, char** argv) -> int {
if (argc != 2) {
std::cerr << "Usage: 3.config PATH" << std::endl;
return 1;
}
/// Here we are going to build the logger using registry. The registry's responsibility is to
/// track registered handlers, formatter and sinks, but for now we're not going to register
/// anything else, since there are predefined types.
auto inner = blackhole::registry::configured()
/// Specify the concrete builder type we want to use. It may be JSON, XML, YAML or whatever
/// else.
->builder<blackhole::config::json_t>(std::ifstream(argv[1]))
/// Build the logger named "root".
.build("root");
/// Wrap the logger with facade to obtain an ability to format messages and provide attributes.
auto log = blackhole::logger_facade<blackhole::root_logger_t>(inner);
log.log(severity::debug, "{} {} HTTP/1.1 {} {}", "GET", "/static/image.png", 404, 347);
log.log(severity::info, "nginx/1.6 configured", {
{"elapsed", 32.5}
});
log.log(severity::warning, "client stopped connection before send body completed");
log.log(severity::error, "file does not exist: {}", "/var/www/favicon.ico", blackhole::attribute_list{
{"Cache", true},
{"Cache-Duration", 10},
{"User-Agent", "Mozilla Firefox"}
});
return 0;
}
The library can be successfully compiled and used without RTTI (with -fno-rtti flag).
gmtime
- manual std::tm
generation without mutex shit.That's the first question I ask myself when seeing yet another silver-bullet library.
First of all, we required a logger with attributes support. Here boost::log
was fine, but it didn't compile in our compilers. Sad. After that we've realized that one of our bottlenecks is located in logging part, that's why boost::log
and log4cxx
weren't fit in our requirements. Thirdly we are developing for stable, but old linux distributives with relatively old compilers that supports only basic part of C++11.
At last, but not least, all that libraries have one fatal disadvantage - NIH.
So here we are.
To be honest, let's describe some popular logging libraries, its advantages and disadvantages as one of them may fit your requirements and you may want to use them instead. It's okay.
Developed by another crazy Russian programmer using dark template magic and Vodka (not sure what was first). It's a perfect and powerful library, seriously.
Pros:
apt-get install
instead.Cons:
log4cxx
have, but not enough for us.Logging framework for C++ patterned after Apache log4j. Yeah, Java.
Pros:
Cons:
Extremely ultra bloody fucking fast logging library. At least the documentation says that. Faster than speed of light!
But everyone knows that even the light is unable to leave from blackhole.
Pros:
Cons:
First of all, the entire library was completely rewritten for performance reasons.
boost::format
into the Hell. It's hard to find a slower library for formatting both in compilation stage and runtime. Instead, the perfect cppformat library with an own compile-time constexpr extensions is used.message
, severity
, timestmap
etc.Each feature and fix is developed in a separate branch. Bugs which are discovered during development of a certain feature, may be fixed in the same branch as their parent issue. This is also true for small features.
master
: master branch - contains a stable, working version of VM code.develop
: development branch - all fixes and features are first merged here.issue/<number>/<slug>
or issue/<slug>
: for issues (both enhancement and bug fixes).