Home

Awesome

Warning

OpenCensus and OpenTracing have merged to form OpenTelemetry, which serves as the next major version of OpenCensus and OpenTracing.

OpenTelemetry has now reached feature parity with OpenCensus, with tracing and metrics SDKs available in .NET, Golang, Java, NodeJS, and Python. All OpenCensus Github repositories, except census-instrumentation/opencensus-python, will be archived on July 31st, 2023. We encourage users to migrate to OpenTelemetry by this date.

To help you gradually migrate your instrumentation to OpenTelemetry, bridges are available in Java, Go, Python, and JS. Read the full blog post to learn more.

OpenCensus Erlang library

Version: 0.5.0

Erlang stats collection and distributed tracing framework

CircleCI codecov Hex.pm Hex.pm

<a name="Using_with_Rebar3_project">Using with Rebar3 project</a>

Add as dependency to rebar.config:

{deps, [opencensus]}.

Or to use the latest from git master branch:

{deps, [{opencensus, {git, "https://github.com/census-instrumentation/opencensus-erlang.git", {branch, "master"}}}]}.

<a name="Tags">Tags</a>

Tags represent propagated key-value pairs. They are propagated using the pdict or Ctx record in the same process and can be encoded to be transmitted on the wire.

ocp:update_tags(#{http_server_method => "GET"}).

<a name="Tracing">Tracing</a>

<a name="Creating_Spans">Creating Spans</a>

Span data is stored and manipulated in an ETS table. Span context must be tracked in order to create child span's, propagate span context across process or node boundaries and for the library to which span data to manipulate in the context it is called.

opencensus provides two methods for tracking this context, the process dictionary and a variable holding a ctx record.

With ocp the span context is tracked in the current process`s process dictionary.

some_fun() ->
  ocp:with_child_span(<<"some_fun/0">>,
                      fun() ->
                          ... body ..
                      end).

More details on working with spans can be found here and in the modules documentation for ocp, oc_trace and oc_span.

<a name="Propagating_Span_Context">Propagating Span Context</a>

Builtin support for encoding and decoding span context from HTTP headers comes in two formats:

Additionally oc_propagation_binary will encode and decode a binary form used for GRPC and other binary protocols.

<a name="Samplers">Samplers</a>

oc_sampler_never: Never enable a new trace, but keeps a trace enabled if its propagated context is enabled.

oc_sampler_always: Enable every new trace for sampling.

oc_sampler_probability: Takes a probability, default 0.5, that any new trace will be sampled.

<a name="Reporters">Reporters</a>

Zipkin: Zipkin v2 reporter.

Google Cloud Trace: Support for v1 in master, v2 and grpc coming soon;

Prometheus: Exports spans as Prometheus metrics.

DataDog: Export spans to DataDog APM

<a name="Sweeper">Cleaning Up Abandoned Spans</a>

Active spans have their data stored in an ETS table. When a span is finished it is removed from the active spans table and moved to a table handled by the reporter process. If a span isn't finished, either because of a mistake in the code creating and finishing spans, or the process with open spans crashes before being able to finish the spans, there would be a memory leak.

The oc_span_sweeper process checks for active spans which started greater than a configurable (span_ttl) duration, with a default of 5 minutes. There are 4 strategies for handling a span that is older than the time to live (strategy):

An example configuration in sys.config to run a check every 5 minutes, dropping active spans older than 5 minutes can be found in the example project helloworld, examples/helloworld/config/sys.config, the sweeper snippet looks like:

{sweeper, #{interval => 300000,
            strategy => drop,
            span_ttl => 300000}}

To disable sweeping set interval to infinity.

<a name="Logging">Logging</a>

OTP-21 includes a new logging framework. When a context is created with a span (for example ocp:with_child_span/1 or oc_trace:with_child_span/2) opencensus will update the current process's logger metadata to include the trace_id, span_id and trace_options with the latest ids under the key span_ctx, trace_options will be 1 if the trace is enabled. To use these with the default formatter you can create a custom template that includes them if they exist like so:

{logger_formatter,
  #{template => [time, " ", pid, " ",
                 {[span_ctx, trace_id], ["trace_id=", [span_ctx, trace_id], " "], []},
                 {[span_ctx, span_id], ["span_id=", [span_ctx, span_id], " "], []},
                 {[span_ctx, trace_options], ["trace_options=", [span_ctx, trace_options], " "], []},
                 msg, "\n"]}}

<a name="Stats">Stats</a>

OpenCensus stats collection happens in two stages:

<a name="Defining_Measures">Defining Measures</a>

oc_stat_measure:new('opencensus.io/http/server/server_latency', "Time between first byte of "
                    "request headers read to last byte of response sent, or terminal error.",
                    milli_seconds).

<a name="Recording">Recording</a>

Measurements are data points associated with a measure. ocp:record implicitly tags the set of Measurements with the tags from the pdict context:

ocp:record('opencensus.io/http/server/server_latency', ServerLatency),

<a name="Views">Views</a>

Views are how Measures are aggregated. You can think of them as queries over the set of recorded data points (measurements).

Views have two parts: the tags to group by and the aggregation type used.

Currently four types of aggregations are supported:

Here we create a view with the distribution aggregation over our measure:

oc_stat_view:subscribe(#{name => "opencensus.io/http/server/server_latency",
                         description => "Latency distribution of HTTP requests",
                         tags => [http_server_method, http_server_path],
                         measure => 'opencensus.io/http/server/server_blatency',
                         aggregation => {oc_stat_aggregation_distribution, 
                                         [{buckets, [0, 100, 300, 650, 800, 1000]}]}})

<a name="Stat Exporters">Stat Exporters</a>

Prometheus: Exports stat views as Prometheus metrics. Simply register as a collector:

prometheus_registry:register_collector(oc_stat_exporter_prometheus)

DogStatsD: Export stat views as DataDog metrics.

Development

$ rebar3 compile

Running tests:

$ rebar3 ct