--- sort: 22 --- # Key concepts ## Data model ### What is a metric Simply put, `metric` is a numeric measure or observation of something. The most common use-cases for metrics are: - check how the system behaves at the particular time period; - correlate behavior changes to other measurements; - observe or forecast trends; - trigger events (alerts) if the metric exceeds a threshold. ### Structure of a metric Let's start with an example. To track how many requests our application serves, we'll define a metric with the name `requests_total`. You can be more specific here by saying `requests_success_total` (for only successful requests) or `request_errors_total` (for requests which failed). Choosing a metric name is very important and supposed to clarify what is actually measured to every person who reads it, just like **variable names** in programming. Every metric can contain additional meta-information in the form of label-value pairs: ``` requests_total{path="/", code="200"} requests_total{path="/", code="403"} ``` The meta-information - a set of `labels` in curly braces - gives us a context for which `path` and with what `code` the `request` was served. Label-value pairs are always of a `string` type. VictoriaMetrics data model is schemaless, which means there is no need to define metric names or their labels in advance. User is free to add or change ingested metrics anytime. Actually, the metric's name is also a label with a special name `__name__`. So the following two series are identical: ``` requests_total{path="/", code="200"} {__name__="requests_total", path="/", code="200"} ``` #### Time series A combination of a metric name and its labels defines a `time series`. For example, `requests_total{path="/", code="200"}` and `requests_total{path="/", code="403"}` are two different time series because they have different values for `code` label. The number of unique time series has an impact on database resource usage. See [what is an active time series](https://docs.victoriametrics.com/FAQ.html#what-is-an-active-time-series) and [what is high churn rate](https://docs.victoriametrics.com/FAQ.html#what-is-high-churn-rate) docs for details. #### Cardinality The number of unique [time series](#time-series) is named `cardinality`. Too big number of unique time series is named `high cardinality`. High cardinality may result in increased resource usage at VictoriaMetrics. See [these docs](https://docs.victoriametrics.com/FAQ.html#what-is-high-cardinality) for more details. #### Raw samples Every unique time series may consist of an arbitrary number of `(value, timestamp)` data points (aka `raw samples`) sorted by `timestamp`. The `value` is a [double-precision floating-point number](https://en.wikipedia.org/wiki/Double-precision_floating-point_format). The `timestamp` is a [Unix timestamp](https://en.wikipedia.org/wiki/Unix_time) with millisecond precision. Below is an example of a single raw sample in [Prometheus text exposition format](https://github.com/prometheus/docs/blob/main/content/docs/instrumenting/exposition_formats.md#text-based-format): ``` requests_total{path="/", code="200"} 123 4567890 ``` - The `requests_total{path="/", code="200"}` identifies the associated time series for the given sample. - The `123` is a sample value. - The `4567890` is an optional timestamp for the sample. If it is missing, then the current timestamp is used when storing the sample in VictoriaMetrics. ### Types of metrics Internally, VictoriaMetrics does not have the notion of a metric type. The concept of a metric type exists specifically to help users to understand how the metric was measured. There are 4 common metric types. #### Counter Counter is a metric, which counts some events. Its value increases or stays the same over time. It cannot decrease in general case. The only exception is e.g. `counter reset`, when the metric resets to zero. The `counter reset` can occur when the service, which exposes the counter, restarts. So, the `counter` metric shows the number of observed events since the service start. In programming, `counter` is a variable that you **increment** each time something happens. `vm_http_requests_total` is a typical example of a counter. The interpretation of a graph above is that time series `vm_http_requests_total{instance="localhost:8428", job="victoriametrics", path="api/v1/query_range"}` was rapidly changing from 1:38 pm to 1:39 pm, then there were no changes until 1:41 pm. Counter is used for measuring the number of events, like the number of requests, errors, logs, messages, etc. The most common [MetricsQL](#metricsql) functions used with counters are: * [rate](https://docs.victoriametrics.com/MetricsQL.html#rate) - calculates the average per-second speed of metric's change. For example, `rate(requests_total)` shows how many requests are served per second on average; * [increase](https://docs.victoriametrics.com/MetricsQL.html#increase) - calculates the growth of a metric on the given time period specified in square brackets. For example, `increase(requests_total[1h])` shows the number of requests served over the last hour. It is OK to have fractional counters. For example, `request_duration_seconds_sum` counter may sum the durations of all the requests. Every duration may have a fractional value in seconds, e.g. `0.5` of a second. So the cumulative sum of all the request durations may be fractional too. It is recommended to put `_total`, `_sum` or `_count` suffix to `counter` metric names, so such metrics can be easily differentiated by humans from other metric types. #### Gauge Gauge is used for measuring a value that can go up and down: The metric `process_resident_memory_anon_bytes` on the graph shows the memory usage of the application at every given time. It is changing frequently, going up and down showing how the process allocates and frees the memory. In programming, `gauge` is a variable to which you **set** a specific value as it changes. Gauge is used in the following scenarios: * measuring temperature, memory usage, disk usage etc; * storing the state of some process. For example, gauge `config_reloaded_successful` can be set to `1` if everything is good, and to `0` if configuration failed to reload; * storing the timestamp when the event happened. For example, `config_last_reload_success_timestamp_seconds` can store the timestamp of the last successful configuration reload. The most common [MetricsQL](#metricsql) functions used with gauges are [aggregation functions](#aggregation-and-grouping-functions) and [rollup functions](https://docs.victoriametrics.com/MetricsQL.html#rollup-functions). #### Histogram Historgram is a set of [counter](#counter) metrics with different `vmrange` or `le` labels. The `vmrange` or `le` labels define measurement boundaries of a particular bucket. When the observed measurement hits a particular bucket, then the corresponding counter is incremented. Histogram buckets usually have `_bucket` suffix in their names. For example, VictoriaMetrics tracks the distribution of rows processed per query with the `vm_rows_read_per_query` histogram. The exposition format for this histogram has the following form: ``` vm_rows_read_per_query_bucket{vmrange="4.084e+02...4.642e+02"} 2 vm_rows_read_per_query_bucket{vmrange="5.275e+02...5.995e+02"} 1 vm_rows_read_per_query_bucket{vmrange="8.799e+02...1.000e+03"} 1 vm_rows_read_per_query_bucket{vmrange="1.468e+03...1.668e+03"} 3 vm_rows_read_per_query_bucket{vmrange="1.896e+03...2.154e+03"} 4 vm_rows_read_per_query_sum 15582 vm_rows_read_per_query_count 11 ``` The `vm_rows_read_per_query_bucket{vmrange="4.084e+02...4.642e+02"} 2` line means that there were 2 queries with the number of rows in the range `(408.4 - 464.2]` since the last VictoriaMetrics start. The counters ending with `_bucket` suffix allow estimating arbitrary percentile for the observed measurement with the help of [histogram_quantile](https://docs.victoriametrics.com/MetricsQL.html#histogram_quantile) function. For example, the following query returns the estimated 99th percentile on the number of rows read per each query during the last hour (see `1h` in square brackets): ```metricsql histogram_quantile(0.99, sum(increase(vm_rows_read_per_query_bucket[1h])) by (vmrange)) ``` This query works in the following way: 1. The `increase(vm_rows_read_per_query_bucket[1h])` calculates per-bucket per-instance number of events over the last hour. 2. The `sum(...) by (vmrange)` calculates per-bucket events by summing per-instance buckets with the same `vmrange` values. 3. The `histogram_quantile(0.99, ...)` calculates 99th percentile over `vmrange` buckets returned at step 2. Histogram metric type exposes two additional counters ending with `_sum` and `_count` suffixes: - the `vm_rows_read_per_query_sum` is a sum of all the observed measurements, e.g. the sum of rows served by all the queries since the last VictoriaMetrics start. - the `vm_rows_read_per_query_count` is the total number of observed events, e.g. the total number of observed queries since the last VictoriaMetrics start. These counters allow calculating the average measurement value on a particular lookbehind window. For example, the following query calculates the average number of rows read per query during the last 5 minutes (see `5m` in square brackets): ```metricsql increase(vm_rows_read_per_query_sum[5m]) / increase(vm_rows_read_per_query_count[5m]) ``` The `vm_rows_read_per_query` histogram may be used in Go application in the following way by using the [github.com/VictoriaMetrics/metrics](https://github.com/VictoriaMetrics/metrics) package: ```go // define the histogram rowsReadPerQuery := metrics.NewHistogram(`vm_rows_read_per_query`) // use the histogram during processing for _, query := range queries { rowsReadPerQuery.Update(float64(len(query.Rows))) } ``` Now let's see what happens each time when `rowsReadPerQuery.Update` is called: * counter `vm_rows_read_per_query_sum` is incremented by value of `len(query.Rows)` expression; * counter `vm_rows_read_per_query_count` increments by 1; * counter `vm_rows_read_per_query_bucket` gets incremented only if observed value is within the range (`bucket`) defined in `vmrange`. Such a combination of `counter` metrics allows plotting [Heatmaps in Grafana](https://grafana.com/docs/grafana/latest/visualizations/heatmap/) and calculating [quantiles](https://prometheus.io/docs/practices/histograms/#quantiles): Grafana doesn't understand buckets with `vmrange` labels, so the [prometheus_buckets](https://docs.victoriametrics.com/MetricsQL.html#prometheus_buckets) function must be used for converting buckets with `vmrange` labels to buckets with `le` labels before building heatmaps in Grafana. Histograms are usually used for measuring the distribution of latency, sizes of elements (batch size, for example) etc. There are two implementations of a histogram supported by VictoriaMetrics: 1. [Prometheus histogram](https://prometheus.io/docs/practices/histograms/). The canonical histogram implementation is supported by most of the [client libraries for metrics instrumentation](https://prometheus.io/docs/instrumenting/clientlibs/). Prometheus histogram requires a user to define ranges (`buckets`) statically. 2. [VictoriaMetrics histogram](https://valyala.medium.com/improving-histogram-usability-for-prometheus-and-grafana-bc7e5df0e350) supported by [VictoriaMetrics/metrics](https://github.com/VictoriaMetrics/metrics) instrumentation library. Victoriametrics histogram automatically handles bucket boundaries, so users don't need to think about them. We recommend reading the following articles before you start using histograms: 1. [Prometheus histogram](https://prometheus.io/docs/concepts/metric_types/#histogram) 2. [Histograms and summaries](https://prometheus.io/docs/practices/histograms/) 3. [How does a Prometheus Histogram work?](https://www.robustperception.io/how-does-a-prometheus-histogram-work) 4. [Improving histogram usability for Prometheus and Grafana](https://valyala.medium.com/improving-histogram-usability-for-prometheus-and-grafana-bc7e5df0e350) #### Summary Summary metric type is quite similar to [histogram](#histogram) and is used for [quantiles](https://prometheus.io/docs/practices/histograms/#quantiles) calculations. The main difference is that calculations are made on the client-side, so metrics exposition format already contains pre-defined quantiles: ``` go_gc_duration_seconds{quantile="0"} 0 go_gc_duration_seconds{quantile="0.25"} 0 go_gc_duration_seconds{quantile="0.5"} 0 go_gc_duration_seconds{quantile="0.75"} 8.0696e-05 go_gc_duration_seconds{quantile="1"} 0.001222168 go_gc_duration_seconds_sum 0.015077078 go_gc_duration_seconds_count 83 ``` The visualisation of summaries is pretty straightforward: Such an approach makes summaries easier to use but also puts significant limitations compared to [histograms](#histogram): - It is impossible to calculate quantile over multiple summary metrics, e.g. `sum(go_gc_duration_seconds{quantile="0.75"})`, `avg(go_gc_duration_seconds{quantile="0.75"})` or `max(go_gc_duration_seconds{quantile="0.75"})` won't return the expected 75th percentile over `go_gc_duration_seconds` metrics collected from multiple instances of the application. See [this article](https://latencytipoftheday.blogspot.de/2014/06/latencytipoftheday-you-cant-average.html) for details. - It is impossible to calculate quantiles other than the already pre-calculated quantiles. - It is impossible to calculate quantiles for measurements collected over an arbitrary time range. Usually, `summary` quantiles are calculated over a fixed time range such as the last 5 minutes. Summaries are usually used for tracking the pre-defined percentiles for latency, sizes of elements (batch size, for example) etc. ### Instrumenting application with metrics As was said at the beginning of the [types of metrics](#types-of-metrics) section, metric type defines how it was measured. VictoriaMetrics TSDB doesn't know about metric types. All it sees are metric names, labels, values, and timestamps. What are these metrics, what do they measure, and how - all this depends on the application which emits them. To instrument your application with metrics compatible with VictoriaMetrics we recommend using [github.com/VictoriaMetrics/metrics](https://github.com/VictoriaMetrics/metrics) package. See more details on how to use it in [this article](https://victoriametrics.medium.com/how-to-monitor-go-applications-with-victoriametrics-c04703110870). VictoriaMetrics is also compatible with [Prometheus client libraries for metrics instrumentation](https://prometheus.io/docs/instrumenting/clientlibs/). #### Naming We recommend following [Prometheus naming convention for metrics](https://prometheus.io/docs/practices/naming/). There are no strict restrictions, so any metric name and labels are accepted by VictoriaMetrics. But the convention helps to keep names meaningful, descriptive, and clear to other people. Following convention is a good practice. #### Labels Every measurement can contain an arbitrary number of `key="value"` labels. The good practice is to keep this number limited. Otherwise, it would be difficult to deal with measurements containing a big number of labels. By default, VictoriaMetrics limits the number of labels per measurement to `30` and drops other labels. This limit can be changed via `-maxLabelsPerTimeseries` command-line flag if necessary (but this isn't recommended). Every label value can contain an arbitrary string value. The good practice is to use short and meaningful label values to describe the attribute of the metric, not to tell the story about it. For example, label-value pair `environment="prod"` is ok, but `log_message="long log message with a lot of details..."` is not ok. By default, VcitoriaMetrics limits label's value size with 16kB. This limit can be changed via `-maxLabelValueLen` command-line flag. It is very important to keep under control the number of unique label values, since every unique label value leads to a new [time series](#time-series). Try to avoid using volatile label values such as session ID or query ID in order to avoid excessive resource usage and database slowdown. ## Write data VictoriaMetrics supports both models used in modern monitoring applications: [push](#push-model) and [pull](#pull-model). ### Push model Client regularly sends the collected metrics to the server in the push model: The client (application) decides when and where to send its metrics. VictoriaMetrics supports the following protocols for data ingestion (aka `push protocols`): * [Prometheus remote write API](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#prometheus-setup). * [Prometheus text exposition format](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#how-to-import-data-in-prometheus-exposition-format). * [DataDog protocol](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#how-to-send-data-from-datadog-agent). * [InfluxDB line protocol](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf) over HTTP, TCP and UDP. * [Graphite plaintext protocol](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#how-to-send-data-from-graphite-compatible-agents-such-as-statsd) with [tags](https://graphite.readthedocs.io/en/latest/tags.html#carbon). * [OpenTSDB put message](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#sending-data-via-telnet-put-protocol). * [HTTP OpenTSDB /api/put requests](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#sending-opentsdb-data-via-http-apiput-requests). * [JSON line format](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#how-to-import-data-in-json-line-format). * [Arbitrary CSV data](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#how-to-import-csv-data). All the protocols are fully compatible with VictoriaMetrics [data model](#data-model) and can be used in production. We recommend using the [github.com/VictoriaMetrics/metrics](https://github.com/VictoriaMetrics/metrics) package for pushing application metrics to VictoriaMetrics. It is also possible to use already existing clients compatible with the protocols listed above like [Telegraf](https://github.com/influxdata/telegraf) for [InfluxDB line protocol](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#how-to-send-data-from-influxdb-compatible-agents-such-as-telegraf). Creating custom clients or instrumenting the application for metrics writing is as easy as sending a POST request: ```console curl -d '{"metric":{"__name__":"foo","job":"node_exporter"},"values":[0,1,2],"timestamps":[1549891472010,1549891487724,1549891503438]}' -X POST 'http://localhost:8428/api/v1/import' ``` It is allowed to push/write metrics to [single-node VictoriaMetrics](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html), to [cluster component vminsert](https://docs.victoriametrics.com/Cluster-VictoriaMetrics.html#architecture-overview) and to [vmagent](https://docs.victoriametrics.com/vmagent.html). The pros of push model: * Simpler configuration at VictoriaMetrics side - there is no need to configure VictoriaMetrics with locations of the monitored applications. There is no need in complex [service discovery schemes](https://docs.victoriametrics.com/sd_configs.html). * Simpler security setup - there is no need to set up access from VictoriaMetrics to each monitored application. See [Foiled by the Firewall: A Tale of Transition From Prometheus to VictoriaMetrics](https://www.percona.com/blog/2020/12/01/foiled-by-the-firewall-a-tale-of-transition-from-prometheus-to-victoriametrics/) elaborating more on why Percona switched from pull to push model. The cons of push protocol: * Increased configuration complexity for monitored applications. Every application needs to be individually configured with the address of the monitoring system for metrics delivery. It also needs to be configured with the interval between metric pushes and the strategy in case of metric delivery failure. * Non-trivial setup for metrics' delivery into multiple monitoring systems. * It may be hard to tell whether the application went down or just stopped sending metrics for a different reason. * Applications can overload the monitoring system by pushing metrics at too short intervals. ### Pull model Pull model is an approach popularized by [Prometheus](https://prometheus.io/), where the monitoring system decides when and where to pull metrics from: In pull model, the monitoring system needs to be aware of all the applications it needs to monitor. The metrics are scraped (pulled) from the known applications (aka `scrape targets`) via HTTP protocol on a regular basis (aka `scrape_interval`). VictoriaMetrics supports discovering Prometheus-compatible targets and scraping metrics from them in the same way as Prometheus does - see [these docs](https://docs.victoriametrics.com/#how-to-scrape-prometheus-exporters-such-as-node-exporter). Metrics scraping is supported by [single-node VictoriaMetrics](https://docs.victoriametrics.com/#how-to-scrape-prometheus-exporters-such-as-node-exporter) and by [vmagent](https://docs.victoriametrics.com/vmagent.html). The pros of the pull model: * Easier to debug - VictoriaMetrics knows about all the monitored applications (aka `scrape targets`). The `up == 0` query instantly shows unavailable scrape targets. The actual information about scrape targets is available at `http://victoriametrics:8428/targets` and `http://vmagent:8429/targets`. * Monitoring system controls the frequency of metrics' scrape, so it is easier to control its load. * Applications aren't aware of the monitoring system and don't need to implement the logic for metrics delivery. The cons of the pull model: * Harder security setup - monitoring system needs to have access to applications it monitors. * Pull model needs non-trivial [service discovery schemes](https://docs.victoriametrics.com/sd_configs.html). ### Common approaches for data collection VictoriaMetrics supports both [push](#push-model) and [pull](#pull-model) models for data collection. Many installations use exclusively one of these models, or both at once. The most common approach for data collection is using both models: In this approach the additional component is used - [vmagent](https://docs.victoriametrics.com/vmagent.html). Vmagent is a lightweight agent whose main purpose is to collect, filter, relabel and deliver metrics to VictoriaMetrics. It supports all [push](#push-model) and [pull](#pull-model) protocols mentioned above. The basic monitoring setup of VictoriaMetrics and vmagent is described in the [example docker-compose manifest](https://github.com/VictoriaMetrics/VictoriaMetrics/tree/master/deployment/docker). In this example vmagent [scrapes a list of targets](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/deployment/docker/prometheus.yml) and [forwards collected data to VictoriaMetrics](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/9d7da130b5a873be334b38c8d8dec702c9e8fac5/deployment/docker/docker-compose.yml#L15). VictoriaMetrics is then used as a [datasource for Grafana](https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/deployment/docker/provisioning/datasources/datasource.yml) installation for querying collected data. VictoriaMetrics components allow building more advanced topologies. For example, vmagents can push metrics from separate datacenters to the central VictoriaMetrics: VictoriaMetrics in this example may be either [single-node VictoriaMetrics](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html) or [VictoriaMetrics Cluster](https://docs.victoriametrics.com/Cluster-VictoriaMetrics.html). Vmagent also allows [replicating the same data to multiple destinations](https://docs.victoriametrics.com/vmagent.html#replication-and-high-availability). ## Query data VictoriaMetrics provides an [HTTP API](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#prometheus-querying-api-usage) for serving read queries. The API is used in various integrations such as [Grafana](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#grafana-setup). The same API is also used by [VMUI](https://docs.victoriametrics.com/Single-server-VictoriaMetrics.html#vmui) - a graphical User Interface for querying and visualizing metrics. The API consists of two main handlers for serving [instant queries](#instant-query) and [range queries](#range-query). ### Instant query Instant query executes the query expression at the given timestamp: ``` GET | POST /api/v1/query?query=...&time=...&step=... ``` Params: * `query` - [MetricsQL](https://docs.victoriametrics.com/MetricsQL.html) expression. * `time` - optional timestamp when to evaluate the `query`. If `time` is skipped, then the current timestamp is used. The `time` param can be specified in [multiple allowed formats](https://docs.victoriametrics.com/#timestamp-formats). * `step` - optional max lookback window for searching for raw samples when executing the `query`. If `step` is skipped, then it is set to `5m` (5 minutes) by default. To understand how instant queries work, let's begin with a data sample: ``` foo_bar 1.00 1652169600000 # 2022-05-10 10:00:00 foo_bar 2.00 1652169660000 # 2022-05-10 10:01:00 foo_bar 3.00 1652169720000 # 2022-05-10 10:02:00 foo_bar 5.00 1652169840000 # 2022-05-10 10:04:00, one point missed foo_bar 5.50 1652169960000 # 2022-05-10 10:06:00, one point missed foo_bar 5.50 1652170020000 # 2022-05-10 10:07:00 foo_bar 4.00 1652170080000 # 2022-05-10 10:08:00 foo_bar 3.50 1652170260000 # 2022-05-10 10:11:00, two points missed foo_bar 3.25 1652170320000 # 2022-05-10 10:12:00 foo_bar 3.00 1652170380000 # 2022-05-10 10:13:00 foo_bar 2.00 1652170440000 # 2022-05-10 10:14:00 foo_bar 1.00 1652170500000 # 2022-05-10 10:15:00 foo_bar 4.00 1652170560000 # 2022-05-10 10:16:00 ``` The data sample contains a list of samples for `foo_bar` time series with time intervals between samples from 1m to 3m. If we plot this data sample on the graph, it will have the following form:
To get the value of `foo_bar` metric at some specific moment of time, for example `2022-05-10 10:03:00`, in VictoriaMetrics we need to issue an **instant query**: ```console curl "http://