VictoriaMetrics/app/vmagent
2020-03-09 13:31:22 +02:00
..
common app/vmagent: initial implementation for vmagent 2020-02-23 17:31:54 +02:00
deployment app/vmagent: initial implementation for vmagent 2020-02-23 17:31:54 +02:00
graphite app/vmagent: initial implementation for vmagent 2020-02-23 17:31:54 +02:00
influx app/vmagent: add ability to accept Influx line protocol data via TCP and UDP 2020-02-25 19:18:01 +02:00
opentsdb app/vmagent: initial implementation for vmagent 2020-02-23 17:31:54 +02:00
opentsdbhttp app/vmagent: initial implementation for vmagent 2020-02-23 17:31:54 +02:00
promremotewrite app/vmagent: initial implementation for vmagent 2020-02-23 17:31:54 +02:00
remotewrite app/vmagent: properly apply -remoteWrite.sendTimeout to fasthttp.HostClient 2020-03-09 13:31:22 +02:00
vmimport app/vmagent: initial implementation for vmagent 2020-02-23 17:31:54 +02:00
main.go app/vmagent: allow setting -httpListenAddr to empty string in order to disable listening for http requests 2020-02-26 20:58:26 +02:00
Makefile Makefile: add build and test rules with enabled race detector. These rules have -race suffix 2020-03-05 12:05:16 +02:00
README.md app/vmagent/README.md: small fixes 2020-03-04 18:15:24 +02:00
vmagent.png app/vmagent: update docs 2020-02-25 00:09:53 +02:00

vmagent

vmagent is a tiny but brave agent, which helps you collecting metrics from various sources and storing them to VictoriaMetrics.

vmagent

Motivation

While VictoriaMetrics provides an efficient solution to store and observe metrics, our users needed something fast and RAM friendly to scrape metrics from Prometheus-compatible exporters to VictoriaMetrics. Also, we found that users infrastructure is like snowflakes - never alike, and we decided to add more flexibility to vmagent (like the ability to push metrics instead of pulling them). We did our best and plan to do even more.

Features

  • Can be used as drop-in replacement for Prometheus for scraping targets such as node_exporter. See Quick Start for details.
  • Can add, remove and modify labels via Prometheus relabeling. See these docs for details.
  • Accepts data via all the ingestion protocols supported by VictoriaMetrics:
    • Influx line protocol via http://<vmagent>:8429/write. See these docs.
    • JSON lines import protocol via http://<vmagent>:8429/api/v1/import. See these docs.
    • Graphite plaintext protocol if -graphiteListenAddr command-line flag is set. See these docs.
    • OpenTSDB telnet and http protocols if -opentsdbListenAddr command-line flag is set. See these docs.
    • Prometheus remote write protocol via http://<vmagent>:8429/api/v1/write.
  • Can replicate collected metrics simultaneously to multiple remote storage systems.
  • Works in environments with unstable connections to remote storage. If the remote storage is unavailable, the collected metrics are buffered at -remoteWrite.tmpDataPath. The buffered metrics are sent to remote storage as soon as connection to remote storage is recovered. The maximum disk usage for the buffer can be limited with -remoteWrite.maxDiskUsagePerURL.
  • Uses lower amounts of RAM, CPU, disk IO and network bandwidth comparing to Prometheus.

Quick Start

Just download vmutils-* archive from releases page, unpack it and pass the following flags to vmagent binary in order to start scraping Prometheus targets:

  • -promscrape.config with the path to Prometheus config file (it is usually located at /etc/prometheus/prometheus.yml)
  • -remoteWrite.url with the remote storage endpoint such as VictoriaMetrics. Multiple -remoteWrite.url args can be set in parallel in order to replicate data concurrently to multiple remote storage systems.

Example command line:

/path/to/vmagent -promscrape.config=/path/to/prometheus.yml -remoteWrite.url=https://victoria-metrics-host:8428/api/v1/write

If you need collecting only Influx data, then the following command line would be enough:

/path/to/vmagent -remoteWrite.url=https://victoria-metrics-host:8428/api/v1/write

Then send Influx data to http://vmagent-host:8429/write. See these docs for more details.

vmagent is also available in docker images.

Pass -help to vmagent in order to see the full list of supported command-line flags with their descriptions.

Use cases

IoT and Edge monitoring

vmagent can run and collect metrics in IoT and industrial networks with unreliable or scheduled connections to the remote storage. It buffers the collected data in local files until the connection to remote storage becomes available and then sends the buffered data to the remote storage. It re-tries sending the data to remote storage on any errors. The maximum buffer size can be limited with -remoteWrite.maxDiskUsagePerURL.

vmagent works on various architectures from IoT world - 32-bit arm, 64-bit arm, ppc64, 386, amd64. See the corresponding Makefile rules for details.

Drop-in replacement for Prometheus

If you use Prometheus only for scraping metrics from various targets and forwarding these metrics to remote storage, then vmagent can replace such Prometheus setup. Usually vmagent requires lower amounts of RAM, CPU and network bandwidth comparing to Prometheus for such setup. See these docs for details.

Replication and high availability

vmagent replicates the collected metrics among multiple remote storage instances configured via -remoteWrite.url args. If a single remote storage instance temporarily goes out of service, then the collected data remains available in another remote storage instances. vmagent buffers the collected data in files at -remoteWrite.tmpDataPath until the remote storage becomes available again. Then it sends the buffered data to the remote storage in order to prevent data gaps in the remote storage.

Relabeling and filtering

vmagent can add, remove or update labels on the collected data before sending it to remote storage. Additionally, it can remove unneeded samples via Prometheus-like relabeling before sending the collected data to remote storage. See these docs for details.

Splitting data streams among multiple systems

vmagent supports splitting of the collected data among muliple destinations with the help of -remoteWrite.urlRelabelConfig, which is applied independently for each configured -remoteWrite.url destination. For instance, it is possible to replicate or split data among long-term remote storage, short-term remote storage and real-time analytical system built on top of Kafka. Note that each destination can receive its own subset of the collected data thanks to per-destination relabeling via -remoteWrite.urlRelabelConfig.

How to collect metrics in Prometheus format

Pass the path to prometheus.yml to -promscrape.config command-line flag. vmagent takes into account the following sections from Prometheus config file:

  • global
  • scrape_configs

All the other sections are ignored, including remote_write section. Use -remoteWrite.* command-line flags instead for configuring remote write settings.

The following scrape types in scrape_config section are supported:

  • static_configs - for scraping statically defined targets. See these docs for details.
  • file_sd_configs - for scraping targets defined in external files aka file-based service discover. See these docs for details.

The following service discovery mechanisms will be added to vmagent soon:

File feature requests at our issue tracker if you need other service discovery mechanisms to be supported by vmagent.

Adding labels to metrics

Labels can be added to metrics via the following mechanisms:

  • Via global -> external_labels section in -promscrape.config file. These labels are added only to metrics scraped from targets configured in -promscrape.config file.
  • Via -remoteWrite.label command-line flag. These labels are added to all the collected metrics before sending them to -remoteWrite.url.

Relabeling

vmagent supports Prometheus relabeling. Additionally it provides the following extra actions:

  • replace_all: replaces all the occurences of regex in the values of source_labels with the replacement and stores the result in the target_label.
  • labelmap_all: replaces all the occurences of regex in all the label names with the replacement.

The relabeling can be defined in the following places:

  • At scrape_config -> relabel_configs section in -promscrape.config file. This relabeling is applied to targets when parsing the file during vmagent startup or during config reload after sending SIGHUP signal to vmagent via kill -HUP.
  • At scrape_config -> metric_relabel_configs section in -promscrape.config file. This relabeling is applied to metrics after each scrape for the configured targets.
  • At -remoteWrite.relabelConfig file. This relabeling is aplied to all the collected metrics before sending them to remote storage.
  • At -remoteWrite.urlRelabelConfig files. This relabeling is applied to metrics before sending them to the corresponding -remoteWrite.url.

Read more about relabeling in the following articles:

Monitoring

vmagent exports various metrics in Prometheus exposition format at http://vmagent-host:8429/metrics page. It is recommended setting up regular scraping of this page either via vmagent itself or via Prometheus, so the exported metrics could be analyzed later.

vmagent also exports target statuses at http://vmagent-host:8429/targets page in plaintext format.

Troubleshooting

  • It is recommended increasing the maximum number of open files in the system (ulimit -n) when scraping big number of targets, since vmagent establishes at least a single TCP connection per each target.

  • It is recommended increasing -remoteWrite.queues if vmagent collects more than 100K samples per second and vmagent_remotewrite_pending_data_bytes metric exported by vmagent at /metrics page constantly grows.

  • vmagent buffers scraped data at -remoteWrite.tmpDataPath directory until it is sent to -remoteWrite.url. The directory can grow big when remote storage is unvailable during extended periods of time and if -remoteWrite.maxDiskUsagePerURL isn't set. If you don't want sending all the data from the directory to remote storage, just stop vmagent and delete the directory.

How to build from sources

It is recommended using binary releases - vmagent is located in vmutils-* archives there.

Development build

  1. Install Go. The minimum supported version is Go 1.12.
  2. Run make vmagent from the root folder of the repository. It builds vmagent binary and puts it into the bin folder.

Production build

  1. Install docker.
  2. Run make vmagent-prod from the root folder of the repository. It builds vmagent-prod binary and puts it into the bin folder.

Building docker images

Run make package-vmagent. It builds victoriametrics/vmagent:<PKG_TAG> docker image locally. <PKG_TAG> is auto-generated image tag, which depends on source code in the repository. The <PKG_TAG> may be manually set via PKG_TAG=foobar make package-vmagent.