package datadogv2 import ( "net/http" "github.com/VictoriaMetrics/VictoriaMetrics/app/vminsert/netstorage" "github.com/VictoriaMetrics/VictoriaMetrics/app/vminsert/relabel" "github.com/VictoriaMetrics/VictoriaMetrics/lib/auth" "github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal" parserCommon "github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/common" "github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/datadogutils" "github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/datadogv2" "github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/datadogv2/stream" "github.com/VictoriaMetrics/VictoriaMetrics/lib/storage" "github.com/VictoriaMetrics/VictoriaMetrics/lib/tenantmetrics" "github.com/VictoriaMetrics/metrics" ) var ( rowsInserted = metrics.NewCounter(`vm_rows_inserted_total{type="datadogv2"}`) rowsTenantInserted = tenantmetrics.NewCounterMap(`vm_tenant_inserted_rows_total{type="datadogv2"}`) rowsPerInsert = metrics.NewHistogram(`vm_rows_per_insert{type="datadogv2"}`) ) // InsertHandlerForHTTP processes remote write for DataDog POST /api/v2/series request. // // See https://docs.datadoghq.com/api/latest/metrics/#submit-metrics func InsertHandlerForHTTP(at *auth.Token, req *http.Request) error { extraLabels, err := parserCommon.GetExtraLabels(req) if err != nil { return err } ct := req.Header.Get("Content-Type") ce := req.Header.Get("Content-Encoding") return stream.Parse(req.Body, ce, ct, func(series []datadogv2.Series) error { return insertRows(at, series, extraLabels) }) } func insertRows(at *auth.Token, series []datadogv2.Series, extraLabels []prompbmarshal.Label) error { ctx := netstorage.GetInsertCtx() defer netstorage.PutInsertCtx(ctx) ctx.Reset() rowsTotal := 0 perTenantRows := make(map[auth.Token]int) hasRelabeling := relabel.HasRelabeling() for i := range series { ss := &series[i] rowsTotal += len(ss.Points) ctx.Labels = ctx.Labels[:0] ctx.AddLabel("", ss.Metric) for _, rs := range ss.Resources { ctx.AddLabel(rs.Type, rs.Name) } for _, tag := range ss.Tags { name, value := datadogutils.SplitTag(tag) if name == "host" { name = "exported_host" } ctx.AddLabel(name, value) } if ss.SourceTypeName != "" { ctx.AddLabel("source_type_name", ss.SourceTypeName) } for j := range extraLabels { label := &extraLabels[j] ctx.AddLabel(label.Name, label.Value) } if hasRelabeling { ctx.ApplyRelabeling() } if len(ctx.Labels) == 0 { // Skip metric without labels. continue } ctx.SortLabelsIfNeeded() atLocal := ctx.GetLocalAuthToken(at) ctx.MetricNameBuf = storage.MarshalMetricNameRaw(ctx.MetricNameBuf[:0], atLocal.AccountID, atLocal.ProjectID, ctx.Labels) storageNodeIdx := ctx.GetStorageNodeIdx(atLocal, ctx.Labels) for _, pt := range ss.Points { timestamp := pt.Timestamp * 1000 value := pt.Value if err := ctx.WriteDataPointExt(storageNodeIdx, ctx.MetricNameBuf, timestamp, value); err != nil { return err } } perTenantRows[*atLocal] += len(ss.Points) } rowsInserted.Add(rowsTotal) rowsTenantInserted.MultiAdd(perTenantRows) rowsPerInsert.Update(float64(rowsTotal)) return ctx.FlushBufs() }