mirror of
https://github.com/VictoriaMetrics/VictoriaMetrics.git
synced 2024-11-21 14:44:00 +00:00
lib/streamaggr: properly drop samples on the first incomplete interval
Previously samples were dropped on the first incomplete interval and the next complete interval. Also make sure that the de-duplication is performed just before flushing the aggregate state. This should help the case then dedup_interval = interval.
This commit is contained in:
parent
51745ec5ff
commit
2ffef39bb3
2 changed files with 54 additions and 36 deletions
|
@ -82,16 +82,22 @@ type Options struct {
|
||||||
// The last sample per each series is left per each DedupInterval if DedupInterval > 0.
|
// The last sample per each series is left per each DedupInterval if DedupInterval > 0.
|
||||||
//
|
//
|
||||||
// By default deduplication is disabled.
|
// By default deduplication is disabled.
|
||||||
|
//
|
||||||
|
// The deduplication can be set up individually per each aggregation via dedup_interval option.
|
||||||
DedupInterval time.Duration
|
DedupInterval time.Duration
|
||||||
|
|
||||||
// NoAlignFlushToInterval disables alignment of flushes to the aggregation interval.
|
// NoAlignFlushToInterval disables alignment of flushes to the aggregation interval.
|
||||||
//
|
//
|
||||||
// By default flushes are aligned to aggregation interval.
|
// By default flushes are aligned to aggregation interval.
|
||||||
|
//
|
||||||
|
// The alignment of flushes can be disabled individually per each aggregation via no_align_flush_to_interval option.
|
||||||
NoAlignFlushToInterval bool
|
NoAlignFlushToInterval bool
|
||||||
|
|
||||||
// FlushOnShutdown enables flush of incomplete state on shutdown.
|
// FlushOnShutdown enables flush of incomplete state on start and shutdown.
|
||||||
//
|
//
|
||||||
// By default incomplete state is dropped on shutdown.
|
// By default incomplete state is dropped on shutdown.
|
||||||
|
//
|
||||||
|
// The flush of incomplete state can be enabled individually per each aggregation via flush_on_shutdown option.
|
||||||
FlushOnShutdown bool
|
FlushOnShutdown bool
|
||||||
|
|
||||||
// KeepMetricNames instructs to leave metric names as is for the output time series without adding any suffix.
|
// KeepMetricNames instructs to leave metric names as is for the output time series without adding any suffix.
|
||||||
|
@ -100,6 +106,7 @@ type Options struct {
|
||||||
//
|
//
|
||||||
// input_name:<interval>[_by_<by_labels>][_without_<without_labels>]_<output>
|
// input_name:<interval>[_by_<by_labels>][_without_<without_labels>]_<output>
|
||||||
//
|
//
|
||||||
|
// This option can be overriden individually per each aggregation via keep_metric_names option.
|
||||||
KeepMetricNames bool
|
KeepMetricNames bool
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -601,20 +608,22 @@ func (a *aggregator) runFlusher(pushFunc PushFunc, alignFlushToInterval, skipInc
|
||||||
if d < interval {
|
if d < interval {
|
||||||
sleep(d)
|
sleep(d)
|
||||||
}
|
}
|
||||||
flushTickerCh <- time.NewTicker(interval)
|
t := time.NewTicker(interval)
|
||||||
|
if skipIncompleteFlush {
|
||||||
|
a.dedupFlush(dedupInterval)
|
||||||
|
a.flush(nil, interval)
|
||||||
|
}
|
||||||
|
flushTickerCh <- t
|
||||||
}()
|
}()
|
||||||
|
|
||||||
var flushTickerC <-chan time.Time
|
var flushTickerC <-chan time.Time
|
||||||
var dedupFlushTickerC <-chan time.Time
|
var dedupFlushTickerC <-chan time.Time
|
||||||
isFirstFlush := true
|
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-a.stopCh:
|
case <-a.stopCh:
|
||||||
if !skipIncompleteFlush {
|
if !skipIncompleteFlush {
|
||||||
if dedupInterval > 0 {
|
a.dedupFlush(dedupInterval)
|
||||||
a.dedupFlush()
|
a.flush(pushFunc, interval)
|
||||||
}
|
|
||||||
a.flush(pushFunc)
|
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
case flushTicker := <-flushTickerCh:
|
case flushTicker := <-flushTickerCh:
|
||||||
|
@ -624,43 +633,42 @@ func (a *aggregator) runFlusher(pushFunc PushFunc, alignFlushToInterval, skipInc
|
||||||
dedupFlushTickerC = dedupFlushTicker.C
|
dedupFlushTickerC = dedupFlushTicker.C
|
||||||
defer dedupFlushTicker.Stop()
|
defer dedupFlushTicker.Stop()
|
||||||
case <-flushTickerC:
|
case <-flushTickerC:
|
||||||
if isFirstFlush {
|
select {
|
||||||
isFirstFlush = false
|
|
||||||
if alignFlushToInterval && skipIncompleteFlush {
|
|
||||||
a.flush(nil)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
}
|
|
||||||
startTime := time.Now()
|
|
||||||
a.flush(pushFunc)
|
|
||||||
d := time.Since(startTime)
|
|
||||||
a.flushDuration.Update(d.Seconds())
|
|
||||||
if d > interval {
|
|
||||||
a.flushTimeouts.Inc()
|
|
||||||
logger.Warnf("stream aggregation couldn't be finished in the configured interval=%s; it took %s; "+
|
|
||||||
"possible solutions: increase interval; use match filter matching smaller number of series; "+
|
|
||||||
"reduce samples' ingestion rate to stream aggregation", interval, d)
|
|
||||||
}
|
|
||||||
case <-dedupFlushTickerC:
|
case <-dedupFlushTickerC:
|
||||||
|
// flush deduplicated samples if needed before flushing the aggregated samples
|
||||||
|
a.dedupFlush(dedupInterval)
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
a.flush(pushFunc, interval)
|
||||||
|
case <-dedupFlushTickerC:
|
||||||
|
a.dedupFlush(dedupInterval)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *aggregator) dedupFlush(dedupInterval time.Duration) {
|
||||||
|
if dedupInterval <= 0 {
|
||||||
|
// The de-duplication is disabled.
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
startTime := time.Now()
|
startTime := time.Now()
|
||||||
a.dedupFlush()
|
|
||||||
|
a.da.flush(a.pushSamples)
|
||||||
|
|
||||||
d := time.Since(startTime)
|
d := time.Since(startTime)
|
||||||
a.dedupFlushDuration.Update(d.Seconds())
|
a.dedupFlushDuration.Update(d.Seconds())
|
||||||
if d > dedupInterval {
|
if d > dedupInterval {
|
||||||
a.dedupFlushTimeouts.Inc()
|
a.dedupFlushTimeouts.Inc()
|
||||||
logger.Warnf("stream aggregation deduplication couldn't be finished in the configured dedup_interval=%s; it took %s; "+
|
logger.Warnf("deduplication couldn't be finished in the configured dedup_interval=%s; it took %.03f; "+
|
||||||
"possible solutions: increase dedup_interval; use match filter matching smaller number of series; "+
|
"possible solutions: increase dedup_interval; use match filter matching smaller number of series; "+
|
||||||
"reduce samples' ingestion rate to stream aggregation", dedupInterval, d)
|
"reduce samples' ingestion rate to stream aggregation", dedupInterval, d.Seconds())
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *aggregator) dedupFlush() {
|
func (a *aggregator) flush(pushFunc PushFunc, interval time.Duration) {
|
||||||
a.da.flush(a.pushSamples)
|
startTime := time.Now()
|
||||||
}
|
|
||||||
|
|
||||||
func (a *aggregator) flush(pushFunc PushFunc) {
|
|
||||||
var wg sync.WaitGroup
|
var wg sync.WaitGroup
|
||||||
for _, as := range a.aggrStates {
|
for _, as := range a.aggrStates {
|
||||||
flushConcurrencyCh <- struct{}{}
|
flushConcurrencyCh <- struct{}{}
|
||||||
|
@ -679,6 +687,15 @@ func (a *aggregator) flush(pushFunc PushFunc) {
|
||||||
}(as)
|
}(as)
|
||||||
}
|
}
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
|
|
||||||
|
d := time.Since(startTime)
|
||||||
|
a.flushDuration.Update(d.Seconds())
|
||||||
|
if d > interval {
|
||||||
|
a.flushTimeouts.Inc()
|
||||||
|
logger.Warnf("stream aggregation couldn't be finished in the configured interval=%s; it took %.03f; "+
|
||||||
|
"possible solutions: increase interval; use match filter matching smaller number of series; "+
|
||||||
|
"reduce samples' ingestion rate to stream aggregation", interval, d.Seconds())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
var flushConcurrencyCh = make(chan struct{}, cgroup.AvailableCPUs())
|
var flushConcurrencyCh = make(chan struct{}, cgroup.AvailableCPUs())
|
||||||
|
|
|
@ -4,6 +4,7 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"strings"
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal"
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal"
|
||||||
)
|
)
|
||||||
|
@ -55,7 +56,7 @@ func benchmarkAggregatorsFlushSerial(b *testing.B, output string) {
|
||||||
for i := 0; i < b.N; i++ {
|
for i := 0; i < b.N; i++ {
|
||||||
matchIdxs = a.Push(benchSeries, matchIdxs)
|
matchIdxs = a.Push(benchSeries, matchIdxs)
|
||||||
for _, aggr := range a.as {
|
for _, aggr := range a.as {
|
||||||
aggr.flush(pushFunc)
|
aggr.flush(pushFunc, time.Hour)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue