2023-07-27 20:26:45 +00:00
package stream
import (
"fmt"
"io"
"strconv"
"sync"
2023-12-15 10:16:25 +00:00
"github.com/VictoriaMetrics/metrics"
2023-07-27 20:26:45 +00:00
"github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/decimal"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fasttime"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/common"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/opentelemetry/pb"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/writeconcurrencylimiter"
)
// ParseStream parses OpenTelemetry protobuf or json data from r and calls callback for the parsed rows.
//
// callback shouldn't hold tss items after returning.
2024-02-29 12:38:21 +00:00
//
// optional processBody can be used for pre-processing the read request body from r before parsing it in OpenTelemetry format.
func ParseStream ( r io . Reader , isGzipped bool , processBody func ( [ ] byte ) ( [ ] byte , error ) , callback func ( tss [ ] prompbmarshal . TimeSeries ) error ) error {
2023-07-27 20:26:45 +00:00
wcr := writeconcurrencylimiter . GetReader ( r )
defer writeconcurrencylimiter . PutReader ( wcr )
r = wcr
if isGzipped {
zr , err := common . GetGzipReader ( r )
if err != nil {
return fmt . Errorf ( "cannot read gzip-compressed OpenTelemetry protocol data: %w" , err )
}
defer common . PutGzipReader ( zr )
r = zr
}
wr := getWriteContext ( )
defer putWriteContext ( wr )
2024-02-29 12:03:24 +00:00
req , err := wr . readAndUnpackRequest ( r , processBody )
2023-07-27 20:26:45 +00:00
if err != nil {
return fmt . Errorf ( "cannot unpack OpenTelemetry metrics: %w" , err )
}
wr . parseRequestToTss ( req )
if err := callback ( wr . tss ) ; err != nil {
return fmt . Errorf ( "error when processing OpenTelemetry samples: %w" , err )
}
return nil
}
func ( wr * writeContext ) appendSamplesFromScopeMetrics ( sc * pb . ScopeMetrics ) {
for _ , m := range sc . Metrics {
if len ( m . Name ) == 0 {
// skip metrics without names
continue
}
2024-03-29 12:51:24 +00:00
metricName := sanitizeMetricName ( m )
2024-01-14 19:17:37 +00:00
switch {
case m . Gauge != nil :
for _ , p := range m . Gauge . DataPoints {
2024-03-29 12:51:24 +00:00
wr . appendSampleFromNumericPoint ( metricName , p )
2023-07-27 20:26:45 +00:00
}
2024-01-14 19:17:37 +00:00
case m . Sum != nil :
if m . Sum . AggregationTemporality != pb . AggregationTemporalityCumulative {
2023-07-27 20:26:45 +00:00
rowsDroppedUnsupportedSum . Inc ( )
continue
}
2024-01-14 19:17:37 +00:00
for _ , p := range m . Sum . DataPoints {
2024-03-29 12:51:24 +00:00
wr . appendSampleFromNumericPoint ( metricName , p )
2023-07-27 20:26:45 +00:00
}
2024-01-14 19:17:37 +00:00
case m . Summary != nil :
for _ , p := range m . Summary . DataPoints {
2024-03-29 12:51:24 +00:00
wr . appendSamplesFromSummary ( metricName , p )
2023-07-27 20:26:45 +00:00
}
2024-01-14 19:17:37 +00:00
case m . Histogram != nil :
if m . Histogram . AggregationTemporality != pb . AggregationTemporalityCumulative {
2023-07-27 20:26:45 +00:00
rowsDroppedUnsupportedHistogram . Inc ( )
continue
}
2024-01-14 19:17:37 +00:00
for _ , p := range m . Histogram . DataPoints {
2024-03-29 12:51:24 +00:00
wr . appendSamplesFromHistogram ( metricName , p )
2023-07-27 20:26:45 +00:00
}
default :
rowsDroppedUnsupportedMetricType . Inc ( )
2024-03-29 12:51:24 +00:00
logger . Warnf ( "unsupported type for metric %q" , metricName )
2023-07-27 20:26:45 +00:00
}
}
}
// appendSampleFromNumericPoint appends p to wr.tss
func ( wr * writeContext ) appendSampleFromNumericPoint ( metricName string , p * pb . NumberDataPoint ) {
var v float64
2024-01-14 19:17:37 +00:00
switch {
case p . IntValue != nil :
v = float64 ( * p . IntValue )
case p . DoubleValue != nil :
v = * p . DoubleValue
2023-07-27 20:26:45 +00:00
}
t := int64 ( p . TimeUnixNano / 1e6 )
isStale := ( p . Flags ) & uint32 ( 1 ) != 0
wr . pointLabels = appendAttributesToPromLabels ( wr . pointLabels [ : 0 ] , p . Attributes )
wr . appendSample ( metricName , t , v , isStale )
}
// appendSamplesFromSummary appends summary p to wr.tss
func ( wr * writeContext ) appendSamplesFromSummary ( metricName string , p * pb . SummaryDataPoint ) {
t := int64 ( p . TimeUnixNano / 1e6 )
isStale := ( p . Flags ) & uint32 ( 1 ) != 0
wr . pointLabels = appendAttributesToPromLabels ( wr . pointLabels [ : 0 ] , p . Attributes )
wr . appendSample ( metricName + "_sum" , t , p . Sum , isStale )
wr . appendSample ( metricName + "_count" , t , float64 ( p . Count ) , isStale )
for _ , q := range p . QuantileValues {
qValue := strconv . FormatFloat ( q . Quantile , 'f' , - 1 , 64 )
wr . appendSampleWithExtraLabel ( metricName , "quantile" , qValue , t , q . Value , isStale )
}
}
// appendSamplesFromHistogram appends histogram p to wr.tss
func ( wr * writeContext ) appendSamplesFromHistogram ( metricName string , p * pb . HistogramDataPoint ) {
if len ( p . BucketCounts ) == 0 {
// nothing to append
return
}
if len ( p . BucketCounts ) != len ( p . ExplicitBounds ) + 1 {
// fast path, broken data format
logger . Warnf ( "opentelemetry bad histogram format: %q, size of buckets: %d, size of bounds: %d" , metricName , len ( p . BucketCounts ) , len ( p . ExplicitBounds ) )
return
}
t := int64 ( p . TimeUnixNano / 1e6 )
isStale := ( p . Flags ) & uint32 ( 1 ) != 0
wr . pointLabels = appendAttributesToPromLabels ( wr . pointLabels [ : 0 ] , p . Attributes )
2023-08-12 12:09:18 +00:00
wr . appendSample ( metricName + "_count" , t , float64 ( p . Count ) , isStale )
if p . Sum == nil {
// fast path, convert metric as simple counter.
// given buckets cannot be used for histogram functions.
// Negative threshold buckets MAY be used, but then the Histogram MetricPoint MUST NOT contain a sum value as it would no longer be a counter semantically.
// https://github.com/OpenObservability/OpenMetrics/blob/main/specification/OpenMetrics.md#histogram
return
}
2023-07-27 20:26:45 +00:00
wr . appendSample ( metricName + "_sum" , t , * p . Sum , isStale )
var cumulative uint64
for index , bound := range p . ExplicitBounds {
cumulative += p . BucketCounts [ index ]
boundLabelValue := strconv . FormatFloat ( bound , 'f' , - 1 , 64 )
wr . appendSampleWithExtraLabel ( metricName + "_bucket" , "le" , boundLabelValue , t , float64 ( cumulative ) , isStale )
}
cumulative += p . BucketCounts [ len ( p . BucketCounts ) - 1 ]
wr . appendSampleWithExtraLabel ( metricName + "_bucket" , "le" , "+Inf" , t , float64 ( cumulative ) , isStale )
}
// appendSample appends sample with the given metricName to wr.tss
func ( wr * writeContext ) appendSample ( metricName string , t int64 , v float64 , isStale bool ) {
wr . appendSampleWithExtraLabel ( metricName , "" , "" , t , v , isStale )
}
// appendSampleWithExtraLabel appends sample with the given metricName and the given (labelName=labelValue) extra label to wr.tss
func ( wr * writeContext ) appendSampleWithExtraLabel ( metricName , labelName , labelValue string , t int64 , v float64 , isStale bool ) {
if isStale {
v = decimal . StaleNaN
}
if t <= 0 {
// Set the current timestamp if t isn't set.
t = int64 ( fasttime . UnixTimestamp ( ) ) * 1000
}
labelsPool := wr . labelsPool
labelsLen := len ( labelsPool )
labelsPool = append ( labelsPool , prompbmarshal . Label {
Name : "__name__" ,
Value : metricName ,
} )
labelsPool = append ( labelsPool , wr . baseLabels ... )
labelsPool = append ( labelsPool , wr . pointLabels ... )
if labelName != "" && labelValue != "" {
labelsPool = append ( labelsPool , prompbmarshal . Label {
Name : labelName ,
Value : labelValue ,
} )
}
samplesPool := wr . samplesPool
samplesLen := len ( samplesPool )
samplesPool = append ( samplesPool , prompbmarshal . Sample {
Timestamp : t ,
Value : v ,
} )
wr . tss = append ( wr . tss , prompbmarshal . TimeSeries {
Labels : labelsPool [ labelsLen : ] ,
Samples : samplesPool [ samplesLen : ] ,
} )
wr . labelsPool = labelsPool
wr . samplesPool = samplesPool
rowsRead . Inc ( )
}
// appendAttributesToPromLabels appends attributes to dst and returns the result.
func appendAttributesToPromLabels ( dst [ ] prompbmarshal . Label , attributes [ ] * pb . KeyValue ) [ ] prompbmarshal . Label {
for _ , at := range attributes {
dst = append ( dst , prompbmarshal . Label {
2024-03-29 12:51:24 +00:00
Name : sanitizeLabelName ( at . Key ) ,
2023-07-27 20:26:45 +00:00
Value : at . Value . FormatString ( ) ,
} )
}
return dst
}
type writeContext struct {
// bb holds the original data (json or protobuf), which must be parsed.
bb bytesutil . ByteBuffer
// tss holds parsed time series
tss [ ] prompbmarshal . TimeSeries
// baseLabels are labels, which must be added to all the ingested samples
baseLabels [ ] prompbmarshal . Label
// pointLabels are labels, which must be added to the ingested OpenTelemetry points
pointLabels [ ] prompbmarshal . Label
// pools are used for reducing memory allocations when parsing time series
labelsPool [ ] prompbmarshal . Label
samplesPool [ ] prompbmarshal . Sample
}
func ( wr * writeContext ) reset ( ) {
wr . bb . Reset ( )
2024-04-20 19:00:00 +00:00
clear ( wr . tss )
wr . tss = wr . tss [ : 0 ]
2023-07-27 20:26:45 +00:00
wr . baseLabels = resetLabels ( wr . baseLabels )
wr . pointLabels = resetLabels ( wr . pointLabels )
wr . labelsPool = resetLabels ( wr . labelsPool )
wr . samplesPool = wr . samplesPool [ : 0 ]
}
func resetLabels ( labels [ ] prompbmarshal . Label ) [ ] prompbmarshal . Label {
2024-04-20 19:00:00 +00:00
clear ( labels )
2023-07-27 20:26:45 +00:00
return labels [ : 0 ]
}
2024-02-29 12:38:21 +00:00
func ( wr * writeContext ) readAndUnpackRequest ( r io . Reader , processBody func ( [ ] byte ) ( [ ] byte , error ) ) ( * pb . ExportMetricsServiceRequest , error ) {
2023-07-27 20:26:45 +00:00
if _ , err := wr . bb . ReadFrom ( r ) ; err != nil {
return nil , fmt . Errorf ( "cannot read request: %w" , err )
}
var req pb . ExportMetricsServiceRequest
2024-02-29 12:03:24 +00:00
if processBody != nil {
2024-02-29 12:38:21 +00:00
data , err := processBody ( wr . bb . B )
if err != nil {
2024-02-29 12:03:24 +00:00
return nil , fmt . Errorf ( "cannot process request body: %w" , err )
}
2024-02-29 12:38:21 +00:00
wr . bb . B = append ( wr . bb . B [ : 0 ] , data ... )
2024-02-29 12:03:24 +00:00
}
2024-01-14 19:17:37 +00:00
if err := req . UnmarshalProtobuf ( wr . bb . B ) ; err != nil {
2023-07-27 20:26:45 +00:00
return nil , fmt . Errorf ( "cannot unmarshal request from %d bytes: %w" , len ( wr . bb . B ) , err )
}
return & req , nil
}
func ( wr * writeContext ) parseRequestToTss ( req * pb . ExportMetricsServiceRequest ) {
for _ , rm := range req . ResourceMetrics {
2023-12-17 17:05:22 +00:00
var attributes [ ] * pb . KeyValue
if rm . Resource != nil {
attributes = rm . Resource . Attributes
2023-07-27 20:26:45 +00:00
}
2023-12-17 17:05:22 +00:00
wr . baseLabels = appendAttributesToPromLabels ( wr . baseLabels [ : 0 ] , attributes )
2023-07-27 20:26:45 +00:00
for _ , sc := range rm . ScopeMetrics {
wr . appendSamplesFromScopeMetrics ( sc )
}
}
}
var wrPool sync . Pool
func getWriteContext ( ) * writeContext {
v := wrPool . Get ( )
if v == nil {
return & writeContext { }
}
return v . ( * writeContext )
}
func putWriteContext ( wr * writeContext ) {
wr . reset ( )
wrPool . Put ( wr )
}
var (
rowsRead = metrics . NewCounter ( ` vm_protoparser_rows_read_total { type="opentelemetry"} ` )
rowsDroppedUnsupportedHistogram = metrics . NewCounter ( ` vm_protoparser_rows_dropped_total { type="opentelemetry",reason="unsupported_histogram_aggregation"} ` )
rowsDroppedUnsupportedSum = metrics . NewCounter ( ` vm_protoparser_rows_dropped_total { type="opentelemetry",reason="unsupported_sum_aggregation"} ` )
rowsDroppedUnsupportedMetricType = metrics . NewCounter ( ` vm_protoparser_rows_dropped_total { type="opentelemetry",reason="unsupported_metric_type"} ` )
)