2023-07-27 22:26:45 +02:00
package stream
import (
"fmt"
"io"
"strconv"
"sync"
2023-12-15 11:16:25 +01:00
"github.com/VictoriaMetrics/metrics"
2023-07-27 22:26:45 +02:00
"github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/decimal"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fasttime"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/prompbmarshal"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/common"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/protoparser/opentelemetry/pb"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/writeconcurrencylimiter"
)
// ParseStream parses OpenTelemetry protobuf or json data from r and calls callback for the parsed rows.
//
// callback shouldn't hold tss items after returning.
func ParseStream ( r io . Reader , isGzipped bool , callback func ( tss [ ] prompbmarshal . TimeSeries ) error ) error {
wcr := writeconcurrencylimiter . GetReader ( r )
defer writeconcurrencylimiter . PutReader ( wcr )
r = wcr
if isGzipped {
zr , err := common . GetGzipReader ( r )
if err != nil {
return fmt . Errorf ( "cannot read gzip-compressed OpenTelemetry protocol data: %w" , err )
}
defer common . PutGzipReader ( zr )
r = zr
}
wr := getWriteContext ( )
defer putWriteContext ( wr )
req , err := wr . readAndUnpackRequest ( r )
if err != nil {
return fmt . Errorf ( "cannot unpack OpenTelemetry metrics: %w" , err )
}
wr . parseRequestToTss ( req )
if err := callback ( wr . tss ) ; err != nil {
return fmt . Errorf ( "error when processing OpenTelemetry samples: %w" , err )
}
return nil
}
func ( wr * writeContext ) appendSamplesFromScopeMetrics ( sc * pb . ScopeMetrics ) {
for _ , m := range sc . Metrics {
if len ( m . Name ) == 0 {
// skip metrics without names
continue
}
switch t := m . Data . ( type ) {
case * pb . Metric_Gauge :
for _ , p := range t . Gauge . DataPoints {
wr . appendSampleFromNumericPoint ( m . Name , p )
}
case * pb . Metric_Sum :
if t . Sum . AggregationTemporality != pb . AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE {
rowsDroppedUnsupportedSum . Inc ( )
continue
}
for _ , p := range t . Sum . DataPoints {
wr . appendSampleFromNumericPoint ( m . Name , p )
}
case * pb . Metric_Summary :
for _ , p := range t . Summary . DataPoints {
wr . appendSamplesFromSummary ( m . Name , p )
}
case * pb . Metric_Histogram :
if t . Histogram . AggregationTemporality != pb . AggregationTemporality_AGGREGATION_TEMPORALITY_CUMULATIVE {
rowsDroppedUnsupportedHistogram . Inc ( )
continue
}
for _ , p := range t . Histogram . DataPoints {
wr . appendSamplesFromHistogram ( m . Name , p )
}
default :
rowsDroppedUnsupportedMetricType . Inc ( )
logger . Warnf ( "unsupported type %T for metric %q" , t , m . Name )
}
}
}
// appendSampleFromNumericPoint appends p to wr.tss
func ( wr * writeContext ) appendSampleFromNumericPoint ( metricName string , p * pb . NumberDataPoint ) {
var v float64
switch t := p . Value . ( type ) {
case * pb . NumberDataPoint_AsInt :
v = float64 ( t . AsInt )
case * pb . NumberDataPoint_AsDouble :
v = t . AsDouble
}
t := int64 ( p . TimeUnixNano / 1e6 )
isStale := ( p . Flags ) & uint32 ( 1 ) != 0
wr . pointLabels = appendAttributesToPromLabels ( wr . pointLabels [ : 0 ] , p . Attributes )
wr . appendSample ( metricName , t , v , isStale )
}
// appendSamplesFromSummary appends summary p to wr.tss
func ( wr * writeContext ) appendSamplesFromSummary ( metricName string , p * pb . SummaryDataPoint ) {
t := int64 ( p . TimeUnixNano / 1e6 )
isStale := ( p . Flags ) & uint32 ( 1 ) != 0
wr . pointLabels = appendAttributesToPromLabels ( wr . pointLabels [ : 0 ] , p . Attributes )
wr . appendSample ( metricName + "_sum" , t , p . Sum , isStale )
wr . appendSample ( metricName + "_count" , t , float64 ( p . Count ) , isStale )
for _ , q := range p . QuantileValues {
qValue := strconv . FormatFloat ( q . Quantile , 'f' , - 1 , 64 )
wr . appendSampleWithExtraLabel ( metricName , "quantile" , qValue , t , q . Value , isStale )
}
}
// appendSamplesFromHistogram appends histogram p to wr.tss
func ( wr * writeContext ) appendSamplesFromHistogram ( metricName string , p * pb . HistogramDataPoint ) {
if len ( p . BucketCounts ) == 0 {
// nothing to append
return
}
if len ( p . BucketCounts ) != len ( p . ExplicitBounds ) + 1 {
// fast path, broken data format
logger . Warnf ( "opentelemetry bad histogram format: %q, size of buckets: %d, size of bounds: %d" , metricName , len ( p . BucketCounts ) , len ( p . ExplicitBounds ) )
return
}
t := int64 ( p . TimeUnixNano / 1e6 )
isStale := ( p . Flags ) & uint32 ( 1 ) != 0
wr . pointLabels = appendAttributesToPromLabels ( wr . pointLabels [ : 0 ] , p . Attributes )
2023-08-12 14:09:18 +02:00
wr . appendSample ( metricName + "_count" , t , float64 ( p . Count ) , isStale )
if p . Sum == nil {
// fast path, convert metric as simple counter.
// given buckets cannot be used for histogram functions.
// Negative threshold buckets MAY be used, but then the Histogram MetricPoint MUST NOT contain a sum value as it would no longer be a counter semantically.
// https://github.com/OpenObservability/OpenMetrics/blob/main/specification/OpenMetrics.md#histogram
return
}
2023-07-27 22:26:45 +02:00
wr . appendSample ( metricName + "_sum" , t , * p . Sum , isStale )
var cumulative uint64
for index , bound := range p . ExplicitBounds {
cumulative += p . BucketCounts [ index ]
boundLabelValue := strconv . FormatFloat ( bound , 'f' , - 1 , 64 )
wr . appendSampleWithExtraLabel ( metricName + "_bucket" , "le" , boundLabelValue , t , float64 ( cumulative ) , isStale )
}
cumulative += p . BucketCounts [ len ( p . BucketCounts ) - 1 ]
wr . appendSampleWithExtraLabel ( metricName + "_bucket" , "le" , "+Inf" , t , float64 ( cumulative ) , isStale )
}
// appendSample appends sample with the given metricName to wr.tss
func ( wr * writeContext ) appendSample ( metricName string , t int64 , v float64 , isStale bool ) {
wr . appendSampleWithExtraLabel ( metricName , "" , "" , t , v , isStale )
}
// appendSampleWithExtraLabel appends sample with the given metricName and the given (labelName=labelValue) extra label to wr.tss
func ( wr * writeContext ) appendSampleWithExtraLabel ( metricName , labelName , labelValue string , t int64 , v float64 , isStale bool ) {
if isStale {
v = decimal . StaleNaN
}
if t <= 0 {
// Set the current timestamp if t isn't set.
t = int64 ( fasttime . UnixTimestamp ( ) ) * 1000
}
labelsPool := wr . labelsPool
labelsLen := len ( labelsPool )
labelsPool = append ( labelsPool , prompbmarshal . Label {
Name : "__name__" ,
Value : metricName ,
} )
labelsPool = append ( labelsPool , wr . baseLabels ... )
labelsPool = append ( labelsPool , wr . pointLabels ... )
if labelName != "" && labelValue != "" {
labelsPool = append ( labelsPool , prompbmarshal . Label {
Name : labelName ,
Value : labelValue ,
} )
}
samplesPool := wr . samplesPool
samplesLen := len ( samplesPool )
samplesPool = append ( samplesPool , prompbmarshal . Sample {
Timestamp : t ,
Value : v ,
} )
wr . tss = append ( wr . tss , prompbmarshal . TimeSeries {
Labels : labelsPool [ labelsLen : ] ,
Samples : samplesPool [ samplesLen : ] ,
} )
wr . labelsPool = labelsPool
wr . samplesPool = samplesPool
rowsRead . Inc ( )
}
// appendAttributesToPromLabels appends attributes to dst and returns the result.
func appendAttributesToPromLabels ( dst [ ] prompbmarshal . Label , attributes [ ] * pb . KeyValue ) [ ] prompbmarshal . Label {
for _ , at := range attributes {
dst = append ( dst , prompbmarshal . Label {
Name : at . Key ,
Value : at . Value . FormatString ( ) ,
} )
}
return dst
}
type writeContext struct {
// bb holds the original data (json or protobuf), which must be parsed.
bb bytesutil . ByteBuffer
// tss holds parsed time series
tss [ ] prompbmarshal . TimeSeries
// baseLabels are labels, which must be added to all the ingested samples
baseLabels [ ] prompbmarshal . Label
// pointLabels are labels, which must be added to the ingested OpenTelemetry points
pointLabels [ ] prompbmarshal . Label
// pools are used for reducing memory allocations when parsing time series
labelsPool [ ] prompbmarshal . Label
samplesPool [ ] prompbmarshal . Sample
}
func ( wr * writeContext ) reset ( ) {
wr . bb . Reset ( )
tss := wr . tss
for i := range tss {
ts := & tss [ i ]
ts . Labels = nil
ts . Samples = nil
}
wr . tss = tss [ : 0 ]
wr . baseLabels = resetLabels ( wr . baseLabels )
wr . pointLabels = resetLabels ( wr . pointLabels )
wr . labelsPool = resetLabels ( wr . labelsPool )
wr . samplesPool = wr . samplesPool [ : 0 ]
}
func resetLabels ( labels [ ] prompbmarshal . Label ) [ ] prompbmarshal . Label {
for i := range labels {
label := & labels [ i ]
label . Name = ""
label . Value = ""
}
return labels [ : 0 ]
}
func ( wr * writeContext ) readAndUnpackRequest ( r io . Reader ) ( * pb . ExportMetricsServiceRequest , error ) {
if _ , err := wr . bb . ReadFrom ( r ) ; err != nil {
return nil , fmt . Errorf ( "cannot read request: %w" , err )
}
var req pb . ExportMetricsServiceRequest
if err := req . UnmarshalVT ( wr . bb . B ) ; err != nil {
return nil , fmt . Errorf ( "cannot unmarshal request from %d bytes: %w" , len ( wr . bb . B ) , err )
}
return & req , nil
}
func ( wr * writeContext ) parseRequestToTss ( req * pb . ExportMetricsServiceRequest ) {
for _ , rm := range req . ResourceMetrics {
2023-12-17 18:05:22 +01:00
var attributes [ ] * pb . KeyValue
if rm . Resource != nil {
attributes = rm . Resource . Attributes
2023-07-27 22:26:45 +02:00
}
2023-12-17 18:05:22 +01:00
wr . baseLabels = appendAttributesToPromLabels ( wr . baseLabels [ : 0 ] , attributes )
2023-07-27 22:26:45 +02:00
for _ , sc := range rm . ScopeMetrics {
wr . appendSamplesFromScopeMetrics ( sc )
}
}
}
var wrPool sync . Pool
func getWriteContext ( ) * writeContext {
v := wrPool . Get ( )
if v == nil {
return & writeContext { }
}
return v . ( * writeContext )
}
func putWriteContext ( wr * writeContext ) {
wr . reset ( )
wrPool . Put ( wr )
}
var (
rowsRead = metrics . NewCounter ( ` vm_protoparser_rows_read_total { type="opentelemetry"} ` )
rowsDroppedUnsupportedHistogram = metrics . NewCounter ( ` vm_protoparser_rows_dropped_total { type="opentelemetry",reason="unsupported_histogram_aggregation"} ` )
rowsDroppedUnsupportedSum = metrics . NewCounter ( ` vm_protoparser_rows_dropped_total { type="opentelemetry",reason="unsupported_sum_aggregation"} ` )
rowsDroppedUnsupportedMetricType = metrics . NewCounter ( ` vm_protoparser_rows_dropped_total { type="opentelemetry",reason="unsupported_metric_type"} ` )
)