2024-05-25 21:36:16 +02:00
|
|
|
package logstorage
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
2024-05-29 01:52:13 +02:00
|
|
|
"slices"
|
2024-05-25 21:36:16 +02:00
|
|
|
"unsafe"
|
|
|
|
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil"
|
|
|
|
)
|
|
|
|
|
|
|
|
// pipePackJSON processes '| pack_json ...' pipe.
|
|
|
|
//
|
|
|
|
// See https://docs.victoriametrics.com/victorialogs/logsql/#pack_json-pipe
|
|
|
|
type pipePackJSON struct {
|
|
|
|
resultField string
|
2024-05-29 01:52:13 +02:00
|
|
|
|
|
|
|
fields []string
|
2024-05-25 21:36:16 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
func (pp *pipePackJSON) String() string {
|
|
|
|
s := "pack_json"
|
2024-05-29 01:52:13 +02:00
|
|
|
if len(pp.fields) > 0 {
|
|
|
|
s += " fields (" + fieldsToString(pp.fields) + ")"
|
|
|
|
}
|
2024-05-25 21:36:16 +02:00
|
|
|
if !isMsgFieldName(pp.resultField) {
|
|
|
|
s += " as " + quoteTokenIfNeeded(pp.resultField)
|
|
|
|
}
|
|
|
|
return s
|
|
|
|
}
|
|
|
|
|
|
|
|
func (pp *pipePackJSON) updateNeededFields(neededFields, unneededFields fieldsSet) {
|
|
|
|
if neededFields.contains("*") {
|
|
|
|
if !unneededFields.contains(pp.resultField) {
|
2024-05-29 01:52:13 +02:00
|
|
|
if len(pp.fields) > 0 {
|
|
|
|
unneededFields.removeFields(pp.fields)
|
|
|
|
} else {
|
|
|
|
unneededFields.reset()
|
|
|
|
}
|
2024-05-25 21:36:16 +02:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if neededFields.contains(pp.resultField) {
|
2024-05-29 01:52:13 +02:00
|
|
|
if len(pp.fields) > 0 {
|
|
|
|
neededFields.addFields(pp.fields)
|
|
|
|
} else {
|
|
|
|
neededFields.add("*")
|
|
|
|
}
|
2024-05-25 21:36:16 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (pp *pipePackJSON) optimize() {
|
|
|
|
// nothing to do
|
|
|
|
}
|
|
|
|
|
|
|
|
func (pp *pipePackJSON) hasFilterInWithQuery() bool {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2024-05-26 02:01:32 +02:00
|
|
|
func (pp *pipePackJSON) initFilterInValues(_ map[string][]string, _ getFieldValuesFunc) (pipe, error) {
|
2024-05-25 21:36:16 +02:00
|
|
|
return pp, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (pp *pipePackJSON) newPipeProcessor(workersCount int, _ <-chan struct{}, _ func(), ppNext pipeProcessor) pipeProcessor {
|
|
|
|
return &pipePackJSONProcessor{
|
|
|
|
pp: pp,
|
|
|
|
ppNext: ppNext,
|
|
|
|
|
|
|
|
shards: make([]pipePackJSONProcessorShard, workersCount),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
type pipePackJSONProcessor struct {
|
|
|
|
pp *pipePackJSON
|
|
|
|
ppNext pipeProcessor
|
|
|
|
|
|
|
|
shards []pipePackJSONProcessorShard
|
|
|
|
}
|
|
|
|
|
|
|
|
type pipePackJSONProcessorShard struct {
|
|
|
|
pipePackJSONProcessorShardNopad
|
|
|
|
|
|
|
|
// The padding prevents false sharing on widespread platforms with 128 mod (cache line size) = 0 .
|
|
|
|
_ [128 - unsafe.Sizeof(pipePackJSONProcessorShardNopad{})%128]byte
|
|
|
|
}
|
|
|
|
|
|
|
|
type pipePackJSONProcessorShardNopad struct {
|
|
|
|
rc resultColumn
|
|
|
|
|
|
|
|
buf []byte
|
|
|
|
fields []Field
|
2024-05-29 01:52:13 +02:00
|
|
|
|
|
|
|
cs []*blockResultColumn
|
2024-05-25 21:36:16 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
func (ppp *pipePackJSONProcessor) writeBlock(workerID uint, br *blockResult) {
|
|
|
|
if len(br.timestamps) == 0 {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
shard := &ppp.shards[workerID]
|
|
|
|
|
|
|
|
shard.rc.name = ppp.pp.resultField
|
|
|
|
|
2024-05-29 01:52:13 +02:00
|
|
|
cs := shard.cs[:0]
|
|
|
|
if len(ppp.pp.fields) == 0 {
|
|
|
|
csAll := br.getColumns()
|
|
|
|
cs = append(cs, csAll...)
|
|
|
|
} else {
|
|
|
|
for _, f := range ppp.pp.fields {
|
|
|
|
c := br.getColumnByName(f)
|
|
|
|
cs = append(cs, c)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
shard.cs = cs
|
2024-05-25 21:36:16 +02:00
|
|
|
|
|
|
|
buf := shard.buf[:0]
|
|
|
|
fields := shard.fields
|
|
|
|
for rowIdx := range br.timestamps {
|
|
|
|
fields = fields[:0]
|
|
|
|
for _, c := range cs {
|
|
|
|
v := c.getValueAtRow(br, rowIdx)
|
|
|
|
fields = append(fields, Field{
|
|
|
|
Name: c.name,
|
|
|
|
Value: v,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
bufLen := len(buf)
|
2024-06-04 00:59:25 +02:00
|
|
|
buf = MarshalFieldsToJSON(buf, fields)
|
2024-05-25 21:36:16 +02:00
|
|
|
v := bytesutil.ToUnsafeString(buf[bufLen:])
|
|
|
|
shard.rc.addValue(v)
|
|
|
|
}
|
2024-05-25 22:13:32 +02:00
|
|
|
shard.fields = fields
|
2024-05-25 21:36:16 +02:00
|
|
|
|
|
|
|
br.addResultColumn(&shard.rc)
|
|
|
|
ppp.ppNext.writeBlock(workerID, br)
|
|
|
|
|
|
|
|
shard.rc.reset()
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ppp *pipePackJSONProcessor) flush() error {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func parsePackJSON(lex *lexer) (*pipePackJSON, error) {
|
|
|
|
if !lex.isKeyword("pack_json") {
|
|
|
|
return nil, fmt.Errorf("unexpected token: %q; want %q", lex.token, "pack_json")
|
|
|
|
}
|
|
|
|
lex.nextToken()
|
|
|
|
|
2024-05-29 01:52:13 +02:00
|
|
|
var fields []string
|
|
|
|
if lex.isKeyword("fields") {
|
|
|
|
lex.nextToken()
|
|
|
|
fs, err := parseFieldNamesInParens(lex)
|
|
|
|
if err != nil {
|
|
|
|
return nil, fmt.Errorf("cannot parse fields: %w", err)
|
|
|
|
}
|
|
|
|
if slices.Contains(fs, "*") {
|
|
|
|
fs = nil
|
|
|
|
}
|
|
|
|
fields = fs
|
|
|
|
}
|
|
|
|
|
2024-05-25 21:36:16 +02:00
|
|
|
// parse optional 'as ...` part
|
|
|
|
resultField := "_msg"
|
|
|
|
if lex.isKeyword("as") {
|
|
|
|
lex.nextToken()
|
2024-05-29 01:52:13 +02:00
|
|
|
}
|
|
|
|
if !lex.isKeyword("|", ")", "") {
|
2024-05-25 21:36:16 +02:00
|
|
|
field, err := parseFieldName(lex)
|
|
|
|
if err != nil {
|
|
|
|
return nil, fmt.Errorf("cannot parse result field for 'pack_json': %w", err)
|
|
|
|
}
|
|
|
|
resultField = field
|
|
|
|
}
|
|
|
|
|
|
|
|
pp := &pipePackJSON{
|
|
|
|
resultField: resultField,
|
2024-05-29 01:52:13 +02:00
|
|
|
fields: fields,
|
2024-05-25 21:36:16 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
return pp, nil
|
|
|
|
}
|