2019-05-22 23:16:55 +02:00
package storage
import (
"bytes"
"errors"
"fmt"
"io"
"path/filepath"
"sort"
"sync"
"sync/atomic"
"time"
"unsafe"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/encoding"
2019-06-11 00:56:37 +02:00
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fs"
2019-05-22 23:16:55 +02:00
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/memory"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/mergeset"
2019-09-24 20:10:22 +02:00
"github.com/VictoriaMetrics/VictoriaMetrics/lib/uint64set"
2019-08-13 20:35:19 +02:00
"github.com/VictoriaMetrics/VictoriaMetrics/lib/workingsetcache"
2019-05-22 23:16:55 +02:00
"github.com/VictoriaMetrics/fastcache"
xxhash "github.com/cespare/xxhash/v2"
)
const (
// Prefix for MetricName->TSID entries.
nsPrefixMetricNameToTSID = 0
// Prefix for Tag->MetricID entries.
2019-09-20 18:46:47 +02:00
nsPrefixTagToMetricIDs = 1
2019-05-22 23:16:55 +02:00
// Prefix for MetricID->TSID entries.
nsPrefixMetricIDToTSID = 2
// Prefix for MetricID->MetricName entries.
nsPrefixMetricIDToMetricName = 3
// Prefix for deleted MetricID entries.
2019-09-25 12:47:06 +02:00
nsPrefixDeletedMetricID = 4
2019-05-22 23:16:55 +02:00
// Prefix for Date->MetricID entries.
nsPrefixDateToMetricID = 5
2019-11-09 22:17:42 +01:00
// Prefix for (Date,Tag)->MetricID entries.
nsPrefixDateTagToMetricIDs = 6
2019-05-22 23:16:55 +02:00
)
2019-09-25 12:47:06 +02:00
func shouldCacheBlock ( item [ ] byte ) bool {
if len ( item ) == 0 {
return true
}
// Do not cache items starting from
switch item [ 0 ] {
2019-11-09 22:17:42 +01:00
case nsPrefixTagToMetricIDs , nsPrefixDateTagToMetricIDs :
// Do not cache blocks with tag->metricIDs and (date,tag)->metricIDs items, since:
2019-09-25 12:47:06 +02:00
// - these blocks are scanned sequentially, so the overhead
// on their unmarshaling is amortized by the sequential scan.
// - these blocks can occupy high amounts of RAM in cache
// and evict other frequently accessed blocks.
return false
case nsPrefixDeletedMetricID :
// Do not cache blocks with deleted metricIDs,
// since these metricIDs are loaded only once during app start.
return false
default :
return true
}
}
2019-05-22 23:16:55 +02:00
// indexDB represents an index db.
type indexDB struct {
2019-10-17 17:22:56 +02:00
// Atomic counters must go at the top of the structure in order to properly align by 8 bytes on 32-bit archs.
// See https://github.com/VictoriaMetrics/VictoriaMetrics/issues/212 .
2019-05-22 23:16:55 +02:00
refCount uint64
2019-10-17 17:22:56 +02:00
2019-11-08 18:57:57 +01:00
// The counter for newly created time series. It can be used for determining time series churn rate.
newTimeseriesCreated uint64
2019-10-17 17:22:56 +02:00
// The number of missing MetricID -> TSID entries.
// High rate for this value means corrupted indexDB.
missingTSIDsForMetricID uint64
// The number of calls to search for metric ids for recent hours.
recentHourMetricIDsSearchCalls uint64
// The number of cache hits during search for metric ids in recent hours.
recentHourMetricIDsSearchHits uint64
// The number of searches for metric ids by days.
dateMetricIDsSearchCalls uint64
// The number of successful searches for metric ids by days.
dateMetricIDsSearchHits uint64
2019-11-09 22:17:42 +01:00
// The number of calls for date range searches.
dateRangeSearchCalls uint64
// The number of hits for date range searches.
dateRangeSearchHits uint64
2019-12-02 19:44:18 +01:00
// missingMetricNamesForMetricID is a counter of missing MetricID -> MetricName entries.
// High rate may mean corrupted indexDB due to unclean shutdown.
// The db must be automatically recovered after that.
missingMetricNamesForMetricID uint64
2019-10-17 17:22:56 +02:00
mustDrop uint64
2019-11-09 22:17:42 +01:00
// Start date fully covered by per-day inverted index.
startDateForPerDayInvertedIndex uint64
2019-10-17 17:22:56 +02:00
name string
tb * mergeset . Table
2019-05-22 23:16:55 +02:00
extDB * indexDB
extDBLock sync . Mutex
// Cache for fast TagFilters -> TSIDs lookup.
2019-08-13 20:35:19 +02:00
tagCache * workingsetcache . Cache
2019-05-22 23:16:55 +02:00
// Cache for fast MetricID -> TSID lookup.
2019-08-13 20:35:19 +02:00
metricIDCache * workingsetcache . Cache
2019-05-22 23:16:55 +02:00
// Cache for fast MetricID -> MetricName lookup.
2019-08-13 20:35:19 +02:00
metricNameCache * workingsetcache . Cache
2019-05-22 23:16:55 +02:00
2020-03-30 23:44:41 +02:00
// Cache for useless TagFilters entries, which have no tag filters
2019-06-10 13:02:44 +02:00
// matching low number of metrics.
2019-08-13 20:35:19 +02:00
uselessTagFiltersCache * workingsetcache . Cache
2019-06-10 13:02:44 +02:00
2020-03-30 23:44:41 +02:00
// Cache for (date, tagFilter) -> metricIDsLen, which is used for reducing
// the amount of work when matching a set of filters.
metricIDsPerDateTagFilterCache * workingsetcache . Cache
2019-05-22 23:16:55 +02:00
indexSearchPool sync . Pool
2019-09-24 20:10:22 +02:00
// An inmemory set of deleted metricIDs.
2019-05-22 23:16:55 +02:00
//
2019-09-24 20:10:22 +02:00
// The set holds deleted metricIDs for the current db and for the extDB.
2019-05-22 23:16:55 +02:00
//
2019-09-24 20:10:22 +02:00
// It is safe to keep the set in memory even for big number of deleted
// metricIDs, since it usually requires 1 bit per deleted metricID.
2019-05-22 23:16:55 +02:00
deletedMetricIDs atomic . Value
deletedMetricIDsUpdateLock sync . Mutex
2019-06-09 18:06:53 +02:00
// Global lists of metric ids for the current and the previous hours.
// They are used for fast lookups on small time ranges covering
// up to two last hours.
currHourMetricIDs * atomic . Value
prevHourMetricIDs * atomic . Value
2019-05-22 23:16:55 +02:00
}
// openIndexDB opens index db from the given path with the given caches.
2019-08-13 20:35:19 +02:00
func openIndexDB ( path string , metricIDCache , metricNameCache * workingsetcache . Cache , currHourMetricIDs , prevHourMetricIDs * atomic . Value ) ( * indexDB , error ) {
2019-06-25 19:09:57 +02:00
if metricIDCache == nil {
logger . Panicf ( "BUG: metricIDCache must be non-nil" )
}
if metricNameCache == nil {
logger . Panicf ( "BUG: metricNameCache must be non-nil" )
}
if currHourMetricIDs == nil {
logger . Panicf ( "BUG: currHourMetricIDs must be non-nil" )
}
if prevHourMetricIDs == nil {
logger . Panicf ( "BUG: prevHourMetricIDs must be non-nil" )
}
2019-09-20 18:46:47 +02:00
tb , err := mergeset . OpenTable ( path , invalidateTagCache , mergeTagToMetricIDsRows )
2019-05-22 23:16:55 +02:00
if err != nil {
return nil , fmt . Errorf ( "cannot open indexDB %q: %s" , path , err )
}
name := filepath . Base ( path )
// Do not persist tagCache in files, since it is very volatile.
mem := memory . Allowed ( )
db := & indexDB {
refCount : 1 ,
tb : tb ,
name : name ,
2020-03-30 23:44:41 +02:00
tagCache : workingsetcache . New ( mem / 32 , time . Hour ) ,
metricIDCache : metricIDCache ,
metricNameCache : metricNameCache ,
uselessTagFiltersCache : workingsetcache . New ( mem / 128 , time . Hour ) ,
metricIDsPerDateTagFilterCache : workingsetcache . New ( mem / 128 , time . Hour ) ,
2019-06-10 13:02:44 +02:00
2019-06-09 18:06:53 +02:00
currHourMetricIDs : currHourMetricIDs ,
prevHourMetricIDs : prevHourMetricIDs ,
2019-05-22 23:16:55 +02:00
}
is := db . getIndexSearch ( )
dmis , err := is . loadDeletedMetricIDs ( )
db . putIndexSearch ( is )
if err != nil {
return nil , fmt . Errorf ( "cannot load deleted metricIDs: %s" , err )
}
db . setDeletedMetricIDs ( dmis )
2019-11-09 22:17:42 +01:00
is = db . getIndexSearch ( )
date , err := is . getStartDateForPerDayInvertedIndex ( )
db . putIndexSearch ( is )
if err != nil {
return nil , fmt . Errorf ( "cannot obtain start date for per-day inverted index: %s" , err )
}
db . startDateForPerDayInvertedIndex = date
2019-05-22 23:16:55 +02:00
return db , nil
}
// IndexDBMetrics contains essential metrics for indexDB.
type IndexDBMetrics struct {
TagCacheSize uint64
2019-07-09 23:47:29 +02:00
TagCacheSizeBytes uint64
2019-05-22 23:16:55 +02:00
TagCacheRequests uint64
TagCacheMisses uint64
2019-06-10 13:02:44 +02:00
UselessTagFiltersCacheSize uint64
2019-07-09 23:47:29 +02:00
UselessTagFiltersCacheSizeBytes uint64
2019-06-10 13:02:44 +02:00
UselessTagFiltersCacheRequests uint64
UselessTagFiltersCacheMisses uint64
2019-05-22 23:16:55 +02:00
DeletedMetricsCount uint64
IndexDBRefCount uint64
2019-11-08 18:57:57 +01:00
NewTimeseriesCreated uint64
2019-05-22 23:16:55 +02:00
MissingTSIDsForMetricID uint64
2019-06-09 18:06:53 +02:00
RecentHourMetricIDsSearchCalls uint64
RecentHourMetricIDsSearchHits uint64
DateMetricIDsSearchCalls uint64
DateMetricIDsSearchHits uint64
2019-11-09 22:17:42 +01:00
DateRangeSearchCalls uint64
DateRangeSearchHits uint64
2019-12-02 19:44:18 +01:00
MissingMetricNamesForMetricID uint64
2019-11-06 13:24:48 +01:00
IndexBlocksWithMetricIDsProcessed uint64
IndexBlocksWithMetricIDsIncorrectOrder uint64
2019-05-22 23:16:55 +02:00
mergeset . TableMetrics
}
func ( db * indexDB ) scheduleToDrop ( ) {
atomic . AddUint64 ( & db . mustDrop , 1 )
}
// UpdateMetrics updates m with metrics from the db.
func ( db * indexDB ) UpdateMetrics ( m * IndexDBMetrics ) {
var cs fastcache . Stats
2019-06-10 13:02:44 +02:00
cs . Reset ( )
2019-05-22 23:16:55 +02:00
db . tagCache . UpdateStats ( & cs )
m . TagCacheSize += cs . EntriesCount
2019-07-09 23:47:29 +02:00
m . TagCacheSizeBytes += cs . BytesSize
2019-05-22 23:16:55 +02:00
m . TagCacheRequests += cs . GetBigCalls
m . TagCacheMisses += cs . Misses
2019-06-10 13:02:44 +02:00
cs . Reset ( )
db . uselessTagFiltersCache . UpdateStats ( & cs )
m . UselessTagFiltersCacheSize += cs . EntriesCount
2019-07-09 23:47:29 +02:00
m . UselessTagFiltersCacheSizeBytes += cs . BytesSize
2019-11-08 12:16:40 +01:00
m . UselessTagFiltersCacheRequests += cs . GetCalls
2019-06-10 13:02:44 +02:00
m . UselessTagFiltersCacheMisses += cs . Misses
2019-09-24 20:10:22 +02:00
m . DeletedMetricsCount += uint64 ( db . getDeletedMetricIDs ( ) . Len ( ) )
2019-05-22 23:16:55 +02:00
m . IndexDBRefCount += atomic . LoadUint64 ( & db . refCount )
2019-11-08 18:57:57 +01:00
m . NewTimeseriesCreated += atomic . LoadUint64 ( & db . newTimeseriesCreated )
2019-05-22 23:16:55 +02:00
m . MissingTSIDsForMetricID += atomic . LoadUint64 ( & db . missingTSIDsForMetricID )
2019-06-09 18:06:53 +02:00
m . RecentHourMetricIDsSearchCalls += atomic . LoadUint64 ( & db . recentHourMetricIDsSearchCalls )
m . RecentHourMetricIDsSearchHits += atomic . LoadUint64 ( & db . recentHourMetricIDsSearchHits )
m . DateMetricIDsSearchCalls += atomic . LoadUint64 ( & db . dateMetricIDsSearchCalls )
m . DateMetricIDsSearchHits += atomic . LoadUint64 ( & db . dateMetricIDsSearchHits )
2019-05-22 23:16:55 +02:00
2019-11-09 22:17:42 +01:00
m . DateRangeSearchCalls += atomic . LoadUint64 ( & db . dateRangeSearchCalls )
m . DateRangeSearchHits += atomic . LoadUint64 ( & db . dateRangeSearchHits )
2019-12-02 19:44:18 +01:00
m . MissingMetricNamesForMetricID += atomic . LoadUint64 ( & db . missingMetricNamesForMetricID )
2019-11-06 13:24:48 +01:00
m . IndexBlocksWithMetricIDsProcessed = atomic . LoadUint64 ( & indexBlocksWithMetricIDsProcessed )
m . IndexBlocksWithMetricIDsIncorrectOrder = atomic . LoadUint64 ( & indexBlocksWithMetricIDsIncorrectOrder )
2019-05-22 23:16:55 +02:00
db . tb . UpdateMetrics ( & m . TableMetrics )
db . doExtDB ( func ( extDB * indexDB ) {
extDB . tb . UpdateMetrics ( & m . TableMetrics )
m . IndexDBRefCount += atomic . LoadUint64 ( & extDB . refCount )
} )
}
func ( db * indexDB ) doExtDB ( f func ( extDB * indexDB ) ) bool {
db . extDBLock . Lock ( )
extDB := db . extDB
if extDB != nil {
extDB . incRef ( )
}
db . extDBLock . Unlock ( )
if extDB == nil {
return false
}
f ( extDB )
extDB . decRef ( )
return true
}
// SetExtDB sets external db to search.
//
// It decrements refCount for the previous extDB.
func ( db * indexDB ) SetExtDB ( extDB * indexDB ) {
// Add deleted metricIDs from extDB to db.
if extDB != nil {
dmisExt := extDB . getDeletedMetricIDs ( )
2019-11-03 23:34:24 +01:00
db . updateDeletedMetricIDs ( dmisExt )
2019-05-22 23:16:55 +02:00
}
db . extDBLock . Lock ( )
prevExtDB := db . extDB
db . extDB = extDB
db . extDBLock . Unlock ( )
if prevExtDB != nil {
prevExtDB . decRef ( )
}
}
// MustClose closes db.
func ( db * indexDB ) MustClose ( ) {
db . decRef ( )
}
func ( db * indexDB ) incRef ( ) {
atomic . AddUint64 ( & db . refCount , 1 )
}
func ( db * indexDB ) decRef ( ) {
n := atomic . AddUint64 ( & db . refCount , ^ uint64 ( 0 ) )
2019-06-20 13:37:46 +02:00
if int64 ( n ) < 0 {
2019-05-22 23:16:55 +02:00
logger . Panicf ( "BUG: negative refCount: %d" , n )
}
if n > 0 {
return
}
tbPath := db . tb . Path ( )
db . tb . MustClose ( )
db . SetExtDB ( nil )
2019-06-25 13:39:17 +02:00
// Free space occupied by caches owned by db.
2019-08-13 20:35:19 +02:00
db . tagCache . Stop ( )
db . uselessTagFiltersCache . Stop ( )
2020-03-30 23:44:41 +02:00
db . metricIDsPerDateTagFilterCache . Stop ( )
2019-06-25 13:39:17 +02:00
db . tagCache = nil
db . metricIDCache = nil
db . metricNameCache = nil
db . uselessTagFiltersCache = nil
2020-03-30 23:44:41 +02:00
db . metricIDsPerDateTagFilterCache = nil
2019-06-25 13:39:17 +02:00
2019-05-22 23:16:55 +02:00
if atomic . LoadUint64 ( & db . mustDrop ) == 0 {
return
}
logger . Infof ( "dropping indexDB %q" , tbPath )
2019-06-12 00:53:43 +02:00
fs . MustRemoveAll ( tbPath )
2019-05-22 23:16:55 +02:00
logger . Infof ( "indexDB %q has been dropped" , tbPath )
}
func ( db * indexDB ) getFromTagCache ( key [ ] byte ) ( [ ] TSID , bool ) {
2019-08-14 00:50:20 +02:00
compressedBuf := tagBufPool . Get ( )
defer tagBufPool . Put ( compressedBuf )
compressedBuf . B = db . tagCache . GetBig ( compressedBuf . B [ : 0 ] , key )
if len ( compressedBuf . B ) == 0 {
2019-05-22 23:16:55 +02:00
return nil , false
}
2019-08-14 00:50:20 +02:00
buf := tagBufPool . Get ( )
defer tagBufPool . Put ( buf )
var err error
buf . B , err = encoding . DecompressZSTD ( buf . B [ : 0 ] , compressedBuf . B )
if err != nil {
logger . Panicf ( "FATAL: cannot decompress tsids from tagCache: %s" , err )
}
tsids , err := unmarshalTSIDs ( nil , buf . B )
2019-05-22 23:16:55 +02:00
if err != nil {
logger . Panicf ( "FATAL: cannot unmarshal tsids from tagCache: %s" , err )
}
return tsids , true
}
2019-08-14 00:50:20 +02:00
var tagBufPool bytesutil . ByteBufferPool
2019-05-22 23:16:55 +02:00
func ( db * indexDB ) putToTagCache ( tsids [ ] TSID , key [ ] byte ) {
2019-08-14 00:50:20 +02:00
buf := tagBufPool . Get ( )
buf . B = marshalTSIDs ( buf . B [ : 0 ] , tsids )
compressedBuf := tagBufPool . Get ( )
compressedBuf . B = encoding . CompressZSTDLevel ( compressedBuf . B [ : 0 ] , buf . B , 1 )
tagBufPool . Put ( buf )
db . tagCache . SetBig ( key , compressedBuf . B )
tagBufPool . Put ( compressedBuf )
2019-05-22 23:16:55 +02:00
}
func ( db * indexDB ) getFromMetricIDCache ( dst * TSID , metricID uint64 ) error {
// There is no need in checking for deleted metricIDs here, since they
// must be checked by the caller.
buf := ( * [ unsafe . Sizeof ( * dst ) ] byte ) ( unsafe . Pointer ( dst ) )
key := ( * [ unsafe . Sizeof ( metricID ) ] byte ) ( unsafe . Pointer ( & metricID ) )
tmp := db . metricIDCache . Get ( buf [ : 0 ] , key [ : ] )
if len ( tmp ) == 0 {
// The TSID for the given metricID wasn't found in the cache.
return io . EOF
}
if & tmp [ 0 ] != & buf [ 0 ] || len ( tmp ) != len ( buf ) {
return fmt . Errorf ( "corrupted MetricID->TSID cache: unexpected size for metricID=%d value; got %d bytes; want %d bytes" , metricID , len ( tmp ) , len ( buf ) )
}
return nil
}
func ( db * indexDB ) putToMetricIDCache ( metricID uint64 , tsid * TSID ) {
buf := ( * [ unsafe . Sizeof ( * tsid ) ] byte ) ( unsafe . Pointer ( tsid ) )
key := ( * [ unsafe . Sizeof ( metricID ) ] byte ) ( unsafe . Pointer ( & metricID ) )
db . metricIDCache . Set ( key [ : ] , buf [ : ] )
}
func ( db * indexDB ) getMetricNameFromCache ( dst [ ] byte , metricID uint64 ) [ ] byte {
// There is no need in checking for deleted metricIDs here, since they
// must be checked by the caller.
key := ( * [ unsafe . Sizeof ( metricID ) ] byte ) ( unsafe . Pointer ( & metricID ) )
return db . metricNameCache . Get ( dst , key [ : ] )
}
func ( db * indexDB ) putMetricNameToCache ( metricID uint64 , metricName [ ] byte ) {
key := ( * [ unsafe . Sizeof ( metricID ) ] byte ) ( unsafe . Pointer ( & metricID ) )
db . metricNameCache . Set ( key [ : ] , metricName )
}
2019-11-06 12:39:48 +01:00
func marshalTagFiltersKey ( dst [ ] byte , tfss [ ] * TagFilters , tr TimeRange , versioned bool ) [ ] byte {
2019-06-25 12:08:56 +02:00
prefix := ^ uint64 ( 0 )
if versioned {
prefix = atomic . LoadUint64 ( & tagFiltersKeyGen )
}
2019-11-08 12:16:40 +01:00
const cacheGranularityMs = 1000 * 10
2019-11-06 12:39:48 +01:00
startTime := ( uint64 ( tr . MinTimestamp ) / cacheGranularityMs ) * cacheGranularityMs
endTime := ( uint64 ( tr . MaxTimestamp ) / cacheGranularityMs ) * cacheGranularityMs
2019-05-22 23:16:55 +02:00
dst = encoding . MarshalUint64 ( dst , prefix )
2019-11-06 12:39:48 +01:00
dst = encoding . MarshalUint64 ( dst , startTime )
dst = encoding . MarshalUint64 ( dst , endTime )
2019-05-22 23:16:55 +02:00
for _ , tfs := range tfss {
dst = append ( dst , 0 ) // separator between tfs groups.
for i := range tfs . tfs {
dst = tfs . tfs [ i ] . Marshal ( dst )
}
}
return dst
}
func marshalTSIDs ( dst [ ] byte , tsids [ ] TSID ) [ ] byte {
dst = encoding . MarshalUint64 ( dst , uint64 ( len ( tsids ) ) )
for i := range tsids {
dst = tsids [ i ] . Marshal ( dst )
}
return dst
}
func unmarshalTSIDs ( dst [ ] TSID , src [ ] byte ) ( [ ] TSID , error ) {
if len ( src ) < 8 {
return dst , fmt . Errorf ( "cannot unmarshal the number of tsids from %d bytes; require at least %d bytes" , len ( src ) , 8 )
}
n := encoding . UnmarshalUint64 ( src )
src = src [ 8 : ]
dstLen := len ( dst )
if nn := dstLen + int ( n ) - cap ( dst ) ; nn > 0 {
dst = append ( dst [ : cap ( dst ) ] , make ( [ ] TSID , nn ) ... )
}
dst = dst [ : dstLen + int ( n ) ]
for i := 0 ; i < int ( n ) ; i ++ {
tail , err := dst [ dstLen + i ] . Unmarshal ( src )
if err != nil {
return dst , fmt . Errorf ( "cannot unmarshal tsid #%d out of %d: %s" , i , n , err )
}
src = tail
}
if len ( src ) > 0 {
return dst , fmt . Errorf ( "non-zero tail left after unmarshaling %d tsids; len(tail)=%d" , n , len ( src ) )
}
return dst , nil
}
2019-08-29 13:39:05 +02:00
func invalidateTagCache ( ) {
2019-05-22 23:16:55 +02:00
// This function must be fast, since it is called each
// time new timeseries is added.
atomic . AddUint64 ( & tagFiltersKeyGen , 1 )
}
var tagFiltersKeyGen uint64
// getTSIDByNameNoCreate fills the dst with TSID for the given metricName.
//
// It returns io.EOF if the given mn isn't found locally.
func ( db * indexDB ) getTSIDByNameNoCreate ( dst * TSID , metricName [ ] byte ) error {
is := db . getIndexSearch ( )
err := is . getTSIDByMetricName ( dst , metricName )
db . putIndexSearch ( is )
if err == nil {
return nil
}
if err != io . EOF {
return fmt . Errorf ( "cannot search TSID by MetricName %q: %s" , metricName , err )
}
// Do not search for the TSID in the external storage,
// since this function is already called by another indexDB instance.
// The TSID for the given mn wasn't found.
return io . EOF
}
type indexSearch struct {
db * indexDB
ts mergeset . TableSearch
kb bytesutil . ByteBuffer
2019-09-20 18:46:47 +02:00
mp tagToMetricIDsRowParser
2019-05-22 23:16:55 +02:00
// tsidByNameMisses and tsidByNameSkips is used for a performance
// hack in GetOrCreateTSIDByName. See the comment there.
tsidByNameMisses int
tsidByNameSkips int
}
// GetOrCreateTSIDByName fills the dst with TSID for the given metricName.
func ( is * indexSearch ) GetOrCreateTSIDByName ( dst * TSID , metricName [ ] byte ) error {
// A hack: skip searching for the TSID after many serial misses.
// This should improve insertion performance for big batches
// of new time series.
if is . tsidByNameMisses < 100 {
err := is . getTSIDByMetricName ( dst , metricName )
if err == nil {
is . tsidByNameMisses = 0
return nil
}
if err != io . EOF {
return fmt . Errorf ( "cannot search TSID by MetricName %q: %s" , metricName , err )
}
is . tsidByNameMisses ++
} else {
is . tsidByNameSkips ++
if is . tsidByNameSkips > 10000 {
is . tsidByNameSkips = 0
is . tsidByNameMisses = 0
}
}
// TSID for the given name wasn't found. Create it.
// It is OK if duplicate TSID for mn is created by concurrent goroutines.
// Metric results will be merged by mn after TableSearch.
if err := is . db . createTSIDByName ( dst , metricName ) ; err != nil {
return fmt . Errorf ( "cannot create TSID by MetricName %q: %s" , metricName , err )
}
return nil
}
func ( db * indexDB ) getIndexSearch ( ) * indexSearch {
v := db . indexSearchPool . Get ( )
if v == nil {
v = & indexSearch {
db : db ,
}
}
is := v . ( * indexSearch )
2019-09-25 12:47:06 +02:00
is . ts . Init ( db . tb , shouldCacheBlock )
2019-05-22 23:16:55 +02:00
return is
}
func ( db * indexDB ) putIndexSearch ( is * indexSearch ) {
is . ts . MustClose ( )
is . kb . Reset ( )
2019-09-20 18:46:47 +02:00
is . mp . Reset ( )
2019-05-22 23:16:55 +02:00
// Do not reset tsidByNameMisses and tsidByNameSkips,
// since they are used in GetOrCreateTSIDByName across call boundaries.
db . indexSearchPool . Put ( is )
}
func ( db * indexDB ) createTSIDByName ( dst * TSID , metricName [ ] byte ) error {
mn := GetMetricName ( )
defer PutMetricName ( mn )
if err := mn . Unmarshal ( metricName ) ; err != nil {
return fmt . Errorf ( "cannot unmarshal metricName %q: %s" , metricName , err )
}
if err := db . generateTSID ( dst , metricName , mn ) ; err != nil {
return fmt . Errorf ( "cannot generate TSID: %s" , err )
}
2019-11-08 12:16:40 +01:00
db . putMetricNameToCache ( dst . MetricID , metricName )
2019-05-22 23:16:55 +02:00
if err := db . createIndexes ( dst , mn ) ; err != nil {
return fmt . Errorf ( "cannot create indexes: %s" , err )
}
2019-08-29 13:39:05 +02:00
// There is no need in invalidating tag cache, since it is invalidated
// on db.tb flush via invalidateTagCache flushCallback passed to OpenTable.
2019-05-22 23:16:55 +02:00
2019-11-08 18:57:57 +01:00
atomic . AddUint64 ( & db . newTimeseriesCreated , 1 )
2019-05-22 23:16:55 +02:00
return nil
}
func ( db * indexDB ) generateTSID ( dst * TSID , metricName [ ] byte , mn * MetricName ) error {
// Search the TSID in the external storage.
// This is usually the db from the previous period.
var err error
if db . doExtDB ( func ( extDB * indexDB ) {
err = extDB . getTSIDByNameNoCreate ( dst , metricName )
} ) {
if err == nil {
// The TSID has been found in the external storage.
return nil
}
if err != io . EOF {
return fmt . Errorf ( "external search failed: %s" , err )
}
}
2019-09-20 18:46:47 +02:00
// The TSID wasn't found in the external storage.
2019-05-22 23:16:55 +02:00
// Generate it locally.
dst . MetricGroupID = xxhash . Sum64 ( mn . MetricGroup )
if len ( mn . Tags ) > 0 {
dst . JobID = uint32 ( xxhash . Sum64 ( mn . Tags [ 0 ] . Value ) )
}
if len ( mn . Tags ) > 1 {
dst . InstanceID = uint32 ( xxhash . Sum64 ( mn . Tags [ 1 ] . Value ) )
}
dst . MetricID = getUniqueUint64 ( )
return nil
}
func ( db * indexDB ) createIndexes ( tsid * TSID , mn * MetricName ) error {
// The order of index items is important.
// It guarantees index consistency.
items := getIndexItems ( )
// Create MetricName -> TSID index.
items . B = append ( items . B , nsPrefixMetricNameToTSID )
items . B = mn . Marshal ( items . B )
items . B = append ( items . B , kvSeparatorChar )
items . B = tsid . Marshal ( items . B )
items . Next ( )
// Create MetricID -> MetricName index.
items . B = marshalCommonPrefix ( items . B , nsPrefixMetricIDToMetricName )
items . B = encoding . MarshalUint64 ( items . B , tsid . MetricID )
items . B = mn . Marshal ( items . B )
items . Next ( )
// Create MetricID -> TSID index.
items . B = marshalCommonPrefix ( items . B , nsPrefixMetricIDToTSID )
items . B = encoding . MarshalUint64 ( items . B , tsid . MetricID )
items . B = tsid . Marshal ( items . B )
items . Next ( )
commonPrefix := kbPool . Get ( )
2019-09-20 18:46:47 +02:00
commonPrefix . B = marshalCommonPrefix ( commonPrefix . B [ : 0 ] , nsPrefixTagToMetricIDs )
2019-05-22 23:16:55 +02:00
// Create MetricGroup -> MetricID index.
items . B = append ( items . B , commonPrefix . B ... )
items . B = marshalTagValue ( items . B , nil )
items . B = marshalTagValue ( items . B , mn . MetricGroup )
items . B = encoding . MarshalUint64 ( items . B , tsid . MetricID )
items . Next ( )
// For each tag create tag -> MetricID index.
for i := range mn . Tags {
tag := & mn . Tags [ i ]
items . B = append ( items . B , commonPrefix . B ... )
items . B = tag . Marshal ( items . B )
items . B = encoding . MarshalUint64 ( items . B , tsid . MetricID )
items . Next ( )
}
kbPool . Put ( commonPrefix )
err := db . tb . AddItems ( items . Items )
putIndexItems ( items )
return err
}
type indexItems struct {
B [ ] byte
Items [ ] [ ] byte
start int
}
func ( ii * indexItems ) reset ( ) {
ii . B = ii . B [ : 0 ]
ii . Items = ii . Items [ : 0 ]
ii . start = 0
}
func ( ii * indexItems ) Next ( ) {
ii . Items = append ( ii . Items , ii . B [ ii . start : ] )
ii . start = len ( ii . B )
}
func getIndexItems ( ) * indexItems {
v := indexItemsPool . Get ( )
if v == nil {
return & indexItems { }
}
return v . ( * indexItems )
}
func putIndexItems ( ii * indexItems ) {
ii . reset ( )
indexItemsPool . Put ( ii )
}
var indexItemsPool sync . Pool
// SearchTagKeys returns all the tag keys.
func ( db * indexDB ) SearchTagKeys ( maxTagKeys int ) ( [ ] string , error ) {
// TODO: cache results?
tks := make ( map [ string ] struct { } )
is := db . getIndexSearch ( )
err := is . searchTagKeys ( tks , maxTagKeys )
db . putIndexSearch ( is )
if err != nil {
return nil , err
}
ok := db . doExtDB ( func ( extDB * indexDB ) {
is := extDB . getIndexSearch ( )
err = is . searchTagKeys ( tks , maxTagKeys )
extDB . putIndexSearch ( is )
} )
if ok && err != nil {
return nil , err
}
keys := make ( [ ] string , 0 , len ( tks ) )
for key := range tks {
keys = append ( keys , key )
}
// Do not sort keys, since they must be sorted by vmselect.
return keys , nil
}
func ( is * indexSearch ) searchTagKeys ( tks map [ string ] struct { } , maxTagKeys int ) error {
ts := & is . ts
kb := & is . kb
2019-09-20 18:46:47 +02:00
mp := & is . mp
mp . Reset ( )
2019-05-22 23:16:55 +02:00
dmis := is . db . getDeletedMetricIDs ( )
2019-09-20 18:46:47 +02:00
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixTagToMetricIDs )
prefix := kb . B
ts . Seek ( prefix )
2019-05-22 23:16:55 +02:00
for len ( tks ) < maxTagKeys && ts . NextItem ( ) {
item := ts . Item
2019-09-20 18:46:47 +02:00
if ! bytes . HasPrefix ( item , prefix ) {
2019-05-22 23:16:55 +02:00
break
}
2019-11-09 22:17:42 +01:00
if err := mp . Init ( item , nsPrefixTagToMetricIDs ) ; err != nil {
2019-09-20 18:46:47 +02:00
return err
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
if mp . IsDeletedTag ( dmis ) {
2019-05-22 23:16:55 +02:00
continue
}
// Store tag key.
2019-09-20 18:46:47 +02:00
tks [ string ( mp . Tag . Key ) ] = struct { } { }
2019-05-22 23:16:55 +02:00
// Search for the next tag key.
2019-09-20 18:46:47 +02:00
// The last char in kb.B must be tagSeparatorChar.
// Just increment it in order to jump to the next tag key.
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixTagToMetricIDs )
kb . B = marshalTagValue ( kb . B , mp . Tag . Key )
2019-05-22 23:16:55 +02:00
kb . B [ len ( kb . B ) - 1 ] ++
ts . Seek ( kb . B )
}
if err := ts . Error ( ) ; err != nil {
2019-09-20 18:46:47 +02:00
return fmt . Errorf ( "error during search for prefix %q: %s" , prefix , err )
2019-05-22 23:16:55 +02:00
}
return nil
}
// SearchTagValues returns all the tag values for the given tagKey
func ( db * indexDB ) SearchTagValues ( tagKey [ ] byte , maxTagValues int ) ( [ ] string , error ) {
// TODO: cache results?
tvs := make ( map [ string ] struct { } )
is := db . getIndexSearch ( )
2019-09-20 18:46:47 +02:00
err := is . searchTagValues ( tvs , tagKey , maxTagValues )
2019-05-22 23:16:55 +02:00
db . putIndexSearch ( is )
if err != nil {
return nil , err
}
ok := db . doExtDB ( func ( extDB * indexDB ) {
is := extDB . getIndexSearch ( )
2019-09-20 18:46:47 +02:00
err = is . searchTagValues ( tvs , tagKey , maxTagValues )
2019-05-22 23:16:55 +02:00
extDB . putIndexSearch ( is )
} )
if ok && err != nil {
return nil , err
}
tagValues := make ( [ ] string , 0 , len ( tvs ) )
for tv := range tvs {
tagValues = append ( tagValues , tv )
}
// Do not sort tagValues, since they must be sorted by vmselect.
return tagValues , nil
}
2019-09-20 18:46:47 +02:00
func ( is * indexSearch ) searchTagValues ( tvs map [ string ] struct { } , tagKey [ ] byte , maxTagValues int ) error {
2019-05-22 23:16:55 +02:00
ts := & is . ts
kb := & is . kb
2019-09-20 18:46:47 +02:00
mp := & is . mp
mp . Reset ( )
2019-05-22 23:16:55 +02:00
dmis := is . db . getDeletedMetricIDs ( )
2019-09-20 18:46:47 +02:00
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixTagToMetricIDs )
kb . B = marshalTagValue ( kb . B , tagKey )
prefix := kb . B
2019-05-22 23:16:55 +02:00
ts . Seek ( prefix )
for len ( tvs ) < maxTagValues && ts . NextItem ( ) {
2019-09-20 18:46:47 +02:00
item := ts . Item
if ! bytes . HasPrefix ( item , prefix ) {
2019-05-22 23:16:55 +02:00
break
}
2019-11-09 22:17:42 +01:00
if err := mp . Init ( item , nsPrefixTagToMetricIDs ) ; err != nil {
2019-09-20 18:46:47 +02:00
return err
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
if mp . IsDeletedTag ( dmis ) {
continue
2019-05-22 23:16:55 +02:00
}
// Store tag value
2019-09-20 18:46:47 +02:00
tvs [ string ( mp . Tag . Value ) ] = struct { } { }
2019-05-22 23:16:55 +02:00
2019-12-02 23:29:44 +01:00
if mp . MetricIDsLen ( ) < maxMetricIDsPerRow / 2 {
// There is no need in searching for the next tag value,
// since it is likely it is located in the next row,
// because the current row contains incomplete metricIDs set.
continue
}
2019-05-22 23:16:55 +02:00
// Search for the next tag value.
2019-09-20 18:46:47 +02:00
// The last char in kb.B must be tagSeparatorChar.
2019-12-02 23:29:44 +01:00
// Just increment it in order to jump to the next tag value.
2019-09-20 18:46:47 +02:00
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixTagToMetricIDs )
kb . B = marshalTagValue ( kb . B , mp . Tag . Key )
kb . B = marshalTagValue ( kb . B , mp . Tag . Value )
2019-05-22 23:16:55 +02:00
kb . B [ len ( kb . B ) - 1 ] ++
ts . Seek ( kb . B )
}
if err := ts . Error ( ) ; err != nil {
return fmt . Errorf ( "error when searching for tag name prefix %q: %s" , prefix , err )
}
return nil
}
// GetSeriesCount returns the approximate number of unique timeseries in the db.
//
// It includes the deleted series too and may count the same series
// up to two times - in db and extDB.
func ( db * indexDB ) GetSeriesCount ( ) ( uint64 , error ) {
is := db . getIndexSearch ( )
2019-06-10 11:27:44 +02:00
n , err := is . getSeriesCount ( )
2019-05-22 23:16:55 +02:00
db . putIndexSearch ( is )
if err != nil {
return 0 , err
}
var nExt uint64
ok := db . doExtDB ( func ( extDB * indexDB ) {
is := extDB . getIndexSearch ( )
2019-06-10 11:27:44 +02:00
nExt , err = is . getSeriesCount ( )
2019-05-22 23:16:55 +02:00
extDB . putIndexSearch ( is )
} )
if ok && err != nil {
return 0 , err
}
return n + nExt , nil
}
// searchMetricName appends metric name for the given metricID to dst
// and returns the result.
func ( db * indexDB ) searchMetricName ( dst [ ] byte , metricID uint64 ) ( [ ] byte , error ) {
is := db . getIndexSearch ( )
dst , err := is . searchMetricName ( dst , metricID )
db . putIndexSearch ( is )
if err != io . EOF {
return dst , err
}
// Try searching in the external indexDB.
if db . doExtDB ( func ( extDB * indexDB ) {
is := extDB . getIndexSearch ( )
dst , err = is . searchMetricName ( dst , metricID )
extDB . putIndexSearch ( is )
} ) {
return dst , err
}
// Cannot find MetricName for the given metricID. This may be the case
// when indexDB contains incomplete set of metricID -> metricName entries
// after a snapshot or due to unflushed entries.
2019-12-02 19:44:18 +01:00
atomic . AddUint64 ( & db . missingMetricNamesForMetricID , 1 )
// Mark the metricID as deleted, so it will be created again when new data point
// for the given time series will arrive.
if err := db . deleteMetricIDs ( [ ] uint64 { metricID } ) ; err != nil {
return dst , fmt . Errorf ( "cannot delete metricID for missing metricID->metricName entry; metricID=%d; error: %s" , metricID , err )
}
2019-05-22 23:16:55 +02:00
return dst , io . EOF
}
// DeleteTSIDs marks as deleted all the TSIDs matching the given tfss.
//
// The caller must reset all the caches which may contain the deleted TSIDs.
//
// Returns the number of metrics deleted.
func ( db * indexDB ) DeleteTSIDs ( tfss [ ] * TagFilters ) ( int , error ) {
if len ( tfss ) == 0 {
return 0 , nil
}
// Obtain metricIDs to delete.
2019-11-09 17:48:58 +01:00
tr := TimeRange {
MinTimestamp : 0 ,
2019-11-09 22:05:14 +01:00
MaxTimestamp : ( 1 << 63 ) - 1 ,
2019-11-09 17:48:58 +01:00
}
2019-05-22 23:16:55 +02:00
is := db . getIndexSearch ( )
2019-11-11 12:21:05 +01:00
metricIDs , err := is . searchMetricIDs ( tfss , tr , 2e9 )
2019-05-22 23:16:55 +02:00
db . putIndexSearch ( is )
if err != nil {
return 0 , err
}
2019-12-02 19:44:18 +01:00
if err := db . deleteMetricIDs ( metricIDs ) ; err != nil {
return 0 , err
}
// Delete TSIDs in the extDB.
deletedCount := len ( metricIDs )
if db . doExtDB ( func ( extDB * indexDB ) {
var n int
n , err = extDB . DeleteTSIDs ( tfss )
deletedCount += n
} ) {
if err != nil {
return deletedCount , fmt . Errorf ( "cannot delete tsids in extDB: %s" , err )
}
}
return deletedCount , nil
}
func ( db * indexDB ) deleteMetricIDs ( metricIDs [ ] uint64 ) error {
2019-05-22 23:16:55 +02:00
if len ( metricIDs ) == 0 {
// Nothing to delete
2019-12-02 19:44:18 +01:00
return nil
2019-05-22 23:16:55 +02:00
}
// Mark the found metricIDs as deleted.
items := getIndexItems ( )
for _ , metricID := range metricIDs {
2019-09-25 12:47:06 +02:00
items . B = append ( items . B , nsPrefixDeletedMetricID )
2019-05-22 23:16:55 +02:00
items . B = encoding . MarshalUint64 ( items . B , metricID )
items . Next ( )
}
2019-12-02 19:44:18 +01:00
err := db . tb . AddItems ( items . Items )
2019-05-22 23:16:55 +02:00
putIndexItems ( items )
if err != nil {
2019-12-02 19:44:18 +01:00
return err
2019-05-22 23:16:55 +02:00
}
// atomically add deleted metricIDs to an inmemory map.
2019-11-03 23:34:24 +01:00
dmis := & uint64set . Set { }
for _ , metricID := range metricIDs {
dmis . Add ( metricID )
}
db . updateDeletedMetricIDs ( dmis )
2019-05-22 23:16:55 +02:00
// Reset TagFilters -> TSIDS cache, since it may contain deleted TSIDs.
2019-08-29 13:39:05 +02:00
invalidateTagCache ( )
2019-05-22 23:16:55 +02:00
2019-08-27 13:20:53 +02:00
// Do not reset uselessTagFiltersCache, since the found metricIDs
// on cache miss are filtered out later with deletedMetricIDs.
2019-12-02 19:44:18 +01:00
return nil
2019-05-22 23:16:55 +02:00
}
2019-09-24 20:10:22 +02:00
func ( db * indexDB ) getDeletedMetricIDs ( ) * uint64set . Set {
return db . deletedMetricIDs . Load ( ) . ( * uint64set . Set )
2019-05-22 23:16:55 +02:00
}
2019-09-24 20:10:22 +02:00
func ( db * indexDB ) setDeletedMetricIDs ( dmis * uint64set . Set ) {
2019-05-22 23:16:55 +02:00
db . deletedMetricIDs . Store ( dmis )
}
2019-11-03 23:34:24 +01:00
func ( db * indexDB ) updateDeletedMetricIDs ( metricIDs * uint64set . Set ) {
2019-05-22 23:16:55 +02:00
db . deletedMetricIDsUpdateLock . Lock ( )
dmisOld := db . getDeletedMetricIDs ( )
2019-09-24 20:10:22 +02:00
dmisNew := dmisOld . Clone ( )
2019-11-03 23:34:24 +01:00
dmisNew . Union ( metricIDs )
2019-05-22 23:16:55 +02:00
db . setDeletedMetricIDs ( dmisNew )
db . deletedMetricIDsUpdateLock . Unlock ( )
}
2019-11-09 22:17:42 +01:00
func ( is * indexSearch ) getStartDateForPerDayInvertedIndex ( ) ( uint64 , error ) {
minDate := uint64 ( timestampFromTime ( time . Now ( ) ) ) / msecPerDay
kb := & is . kb
ts := & is . ts
kb . B = append ( kb . B [ : 0 ] , nsPrefixDateTagToMetricIDs )
prefix := kb . B
ts . Seek ( kb . B )
2019-11-09 23:38:49 +01:00
if ts . NextItem ( ) {
2019-11-09 22:17:42 +01:00
item := ts . Item
if ! bytes . HasPrefix ( item , prefix ) {
2019-11-09 23:38:49 +01:00
// The databse doesn't contain per-day inverted index yet.
2020-03-30 23:44:41 +02:00
return minDate , nil
2019-11-09 22:17:42 +01:00
}
suffix := item [ len ( prefix ) : ]
// Suffix must contain encoded 64-bit date.
if len ( suffix ) < 8 {
return 0 , fmt . Errorf ( "unexpected (date, tag)->metricIDs row len; must be at least 8 bytes; got %d bytes" , len ( suffix ) )
}
minDate = encoding . UnmarshalUint64 ( suffix )
2020-03-30 23:44:41 +02:00
return minDate , nil
2019-11-09 22:17:42 +01:00
}
if err := ts . Error ( ) ; err != nil {
return 0 , err
}
2019-11-16 11:11:22 +01:00
// There are no (date,tag)->metricIDs entries in the database yet.
2020-03-30 23:44:41 +02:00
return minDate , nil
2019-11-09 22:17:42 +01:00
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) loadDeletedMetricIDs ( ) ( * uint64set . Set , error ) {
dmis := & uint64set . Set { }
2019-05-22 23:16:55 +02:00
ts := & is . ts
kb := & is . kb
2019-09-25 12:47:06 +02:00
kb . B = append ( kb . B [ : 0 ] , nsPrefixDeletedMetricID )
2019-05-22 23:16:55 +02:00
ts . Seek ( kb . B )
for ts . NextItem ( ) {
item := ts . Item
if ! bytes . HasPrefix ( item , kb . B ) {
break
}
item = item [ len ( kb . B ) : ]
if len ( item ) != 8 {
return nil , fmt . Errorf ( "unexpected item len; got %d bytes; want %d bytes" , len ( item ) , 8 )
}
metricID := encoding . UnmarshalUint64 ( item )
2019-09-24 20:10:22 +02:00
dmis . Add ( metricID )
2019-05-22 23:16:55 +02:00
}
if err := ts . Error ( ) ; err != nil {
return nil , err
}
return dmis , nil
}
2019-09-23 21:34:04 +02:00
// searchTSIDs returns sorted tsids matching the given tfss over the given tr.
2019-05-22 23:16:55 +02:00
func ( db * indexDB ) searchTSIDs ( tfss [ ] * TagFilters , tr TimeRange , maxMetrics int ) ( [ ] TSID , error ) {
if len ( tfss ) == 0 {
return nil , nil
}
tfKeyBuf := tagFiltersKeyBufPool . Get ( )
defer tagFiltersKeyBufPool . Put ( tfKeyBuf )
2019-11-06 12:39:48 +01:00
tfKeyBuf . B = marshalTagFiltersKey ( tfKeyBuf . B [ : 0 ] , tfss , tr , true )
2019-05-22 23:16:55 +02:00
tsids , ok := db . getFromTagCache ( tfKeyBuf . B )
if ok {
// Fast path - tsids found in the cache.
return tsids , nil
}
// Slow path - search for tsids in the db and extDB.
is := db . getIndexSearch ( )
localTSIDs , err := is . searchTSIDs ( tfss , tr , maxMetrics )
db . putIndexSearch ( is )
if err != nil {
return nil , err
}
var extTSIDs [ ] TSID
if db . doExtDB ( func ( extDB * indexDB ) {
2019-06-25 12:08:56 +02:00
tfKeyExtBuf := tagFiltersKeyBufPool . Get ( )
defer tagFiltersKeyBufPool . Put ( tfKeyExtBuf )
// Data in extDB cannot be changed, so use unversioned keys for tag cache.
2019-11-06 12:39:48 +01:00
tfKeyExtBuf . B = marshalTagFiltersKey ( tfKeyExtBuf . B [ : 0 ] , tfss , tr , false )
2019-06-25 12:08:56 +02:00
tsids , ok := extDB . getFromTagCache ( tfKeyExtBuf . B )
2019-05-22 23:16:55 +02:00
if ok {
extTSIDs = tsids
return
}
is := extDB . getIndexSearch ( )
extTSIDs , err = is . searchTSIDs ( tfss , tr , maxMetrics )
extDB . putIndexSearch ( is )
2019-08-14 01:12:42 +02:00
sort . Slice ( extTSIDs , func ( i , j int ) bool { return extTSIDs [ i ] . Less ( & extTSIDs [ j ] ) } )
extDB . putToTagCache ( extTSIDs , tfKeyExtBuf . B )
2019-05-22 23:16:55 +02:00
} ) {
if err != nil {
return nil , err
}
}
// Merge localTSIDs with extTSIDs.
tsids = mergeTSIDs ( localTSIDs , extTSIDs )
// Sort the found tsids, since they must be passed to TSID search
// in the sorted order.
sort . Slice ( tsids , func ( i , j int ) bool { return tsids [ i ] . Less ( & tsids [ j ] ) } )
// Store TSIDs in the cache.
db . putToTagCache ( tsids , tfKeyBuf . B )
return tsids , err
}
var tagFiltersKeyBufPool bytesutil . ByteBufferPool
func ( is * indexSearch ) getTSIDByMetricName ( dst * TSID , metricName [ ] byte ) error {
dmis := is . db . getDeletedMetricIDs ( )
ts := & is . ts
kb := & is . kb
kb . B = append ( kb . B [ : 0 ] , nsPrefixMetricNameToTSID )
kb . B = append ( kb . B , metricName ... )
kb . B = append ( kb . B , kvSeparatorChar )
ts . Seek ( kb . B )
for ts . NextItem ( ) {
if ! bytes . HasPrefix ( ts . Item , kb . B ) {
// Nothing found.
return io . EOF
}
v := ts . Item [ len ( kb . B ) : ]
tail , err := dst . Unmarshal ( v )
if err != nil {
return fmt . Errorf ( "cannot unmarshal TSID: %s" , err )
}
if len ( tail ) > 0 {
return fmt . Errorf ( "unexpected non-empty tail left after unmarshaling TSID: %X" , tail )
}
2019-09-24 20:10:22 +02:00
if dmis . Len ( ) > 0 {
2019-05-22 23:16:55 +02:00
// Verify whether the dst is marked as deleted.
2019-09-24 20:10:22 +02:00
if dmis . Has ( dst . MetricID ) {
2019-05-25 20:51:11 +02:00
// The dst is deleted. Continue searching.
2019-05-22 23:16:55 +02:00
continue
}
}
// Found valid dst.
return nil
}
if err := ts . Error ( ) ; err != nil {
return fmt . Errorf ( "error when searching TSID by metricName; searchPrefix %q: %s" , kb . B , err )
}
// Nothing found
return io . EOF
}
func ( is * indexSearch ) searchMetricName ( dst [ ] byte , metricID uint64 ) ( [ ] byte , error ) {
metricName := is . db . getMetricNameFromCache ( dst , metricID )
if len ( metricName ) > len ( dst ) {
return metricName , nil
}
ts := & is . ts
kb := & is . kb
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixMetricIDToMetricName )
kb . B = encoding . MarshalUint64 ( kb . B , metricID )
if err := ts . FirstItemWithPrefix ( kb . B ) ; err != nil {
if err == io . EOF {
return dst , err
}
return dst , fmt . Errorf ( "error when searching metricName by metricID; searchPrefix %q: %s" , kb . B , err )
}
v := ts . Item [ len ( kb . B ) : ]
dst = append ( dst , v ... )
// There is no need in verifying whether the given metricID is deleted,
// since the filtering must be performed before calling this func.
is . db . putMetricNameToCache ( metricID , dst )
return dst , nil
}
func mergeTSIDs ( a , b [ ] TSID ) [ ] TSID {
if len ( b ) > len ( a ) {
a , b = b , a
}
if len ( b ) == 0 {
return a
}
m := make ( map [ uint64 ] TSID , len ( a ) )
for i := range a {
tsid := & a [ i ]
m [ tsid . MetricID ] = * tsid
}
for i := range b {
tsid := & b [ i ]
m [ tsid . MetricID ] = * tsid
}
tsids := make ( [ ] TSID , 0 , len ( m ) )
for _ , tsid := range m {
tsids = append ( tsids , tsid )
}
return tsids
}
2020-03-31 11:34:29 +02:00
func ( is * indexSearch ) containsTimeRange ( tr TimeRange ) ( bool , error ) {
ts := & is . ts
kb := & is . kb
// Verify whether the maximum date in `ts` covers tr.MinTimestamp.
minDate := uint64 ( tr . MinTimestamp ) / msecPerDay
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixDateToMetricID )
prefix := kb . B
kb . B = encoding . MarshalUint64 ( kb . B , minDate )
ts . Seek ( kb . B )
if ! ts . NextItem ( ) {
if err := ts . Error ( ) ; err != nil {
return false , fmt . Errorf ( "error when searching for minDate=%d, prefix %q: %s" , minDate , kb . B , err )
}
return false , nil
}
if ! bytes . HasPrefix ( ts . Item , prefix ) {
// minDate exceeds max date from ts.
return false , nil
}
return true , nil
}
2019-05-22 23:16:55 +02:00
func ( is * indexSearch ) searchTSIDs ( tfss [ ] * TagFilters , tr TimeRange , maxMetrics int ) ( [ ] TSID , error ) {
2020-03-31 11:34:29 +02:00
ok , err := is . containsTimeRange ( tr )
if err != nil {
return nil , err
}
if ! ok {
// Fast path - the index doesn't contain data for the given tr.
return nil , nil
}
2019-05-22 23:16:55 +02:00
metricIDs , err := is . searchMetricIDs ( tfss , tr , maxMetrics )
if err != nil {
return nil , err
}
if len ( metricIDs ) == 0 {
// Nothing found.
return nil , nil
}
// Obtain TSID values for the given metricIDs.
tsids := make ( [ ] TSID , len ( metricIDs ) )
i := 0
for _ , metricID := range metricIDs {
// Try obtaining TSIDs from db.tsidCache. This is much faster
// than scanning the mergeset if it contains a lot of metricIDs.
tsid := & tsids [ i ]
err := is . db . getFromMetricIDCache ( tsid , metricID )
if err == nil {
// Fast path - the tsid for metricID is found in cache.
i ++
continue
}
if err != io . EOF {
return nil , err
}
if err := is . getTSIDByMetricID ( tsid , metricID ) ; err != nil {
if err == io . EOF {
// Cannot find TSID for the given metricID.
// This may be the case on incomplete indexDB
// due to snapshot or due to unflushed entries.
// Just increment errors counter and skip it.
atomic . AddUint64 ( & is . db . missingTSIDsForMetricID , 1 )
continue
}
return nil , fmt . Errorf ( "cannot find tsid %d out of %d for metricID %d: %s" , i , len ( metricIDs ) , metricID , err )
}
is . db . putToMetricIDCache ( metricID , tsid )
i ++
}
tsids = tsids [ : i ]
// Do not sort the found tsids, since they will be sorted later.
return tsids , nil
}
func ( is * indexSearch ) getTSIDByMetricID ( dst * TSID , metricID uint64 ) error {
// There is no need in checking for deleted metricIDs here, since they
// must be checked by the caller.
ts := & is . ts
kb := & is . kb
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixMetricIDToTSID )
kb . B = encoding . MarshalUint64 ( kb . B , metricID )
if err := ts . FirstItemWithPrefix ( kb . B ) ; err != nil {
if err == io . EOF {
return err
}
return fmt . Errorf ( "error when searching TSID by metricID; searchPrefix %q: %s" , kb . B , err )
}
v := ts . Item [ len ( kb . B ) : ]
tail , err := dst . Unmarshal ( v )
if err != nil {
return fmt . Errorf ( "cannot unmarshal TSID=%X: %s" , v , err )
}
if len ( tail ) > 0 {
return fmt . Errorf ( "unexpected non-zero tail left after unmarshaling TSID: %X" , tail )
}
return nil
}
2019-06-10 11:27:44 +02:00
func ( is * indexSearch ) getSeriesCount ( ) ( uint64 , error ) {
ts := & is . ts
kb := & is . kb
2019-11-09 17:00:58 +01:00
mp := & is . mp
var metricIDsLen uint64
// Extract the number of series from ((__name__=value): metricIDs) rows
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixTagToMetricIDs )
kb . B = marshalTagValue ( kb . B , nil )
2019-05-22 23:16:55 +02:00
ts . Seek ( kb . B )
for ts . NextItem ( ) {
2019-11-09 17:00:58 +01:00
item := ts . Item
if ! bytes . HasPrefix ( item , kb . B ) {
2019-05-22 23:16:55 +02:00
break
}
2019-11-09 17:00:58 +01:00
tail := item [ len ( kb . B ) : ]
n := bytes . IndexByte ( tail , tagSeparatorChar )
if n < 0 {
return 0 , fmt . Errorf ( "invalid tag->metricIDs line %q: cannot find tagSeparatorChar %d" , item , tagSeparatorChar )
}
tail = tail [ n + 1 : ]
if err := mp . InitOnlyTail ( item , tail ) ; err != nil {
return 0 , err
}
2019-05-22 23:16:55 +02:00
// Take into account deleted timeseries too.
2019-11-09 17:00:58 +01:00
// It is OK if series can be counted multiple times in rare cases -
// the returned number is an estimation.
metricIDsLen += uint64 ( mp . MetricIDsLen ( ) )
2019-05-22 23:16:55 +02:00
}
if err := ts . Error ( ) ; err != nil {
return 0 , fmt . Errorf ( "error when counting unique timeseries: %s" , err )
}
2019-11-09 17:00:58 +01:00
return metricIDsLen , nil
2019-05-22 23:16:55 +02:00
}
2019-06-10 11:57:34 +02:00
// updateMetricIDsByMetricNameMatch matches metricName values for the given srcMetricIDs against tfs
2019-05-22 23:16:55 +02:00
// and adds matching metrics to metricIDs.
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) updateMetricIDsByMetricNameMatch ( metricIDs , srcMetricIDs * uint64set . Set , tfs [ ] * tagFilter ) error {
2019-05-22 23:16:55 +02:00
// sort srcMetricIDs in order to speed up Seek below.
2019-09-24 20:10:22 +02:00
sortedMetricIDs := srcMetricIDs . AppendTo ( nil )
2019-05-22 23:16:55 +02:00
metricName := kbPool . Get ( )
defer kbPool . Put ( metricName )
mn := GetMetricName ( )
defer PutMetricName ( mn )
for _ , metricID := range sortedMetricIDs {
var err error
metricName . B , err = is . searchMetricName ( metricName . B [ : 0 ] , metricID )
if err != nil {
2020-04-14 23:05:01 +02:00
if err == io . EOF {
// It is likely the metricID->metricName entry didn't propagate to inverted index yet.
// Skip this metricID for now.
continue
}
2019-05-22 23:16:55 +02:00
return fmt . Errorf ( "cannot find metricName by metricID %d: %s" , metricID , err )
}
if err := mn . Unmarshal ( metricName . B ) ; err != nil {
return fmt . Errorf ( "cannot unmarshal metricName %q: %s" , metricName . B , err )
}
// Match the mn against tfs.
ok , err := matchTagFilters ( mn , tfs , & is . kb )
if err != nil {
return fmt . Errorf ( "cannot match MetricName %s against tagFilters: %s" , mn , err )
}
if ! ok {
continue
}
2019-09-24 20:10:22 +02:00
metricIDs . Add ( metricID )
2019-05-22 23:16:55 +02:00
}
return nil
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) getTagFilterWithMinMetricIDsCountOptimized ( tfs * TagFilters , tr TimeRange , maxMetrics int ) ( * tagFilter , * uint64set . Set , error ) {
2019-08-19 15:04:12 +02:00
// Try fast path with the minimized number of maxMetrics.
maxMetricsAdjusted := is . adjustMaxMetricsAdaptive ( tr , maxMetrics )
minTf , minMetricIDs , err := is . getTagFilterWithMinMetricIDsCountAdaptive ( tfs , maxMetricsAdjusted )
if err == nil {
return minTf , minMetricIDs , nil
}
if err != errTooManyMetrics {
return nil , nil , err
}
// All the tag filters match too many metrics.
// Slow path: try filtering the matching metrics by time range.
// This should work well for cases when old metrics are constantly substituted
// by big number of new metrics. For example, prometheus-operator creates many new
// metrics for each new deployment.
//
// Allow fetching up to 20*maxMetrics metrics for the given time range
// in the hope these metricIDs will be filtered out by other filters later.
maxTimeRangeMetrics := 20 * maxMetrics
metricIDsForTimeRange , err := is . getMetricIDsForTimeRange ( tr , maxTimeRangeMetrics + 1 )
if err == errMissingMetricIDsForDate {
2019-11-06 12:39:48 +01:00
// Slow path: try to find the tag filter without maxMetrics adjustement.
2019-08-19 15:04:12 +02:00
minTf , minMetricIDs , err = is . getTagFilterWithMinMetricIDsCountAdaptive ( tfs , maxMetrics )
if err == nil {
return minTf , minMetricIDs , nil
}
if err != errTooManyMetrics {
return nil , nil , err
}
return nil , nil , fmt . Errorf ( "cannot find tag filter matching less than %d time series; " +
"either increase -search.maxUniqueTimeseries or use more specific tag filters" , maxMetrics )
}
if err != nil {
return nil , nil , err
}
2019-09-24 20:10:22 +02:00
if metricIDsForTimeRange . Len ( ) <= maxTimeRangeMetrics {
2019-08-19 15:04:12 +02:00
return nil , metricIDsForTimeRange , nil
}
// Slow path: try to select the tag filter without maxMetrics adjustement.
minTf , minMetricIDs , err = is . getTagFilterWithMinMetricIDsCountAdaptive ( tfs , maxMetrics )
if err == nil {
return minTf , minMetricIDs , nil
}
if err != errTooManyMetrics {
return nil , nil , err
}
return nil , nil , fmt . Errorf ( "more than %d time series found on the time range %s; either increase -search.maxUniqueTimeseries or shrink the time range" ,
2019-08-27 13:20:53 +02:00
maxMetrics , tr . String ( ) )
2019-08-19 15:04:12 +02:00
}
const maxDaysForDateMetricIDs = 40
func ( is * indexSearch ) adjustMaxMetricsAdaptive ( tr TimeRange , maxMetrics int ) int {
minDate := uint64 ( tr . MinTimestamp ) / msecPerDay
maxDate := uint64 ( tr . MaxTimestamp ) / msecPerDay
if maxDate - minDate > maxDaysForDateMetricIDs {
// Cannot reduce maxMetrics for the given time range,
// since it is expensive extracting metricIDs for the given tr.
return maxMetrics
}
hmPrev := is . db . prevHourMetricIDs . Load ( ) . ( * hourMetricIDs )
if ! hmPrev . isFull {
return maxMetrics
}
2019-09-24 20:10:22 +02:00
hourMetrics := hmPrev . m . Len ( )
2019-11-03 23:44:12 +01:00
if maxMetrics > hourMetrics {
2019-08-19 15:04:12 +02:00
// It is cheaper to filter on the hour or day metrics if the minimum
2019-11-03 23:44:12 +01:00
// number of matching metrics across tfs exceeds hourMetrics.
return hourMetrics
2019-08-19 15:04:12 +02:00
}
return maxMetrics
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) getTagFilterWithMinMetricIDsCountAdaptive ( tfs * TagFilters , maxMetrics int ) ( * tagFilter , * uint64set . Set , error ) {
2019-06-10 13:02:44 +02:00
kb := & is . kb
2019-07-11 13:37:27 +02:00
kb . B = append ( kb . B [ : 0 ] , uselessMultiTagFiltersKeyPrefix )
2019-06-10 13:02:44 +02:00
kb . B = encoding . MarshalUint64 ( kb . B , uint64 ( maxMetrics ) )
2019-07-11 13:37:27 +02:00
kb . B = tfs . marshal ( kb . B )
2019-06-10 13:02:44 +02:00
if len ( is . db . uselessTagFiltersCache . Get ( nil , kb . B ) ) > 0 {
// Skip useless work below, since the tfs doesn't contain tag filters matching less than maxMetrics metrics.
return nil , nil , errTooManyMetrics
}
2019-06-10 12:25:44 +02:00
// Iteratively increase maxAllowedMetrics up to maxMetrics in order to limit
// the time required for founding the tag filter with minimum matching metrics.
maxAllowedMetrics := 16
if maxAllowedMetrics > maxMetrics {
maxAllowedMetrics = maxMetrics
}
for {
minTf , minMetricIDs , err := is . getTagFilterWithMinMetricIDsCount ( tfs , maxAllowedMetrics )
2019-07-11 13:37:27 +02:00
if err != errTooManyMetrics {
if err != nil {
return nil , nil , err
}
2019-09-24 20:10:22 +02:00
if minMetricIDs . Len ( ) < maxAllowedMetrics {
2019-07-11 13:37:27 +02:00
// Found the tag filter with the minimum number of metrics.
return minTf , minMetricIDs , nil
}
2019-06-10 12:25:44 +02:00
}
// Too many metrics matched.
if maxAllowedMetrics >= maxMetrics {
2019-06-10 13:02:44 +02:00
// The tag filter with minimum matching metrics matches at least maxMetrics metrics.
2019-07-11 13:37:27 +02:00
kb . B = append ( kb . B [ : 0 ] , uselessMultiTagFiltersKeyPrefix )
2019-06-10 13:02:44 +02:00
kb . B = encoding . MarshalUint64 ( kb . B , uint64 ( maxMetrics ) )
2019-07-11 13:37:27 +02:00
kb . B = tfs . marshal ( kb . B )
is . db . uselessTagFiltersCache . Set ( kb . B , uselessTagFilterCacheValue )
2019-06-10 12:25:44 +02:00
return nil , nil , errTooManyMetrics
}
// Increase maxAllowedMetrics and try again.
maxAllowedMetrics *= 4
if maxAllowedMetrics > maxMetrics {
maxAllowedMetrics = maxMetrics
}
}
}
var errTooManyMetrics = errors . New ( "all the tag filters match too many metrics" )
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) getTagFilterWithMinMetricIDsCount ( tfs * TagFilters , maxMetrics int ) ( * tagFilter , * uint64set . Set , error ) {
var minMetricIDs * uint64set . Set
2019-05-22 23:16:55 +02:00
var minTf * tagFilter
2019-07-11 13:37:27 +02:00
kb := & is . kb
uselessTagFilters := 0
2019-05-22 23:16:55 +02:00
for i := range tfs . tfs {
tf := & tfs . tfs [ i ]
if tf . isNegative {
// Skip negative filters.
continue
}
2019-07-11 13:37:27 +02:00
kb . B = append ( kb . B [ : 0 ] , uselessSingleTagFilterKeyPrefix )
2019-09-17 22:12:27 +02:00
kb . B = encoding . MarshalUint64 ( kb . B , uint64 ( maxMetrics ) )
2019-07-11 13:37:27 +02:00
kb . B = tf . Marshal ( kb . B )
if len ( is . db . uselessTagFiltersCache . Get ( nil , kb . B ) ) > 0 {
// Skip useless work below, since the tf matches at least maxMetrics metrics.
uselessTagFilters ++
continue
}
2019-06-10 11:57:34 +02:00
metricIDs , err := is . getMetricIDsForTagFilter ( tf , maxMetrics )
2019-05-22 23:16:55 +02:00
if err != nil {
if err == errFallbackToMetricNameMatch {
// Skip tag filters requiring to scan for too many metrics.
2019-09-18 18:07:46 +02:00
kb . B = append ( kb . B [ : 0 ] , uselessSingleTagFilterKeyPrefix )
kb . B = encoding . MarshalUint64 ( kb . B , uint64 ( maxMetrics ) )
kb . B = tf . Marshal ( kb . B )
is . db . uselessTagFiltersCache . Set ( kb . B , uselessTagFilterCacheValue )
uselessTagFilters ++
2019-05-22 23:16:55 +02:00
continue
}
return nil , nil , fmt . Errorf ( "cannot find MetricIDs for tagFilter %s: %s" , tf , err )
}
2019-09-24 20:10:22 +02:00
if metricIDs . Len ( ) >= maxMetrics {
2019-07-11 13:37:27 +02:00
// The tf matches at least maxMetrics. Skip it
kb . B = append ( kb . B [ : 0 ] , uselessSingleTagFilterKeyPrefix )
2019-09-17 22:12:27 +02:00
kb . B = encoding . MarshalUint64 ( kb . B , uint64 ( maxMetrics ) )
2019-07-11 13:37:27 +02:00
kb . B = tf . Marshal ( kb . B )
is . db . uselessTagFiltersCache . Set ( kb . B , uselessTagFilterCacheValue )
uselessTagFilters ++
continue
}
minMetricIDs = metricIDs
minTf = tf
2019-09-24 20:10:22 +02:00
maxMetrics = minMetricIDs . Len ( )
2019-07-11 13:37:27 +02:00
if maxMetrics <= 1 {
// There is no need in inspecting other filters, since minTf
// already matches 0 or 1 metric.
break
2019-05-22 23:16:55 +02:00
}
}
if minTf != nil {
return minTf , minMetricIDs , nil
}
2019-07-11 13:37:27 +02:00
if uselessTagFilters == len ( tfs . tfs ) {
// All the tag filters return at least maxMetrics entries.
return nil , nil , errTooManyMetrics
}
2019-05-22 23:16:55 +02:00
// There is no positive filter with small number of matching metrics.
2019-09-20 10:53:42 +02:00
// Create it, so it matches all the MetricIDs.
2019-09-17 22:12:27 +02:00
kb . B = append ( kb . B [ : 0 ] , uselessNegativeTagFilterKeyPrefix )
kb . B = encoding . MarshalUint64 ( kb . B , uint64 ( maxMetrics ) )
kb . B = tfs . marshal ( kb . B )
if len ( is . db . uselessTagFiltersCache . Get ( nil , kb . B ) ) > 0 {
return nil , nil , errTooManyMetrics
}
2019-09-24 20:10:22 +02:00
metricIDs := & uint64set . Set { }
2019-09-20 10:53:42 +02:00
if err := is . updateMetricIDsAll ( metricIDs , maxMetrics ) ; err != nil {
2019-05-22 23:16:55 +02:00
return nil , nil , err
}
2019-09-24 20:10:22 +02:00
if metricIDs . Len ( ) >= maxMetrics {
2019-09-17 22:12:27 +02:00
kb . B = append ( kb . B [ : 0 ] , uselessNegativeTagFilterKeyPrefix )
kb . B = encoding . MarshalUint64 ( kb . B , uint64 ( maxMetrics ) )
kb . B = tfs . marshal ( kb . B )
is . db . uselessTagFiltersCache . Set ( kb . B , uselessTagFilterCacheValue )
}
2019-05-22 23:16:55 +02:00
return nil , metricIDs , nil
}
func matchTagFilters ( mn * MetricName , tfs [ ] * tagFilter , kb * bytesutil . ByteBuffer ) ( bool , error ) {
2019-09-20 18:46:47 +02:00
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixTagToMetricIDs )
2019-05-22 23:16:55 +02:00
2019-11-21 20:34:32 +01:00
for i , tf := range tfs {
2019-05-22 23:16:55 +02:00
if len ( tf . key ) == 0 {
// Match against mn.MetricGroup.
b := marshalTagValue ( kb . B , nil )
b = marshalTagValue ( b , mn . MetricGroup )
kb . B = b [ : len ( kb . B ) ]
ok , err := matchTagFilter ( b , tf )
if err != nil {
return false , fmt . Errorf ( "cannot match MetricGroup %q with tagFilter %s: %s" , mn . MetricGroup , tf , err )
}
if ! ok {
2019-11-21 20:34:32 +01:00
// Move failed tf to start.
// This should reduce the amount of useless work for the next mn.
if i > 0 {
tfs [ 0 ] , tfs [ i ] = tfs [ i ] , tfs [ 0 ]
}
2019-05-22 23:16:55 +02:00
return false , nil
}
continue
}
// Search for matching tag name.
tagMatched := false
for j := range mn . Tags {
tag := & mn . Tags [ j ]
if string ( tag . Key ) != string ( tf . key ) {
continue
}
2019-07-30 14:14:09 +02:00
// Found the matching tag name. Match the value.
2019-05-22 23:16:55 +02:00
b := tag . Marshal ( kb . B )
kb . B = b [ : len ( kb . B ) ]
ok , err := matchTagFilter ( b , tf )
if err != nil {
return false , fmt . Errorf ( "cannot match tag %q with tagFilter %s: %s" , tag , tf , err )
}
if ! ok {
2019-11-21 20:34:32 +01:00
// Move failed tf to start.
// This should reduce the amount of useless work for the next mn.
if i > 0 {
tfs [ 0 ] , tfs [ i ] = tfs [ i ] , tfs [ 0 ]
}
2019-05-22 23:16:55 +02:00
return false , nil
}
tagMatched = true
break
}
2019-07-30 14:14:09 +02:00
if ! tagMatched && ! tf . isNegative {
2019-05-22 23:16:55 +02:00
// Matching tag name wasn't found.
2019-11-21 20:34:32 +01:00
// Move failed tf to start.
// This should reduce the amount of useless work for the next mn.
if i > 0 {
tfs [ 0 ] , tfs [ i ] = tfs [ i ] , tfs [ 0 ]
}
2019-05-22 23:16:55 +02:00
return false , nil
}
}
return true , nil
}
func matchTagFilter ( b [ ] byte , tf * tagFilter ) ( bool , error ) {
if ! bytes . HasPrefix ( b , tf . prefix ) {
return tf . isNegative , nil
}
ok , err := tf . matchSuffix ( b [ len ( tf . prefix ) : ] )
if err != nil {
return false , err
}
if ! ok {
return tf . isNegative , nil
}
return ! tf . isNegative , nil
}
func ( is * indexSearch ) searchMetricIDs ( tfss [ ] * TagFilters , tr TimeRange , maxMetrics int ) ( [ ] uint64 , error ) {
2019-09-24 20:10:22 +02:00
metricIDs := & uint64set . Set { }
2019-05-22 23:16:55 +02:00
for _ , tfs := range tfss {
if len ( tfs . tfs ) == 0 {
// Return all the metric ids
2019-09-20 10:53:42 +02:00
if err := is . updateMetricIDsAll ( metricIDs , maxMetrics + 1 ) ; err != nil {
2019-05-22 23:16:55 +02:00
return nil , err
}
2019-09-24 20:10:22 +02:00
if metricIDs . Len ( ) > maxMetrics {
2019-05-22 23:16:55 +02:00
return nil , fmt . Errorf ( "the number or unique timeseries exceeds %d; either narrow down the search or increase -search.maxUniqueTimeseries" , maxMetrics )
}
// Stop the iteration, since we cannot find more metric ids with the remaining tfss.
break
}
2019-06-10 11:49:59 +02:00
if err := is . updateMetricIDsForTagFilters ( metricIDs , tfs , tr , maxMetrics + 1 ) ; err != nil {
2019-05-22 23:16:55 +02:00
return nil , err
}
2019-09-24 20:10:22 +02:00
if metricIDs . Len ( ) > maxMetrics {
2019-05-22 23:16:55 +02:00
return nil , fmt . Errorf ( "the number or matching unique timeseries exceeds %d; either narrow down the search or increase -search.maxUniqueTimeseries" , maxMetrics )
}
}
2019-09-24 20:10:22 +02:00
if metricIDs . Len ( ) == 0 {
2019-05-22 23:16:55 +02:00
// Nothing found
return nil , nil
}
2019-09-24 20:10:22 +02:00
sortedMetricIDs := metricIDs . AppendTo ( nil )
2019-05-22 23:16:55 +02:00
// Filter out deleted metricIDs.
dmis := is . db . getDeletedMetricIDs ( )
2019-09-24 20:10:22 +02:00
if dmis . Len ( ) > 0 {
2019-06-10 11:49:59 +02:00
metricIDsFiltered := sortedMetricIDs [ : 0 ]
for _ , metricID := range sortedMetricIDs {
2019-09-24 20:10:22 +02:00
if ! dmis . Has ( metricID ) {
2019-05-22 23:16:55 +02:00
metricIDsFiltered = append ( metricIDsFiltered , metricID )
}
}
2019-06-10 11:49:59 +02:00
sortedMetricIDs = metricIDsFiltered
2019-05-22 23:16:55 +02:00
}
2019-06-10 11:49:59 +02:00
return sortedMetricIDs , nil
2019-05-22 23:16:55 +02:00
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) updateMetricIDsForTagFilters ( metricIDs * uint64set . Set , tfs * TagFilters , tr TimeRange , maxMetrics int ) error {
2020-03-13 21:42:22 +01:00
err := is . tryUpdatingMetricIDsForDateRange ( metricIDs , tfs , tr , maxMetrics )
if err == nil {
2019-11-09 22:17:42 +01:00
// Fast path: found metricIDs by date range.
return nil
}
2020-03-13 21:42:22 +01:00
if err != errFallbackToMetricNameMatch {
return err
}
2019-11-08 12:16:40 +01:00
2019-11-09 22:17:42 +01:00
// Slow path - try searching over the whole inverted index.
2020-04-01 16:40:18 +02:00
// Sort tag filters for faster ts.Seek below.
sort . Slice ( tfs . tfs , func ( i , j int ) bool {
return tfs . tfs [ i ] . Less ( & tfs . tfs [ j ] )
} )
2019-08-19 15:04:12 +02:00
minTf , minMetricIDs , err := is . getTagFilterWithMinMetricIDsCountOptimized ( tfs , tr , maxMetrics )
2019-06-10 12:25:44 +02:00
if err != nil {
2019-08-19 15:04:12 +02:00
return err
2019-05-22 23:16:55 +02:00
}
// Find intersection of minTf with other tfs.
var tfsPostponed [ ] * tagFilter
2019-09-18 22:24:12 +02:00
successfulIntersects := 0
2019-05-22 23:16:55 +02:00
for i := range tfs . tfs {
tf := & tfs . tfs [ i ]
if tf == minTf {
continue
}
2019-06-10 11:57:34 +02:00
mIDs , err := is . intersectMetricIDsWithTagFilter ( tf , minMetricIDs )
2019-05-22 23:16:55 +02:00
if err == errFallbackToMetricNameMatch {
// The tag filter requires too many index scans. Postpone it,
// so tag filters with lower number of index scans may be applied.
tfsPostponed = append ( tfsPostponed , tf )
continue
}
if err != nil {
return err
}
minMetricIDs = mIDs
2019-09-18 22:24:12 +02:00
successfulIntersects ++
}
if len ( tfsPostponed ) > 0 && successfulIntersects == 0 {
return is . updateMetricIDsByMetricNameMatch ( metricIDs , minMetricIDs , tfsPostponed )
2019-05-22 23:16:55 +02:00
}
for i , tf := range tfsPostponed {
2019-06-10 11:57:34 +02:00
mIDs , err := is . intersectMetricIDsWithTagFilter ( tf , minMetricIDs )
2019-05-22 23:16:55 +02:00
if err == errFallbackToMetricNameMatch {
2019-06-10 11:57:34 +02:00
return is . updateMetricIDsByMetricNameMatch ( metricIDs , minMetricIDs , tfsPostponed [ i : ] )
2019-05-22 23:16:55 +02:00
}
if err != nil {
return err
}
minMetricIDs = mIDs
}
2020-01-17 21:10:36 +01:00
metricIDs . UnionMayOwn ( minMetricIDs )
2019-05-22 23:16:55 +02:00
return nil
}
2019-07-11 13:37:27 +02:00
const (
2019-09-17 22:12:27 +02:00
uselessSingleTagFilterKeyPrefix = 0
uselessMultiTagFiltersKeyPrefix = 1
uselessNegativeTagFilterKeyPrefix = 2
2019-11-05 13:17:59 +01:00
uselessTagIntersectKeyPrefix = 3
2019-07-11 13:37:27 +02:00
)
var uselessTagFilterCacheValue = [ ] byte ( "1" )
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) getMetricIDsForTagFilter ( tf * tagFilter , maxMetrics int ) ( * uint64set . Set , error ) {
2019-05-22 23:16:55 +02:00
if tf . isNegative {
logger . Panicf ( "BUG: isNegative must be false" )
}
2019-09-24 20:10:22 +02:00
metricIDs := & uint64set . Set { }
2019-05-22 23:16:55 +02:00
if len ( tf . orSuffixes ) > 0 {
// Fast path for orSuffixes - seek for rows for each value from orSuffxies.
2019-06-10 11:57:34 +02:00
if err := is . updateMetricIDsForOrSuffixesNoFilter ( tf , maxMetrics , metricIDs ) ; err != nil {
2019-09-20 18:46:47 +02:00
if err == errFallbackToMetricNameMatch {
return nil , err
}
return nil , fmt . Errorf ( "error when searching for metricIDs for tagFilter in fast path: %s; tagFilter=%s" , err , tf )
2019-05-22 23:16:55 +02:00
}
return metricIDs , nil
}
2019-06-27 15:15:25 +02:00
// Slow path - scan for all the rows with the given prefix.
2019-11-05 15:59:31 +01:00
maxLoops := maxMetrics * maxIndexScanSlowLoopsPerMetric
2019-06-27 15:15:25 +02:00
err := is . getMetricIDsForTagFilterSlow ( tf , maxLoops , func ( metricID uint64 ) bool {
2019-09-24 20:10:22 +02:00
metricIDs . Add ( metricID )
return metricIDs . Len ( ) < maxMetrics
2019-06-27 15:15:25 +02:00
} )
if err != nil {
2019-09-20 18:46:47 +02:00
if err == errFallbackToMetricNameMatch {
return nil , err
}
return nil , fmt . Errorf ( "error when searching for metricIDs for tagFilter in slow path: %s; tagFilter=%s" , err , tf )
2019-06-27 15:15:25 +02:00
}
return metricIDs , nil
}
func ( is * indexSearch ) getMetricIDsForTagFilterSlow ( tf * tagFilter , maxLoops int , f func ( metricID uint64 ) bool ) error {
if len ( tf . orSuffixes ) > 0 {
logger . Panicf ( "BUG: the getMetricIDsForTagFilterSlow must be called only for empty tf.orSuffixes; got %s" , tf . orSuffixes )
}
// Scan all the rows with tf.prefix and call f on every tf match.
2019-05-22 23:16:55 +02:00
loops := 0
ts := & is . ts
2019-06-27 15:15:25 +02:00
kb := & is . kb
2019-09-20 18:46:47 +02:00
mp := & is . mp
mp . Reset ( )
var prevMatchingSuffix [ ] byte
2019-06-27 15:15:25 +02:00
var prevMatch bool
2019-09-20 18:46:47 +02:00
prefix := tf . prefix
ts . Seek ( prefix )
2019-06-27 15:15:25 +02:00
for ts . NextItem ( ) {
2019-09-20 18:46:47 +02:00
item := ts . Item
if ! bytes . HasPrefix ( item , prefix ) {
return nil
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
tail := item [ len ( prefix ) : ]
n := bytes . IndexByte ( tail , tagSeparatorChar )
if n < 0 {
return fmt . Errorf ( "invalid tag->metricIDs line %q: cannot find tagSeparatorChar=%d" , item , tagSeparatorChar )
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
suffix := tail [ : n + 1 ]
tail = tail [ n + 1 : ]
if err := mp . InitOnlyTail ( item , tail ) ; err != nil {
return err
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
if prevMatch && string ( suffix ) == string ( prevMatchingSuffix ) {
2019-06-27 15:15:25 +02:00
// Fast path: the same tag value found.
// There is no need in checking it again with potentially
// slow tf.matchSuffix, which may call regexp.
2019-09-23 19:40:38 +02:00
loops += mp . MetricIDsLen ( )
2019-09-20 18:46:47 +02:00
if loops > maxLoops {
return errFallbackToMetricNameMatch
}
2019-09-23 19:40:38 +02:00
mp . ParseMetricIDs ( )
2019-09-20 18:46:47 +02:00
for _ , metricID := range mp . MetricIDs {
if ! f ( metricID ) {
return nil
}
2019-06-27 15:15:25 +02:00
}
continue
}
2019-05-22 23:16:55 +02:00
2019-09-20 18:46:47 +02:00
// Slow path: need tf.matchSuffix call.
ok , err := tf . matchSuffix ( suffix )
2019-05-22 23:16:55 +02:00
if err != nil {
2019-09-20 18:46:47 +02:00
return fmt . Errorf ( "error when matching %s against suffix %q: %s" , tf , suffix , err )
2019-05-22 23:16:55 +02:00
}
if ! ok {
2019-06-27 15:15:25 +02:00
prevMatch = false
2019-12-02 23:29:44 +01:00
if mp . MetricIDsLen ( ) < maxMetricIDsPerRow / 2 {
// If the current row contains non-full metricIDs list,
// then it is likely the next row contains the next tag value.
// So skip seeking for the next tag value, since it will be slower than just ts.NextItem call.
continue
}
2019-06-27 15:15:25 +02:00
// Optimization: skip all the metricIDs for the given tag value
2019-09-20 18:46:47 +02:00
kb . B = append ( kb . B [ : 0 ] , item [ : len ( item ) - len ( tail ) ] ... )
2019-06-27 15:15:25 +02:00
// The last char in kb.B must be tagSeparatorChar. Just increment it
// in order to jump to the next tag value.
if len ( kb . B ) == 0 || kb . B [ len ( kb . B ) - 1 ] != tagSeparatorChar || tagSeparatorChar >= 0xff {
return fmt . Errorf ( "data corruption: the last char in k=%X must be %X" , kb . B , tagSeparatorChar )
}
kb . B [ len ( kb . B ) - 1 ] ++
ts . Seek ( kb . B )
2019-05-22 23:16:55 +02:00
continue
}
2019-06-27 15:15:25 +02:00
prevMatch = true
2019-09-20 18:46:47 +02:00
prevMatchingSuffix = append ( prevMatchingSuffix [ : 0 ] , suffix ... )
2019-09-23 19:40:38 +02:00
loops += mp . MetricIDsLen ( )
2019-09-20 18:46:47 +02:00
if loops > maxLoops {
return errFallbackToMetricNameMatch
}
2019-09-23 19:40:38 +02:00
mp . ParseMetricIDs ( )
2019-09-20 18:46:47 +02:00
for _ , metricID := range mp . MetricIDs {
if ! f ( metricID ) {
return nil
}
2019-06-27 15:15:25 +02:00
}
2019-05-22 23:16:55 +02:00
}
if err := ts . Error ( ) ; err != nil {
2019-09-20 18:46:47 +02:00
return fmt . Errorf ( "error when searching for tag filter prefix %q: %s" , prefix , err )
2019-05-22 23:16:55 +02:00
}
2019-06-27 15:15:25 +02:00
return nil
2019-05-22 23:16:55 +02:00
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) updateMetricIDsForOrSuffixesNoFilter ( tf * tagFilter , maxMetrics int , metricIDs * uint64set . Set ) error {
2019-05-22 23:16:55 +02:00
if tf . isNegative {
logger . Panicf ( "BUG: isNegative must be false" )
}
kb := kbPool . Get ( )
defer kbPool . Put ( kb )
for _ , orSuffix := range tf . orSuffixes {
kb . B = append ( kb . B [ : 0 ] , tf . prefix ... )
kb . B = append ( kb . B , orSuffix ... )
kb . B = append ( kb . B , tagSeparatorChar )
2019-06-10 11:57:34 +02:00
if err := is . updateMetricIDsForOrSuffixNoFilter ( kb . B , maxMetrics , metricIDs ) ; err != nil {
2019-05-22 23:16:55 +02:00
return err
}
2019-09-24 20:10:22 +02:00
if metricIDs . Len ( ) >= maxMetrics {
2019-05-22 23:16:55 +02:00
return nil
}
}
return nil
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) updateMetricIDsForOrSuffixesWithFilter ( tf * tagFilter , metricIDs , filter * uint64set . Set ) error {
sortedFilter := filter . AppendTo ( nil )
2019-05-22 23:16:55 +02:00
kb := kbPool . Get ( )
defer kbPool . Put ( kb )
for _ , orSuffix := range tf . orSuffixes {
kb . B = append ( kb . B [ : 0 ] , tf . prefix ... )
kb . B = append ( kb . B , orSuffix ... )
kb . B = append ( kb . B , tagSeparatorChar )
2019-06-10 11:57:34 +02:00
if err := is . updateMetricIDsForOrSuffixWithFilter ( kb . B , metricIDs , sortedFilter , tf . isNegative ) ; err != nil {
2019-05-22 23:16:55 +02:00
return err
}
}
return nil
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) updateMetricIDsForOrSuffixNoFilter ( prefix [ ] byte , maxMetrics int , metricIDs * uint64set . Set ) error {
2019-05-22 23:16:55 +02:00
ts := & is . ts
2019-09-20 18:46:47 +02:00
mp := & is . mp
mp . Reset ( )
2019-05-22 23:16:55 +02:00
maxLoops := maxMetrics * maxIndexScanLoopsPerMetric
loops := 0
ts . Seek ( prefix )
2019-09-24 20:10:22 +02:00
for metricIDs . Len ( ) < maxMetrics && ts . NextItem ( ) {
2019-09-20 18:46:47 +02:00
item := ts . Item
if ! bytes . HasPrefix ( item , prefix ) {
return nil
}
if err := mp . InitOnlyTail ( item , item [ len ( prefix ) : ] ) ; err != nil {
return err
}
2019-09-23 19:40:38 +02:00
loops += mp . MetricIDsLen ( )
2019-05-22 23:16:55 +02:00
if loops > maxLoops {
return errFallbackToMetricNameMatch
}
2019-09-23 19:40:38 +02:00
mp . ParseMetricIDs ( )
2019-09-20 18:46:47 +02:00
for _ , metricID := range mp . MetricIDs {
2019-09-24 20:10:22 +02:00
metricIDs . Add ( metricID )
2019-05-22 23:16:55 +02:00
}
}
if err := ts . Error ( ) ; err != nil {
return fmt . Errorf ( "error when searching for tag filter prefix %q: %s" , prefix , err )
}
return nil
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) updateMetricIDsForOrSuffixWithFilter ( prefix [ ] byte , metricIDs * uint64set . Set , sortedFilter [ ] uint64 , isNegative bool ) error {
2019-09-20 18:46:47 +02:00
if len ( sortedFilter ) == 0 {
return nil
}
firstFilterMetricID := sortedFilter [ 0 ]
lastFilterMetricID := sortedFilter [ len ( sortedFilter ) - 1 ]
2019-05-22 23:16:55 +02:00
ts := & is . ts
2019-09-20 18:46:47 +02:00
mp := & is . mp
mp . Reset ( )
maxLoops := len ( sortedFilter ) * maxIndexScanLoopsPerMetric
loops := 0
ts . Seek ( prefix )
var sf [ ] uint64
var metricID uint64
for ts . NextItem ( ) {
item := ts . Item
if ! bytes . HasPrefix ( item , prefix ) {
return nil
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
if err := mp . InitOnlyTail ( item , item [ len ( prefix ) : ] ) ; err != nil {
return err
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
firstMetricID , lastMetricID := mp . FirstAndLastMetricIDs ( )
if lastMetricID < firstFilterMetricID {
// Skip the item, since it contains metricIDs lower
// than metricIDs in sortedFilter.
continue
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
if firstMetricID > lastFilterMetricID {
// Stop searching, since the current item and all the subsequent items
// contain metricIDs higher than metricIDs in sortedFilter.
return nil
2019-05-22 23:16:55 +02:00
}
2019-09-20 18:46:47 +02:00
sf = sortedFilter
2019-09-23 19:40:38 +02:00
loops += mp . MetricIDsLen ( )
2019-09-20 18:46:47 +02:00
if loops > maxLoops {
return errFallbackToMetricNameMatch
}
2019-09-23 19:40:38 +02:00
mp . ParseMetricIDs ( )
2019-09-20 18:46:47 +02:00
for _ , metricID = range mp . MetricIDs {
if len ( sf ) == 0 {
break
}
if metricID > sf [ 0 ] {
2019-11-25 13:01:36 +01:00
n := binarySearchUint64 ( sf , metricID )
2019-09-20 18:46:47 +02:00
sf = sf [ n : ]
if len ( sf ) == 0 {
break
}
2019-06-29 12:48:29 +02:00
}
2019-09-20 18:46:47 +02:00
if metricID < sf [ 0 ] {
continue
}
if isNegative {
2019-09-24 20:10:22 +02:00
metricIDs . Del ( metricID )
2019-09-20 18:46:47 +02:00
} else {
2019-09-24 20:10:22 +02:00
metricIDs . Add ( metricID )
2019-09-20 18:46:47 +02:00
}
sf = sf [ 1 : ]
2019-05-22 23:16:55 +02:00
}
}
if err := ts . Error ( ) ; err != nil {
return fmt . Errorf ( "error when searching for tag filter prefix %q: %s" , prefix , err )
}
return nil
}
2019-11-25 13:01:36 +01:00
func binarySearchUint64 ( a [ ] uint64 , v uint64 ) uint {
2019-11-21 17:31:44 +01:00
// Copy-pasted sort.Search from https://golang.org/src/sort/search.go?s=2246:2286#L49
i , j := uint ( 0 ) , uint ( len ( a ) )
for i < j {
h := ( i + j ) >> 1
if h < uint ( len ( a ) ) && a [ h ] < v {
i = h + 1
} else {
j = h
}
}
return i
}
2019-06-10 11:57:34 +02:00
var errFallbackToMetricNameMatch = errors . New ( "fall back to updateMetricIDsByMetricNameMatch because of too many index scan loops" )
2019-05-22 23:16:55 +02:00
var errMissingMetricIDsForDate = errors . New ( "missing metricIDs for date" )
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) getMetricIDsForTimeRange ( tr TimeRange , maxMetrics int ) ( * uint64set . Set , error ) {
2019-06-09 18:06:53 +02:00
atomic . AddUint64 ( & is . db . recentHourMetricIDsSearchCalls , 1 )
2019-11-09 22:17:42 +01:00
metricIDs , ok := is . getMetricIDsForRecentHours ( tr , maxMetrics )
if ok {
2019-06-09 18:06:53 +02:00
// Fast path: tr covers the current and / or the previous hour.
// Return the full list of metric ids for this time range.
atomic . AddUint64 ( & is . db . recentHourMetricIDsSearchHits , 1 )
return metricIDs , nil
}
// Slow path: collect the metric ids for all the days covering the given tr.
atomic . AddUint64 ( & is . db . dateMetricIDsSearchCalls , 1 )
2019-06-25 11:55:27 +02:00
minDate := uint64 ( tr . MinTimestamp ) / msecPerDay
maxDate := uint64 ( tr . MaxTimestamp ) / msecPerDay
2019-07-12 01:20:29 +02:00
if maxDate - minDate > maxDaysForDateMetricIDs {
2019-05-22 23:16:55 +02:00
// Too much dates must be covered. Give up.
return nil , errMissingMetricIDsForDate
}
2020-03-13 21:42:22 +01:00
if minDate == maxDate {
// Fast path - query on a single day.
metricIDs , err := is . getMetricIDsForDate ( minDate , maxMetrics )
if err != nil {
return nil , err
}
atomic . AddUint64 ( & is . db . dateMetricIDsSearchHits , 1 )
return metricIDs , nil
}
2019-11-09 22:17:42 +01:00
2020-03-13 21:42:22 +01:00
// Slower path - query over multiple days in parallel.
2019-11-09 22:17:42 +01:00
metricIDs = & uint64set . Set { }
var wg sync . WaitGroup
var errGlobal error
var mu sync . Mutex // protects metricIDs + errGlobal from concurrent access below.
2019-05-22 23:16:55 +02:00
for minDate <= maxDate {
2019-11-09 22:17:42 +01:00
wg . Add ( 1 )
2020-03-13 21:42:22 +01:00
go func ( date uint64 ) {
2019-11-09 22:17:42 +01:00
defer wg . Done ( )
2020-03-13 21:42:22 +01:00
isLocal := is . db . getIndexSearch ( )
2019-11-09 22:17:42 +01:00
defer is . db . putIndexSearch ( isLocal )
2020-03-13 21:42:22 +01:00
m , err := isLocal . getMetricIDsForDate ( date , maxMetrics )
2019-11-09 22:17:42 +01:00
mu . Lock ( )
2020-03-13 21:42:22 +01:00
defer mu . Unlock ( )
if errGlobal != nil {
return
2019-11-09 22:17:42 +01:00
}
if err != nil {
errGlobal = err
2020-03-13 21:42:22 +01:00
return
}
if metricIDs . Len ( ) < maxMetrics {
metricIDs . UnionMayOwn ( m )
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
} ( minDate )
2019-05-22 23:16:55 +02:00
minDate ++
}
2019-11-09 22:17:42 +01:00
wg . Wait ( )
if errGlobal != nil {
return nil , errGlobal
}
2019-06-09 18:06:53 +02:00
atomic . AddUint64 ( & is . db . dateMetricIDsSearchHits , 1 )
2019-05-22 23:16:55 +02:00
return metricIDs , nil
}
2020-03-13 21:42:22 +01:00
func ( is * indexSearch ) tryUpdatingMetricIDsForDateRange ( metricIDs * uint64set . Set , tfs * TagFilters , tr TimeRange , maxMetrics int ) error {
2019-11-09 22:17:42 +01:00
atomic . AddUint64 ( & is . db . dateRangeSearchCalls , 1 )
minDate := uint64 ( tr . MinTimestamp ) / msecPerDay
maxDate := uint64 ( tr . MaxTimestamp ) / msecPerDay
if minDate < is . db . startDateForPerDayInvertedIndex || maxDate < minDate {
// Per-day inverted index doesn't cover the selected date range.
2020-03-13 21:42:22 +01:00
return errFallbackToMetricNameMatch
2019-11-09 22:17:42 +01:00
}
if maxDate - minDate > maxDaysForDateMetricIDs {
// Too much dates must be covered. Give up, since it may be slow.
2020-03-13 21:42:22 +01:00
return errFallbackToMetricNameMatch
}
if minDate == maxDate {
// Fast path - query only a single date.
m , err := is . getMetricIDsForDateAndFilters ( minDate , tfs , maxMetrics )
if err != nil {
return err
}
metricIDs . UnionMayOwn ( m )
atomic . AddUint64 ( & is . db . dateRangeSearchHits , 1 )
return nil
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
// Slower path - search for metricIDs for each day in parallel.
2019-11-09 22:17:42 +01:00
var wg sync . WaitGroup
var errGlobal error
2020-03-13 21:42:22 +01:00
var mu sync . Mutex // protects metricIDs + errGlobal vars from concurrent access below
2019-11-09 22:17:42 +01:00
for minDate <= maxDate {
wg . Add ( 1 )
2020-03-13 21:42:22 +01:00
go func ( date uint64 ) {
2019-11-09 22:17:42 +01:00
defer wg . Done ( )
2020-03-13 21:42:22 +01:00
isLocal := is . db . getIndexSearch ( )
2019-11-09 22:17:42 +01:00
defer is . db . putIndexSearch ( isLocal )
2020-03-13 21:42:22 +01:00
m , err := isLocal . getMetricIDsForDateAndFilters ( date , tfs , maxMetrics )
2019-11-09 22:17:42 +01:00
mu . Lock ( )
2020-03-13 21:42:22 +01:00
defer mu . Unlock ( )
if errGlobal != nil {
return
2019-11-09 22:17:42 +01:00
}
if err != nil {
2020-03-13 21:42:22 +01:00
if err == errFallbackToMetricNameMatch {
// The per-date search is too expensive. Probably it is faster to perform global search
// using metric name match.
errGlobal = err
return
}
2019-12-03 13:46:39 +01:00
dateStr := time . Unix ( int64 ( date * 24 * 3600 ) , 0 )
errGlobal = fmt . Errorf ( "cannot search for metricIDs for %s: %s" , dateStr , err )
2020-03-13 21:42:22 +01:00
return
}
if metricIDs . Len ( ) < maxMetrics {
metricIDs . UnionMayOwn ( m )
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
} ( minDate )
2019-11-09 22:17:42 +01:00
minDate ++
}
wg . Wait ( )
if errGlobal != nil {
2020-03-13 21:42:22 +01:00
return errGlobal
2019-11-09 22:17:42 +01:00
}
atomic . AddUint64 ( & is . db . dateRangeSearchHits , 1 )
2020-03-13 21:42:22 +01:00
return nil
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
func ( is * indexSearch ) getMetricIDsForDateAndFilters ( date uint64 , tfs * TagFilters , maxMetrics int ) ( * uint64set . Set , error ) {
2020-03-30 23:44:41 +02:00
// Sort tfs by the number of matching filters from previous queries.
// This way we limit the amount of work below by applying more specific filters at first.
type tagFilterWithCount struct {
tf * tagFilter
count uint64
}
tfsWithCount := make ( [ ] tagFilterWithCount , len ( tfs . tfs ) )
kb := & is . kb
var buf [ ] byte
2019-11-09 22:17:42 +01:00
for i := range tfs . tfs {
tf := & tfs . tfs [ i ]
2020-03-30 23:44:41 +02:00
kb . B = appendDateTagFilterCacheKey ( kb . B [ : 0 ] , date , tf )
buf = is . db . metricIDsPerDateTagFilterCache . Get ( buf [ : 0 ] , kb . B )
count := uint64 ( 0 )
if len ( buf ) == 8 {
count = encoding . UnmarshalUint64 ( buf )
}
tfsWithCount [ i ] = tagFilterWithCount {
tf : tf ,
count : count ,
}
}
sort . Slice ( tfsWithCount , func ( i , j int ) bool {
a , b := & tfsWithCount [ i ] , & tfsWithCount [ j ]
if a . count != b . count {
return a . count < b . count
}
return a . tf . Less ( b . tf )
} )
// Populate metricIDs with the first non-negative filter.
var tfFirst * tagFilter
for i := range tfsWithCount {
tf := tfsWithCount [ i ] . tf
2019-11-09 22:17:42 +01:00
if tf . isNegative {
continue
}
tfFirst = tf
break
}
2020-03-13 21:42:22 +01:00
var metricIDs * uint64set . Set
2019-12-03 13:46:39 +01:00
maxDateMetrics := maxMetrics * 50
2019-11-09 22:17:42 +01:00
if tfFirst == nil {
2020-03-13 21:42:22 +01:00
// All the filters in tfs are negative. Populate all the metricIDs for the given (date),
// so later they can be filtered out with negative filters.
m , err := is . getMetricIDsForDate ( date , maxDateMetrics )
if err != nil {
2019-11-09 22:17:42 +01:00
if err == errMissingMetricIDsForDate {
2020-03-13 21:42:22 +01:00
// Zero time series were written on the given date.
2019-11-09 22:17:42 +01:00
// It is OK, since (date, metricID) entries must exist for the given date
// according to startDateForPerDayInvertedIndex.
2020-03-13 21:42:22 +01:00
return nil , nil
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
return nil , fmt . Errorf ( "cannot obtain all the metricIDs: %s" , err )
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
metricIDs = m
2019-11-09 22:17:42 +01:00
} else {
2020-03-13 21:42:22 +01:00
// Populate metricIDs for the given tfFirst on the given (date)
2019-11-09 22:17:42 +01:00
m , err := is . getMetricIDsForDateTagFilter ( tfFirst , date , tfs . commonPrefix , maxDateMetrics )
if err != nil {
2020-03-13 21:42:22 +01:00
return nil , err
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
metricIDs = m
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
if metricIDs . Len ( ) >= maxDateMetrics {
2019-12-03 13:46:39 +01:00
// Too many time series found by a single tag filter. Fall back to global search.
2020-03-13 21:42:22 +01:00
return nil , errFallbackToMetricNameMatch
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
// Intersect metricIDs with the rest of filters.
2020-03-30 23:44:41 +02:00
for i := range tfsWithCount {
tfWithCount := & tfsWithCount [ i ]
tf := tfWithCount . tf
2019-11-09 22:17:42 +01:00
if tf == tfFirst {
continue
}
2020-03-30 23:44:41 +02:00
if n := uint64 ( metricIDs . Len ( ) ) ; n < 1000 || n < tfWithCount . count / maxIndexScanLoopsPerMetric {
// It should be faster performing metricName match on the remaining filters
// instead of scanning big number of entries in the inverted index for these filters.
tfsRemaining := tfsWithCount [ i : ]
tfsPostponed := make ( [ ] * tagFilter , 0 , len ( tfsRemaining ) )
for j := range tfsRemaining {
tf := tfsRemaining [ j ] . tf
if tf == tfFirst {
continue
}
tfsPostponed = append ( tfsPostponed , tf )
}
var m uint64set . Set
if err := is . updateMetricIDsByMetricNameMatch ( & m , metricIDs , tfsPostponed ) ; err != nil {
return nil , err
}
return & m , nil
}
2019-11-09 22:17:42 +01:00
m , err := is . getMetricIDsForDateTagFilter ( tf , date , tfs . commonPrefix , maxDateMetrics )
if err != nil {
2020-03-13 21:42:22 +01:00
return nil , err
2019-11-09 22:17:42 +01:00
}
if m . Len ( ) >= maxDateMetrics {
2019-12-03 13:46:39 +01:00
// Too many time series found by a single tag filter. Fall back to global search.
2020-03-13 21:42:22 +01:00
return nil , errFallbackToMetricNameMatch
2019-11-09 22:17:42 +01:00
}
if tf . isNegative {
2020-03-13 21:42:22 +01:00
metricIDs . Subtract ( m )
2019-11-09 22:17:42 +01:00
} else {
2020-03-13 21:42:22 +01:00
metricIDs . Intersect ( m )
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
if metricIDs . Len ( ) == 0 {
// Short circuit - there is no need in applying the remaining filters to empty set.
return nil , nil
2019-11-09 22:17:42 +01:00
}
}
2020-03-13 21:42:22 +01:00
return metricIDs , nil
2019-11-09 22:17:42 +01:00
}
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) getMetricIDsForRecentHours ( tr TimeRange , maxMetrics int ) ( * uint64set . Set , bool ) {
2019-06-09 18:06:53 +02:00
minHour := uint64 ( tr . MinTimestamp ) / msecPerHour
maxHour := uint64 ( tr . MaxTimestamp ) / msecPerHour
hmCurr := is . db . currHourMetricIDs . Load ( ) . ( * hourMetricIDs )
if maxHour == hmCurr . hour && minHour == maxHour && hmCurr . isFull {
// The tr fits the current hour.
// Return a copy of hmCurr.m, because the caller may modify
// the returned map.
2019-09-24 20:10:22 +02:00
if hmCurr . m . Len ( ) > maxMetrics {
2019-06-09 18:06:53 +02:00
return nil , false
}
2019-09-24 20:10:22 +02:00
return hmCurr . m . Clone ( ) , true
2019-06-09 18:06:53 +02:00
}
hmPrev := is . db . prevHourMetricIDs . Load ( ) . ( * hourMetricIDs )
if maxHour == hmPrev . hour && minHour == maxHour && hmPrev . isFull {
// The tr fits the previous hour.
// Return a copy of hmPrev.m, because the caller may modify
// the returned map.
2019-09-24 20:10:22 +02:00
if hmPrev . m . Len ( ) > maxMetrics {
2019-06-09 18:06:53 +02:00
return nil , false
}
2019-09-24 20:10:22 +02:00
return hmPrev . m . Clone ( ) , true
2019-06-09 18:06:53 +02:00
}
if maxHour == hmCurr . hour && minHour == hmPrev . hour && hmCurr . isFull && hmPrev . isFull {
// The tr spans the previous and the current hours.
2019-09-24 20:10:22 +02:00
if hmCurr . m . Len ( ) + hmPrev . m . Len ( ) > maxMetrics {
2019-06-09 18:06:53 +02:00
return nil , false
}
2019-09-24 20:10:22 +02:00
metricIDs := hmCurr . m . Clone ( )
2019-11-03 23:34:24 +01:00
metricIDs . Union ( hmPrev . m )
2019-06-09 18:06:53 +02:00
return metricIDs , true
}
return nil , false
}
2019-05-22 23:16:55 +02:00
func ( db * indexDB ) storeDateMetricID ( date , metricID uint64 ) error {
is := db . getIndexSearch ( )
ok , err := is . hasDateMetricID ( date , metricID )
db . putIndexSearch ( is )
if err != nil {
return err
}
if ok {
// Fast path: the (date, metricID) entry already exists in the db.
return nil
}
2019-11-09 22:17:42 +01:00
// Slow path: create (date, metricID) entries.
2019-05-22 23:16:55 +02:00
items := getIndexItems ( )
2019-11-09 22:17:42 +01:00
defer putIndexItems ( items )
items . B = marshalCommonPrefix ( items . B , nsPrefixDateToMetricID )
2019-05-22 23:16:55 +02:00
items . B = encoding . MarshalUint64 ( items . B , date )
items . B = encoding . MarshalUint64 ( items . B , metricID )
items . Next ( )
2019-11-09 22:17:42 +01:00
// Create per-day inverted index entries for metricID.
kb := kbPool . Get ( )
defer kbPool . Put ( kb )
mn := GetMetricName ( )
defer PutMetricName ( mn )
kb . B , err = db . searchMetricName ( kb . B [ : 0 ] , metricID )
if err != nil {
return fmt . Errorf ( "cannot find metricName by metricID %d: %s" , metricID , err )
}
if err = mn . Unmarshal ( kb . B ) ; err != nil {
return fmt . Errorf ( "cannot unmarshal metricName %q obtained by metricID %d: %s" , metricID , kb . B , err )
}
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixDateTagToMetricIDs )
kb . B = encoding . MarshalUint64 ( kb . B , date )
items . B = append ( items . B , kb . B ... )
items . B = marshalTagValue ( items . B , nil )
items . B = marshalTagValue ( items . B , mn . MetricGroup )
items . B = encoding . MarshalUint64 ( items . B , metricID )
items . Next ( )
for i := range mn . Tags {
tag := & mn . Tags [ i ]
items . B = append ( items . B , kb . B ... )
items . B = tag . Marshal ( items . B )
items . B = encoding . MarshalUint64 ( items . B , metricID )
items . Next ( )
}
if err = db . tb . AddItems ( items . Items ) ; err != nil {
return fmt . Errorf ( "cannot add per-day entires for metricID %d: %s" , metricID , err )
}
return nil
2019-05-22 23:16:55 +02:00
}
func ( is * indexSearch ) hasDateMetricID ( date , metricID uint64 ) ( bool , error ) {
ts := & is . ts
kb := & is . kb
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixDateToMetricID )
kb . B = encoding . MarshalUint64 ( kb . B , date )
kb . B = encoding . MarshalUint64 ( kb . B , metricID )
if err := ts . FirstItemWithPrefix ( kb . B ) ; err != nil {
if err == io . EOF {
return false , nil
}
return false , fmt . Errorf ( "error when searching for (date=%d, metricID=%d) entry: %s" , date , metricID , err )
}
if string ( ts . Item ) != string ( kb . B ) {
return false , fmt . Errorf ( "unexpected entry for (date=%d, metricID=%d); got %q; want %q" , date , metricID , ts . Item , kb . B )
}
return true , nil
}
2019-11-09 22:17:42 +01:00
func ( is * indexSearch ) getMetricIDsForDateTagFilter ( tf * tagFilter , date uint64 , commonPrefix [ ] byte , maxMetrics int ) ( * uint64set . Set , error ) {
// Augument tag filter prefix for per-date search instead of global search.
if ! bytes . HasPrefix ( tf . prefix , commonPrefix ) {
logger . Panicf ( "BUG: unexpected tf.prefix %q; must start with commonPrefix %q" , tf . prefix , commonPrefix )
}
kb := kbPool . Get ( )
defer kbPool . Put ( kb )
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixDateTagToMetricIDs )
kb . B = encoding . MarshalUint64 ( kb . B , date )
kb . B = append ( kb . B , tf . prefix [ len ( commonPrefix ) : ] ... )
tfNew := * tf
tfNew . isNegative = false // isNegative for the original tf is handled by the caller.
tfNew . prefix = kb . B
2020-03-30 23:44:41 +02:00
metricIDs , err := is . getMetricIDsForTagFilter ( & tfNew , maxMetrics )
// Store the number of matching metricIDs in the cache in order to sort tag filters
// in ascending number of matching metricIDs on the next search.
is . kb . B = appendDateTagFilterCacheKey ( is . kb . B [ : 0 ] , date , tf )
metricIDsLen := uint64 ( metricIDs . Len ( ) )
if err != nil {
// Set metricIDsLen to maxMetrics, so the given entry will be moved to the end
// of tag filters on the next search.
metricIDsLen = uint64 ( maxMetrics )
}
kb . B = encoding . MarshalUint64 ( kb . B [ : 0 ] , metricIDsLen )
is . db . metricIDsPerDateTagFilterCache . Set ( is . kb . B , kb . B )
return metricIDs , err
}
func appendDateTagFilterCacheKey ( dst [ ] byte , date uint64 , tf * tagFilter ) [ ] byte {
dst = encoding . MarshalUint64 ( dst , date )
dst = tf . Marshal ( dst )
return dst
2019-11-09 22:17:42 +01:00
}
2020-03-13 21:42:22 +01:00
func ( is * indexSearch ) getMetricIDsForDate ( date uint64 , maxMetrics int ) ( * uint64set . Set , error ) {
2019-11-09 22:17:42 +01:00
// Extract all the metricIDs from (date, __name__=value)->metricIDs entries.
kb := kbPool . Get ( )
defer kbPool . Put ( kb )
2020-03-13 21:42:22 +01:00
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixDateTagToMetricIDs )
2019-11-09 22:17:42 +01:00
kb . B = encoding . MarshalUint64 ( kb . B , date )
kb . B = marshalTagValue ( kb . B , nil )
2020-03-13 21:42:22 +01:00
var metricIDs uint64set . Set
if err := is . updateMetricIDsForPrefix ( kb . B , & metricIDs , maxMetrics ) ; err != nil {
return nil , err
}
if metricIDs . Len ( ) == 0 {
// There are no metricIDs for the given date.
// This may be the case for old data where (data, __name__=value)->metricIDs entries weren't available.
return nil , errMissingMetricIDsForDate
}
return & metricIDs , nil
2019-11-09 22:17:42 +01:00
}
2019-11-09 17:00:58 +01:00
2019-11-09 22:17:42 +01:00
func ( is * indexSearch ) updateMetricIDsAll ( metricIDs * uint64set . Set , maxMetrics int ) error {
kb := kbPool . Get ( )
defer kbPool . Put ( kb )
// Extract all the metricIDs from (__name__=value)->metricIDs entries.
2019-11-09 17:00:58 +01:00
kb . B = marshalCommonPrefix ( kb . B [ : 0 ] , nsPrefixTagToMetricIDs )
kb . B = marshalTagValue ( kb . B , nil )
2019-11-09 22:17:42 +01:00
return is . updateMetricIDsForPrefix ( kb . B , metricIDs , maxMetrics )
}
func ( is * indexSearch ) updateMetricIDsForPrefix ( prefix [ ] byte , metricIDs * uint64set . Set , maxMetrics int ) error {
ts := & is . ts
mp := & is . mp
2019-09-20 10:53:42 +02:00
ts . Seek ( prefix )
for ts . NextItem ( ) {
item := ts . Item
if ! bytes . HasPrefix ( item , prefix ) {
return nil
2019-05-22 23:16:55 +02:00
}
2019-09-20 10:53:42 +02:00
tail := item [ len ( prefix ) : ]
2019-11-09 17:00:58 +01:00
n := bytes . IndexByte ( tail , tagSeparatorChar )
if n < 0 {
return fmt . Errorf ( "invalid tag->metricIDs line %q: cannot find tagSeparatorChar %d" , item , tagSeparatorChar )
}
tail = tail [ n + 1 : ]
if err := mp . InitOnlyTail ( item , tail ) ; err != nil {
return err
}
mp . ParseMetricIDs ( )
for _ , metricID := range mp . MetricIDs {
metricIDs . Add ( metricID )
2019-05-22 23:16:55 +02:00
}
2019-09-24 20:10:22 +02:00
if metricIDs . Len ( ) >= maxMetrics {
2019-09-20 10:53:42 +02:00
return nil
}
2019-05-22 23:16:55 +02:00
}
if err := ts . Error ( ) ; err != nil {
2019-09-20 10:53:42 +02:00
return fmt . Errorf ( "error when searching for all metricIDs by prefix %q: %s" , prefix , err )
2019-05-22 23:16:55 +02:00
}
return nil
}
2019-11-05 15:59:31 +01:00
// The maximum number of index scan loops.
2019-06-10 11:57:34 +02:00
// Bigger number of loops is slower than updateMetricIDsByMetricNameMatch
2019-05-22 23:16:55 +02:00
// over the found metrics.
2019-11-05 12:24:22 +01:00
const maxIndexScanLoopsPerMetric = 100
2019-05-22 23:16:55 +02:00
2020-03-30 23:44:41 +02:00
// The maximum number of slow index scan loops.
2019-11-05 15:59:31 +01:00
// Bigger number of loops is slower than updateMetricIDsByMetricNameMatch
// over the found metrics.
const maxIndexScanSlowLoopsPerMetric = 20
2019-09-24 20:10:22 +02:00
func ( is * indexSearch ) intersectMetricIDsWithTagFilter ( tf * tagFilter , filter * uint64set . Set ) ( * uint64set . Set , error ) {
if filter . Len ( ) == 0 {
2019-05-22 23:16:55 +02:00
return nil , nil
}
2019-11-05 13:17:59 +01:00
kb := & is . kb
filterLenRounded := ( uint64 ( filter . Len ( ) ) / 1024 ) * 1024
kb . B = append ( kb . B [ : 0 ] , uselessTagIntersectKeyPrefix )
kb . B = encoding . MarshalUint64 ( kb . B , filterLenRounded )
kb . B = tf . Marshal ( kb . B )
if len ( is . db . uselessTagFiltersCache . Get ( nil , kb . B ) ) > 0 {
// Skip useless work, since the intersection will return
// errFallbackToMetricNameMatc for the given filter.
return nil , errFallbackToMetricNameMatch
}
metricIDs , err := is . intersectMetricIDsWithTagFilterNocache ( tf , filter )
if err == nil {
return metricIDs , err
}
if err != errFallbackToMetricNameMatch {
return nil , err
}
kb . B = append ( kb . B [ : 0 ] , uselessTagIntersectKeyPrefix )
kb . B = encoding . MarshalUint64 ( kb . B , filterLenRounded )
kb . B = tf . Marshal ( kb . B )
is . db . uselessTagFiltersCache . Set ( kb . B , uselessTagFilterCacheValue )
return nil , errFallbackToMetricNameMatch
}
func ( is * indexSearch ) intersectMetricIDsWithTagFilterNocache ( tf * tagFilter , filter * uint64set . Set ) ( * uint64set . Set , error ) {
2019-05-22 23:16:55 +02:00
metricIDs := filter
if ! tf . isNegative {
2019-09-24 20:10:22 +02:00
metricIDs = & uint64set . Set { }
2019-05-22 23:16:55 +02:00
}
if len ( tf . orSuffixes ) > 0 {
// Fast path for orSuffixes - seek for rows for each value from orSuffixes.
2019-06-10 11:57:34 +02:00
if err := is . updateMetricIDsForOrSuffixesWithFilter ( tf , metricIDs , filter ) ; err != nil {
2019-09-20 18:46:47 +02:00
if err == errFallbackToMetricNameMatch {
return nil , err
}
return nil , fmt . Errorf ( "error when intersecting metricIDs for tagFilter in fast path: %s; tagFilter=%s" , err , tf )
2019-05-22 23:16:55 +02:00
}
return metricIDs , nil
}
// Slow path - scan for all the rows with the given prefix.
2019-11-05 15:59:31 +01:00
maxLoops := filter . Len ( ) * maxIndexScanSlowLoopsPerMetric
2019-06-27 15:15:25 +02:00
err := is . getMetricIDsForTagFilterSlow ( tf , maxLoops , func ( metricID uint64 ) bool {
2019-05-22 23:16:55 +02:00
if tf . isNegative {
// filter must be equal to metricIDs
2019-09-24 20:10:22 +02:00
metricIDs . Del ( metricID )
2019-06-27 15:15:25 +02:00
return true
2019-05-22 23:16:55 +02:00
}
2019-09-24 20:10:22 +02:00
if filter . Has ( metricID ) {
metricIDs . Add ( metricID )
2019-05-22 23:16:55 +02:00
}
2019-06-27 15:15:25 +02:00
return true
} )
if err != nil {
2019-09-20 18:46:47 +02:00
if err == errFallbackToMetricNameMatch {
return nil , err
}
return nil , fmt . Errorf ( "error when intersecting metricIDs for tagFilter in slow path: %s; tagFilter=%s" , err , tf )
2019-05-22 23:16:55 +02:00
}
return metricIDs , nil
}
var kbPool bytesutil . ByteBufferPool
// Returns local unique MetricID.
func getUniqueUint64 ( ) uint64 {
return atomic . AddUint64 ( & uniqueUint64 , 1 )
}
// This number mustn't go backwards on restarts, otherwise metricID
// collisions are possible. So don't change time on the server
// between VictoriaMetrics restarts.
var uniqueUint64 = uint64 ( time . Now ( ) . UnixNano ( ) )
func marshalCommonPrefix ( dst [ ] byte , nsPrefix byte ) [ ] byte {
dst = append ( dst , nsPrefix )
return dst
}
2019-09-20 18:46:47 +02:00
func unmarshalCommonPrefix ( src [ ] byte ) ( [ ] byte , byte , error ) {
if len ( src ) < commonPrefixLen {
return nil , 0 , fmt . Errorf ( "cannot unmarshal common prefix from %d bytes; need at least %d bytes; data=%X" , len ( src ) , commonPrefixLen , src )
}
prefix := src [ 0 ]
return src [ commonPrefixLen : ] , prefix , nil
}
// 1 byte for prefix
const commonPrefixLen = 1
type tagToMetricIDsRowParser struct {
2019-11-09 22:17:42 +01:00
// NSPrefix contains the first byte parsed from the row after Init call.
// This is either nsPrefixTagToMetricIDs or nsPrefixDateTagToMetricIDs.
NSPrefix byte
// Date contains parsed date for nsPrefixDateTagToMetricIDs rows after Init call
Date uint64
2019-09-20 18:46:47 +02:00
// MetricIDs contains parsed MetricIDs after ParseMetricIDs call
MetricIDs [ ] uint64
// Tag contains parsed tag after Init call
Tag Tag
// tail contains the remaining unparsed metricIDs
tail [ ] byte
}
func ( mp * tagToMetricIDsRowParser ) Reset ( ) {
2019-11-09 22:17:42 +01:00
mp . NSPrefix = 0
mp . Date = 0
2019-09-20 18:46:47 +02:00
mp . MetricIDs = mp . MetricIDs [ : 0 ]
mp . Tag . Reset ( )
mp . tail = nil
}
// Init initializes mp from b, which should contain encoded tag->metricIDs row.
//
// b cannot be re-used until Reset call.
2019-11-09 22:17:42 +01:00
func ( mp * tagToMetricIDsRowParser ) Init ( b [ ] byte , nsPrefixExpected byte ) error {
tail , nsPrefix , err := unmarshalCommonPrefix ( b )
2019-09-20 18:46:47 +02:00
if err != nil {
return fmt . Errorf ( "invalid tag->metricIDs row %q: %s" , b , err )
}
2019-11-09 22:17:42 +01:00
if nsPrefix != nsPrefixExpected {
return fmt . Errorf ( "invalid prefix for tag->metricIDs row %q; got %d; want %d" , b , nsPrefix , nsPrefixExpected )
}
if nsPrefix == nsPrefixDateTagToMetricIDs {
// unmarshal date.
if len ( tail ) < 8 {
return fmt . Errorf ( "cannot unmarshal date from (date, tag)->metricIDs row %q from %d bytes; want at least 8 bytes" , b , len ( tail ) )
}
mp . Date = encoding . UnmarshalUint64 ( tail )
tail = tail [ 8 : ]
2019-09-20 18:46:47 +02:00
}
2019-11-09 22:17:42 +01:00
mp . NSPrefix = nsPrefix
2019-09-20 18:46:47 +02:00
tail , err = mp . Tag . Unmarshal ( tail )
if err != nil {
return fmt . Errorf ( "cannot unmarshal tag from tag->metricIDs row %q: %s" , b , err )
}
return mp . InitOnlyTail ( b , tail )
}
2019-11-09 22:17:42 +01:00
// MarshalPrefix marshals row prefix without tail to dst.
func ( mp * tagToMetricIDsRowParser ) MarshalPrefix ( dst [ ] byte ) [ ] byte {
dst = marshalCommonPrefix ( dst , mp . NSPrefix )
if mp . NSPrefix == nsPrefixDateTagToMetricIDs {
dst = encoding . MarshalUint64 ( dst , mp . Date )
}
dst = mp . Tag . Marshal ( dst )
return dst
}
2019-09-20 18:46:47 +02:00
// InitOnlyTail initializes mp.tail from tail.
//
// b must contain tag->metricIDs row.
// b cannot be re-used until Reset call.
func ( mp * tagToMetricIDsRowParser ) InitOnlyTail ( b , tail [ ] byte ) error {
if len ( tail ) == 0 {
return fmt . Errorf ( "missing metricID in the tag->metricIDs row %q" , b )
}
if len ( tail ) % 8 != 0 {
return fmt . Errorf ( "invalid tail length in the tag->metricIDs row; got %d bytes; must be multiple of 8 bytes" , len ( tail ) )
}
mp . tail = tail
return nil
}
// EqualPrefix returns true if prefixes for mp and x are equal.
//
// Prefix contains (tag)
func ( mp * tagToMetricIDsRowParser ) EqualPrefix ( x * tagToMetricIDsRowParser ) bool {
2019-11-09 22:17:42 +01:00
if ! mp . Tag . Equal ( & x . Tag ) {
return false
}
return mp . Date == x . Date && mp . NSPrefix == x . NSPrefix
2019-09-20 18:46:47 +02:00
}
// FirstAndLastMetricIDs returns the first and the last metricIDs in the mp.tail.
func ( mp * tagToMetricIDsRowParser ) FirstAndLastMetricIDs ( ) ( uint64 , uint64 ) {
tail := mp . tail
if len ( tail ) < 8 {
logger . Panicf ( "BUG: cannot unmarshal metricID from %d bytes; need 8 bytes" , len ( tail ) )
return 0 , 0
}
firstMetricID := encoding . UnmarshalUint64 ( tail )
lastMetricID := firstMetricID
if len ( tail ) > 8 {
lastMetricID = encoding . UnmarshalUint64 ( tail [ len ( tail ) - 8 : ] )
}
return firstMetricID , lastMetricID
}
2019-09-23 19:40:38 +02:00
// MetricIDsLen returns the number of MetricIDs in the mp.tail
func ( mp * tagToMetricIDsRowParser ) MetricIDsLen ( ) int {
return len ( mp . tail ) / 8
}
2019-09-20 18:46:47 +02:00
// ParseMetricIDs parses MetricIDs from mp.tail into mp.MetricIDs.
func ( mp * tagToMetricIDsRowParser ) ParseMetricIDs ( ) {
tail := mp . tail
mp . MetricIDs = mp . MetricIDs [ : 0 ]
n := len ( tail ) / 8
if n <= cap ( mp . MetricIDs ) {
mp . MetricIDs = mp . MetricIDs [ : n ]
} else {
mp . MetricIDs = append ( mp . MetricIDs [ : cap ( mp . MetricIDs ) ] , make ( [ ] uint64 , n - cap ( mp . MetricIDs ) ) ... )
}
metricIDs := mp . MetricIDs
_ = metricIDs [ n - 1 ]
for i := 0 ; i < n ; i ++ {
if len ( tail ) < 8 {
logger . Panicf ( "BUG: tail cannot be smaller than 8 bytes; got %d bytes; tail=%X" , len ( tail ) , tail )
return
}
metricID := encoding . UnmarshalUint64 ( tail )
metricIDs [ i ] = metricID
tail = tail [ 8 : ]
}
}
// IsDeletedTag verifies whether the tag from mp is deleted according to dmis.
//
// dmis must contain deleted MetricIDs.
2019-09-24 20:10:22 +02:00
func ( mp * tagToMetricIDsRowParser ) IsDeletedTag ( dmis * uint64set . Set ) bool {
if dmis . Len ( ) == 0 {
2019-09-20 18:46:47 +02:00
return false
}
mp . ParseMetricIDs ( )
for _ , metricID := range mp . MetricIDs {
2019-09-24 20:10:22 +02:00
if ! dmis . Has ( metricID ) {
2019-09-20 18:46:47 +02:00
return false
}
}
return true
}
func mergeTagToMetricIDsRows ( data [ ] byte , items [ ] [ ] byte ) ( [ ] byte , [ ] [ ] byte ) {
2019-11-09 22:17:42 +01:00
data , items = mergeTagToMetricIDsRowsInternal ( data , items , nsPrefixTagToMetricIDs )
data , items = mergeTagToMetricIDsRowsInternal ( data , items , nsPrefixDateTagToMetricIDs )
return data , items
}
func mergeTagToMetricIDsRowsInternal ( data [ ] byte , items [ ] [ ] byte , nsPrefix byte ) ( [ ] byte , [ ] [ ] byte ) {
// Perform quick checks whether items contain rows starting from nsPrefix
2019-09-20 18:46:47 +02:00
// based on the fact that items are sorted.
2019-10-08 15:25:24 +02:00
if len ( items ) <= 2 {
// The first and the last row must remain unchanged.
2019-09-20 18:46:47 +02:00
return data , items
}
firstItem := items [ 0 ]
2019-11-09 22:17:42 +01:00
if len ( firstItem ) > 0 && firstItem [ 0 ] > nsPrefix {
2019-09-20 18:46:47 +02:00
return data , items
}
lastItem := items [ len ( items ) - 1 ]
2019-11-09 22:17:42 +01:00
if len ( lastItem ) > 0 && lastItem [ 0 ] < nsPrefix {
2019-09-20 18:46:47 +02:00
return data , items
}
2019-11-09 22:17:42 +01:00
// items contain at least one row starting from nsPrefix. Merge rows with common tag.
2019-09-20 18:46:47 +02:00
tmm := getTagToMetricIDsRowsMerger ( )
2019-10-08 15:25:24 +02:00
tmm . dataCopy = append ( tmm . dataCopy [ : 0 ] , data ... )
tmm . itemsCopy = append ( tmm . itemsCopy [ : 0 ] , items ... )
2019-09-20 18:46:47 +02:00
mp := & tmm . mp
mpPrev := & tmm . mpPrev
2019-10-08 15:25:24 +02:00
dstData := data [ : 0 ]
dstItems := items [ : 0 ]
2019-09-23 19:40:38 +02:00
for i , item := range items {
2019-11-09 22:17:42 +01:00
if len ( item ) == 0 || item [ 0 ] != nsPrefix || i == 0 || i == len ( items ) - 1 {
// Write rows not starting with nsPrefix as-is.
2019-09-23 19:40:38 +02:00
// Additionally write the first and the last row as-is in order to preserve
// sort order for adjancent blocks.
2019-11-09 22:17:42 +01:00
dstData , dstItems = tmm . flushPendingMetricIDs ( dstData , dstItems , mpPrev )
2019-09-20 18:46:47 +02:00
dstData = append ( dstData , item ... )
dstItems = append ( dstItems , dstData [ len ( dstData ) - len ( item ) : ] )
continue
}
2019-11-09 22:17:42 +01:00
if err := mp . Init ( item , nsPrefix ) ; err != nil {
logger . Panicf ( "FATAL: cannot parse row starting with nsPrefix %d during merge: %s" , nsPrefix , err )
2019-09-20 18:46:47 +02:00
}
2019-09-23 23:49:21 +02:00
if mp . MetricIDsLen ( ) >= maxMetricIDsPerRow {
2019-11-09 22:17:42 +01:00
dstData , dstItems = tmm . flushPendingMetricIDs ( dstData , dstItems , mpPrev )
2019-09-23 23:49:21 +02:00
dstData = append ( dstData , item ... )
dstItems = append ( dstItems , dstData [ len ( dstData ) - len ( item ) : ] )
continue
}
2019-11-09 22:17:42 +01:00
if ! mp . EqualPrefix ( mpPrev ) {
2019-09-20 18:46:47 +02:00
dstData , dstItems = tmm . flushPendingMetricIDs ( dstData , dstItems , mpPrev )
}
mp . ParseMetricIDs ( )
tmm . pendingMetricIDs = append ( tmm . pendingMetricIDs , mp . MetricIDs ... )
mpPrev , mp = mp , mpPrev
2019-09-23 23:49:21 +02:00
if len ( tmm . pendingMetricIDs ) >= maxMetricIDsPerRow {
dstData , dstItems = tmm . flushPendingMetricIDs ( dstData , dstItems , mpPrev )
}
2019-09-20 18:46:47 +02:00
}
if len ( tmm . pendingMetricIDs ) > 0 {
2019-10-08 15:25:24 +02:00
logger . Panicf ( "BUG: tmm.pendingMetricIDs must be empty at this point; got %d items: %d" , len ( tmm . pendingMetricIDs ) , tmm . pendingMetricIDs )
}
2019-11-06 13:24:48 +01:00
if ! checkItemsSorted ( dstItems ) {
// Items could become unsorted if initial items contain duplicate metricIDs:
//
// item1: 1, 1, 5
// item2: 1, 4
//
// Items could become the following after the merge:
//
// item1: 1, 5
// item2: 1, 4
//
// i.e. item1 > item2
//
// Leave the original items unmerged, so they can be merged next time.
// This case should be quite rare - if multiple data points are simultaneously inserted
// into the same new time series from multiple concurrent goroutines.
atomic . AddUint64 ( & indexBlocksWithMetricIDsIncorrectOrder , 1 )
2019-10-08 15:25:24 +02:00
dstData = append ( dstData [ : 0 ] , tmm . dataCopy ... )
2019-10-09 11:13:17 +02:00
dstItems = dstItems [ : 0 ]
// tmm.itemsCopy can point to overwritten data, so it must be updated
// to point to real data from tmm.dataCopy.
buf := dstData
for _ , item := range tmm . itemsCopy {
dstItems = append ( dstItems , buf [ : len ( item ) ] )
buf = buf [ len ( item ) : ]
}
2019-11-06 13:24:48 +01:00
if ! checkItemsSorted ( dstItems ) {
logger . Panicf ( "BUG: the original items weren't sorted; items=%q" , dstItems )
2019-10-09 11:13:17 +02:00
}
2019-09-20 18:46:47 +02:00
}
2019-09-23 19:40:38 +02:00
putTagToMetricIDsRowsMerger ( tmm )
2019-11-06 13:24:48 +01:00
atomic . AddUint64 ( & indexBlocksWithMetricIDsProcessed , 1 )
2019-09-24 18:32:06 +02:00
return dstData , dstItems
2019-09-20 18:46:47 +02:00
}
2019-11-06 13:24:48 +01:00
var indexBlocksWithMetricIDsIncorrectOrder uint64
var indexBlocksWithMetricIDsProcessed uint64
func checkItemsSorted ( items [ ] [ ] byte ) bool {
2019-09-26 12:12:24 +02:00
if len ( items ) == 0 {
2019-11-06 13:24:48 +01:00
return true
2019-09-26 12:12:24 +02:00
}
prevItem := items [ 0 ]
for _ , currItem := range items [ 1 : ] {
if string ( prevItem ) > string ( currItem ) {
2019-11-06 13:24:48 +01:00
return false
2019-09-26 12:12:24 +02:00
}
prevItem = currItem
}
2019-11-06 13:24:48 +01:00
return true
2019-09-26 12:12:24 +02:00
}
2019-09-23 23:49:21 +02:00
// maxMetricIDsPerRow limits the number of metricIDs in tag->metricIDs row.
//
// This reduces overhead on index and metaindex in lib/mergeset.
const maxMetricIDsPerRow = 64
2019-09-19 19:00:33 +02:00
type uint64Sorter [ ] uint64
func ( s uint64Sorter ) Len ( ) int { return len ( s ) }
func ( s uint64Sorter ) Less ( i , j int ) bool {
return s [ i ] < s [ j ]
}
func ( s uint64Sorter ) Swap ( i , j int ) {
s [ i ] , s [ j ] = s [ j ] , s [ i ]
}
2019-09-20 18:46:47 +02:00
type tagToMetricIDsRowsMerger struct {
pendingMetricIDs uint64Sorter
mp tagToMetricIDsRowParser
mpPrev tagToMetricIDsRowParser
2019-10-08 15:25:24 +02:00
itemsCopy [ ] [ ] byte
dataCopy [ ] byte
2019-09-20 18:46:47 +02:00
}
2019-09-23 19:40:38 +02:00
func ( tmm * tagToMetricIDsRowsMerger ) Reset ( ) {
tmm . pendingMetricIDs = tmm . pendingMetricIDs [ : 0 ]
tmm . mp . Reset ( )
tmm . mpPrev . Reset ( )
2019-10-08 15:25:24 +02:00
tmm . itemsCopy = tmm . itemsCopy [ : 0 ]
tmm . dataCopy = tmm . dataCopy [ : 0 ]
2019-09-23 19:40:38 +02:00
}
2019-09-20 18:46:47 +02:00
func ( tmm * tagToMetricIDsRowsMerger ) flushPendingMetricIDs ( dstData [ ] byte , dstItems [ ] [ ] byte , mp * tagToMetricIDsRowParser ) ( [ ] byte , [ ] [ ] byte ) {
if len ( tmm . pendingMetricIDs ) == 0 {
2019-11-09 22:17:42 +01:00
// Nothing to flush
return dstData , dstItems
2019-09-20 18:46:47 +02:00
}
2019-09-23 19:40:38 +02:00
// Use sort.Sort instead of sort.Slice in order to reduce memory allocations.
sort . Sort ( & tmm . pendingMetricIDs )
2019-09-25 16:55:13 +02:00
tmm . pendingMetricIDs = removeDuplicateMetricIDs ( tmm . pendingMetricIDs )
2019-09-23 19:40:38 +02:00
2019-09-25 16:55:13 +02:00
// Marshal pendingMetricIDs
2019-09-20 18:46:47 +02:00
dstDataLen := len ( dstData )
2019-11-09 22:17:42 +01:00
dstData = mp . MarshalPrefix ( dstData )
2019-09-23 19:40:38 +02:00
for _ , metricID := range tmm . pendingMetricIDs {
2019-09-20 18:46:47 +02:00
dstData = encoding . MarshalUint64 ( dstData , metricID )
}
dstItems = append ( dstItems , dstData [ dstDataLen : ] )
2019-09-23 19:40:38 +02:00
tmm . pendingMetricIDs = tmm . pendingMetricIDs [ : 0 ]
2019-09-20 18:46:47 +02:00
return dstData , dstItems
}
2019-09-25 16:55:13 +02:00
func removeDuplicateMetricIDs ( sortedMetricIDs [ ] uint64 ) [ ] uint64 {
if len ( sortedMetricIDs ) < 2 {
return sortedMetricIDs
}
prevMetricID := sortedMetricIDs [ 0 ]
hasDuplicates := false
for _ , metricID := range sortedMetricIDs [ 1 : ] {
if prevMetricID == metricID {
hasDuplicates = true
2019-09-25 17:23:13 +02:00
break
2019-09-25 16:55:13 +02:00
}
prevMetricID = metricID
}
if ! hasDuplicates {
return sortedMetricIDs
}
dstMetricIDs := sortedMetricIDs [ : 1 ]
prevMetricID = sortedMetricIDs [ 0 ]
for _ , metricID := range sortedMetricIDs [ 1 : ] {
if prevMetricID == metricID {
continue
}
dstMetricIDs = append ( dstMetricIDs , metricID )
prevMetricID = metricID
}
return dstMetricIDs
}
2019-09-20 18:46:47 +02:00
func getTagToMetricIDsRowsMerger ( ) * tagToMetricIDsRowsMerger {
v := tmmPool . Get ( )
if v == nil {
return & tagToMetricIDsRowsMerger { }
}
return v . ( * tagToMetricIDsRowsMerger )
}
func putTagToMetricIDsRowsMerger ( tmm * tagToMetricIDsRowsMerger ) {
2019-09-23 19:40:38 +02:00
tmm . Reset ( )
2019-09-20 18:46:47 +02:00
tmmPool . Put ( tmm )
}
var tmmPool sync . Pool