package workingsetcache import ( "sync" "sync/atomic" "time" "github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil" "github.com/VictoriaMetrics/VictoriaMetrics/lib/cgroup" "github.com/VictoriaMetrics/fastcache" ) // Cache modes. const ( split = 0 switching = 1 whole = 2 ) const defaultExpireDuration = 10 * time.Minute // Cache is a cache for working set entries. // // The cache evicts inactive entries after the given expireDuration. // Recently accessed entries survive expireDuration. type Cache struct { curr atomic.Value prev atomic.Value // csHistory holds cache stats history csHistory fastcache.Stats // mode indicates whether to use only curr and skip prev. // // This flag is set to switching if curr is filled for more than 50% space. // In this case using prev would result in RAM waste, // it is better to use only curr cache with doubled size. // After the process of switching, this flag will be set to whole. mode uint32 // The maxBytes value passed to New() or to Load(). maxBytes int // mu serializes access to curr, prev and mode // in expirationWatcher and cacheSizeWatcher. mu sync.Mutex wg sync.WaitGroup stopCh chan struct{} } // Load loads the cache from filePath and limits its size to maxBytes // and evicts inactive entries in 20 minutes. // // Stop must be called on the returned cache when it is no longer needed. func Load(filePath string, maxBytes int) *Cache { return LoadWithExpire(filePath, maxBytes, defaultExpireDuration) } // LoadWithExpire loads the cache from filePath and limits its size to maxBytes // and evicts inactive entires after expireDuration. // // Stop must be called on the returned cache when it is no longer needed. func LoadWithExpire(filePath string, maxBytes int, expireDuration time.Duration) *Cache { curr := fastcache.LoadFromFileOrNew(filePath, maxBytes) var cs fastcache.Stats curr.UpdateStats(&cs) if cs.EntriesCount == 0 { curr.Reset() // The cache couldn't be loaded with maxBytes size. // This may mean that the cache is split into curr and prev caches. // Try loading it again with maxBytes / 2 size. // Put the loaded cache into `prev` instead of `curr` // in order to limit the growth of the cache for the current period of time. prev := fastcache.LoadFromFileOrNew(filePath, maxBytes/2) curr := fastcache.New(maxBytes / 2) c := newCacheInternal(curr, prev, split, maxBytes) c.runWatchers(expireDuration) return c } // The cache has been successfully loaded in full. // Set its' mode to `whole`. // There is no need in runWatchers call. prev := fastcache.New(1024) return newCacheInternal(curr, prev, whole, maxBytes) } // New creates new cache with the given maxBytes capacity. // // Stop must be called on the returned cache when it is no longer needed. func New(maxBytes int) *Cache { return NewWithExpire(maxBytes, defaultExpireDuration) } // NewWithExpire creates new cache with the given maxBytes capacity and the given expireDuration // for inactive entries. // // Stop must be called on the returned cache when it is no longer needed. func NewWithExpire(maxBytes int, expireDuration time.Duration) *Cache { curr := fastcache.New(maxBytes / 2) prev := fastcache.New(1024) c := newCacheInternal(curr, prev, split, maxBytes) c.runWatchers(expireDuration) return c } func newCacheInternal(curr, prev *fastcache.Cache, mode, maxBytes int) *Cache { var c Cache c.maxBytes = maxBytes c.curr.Store(curr) c.prev.Store(prev) c.stopCh = make(chan struct{}) c.setMode(mode) return &c } func (c *Cache) runWatchers(expireDuration time.Duration) { c.wg.Add(1) go func() { defer c.wg.Done() c.expirationWatcher(expireDuration) }() c.wg.Add(1) go func() { defer c.wg.Done() c.cacheSizeWatcher() }() } func (c *Cache) expirationWatcher(expireDuration time.Duration) { expireDuration += timeJitter(expireDuration / 10) t := time.NewTicker(expireDuration) defer t.Stop() var csCurr, csPrev fastcache.Stats for { select { case <-c.stopCh: return case <-t.C: } c.mu.Lock() if atomic.LoadUint32(&c.mode) != split { // Stop the expirationWatcher on non-split mode. c.mu.Unlock() return } // Reset prev cache and swap it with the curr cache. prev := c.prev.Load().(*fastcache.Cache) curr := c.curr.Load().(*fastcache.Cache) csCurr.Reset() curr.UpdateStats(&csCurr) csPrev.Reset() prev.UpdateStats(&csPrev) c.prev.Store(curr) prevGetCalls := csCurr.GetCalls updateCacheStatsHistory(&c.csHistory, &csPrev) prev.Reset() c.curr.Store(prev) c.mu.Unlock() // Watch for the usage of the prev cache and drop it whenever it receives // less than 5% of get calls comparing to the curr cache. // This allows saving memory. prev, curr = curr, prev checkInterval := 10 * time.Second checkerT := time.NewTicker(checkInterval) checkerDeadline := time.Now().Add(expireDuration - checkInterval) for time.Now().Before(checkerDeadline) { select { case <-c.stopCh: break case <-checkerT.C: } c.mu.Lock() if atomic.LoadUint32(&c.mode) != split { // Do nothing in non-split mode. c.mu.Unlock() break } csCurr.Reset() curr.UpdateStats(&csCurr) csPrev.Reset() prev.UpdateStats(&csPrev) getCalls := csPrev.GetCalls - prevGetCalls if float64(getCalls) < 0.05*float64(csCurr.GetCalls) { // The majority of requests are served from the curr cache, // so the prev cache can be deleted. prev.Reset() break } } checkerT.Stop() } } func (c *Cache) cacheSizeWatcher() { checkInterval := 1500 * time.Millisecond checkInterval += timeJitter(checkInterval / 10) t := time.NewTicker(checkInterval) defer t.Stop() var maxBytesSize uint64 for { select { case <-c.stopCh: return case <-t.C: } if c.loadMode() != split { continue } var cs fastcache.Stats curr := c.curr.Load().(*fastcache.Cache) curr.UpdateStats(&cs) if cs.BytesSize >= uint64(0.9*float64(cs.MaxBytesSize)) { maxBytesSize = cs.MaxBytesSize break } } // curr cache size exceeds 90% of its capacity. It is better // to double the size of curr cache and stop using prev cache, // since this will result in higher summary cache capacity. // // Do this in the following steps: // 1) switch to mode=switching // 2) move curr cache to prev // 3) create curr cache with doubled size // 4) wait until curr cache size exceeds maxBytesSize, i.e. it is populated with new data // 5) switch to mode=whole // 6) drop prev cache c.mu.Lock() c.setMode(switching) prev := c.prev.Load().(*fastcache.Cache) curr := c.curr.Load().(*fastcache.Cache) c.prev.Store(curr) var cs fastcache.Stats prev.UpdateStats(&cs) updateCacheStatsHistory(&c.csHistory, &cs) prev.Reset() // use c.maxBytes instead of maxBytesSize*2 for creating new cache, since otherwise the created cache // couldn't be loaded from file with c.maxBytes limit after saving with maxBytesSize*2 limit. c.curr.Store(fastcache.New(c.maxBytes)) c.mu.Unlock() for { select { case <-c.stopCh: return case <-t.C: } var cs fastcache.Stats curr := c.curr.Load().(*fastcache.Cache) curr.UpdateStats(&cs) if cs.BytesSize >= maxBytesSize { break } } c.mu.Lock() c.setMode(whole) prev = c.prev.Load().(*fastcache.Cache) c.prev.Store(fastcache.New(1024)) cs.Reset() prev.UpdateStats(&cs) updateCacheStatsHistory(&c.csHistory, &cs) prev.Reset() c.mu.Unlock() } // Save saves the cache to filePath. func (c *Cache) Save(filePath string) error { curr := c.curr.Load().(*fastcache.Cache) concurrency := cgroup.AvailableCPUs() return curr.SaveToFileConcurrent(filePath, concurrency) } // Stop stops the cache. // // The cache cannot be used after the Stop call. func (c *Cache) Stop() { close(c.stopCh) c.wg.Wait() c.Reset() } // Reset resets the cache. func (c *Cache) Reset() { prev := c.prev.Load().(*fastcache.Cache) prev.Reset() curr := c.curr.Load().(*fastcache.Cache) curr.Reset() // Reset the mode to `split` in the hope the working set size becomes smaller after the reset. c.setMode(split) } func (c *Cache) setMode(mode int) { atomic.StoreUint32(&c.mode, uint32(mode)) } func (c *Cache) loadMode() int { return int(atomic.LoadUint32(&c.mode)) } // UpdateStats updates fcs with cache stats. func (c *Cache) UpdateStats(fcs *fastcache.Stats) { updateCacheStatsHistory(fcs, &c.csHistory) var cs fastcache.Stats curr := c.curr.Load().(*fastcache.Cache) curr.UpdateStats(&cs) updateCacheStats(fcs, &cs) prev := c.prev.Load().(*fastcache.Cache) cs.Reset() prev.UpdateStats(&cs) updateCacheStats(fcs, &cs) } func updateCacheStats(dst, src *fastcache.Stats) { dst.GetCalls += src.GetCalls dst.SetCalls += src.SetCalls dst.Misses += src.Misses dst.Collisions += src.Collisions dst.Corruptions += src.Corruptions dst.EntriesCount += src.EntriesCount dst.BytesSize += src.BytesSize dst.MaxBytesSize += src.MaxBytesSize } func updateCacheStatsHistory(dst, src *fastcache.Stats) { atomic.AddUint64(&dst.GetCalls, atomic.LoadUint64(&src.GetCalls)) atomic.AddUint64(&dst.SetCalls, atomic.LoadUint64(&src.SetCalls)) atomic.AddUint64(&dst.Misses, atomic.LoadUint64(&src.Misses)) atomic.AddUint64(&dst.Collisions, atomic.LoadUint64(&src.Collisions)) atomic.AddUint64(&dst.Corruptions, atomic.LoadUint64(&src.Corruptions)) // Do not add EntriesCount, BytesSize and MaxBytesSize, since these metrics // are calculated from c.curr and c.prev caches. } // Get appends the found value for the given key to dst and returns the result. func (c *Cache) Get(dst, key []byte) []byte { curr := c.curr.Load().(*fastcache.Cache) result := curr.Get(dst, key) if len(result) > len(dst) { // Fast path - the entry is found in the current cache. return result } if c.loadMode() == whole { // Nothing found. return result } // Search for the entry in the previous cache. prev := c.prev.Load().(*fastcache.Cache) result = prev.Get(dst, key) if len(result) <= len(dst) { // Nothing found. return result } // Cache the found entry in the current cache. curr.Set(key, result[len(dst):]) return result } // Has verifies whether the cache contains the given key. func (c *Cache) Has(key []byte) bool { curr := c.curr.Load().(*fastcache.Cache) if curr.Has(key) { return true } if c.loadMode() == whole { return false } prev := c.prev.Load().(*fastcache.Cache) if !prev.Has(key) { return false } // Cache the found entry in the current cache. tmpBuf := tmpBufPool.Get() tmpBuf.B = prev.Get(tmpBuf.B, key) curr.Set(key, tmpBuf.B) tmpBufPool.Put(tmpBuf) return true } var tmpBufPool bytesutil.ByteBufferPool // Set sets the given value for the given key. func (c *Cache) Set(key, value []byte) { curr := c.curr.Load().(*fastcache.Cache) curr.Set(key, value) } // GetBig appends the found value for the given key to dst and returns the result. func (c *Cache) GetBig(dst, key []byte) []byte { curr := c.curr.Load().(*fastcache.Cache) result := curr.GetBig(dst, key) if len(result) > len(dst) { // Fast path - the entry is found in the current cache. return result } if c.loadMode() == whole { // Nothing found. return result } // Search for the entry in the previous cache. prev := c.prev.Load().(*fastcache.Cache) result = prev.GetBig(dst, key) if len(result) <= len(dst) { // Nothing found. return result } // Cache the found entry in the current cache. curr.SetBig(key, result[len(dst):]) return result } // SetBig sets the given value for the given key. func (c *Cache) SetBig(key, value []byte) { curr := c.curr.Load().(*fastcache.Cache) curr.SetBig(key, value) } func timeJitter(d time.Duration) time.Duration { n := float64(time.Now().UnixNano()%1e9) / 1e9 return time.Duration(float64(d) * n) }