2020-02-23 12:35:47 +01:00
|
|
|
package persistentqueue
|
|
|
|
|
|
|
|
import (
|
|
|
|
"encoding/json"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"io/ioutil"
|
|
|
|
"os"
|
|
|
|
"regexp"
|
|
|
|
"strconv"
|
|
|
|
"sync"
|
|
|
|
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil"
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/encoding"
|
2020-09-16 16:30:04 +02:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fasttime"
|
2020-02-23 12:35:47 +01:00
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/filestream"
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fs"
|
|
|
|
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
|
2020-03-03 18:48:46 +01:00
|
|
|
"github.com/VictoriaMetrics/metrics"
|
2020-02-23 12:35:47 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
// MaxBlockSize is the maximum size of the block persistent queue can work with.
|
|
|
|
const MaxBlockSize = 32 * 1024 * 1024
|
|
|
|
|
|
|
|
const defaultChunkFileSize = (MaxBlockSize + 8) * 16
|
|
|
|
|
|
|
|
var chunkFileNameRegex = regexp.MustCompile("^[0-9A-F]{16}$")
|
|
|
|
|
|
|
|
// Queue represents persistent queue.
|
|
|
|
type Queue struct {
|
2020-03-03 18:48:46 +01:00
|
|
|
chunkFileSize uint64
|
|
|
|
maxBlockSize uint64
|
|
|
|
maxPendingBytes uint64
|
2020-02-23 12:35:47 +01:00
|
|
|
|
|
|
|
dir string
|
|
|
|
name string
|
|
|
|
|
|
|
|
// mu protects all the fields below.
|
|
|
|
mu sync.Mutex
|
|
|
|
|
|
|
|
// cond is used for notifying blocked readers when new data has been added
|
|
|
|
// or when MustClose is called.
|
|
|
|
cond sync.Cond
|
|
|
|
|
|
|
|
reader *filestream.Reader
|
|
|
|
readerPath string
|
|
|
|
readerOffset uint64
|
|
|
|
readerLocalOffset uint64
|
|
|
|
|
|
|
|
writer *filestream.Writer
|
|
|
|
writerPath string
|
|
|
|
writerOffset uint64
|
|
|
|
writerLocalOffset uint64
|
|
|
|
writerFlushedOffset uint64
|
|
|
|
|
2020-09-16 16:30:04 +02:00
|
|
|
lastMetainfoFlushTime uint64
|
|
|
|
|
2020-02-23 12:35:47 +01:00
|
|
|
mustStop bool
|
2020-03-03 18:48:46 +01:00
|
|
|
|
|
|
|
blocksDropped *metrics.Counter
|
|
|
|
bytesDropped *metrics.Counter
|
|
|
|
|
|
|
|
blocksWritten *metrics.Counter
|
|
|
|
bytesWritten *metrics.Counter
|
|
|
|
|
|
|
|
blocksRead *metrics.Counter
|
|
|
|
bytesRead *metrics.Counter
|
2020-07-13 20:54:09 +02:00
|
|
|
|
|
|
|
bytesPending *metrics.Gauge
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
2020-02-28 18:57:39 +01:00
|
|
|
// ResetIfEmpty resets q if it is empty.
|
|
|
|
//
|
|
|
|
// This is needed in order to remove chunk file associated with empty q.
|
|
|
|
func (q *Queue) ResetIfEmpty() {
|
|
|
|
q.mu.Lock()
|
|
|
|
defer q.mu.Unlock()
|
|
|
|
|
|
|
|
if q.readerOffset != q.writerOffset {
|
|
|
|
// The queue isn't empty.
|
|
|
|
return
|
|
|
|
}
|
|
|
|
if q.readerOffset < 16*1024*1024 {
|
|
|
|
// The file is too small to drop. Leave it as is in order to reduce filesystem load.
|
|
|
|
return
|
|
|
|
}
|
|
|
|
if q.readerPath != q.writerPath {
|
|
|
|
logger.Panicf("BUG: readerPath=%q doesn't match writerPath=%q", q.readerPath, q.writerPath)
|
|
|
|
}
|
|
|
|
|
|
|
|
q.reader.MustClose()
|
|
|
|
q.writer.MustClose()
|
|
|
|
fs.MustRemoveAll(q.readerPath)
|
|
|
|
|
|
|
|
q.writerOffset = 0
|
|
|
|
q.writerLocalOffset = 0
|
|
|
|
q.writerFlushedOffset = 0
|
|
|
|
|
|
|
|
q.readerOffset = 0
|
|
|
|
q.readerLocalOffset = 0
|
|
|
|
|
|
|
|
q.writerPath = q.chunkFilePath(q.writerOffset)
|
|
|
|
w, err := filestream.Create(q.writerPath, false)
|
|
|
|
if err != nil {
|
|
|
|
logger.Panicf("FATAL: cannot create chunk file %q: %s", q.writerPath, err)
|
|
|
|
}
|
|
|
|
q.writer = w
|
|
|
|
|
|
|
|
q.readerPath = q.writerPath
|
|
|
|
r, err := filestream.Open(q.readerPath, true)
|
|
|
|
if err != nil {
|
|
|
|
logger.Panicf("FATAL: cannot open chunk file %q: %s", q.readerPath, err)
|
|
|
|
}
|
|
|
|
q.reader = r
|
|
|
|
|
2020-09-16 16:30:04 +02:00
|
|
|
if err := q.flushMetainfoLocked(); err != nil {
|
2020-02-28 18:57:39 +01:00
|
|
|
logger.Panicf("FATAL: cannot flush metainfo: %s", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-23 12:35:47 +01:00
|
|
|
// GetPendingBytes returns the number of pending bytes in the queue.
|
|
|
|
func (q *Queue) GetPendingBytes() uint64 {
|
|
|
|
q.mu.Lock()
|
|
|
|
n := q.writerOffset - q.readerOffset
|
|
|
|
q.mu.Unlock()
|
|
|
|
return n
|
|
|
|
}
|
|
|
|
|
|
|
|
// MustOpen opens persistent queue from the given path.
|
2020-03-03 18:48:46 +01:00
|
|
|
//
|
|
|
|
// If maxPendingBytes is greater than 0, then the max queue size is limited by this value.
|
|
|
|
// The oldest data is deleted when queue size exceeds maxPendingBytes.
|
|
|
|
func MustOpen(path, name string, maxPendingBytes int) *Queue {
|
|
|
|
if maxPendingBytes < 0 {
|
|
|
|
maxPendingBytes = 0
|
|
|
|
}
|
|
|
|
return mustOpen(path, name, defaultChunkFileSize, MaxBlockSize, uint64(maxPendingBytes))
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
2020-03-03 18:48:46 +01:00
|
|
|
func mustOpen(path, name string, chunkFileSize, maxBlockSize, maxPendingBytes uint64) *Queue {
|
2020-02-23 12:35:47 +01:00
|
|
|
if chunkFileSize < 8 || chunkFileSize-8 < maxBlockSize {
|
|
|
|
logger.Panicf("BUG: too small chunkFileSize=%d for maxBlockSize=%d; chunkFileSize must fit at least one block", chunkFileSize, maxBlockSize)
|
|
|
|
}
|
|
|
|
if maxBlockSize <= 0 {
|
|
|
|
logger.Panicf("BUG: maxBlockSize must be greater than 0; got %d", maxBlockSize)
|
|
|
|
}
|
2020-03-03 18:48:46 +01:00
|
|
|
q, err := tryOpeningQueue(path, name, chunkFileSize, maxBlockSize, maxPendingBytes)
|
2020-02-23 12:35:47 +01:00
|
|
|
if err != nil {
|
|
|
|
logger.Errorf("cannot open persistent queue at %q: %s; cleaning it up and trying again", path, err)
|
|
|
|
fs.RemoveDirContents(path)
|
2020-03-03 18:48:46 +01:00
|
|
|
q, err = tryOpeningQueue(path, name, chunkFileSize, maxBlockSize, maxPendingBytes)
|
2020-02-23 12:35:47 +01:00
|
|
|
if err != nil {
|
|
|
|
logger.Panicf("FATAL: %s", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return q
|
|
|
|
}
|
|
|
|
|
2020-03-03 18:48:46 +01:00
|
|
|
func tryOpeningQueue(path, name string, chunkFileSize, maxBlockSize, maxPendingBytes uint64) (*Queue, error) {
|
2020-02-23 12:35:47 +01:00
|
|
|
var q Queue
|
|
|
|
q.chunkFileSize = chunkFileSize
|
|
|
|
q.maxBlockSize = maxBlockSize
|
2020-03-03 18:48:46 +01:00
|
|
|
q.maxPendingBytes = maxPendingBytes
|
2020-02-23 12:35:47 +01:00
|
|
|
q.dir = path
|
|
|
|
q.name = name
|
|
|
|
q.cond.L = &q.mu
|
|
|
|
|
2020-03-03 18:48:46 +01:00
|
|
|
q.blocksDropped = metrics.GetOrCreateCounter(fmt.Sprintf(`vm_persistentqueue_blocks_dropped_total{path=%q}`, path))
|
|
|
|
q.bytesDropped = metrics.GetOrCreateCounter(fmt.Sprintf(`vm_persistentqueue_bytes_dropped_total{path=%q}`, path))
|
|
|
|
q.blocksWritten = metrics.GetOrCreateCounter(fmt.Sprintf(`vm_persistentqueue_blocks_written_total{path=%q}`, path))
|
|
|
|
q.bytesWritten = metrics.GetOrCreateCounter(fmt.Sprintf(`vm_persistentqueue_bytes_written_total{path=%q}`, path))
|
|
|
|
q.blocksRead = metrics.GetOrCreateCounter(fmt.Sprintf(`vm_persistentqueue_blocks_read_total{path=%q}`, path))
|
|
|
|
q.bytesRead = metrics.GetOrCreateCounter(fmt.Sprintf(`vm_persistentqueue_bytes_read_total{path=%q}`, path))
|
2020-07-13 20:54:09 +02:00
|
|
|
q.bytesPending = metrics.GetOrCreateGauge(fmt.Sprintf(`vm_persistentqueue_bytes_pending{path=%q}`, path), func() float64 {
|
|
|
|
return float64(q.GetPendingBytes())
|
|
|
|
})
|
2020-03-03 18:48:46 +01:00
|
|
|
|
2020-02-23 12:35:47 +01:00
|
|
|
cleanOnError := func() {
|
|
|
|
if q.reader != nil {
|
|
|
|
q.reader.MustClose()
|
|
|
|
}
|
|
|
|
if q.writer != nil {
|
|
|
|
q.writer.MustClose()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := fs.MkdirAllIfNotExist(path); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return nil, fmt.Errorf("cannot create directory %q: %w", path, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// Read metainfo.
|
|
|
|
var mi metainfo
|
|
|
|
metainfoPath := q.metainfoPath()
|
|
|
|
if err := mi.ReadFromFile(metainfoPath); err != nil {
|
|
|
|
if !os.IsNotExist(err) {
|
|
|
|
logger.Errorf("cannot read metainfo for persistent queue from %q: %s; re-creating %q", metainfoPath, err, path)
|
|
|
|
}
|
|
|
|
|
|
|
|
// path contents is broken or missing. Re-create it from scratch.
|
|
|
|
fs.RemoveDirContents(path)
|
|
|
|
mi.Reset()
|
|
|
|
mi.Name = q.name
|
|
|
|
if err := mi.WriteToFile(metainfoPath); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return nil, fmt.Errorf("cannot create %q: %w", metainfoPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// Create initial chunk file.
|
|
|
|
filepath := q.chunkFilePath(0)
|
|
|
|
if err := fs.WriteFileAtomically(filepath, nil); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return nil, fmt.Errorf("cannot create %q: %w", filepath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
if mi.Name != q.name {
|
|
|
|
return nil, fmt.Errorf("unexpected queue name; got %q; want %q", mi.Name, q.name)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Locate reader and writer chunks in the path.
|
|
|
|
fis, err := ioutil.ReadDir(path)
|
|
|
|
if err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return nil, fmt.Errorf("cannot read contents of the directory %q: %w", path, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
for _, fi := range fis {
|
|
|
|
fname := fi.Name()
|
|
|
|
filepath := path + "/" + fname
|
|
|
|
if fi.IsDir() {
|
|
|
|
logger.Errorf("skipping unknown directory %q", filepath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if fname == "metainfo.json" {
|
|
|
|
// skip metainfo file
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if !chunkFileNameRegex.MatchString(fname) {
|
|
|
|
logger.Errorf("skipping unknown file %q", filepath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
offset, err := strconv.ParseUint(fname, 16, 64)
|
|
|
|
if err != nil {
|
|
|
|
logger.Panicf("BUG: cannot parse hex %q: %s", fname, err)
|
|
|
|
}
|
|
|
|
if offset%q.chunkFileSize != 0 {
|
2020-09-16 16:30:04 +02:00
|
|
|
logger.Errorf("unexpected offset for chunk file %q: %d; it must be multiple of %d; removing the file", filepath, offset, q.chunkFileSize)
|
2020-02-23 12:35:47 +01:00
|
|
|
fs.MustRemoveAll(filepath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if mi.ReaderOffset >= offset+q.chunkFileSize {
|
|
|
|
logger.Errorf("unexpected chunk file found from the past: %q; removing it", filepath)
|
|
|
|
fs.MustRemoveAll(filepath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if mi.WriterOffset < offset {
|
|
|
|
logger.Errorf("unexpected chunk file found from the future: %q; removing it", filepath)
|
|
|
|
fs.MustRemoveAll(filepath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if mi.ReaderOffset >= offset && mi.ReaderOffset < offset+q.chunkFileSize {
|
|
|
|
// Found the chunk for reading
|
|
|
|
if q.reader != nil {
|
|
|
|
logger.Panicf("BUG: reader is already initialized with readerPath=%q, readerOffset=%d, readerLocalOffset=%d",
|
|
|
|
q.readerPath, q.readerOffset, q.readerLocalOffset)
|
|
|
|
}
|
|
|
|
q.readerPath = filepath
|
|
|
|
q.readerOffset = mi.ReaderOffset
|
|
|
|
q.readerLocalOffset = mi.ReaderOffset % q.chunkFileSize
|
|
|
|
if fileSize := fs.MustFileSize(q.readerPath); fileSize < q.readerLocalOffset {
|
|
|
|
logger.Errorf("chunk file %q size is too small for the given reader offset; file size %d bytes; reader offset: %d bytes; removing the file",
|
|
|
|
q.readerPath, fileSize, q.readerLocalOffset)
|
|
|
|
fs.MustRemoveAll(q.readerPath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
r, err := filestream.OpenReaderAt(q.readerPath, int64(q.readerLocalOffset), true)
|
|
|
|
if err != nil {
|
|
|
|
logger.Errorf("cannot open %q for reading at offset %d: %s; removing this file", q.readerPath, q.readerLocalOffset, err)
|
|
|
|
fs.MustRemoveAll(filepath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
q.reader = r
|
|
|
|
}
|
|
|
|
if mi.WriterOffset >= offset && mi.WriterOffset < offset+q.chunkFileSize {
|
|
|
|
// Found the chunk file for writing
|
|
|
|
if q.writer != nil {
|
|
|
|
logger.Panicf("BUG: writer is already initialized with writerPath=%q, writerOffset=%d, writerLocalOffset=%d",
|
|
|
|
q.writerPath, q.writerOffset, q.writerLocalOffset)
|
|
|
|
}
|
|
|
|
q.writerPath = filepath
|
|
|
|
q.writerOffset = mi.WriterOffset
|
|
|
|
q.writerLocalOffset = mi.WriterOffset % q.chunkFileSize
|
|
|
|
q.writerFlushedOffset = mi.WriterOffset
|
|
|
|
if fileSize := fs.MustFileSize(q.writerPath); fileSize != q.writerLocalOffset {
|
2020-09-16 16:30:04 +02:00
|
|
|
if fileSize < q.writerLocalOffset {
|
|
|
|
logger.Errorf("%q size (%d bytes) is smaller than the writer offset (%d bytes); removing the file",
|
|
|
|
q.writerPath, fileSize, q.writerLocalOffset)
|
|
|
|
fs.MustRemoveAll(q.writerPath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
logger.Warnf("%q size (%d bytes) is bigger than writer offset (%d bytes); "+
|
|
|
|
"this may be the case on unclean shutdown (OOM, `kill -9`, hardware reset); trying to fix it by adjusting fileSize to %d",
|
|
|
|
q.writerPath, fileSize, q.writerLocalOffset, q.writerLocalOffset)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
w, err := filestream.OpenWriterAt(q.writerPath, int64(q.writerLocalOffset), false)
|
|
|
|
if err != nil {
|
|
|
|
logger.Errorf("cannot open %q for writing at offset %d: %s; removing this file", q.writerPath, q.writerLocalOffset, err)
|
|
|
|
fs.MustRemoveAll(filepath)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
q.writer = w
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if q.reader == nil {
|
|
|
|
cleanOnError()
|
|
|
|
return nil, fmt.Errorf("couldn't find chunk file for reading in %q", q.dir)
|
|
|
|
}
|
|
|
|
if q.writer == nil {
|
|
|
|
cleanOnError()
|
|
|
|
return nil, fmt.Errorf("couldn't find chunk file for writing in %q", q.dir)
|
|
|
|
}
|
|
|
|
return &q, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// MustClose closes q.
|
|
|
|
//
|
|
|
|
// It unblocks all the MustReadBlock calls.
|
|
|
|
//
|
|
|
|
// MustWriteBlock mustn't be called during and after the call to MustClose.
|
|
|
|
func (q *Queue) MustClose() {
|
|
|
|
q.mu.Lock()
|
|
|
|
defer q.mu.Unlock()
|
|
|
|
|
|
|
|
// Unblock goroutines blocked on cond in MustReadBlock.
|
|
|
|
q.mustStop = true
|
|
|
|
q.cond.Broadcast()
|
|
|
|
|
|
|
|
// Close writer.
|
|
|
|
q.writer.MustClose()
|
|
|
|
q.writer = nil
|
|
|
|
|
|
|
|
// Close reader.
|
|
|
|
q.reader.MustClose()
|
|
|
|
q.reader = nil
|
|
|
|
|
|
|
|
// Store metainfo
|
2020-09-16 16:30:04 +02:00
|
|
|
if err := q.flushMetainfoLocked(); err != nil {
|
2020-02-23 12:35:47 +01:00
|
|
|
logger.Panicf("FATAL: cannot flush chunked queue metainfo: %s", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (q *Queue) chunkFilePath(offset uint64) string {
|
|
|
|
return fmt.Sprintf("%s/%016X", q.dir, offset)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (q *Queue) metainfoPath() string {
|
|
|
|
return q.dir + "/metainfo.json"
|
|
|
|
}
|
|
|
|
|
|
|
|
// MustWriteBlock writes block to q.
|
|
|
|
//
|
|
|
|
// The block size cannot exceed MaxBlockSize.
|
|
|
|
//
|
|
|
|
// It is safe calling this function from concurrent goroutines.
|
|
|
|
func (q *Queue) MustWriteBlock(block []byte) {
|
|
|
|
if uint64(len(block)) > q.maxBlockSize {
|
|
|
|
logger.Panicf("BUG: too big block to send: %d bytes; it mustn't exceed %d bytes", len(block), q.maxBlockSize)
|
|
|
|
}
|
|
|
|
|
|
|
|
q.mu.Lock()
|
|
|
|
defer q.mu.Unlock()
|
|
|
|
|
|
|
|
if q.mustStop {
|
|
|
|
logger.Panicf("BUG: MustWriteBlock cannot be called after MustClose")
|
|
|
|
}
|
|
|
|
if q.readerOffset > q.writerOffset {
|
|
|
|
logger.Panicf("BUG: readerOffset=%d shouldn't exceed writerOffset=%d", q.readerOffset, q.writerOffset)
|
|
|
|
}
|
2020-03-03 18:48:46 +01:00
|
|
|
if q.maxPendingBytes > 0 {
|
|
|
|
// Drain the oldest blocks until the number of pending bytes becomes enough for the block.
|
|
|
|
blockSize := uint64(len(block) + 8)
|
|
|
|
maxPendingBytes := q.maxPendingBytes
|
|
|
|
if blockSize < maxPendingBytes {
|
|
|
|
maxPendingBytes -= blockSize
|
|
|
|
} else {
|
|
|
|
maxPendingBytes = 0
|
|
|
|
}
|
|
|
|
bb := blockBufPool.Get()
|
|
|
|
for q.writerOffset-q.readerOffset > maxPendingBytes {
|
|
|
|
var err error
|
|
|
|
bb.B, err = q.readBlockLocked(bb.B[:0])
|
|
|
|
if err != nil {
|
|
|
|
logger.Panicf("FATAL: cannot read the oldest block %s", err)
|
|
|
|
}
|
|
|
|
q.blocksDropped.Inc()
|
|
|
|
q.bytesDropped.Add(len(bb.B))
|
|
|
|
}
|
|
|
|
blockBufPool.Put(bb)
|
|
|
|
if blockSize > q.maxPendingBytes {
|
|
|
|
// The block is too big to put it into the queue. Drop it.
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
2020-02-23 12:35:47 +01:00
|
|
|
if err := q.writeBlockLocked(block); err != nil {
|
|
|
|
logger.Panicf("FATAL: %s", err)
|
|
|
|
}
|
2020-05-16 08:31:46 +02:00
|
|
|
|
|
|
|
// Notify blocked reader if any.
|
|
|
|
// See https://github.com/VictoriaMetrics/VictoriaMetrics/pull/484 for details.
|
|
|
|
q.cond.Signal()
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
2020-03-03 18:48:46 +01:00
|
|
|
var blockBufPool bytesutil.ByteBufferPool
|
|
|
|
|
2020-02-23 12:35:47 +01:00
|
|
|
func (q *Queue) writeBlockLocked(block []byte) error {
|
|
|
|
if q.writerLocalOffset+q.maxBlockSize+8 > q.chunkFileSize {
|
|
|
|
// Finalize the current chunk and start new one.
|
|
|
|
q.writer.MustClose()
|
2020-09-16 16:30:04 +02:00
|
|
|
fs.MustSyncPath(q.writerPath)
|
2020-02-23 12:35:47 +01:00
|
|
|
if n := q.writerOffset % q.chunkFileSize; n > 0 {
|
|
|
|
q.writerOffset += (q.chunkFileSize - n)
|
|
|
|
}
|
|
|
|
q.writerFlushedOffset = q.writerOffset
|
|
|
|
q.writerLocalOffset = 0
|
|
|
|
q.writerPath = q.chunkFilePath(q.writerOffset)
|
|
|
|
w, err := filestream.Create(q.writerPath, false)
|
|
|
|
if err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot create chunk file %q: %w", q.writerPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
q.writer = w
|
2020-09-16 16:30:04 +02:00
|
|
|
if err := q.flushMetainfoLocked(); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot flush metainfo: %w", err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
2020-09-16 16:30:04 +02:00
|
|
|
fs.MustSyncPath(q.dir)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// Write block len.
|
|
|
|
blockLen := uint64(len(block))
|
|
|
|
header := headerBufPool.Get()
|
|
|
|
header.B = encoding.MarshalUint64(header.B, blockLen)
|
|
|
|
err := q.write(header.B)
|
|
|
|
headerBufPool.Put(header)
|
|
|
|
if err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot write header with size 8 bytes to %q: %w", q.writerPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// Write block contents.
|
|
|
|
if err := q.write(block); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot write block contents with size %d bytes to %q: %w", len(block), q.writerPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
2020-03-03 18:48:46 +01:00
|
|
|
q.blocksWritten.Inc()
|
|
|
|
q.bytesWritten.Add(len(block))
|
2020-09-16 16:30:04 +02:00
|
|
|
return q.flushMetainfoIfNeededLocked()
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// MustReadBlock appends the next block from q to dst and returns the result.
|
|
|
|
//
|
|
|
|
// false is returned after MustClose call.
|
|
|
|
//
|
|
|
|
// It is safe calling this function from concurrent goroutines.
|
|
|
|
func (q *Queue) MustReadBlock(dst []byte) ([]byte, bool) {
|
|
|
|
q.mu.Lock()
|
|
|
|
defer q.mu.Unlock()
|
|
|
|
|
|
|
|
for {
|
|
|
|
if q.mustStop {
|
|
|
|
return dst, false
|
|
|
|
}
|
|
|
|
if q.readerOffset > q.writerOffset {
|
|
|
|
logger.Panicf("BUG: readerOffset=%d cannot exceed writerOffset=%d", q.readerOffset, q.writerOffset)
|
|
|
|
}
|
|
|
|
if q.readerOffset < q.writerOffset {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
q.cond.Wait()
|
|
|
|
}
|
|
|
|
|
|
|
|
data, err := q.readBlockLocked(dst)
|
|
|
|
if err != nil {
|
2020-09-16 16:30:04 +02:00
|
|
|
// Skip the current chunk, since it may be broken.
|
|
|
|
q.readerOffset += q.chunkFileSize - q.readerOffset%q.chunkFileSize
|
|
|
|
_ = q.flushMetainfoLocked()
|
2020-02-23 12:35:47 +01:00
|
|
|
logger.Panicf("FATAL: %s", err)
|
|
|
|
}
|
|
|
|
return data, true
|
|
|
|
}
|
|
|
|
|
|
|
|
func (q *Queue) readBlockLocked(dst []byte) ([]byte, error) {
|
|
|
|
if q.readerLocalOffset+q.maxBlockSize+8 > q.chunkFileSize {
|
|
|
|
// Remove the current chunk and go to the next chunk.
|
|
|
|
q.reader.MustClose()
|
|
|
|
fs.MustRemoveAll(q.readerPath)
|
|
|
|
if n := q.readerOffset % q.chunkFileSize; n > 0 {
|
|
|
|
q.readerOffset += (q.chunkFileSize - n)
|
|
|
|
}
|
|
|
|
q.readerLocalOffset = 0
|
|
|
|
q.readerPath = q.chunkFilePath(q.readerOffset)
|
|
|
|
r, err := filestream.Open(q.readerPath, true)
|
|
|
|
if err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return dst, fmt.Errorf("cannot open chunk file %q: %w", q.readerPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
q.reader = r
|
2020-09-16 16:30:04 +02:00
|
|
|
if err := q.flushMetainfoLocked(); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return dst, fmt.Errorf("cannot flush metainfo: %w", err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
2020-09-16 16:30:04 +02:00
|
|
|
fs.MustSyncPath(q.dir)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// Read block len.
|
|
|
|
header := headerBufPool.Get()
|
|
|
|
header.B = bytesutil.Resize(header.B, 8)
|
|
|
|
err := q.readFull(header.B)
|
|
|
|
blockLen := encoding.UnmarshalUint64(header.B)
|
|
|
|
headerBufPool.Put(header)
|
|
|
|
if err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return dst, fmt.Errorf("cannot read header with size 8 bytes from %q: %w", q.readerPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
if blockLen > q.maxBlockSize {
|
|
|
|
return dst, fmt.Errorf("too big block size read from %q: %d bytes; cannot exceed %d bytes", q.readerPath, blockLen, q.maxBlockSize)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Read block contents.
|
|
|
|
dstLen := len(dst)
|
|
|
|
dst = bytesutil.Resize(dst, dstLen+int(blockLen))
|
|
|
|
if err := q.readFull(dst[dstLen:]); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return dst, fmt.Errorf("cannot read block contents with size %d bytes from %q: %w", blockLen, q.readerPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
2020-03-03 18:48:46 +01:00
|
|
|
q.blocksRead.Inc()
|
|
|
|
q.bytesRead.Add(int(blockLen))
|
2020-09-16 16:30:04 +02:00
|
|
|
if err := q.flushMetainfoIfNeededLocked(); err != nil {
|
|
|
|
return dst, err
|
|
|
|
}
|
2020-02-23 12:35:47 +01:00
|
|
|
return dst, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (q *Queue) write(buf []byte) error {
|
|
|
|
bufLen := uint64(len(buf))
|
|
|
|
n, err := q.writer.Write(buf)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if uint64(n) != bufLen {
|
|
|
|
return fmt.Errorf("unexpected number of bytes written; got %d bytes; want %d bytes", n, bufLen)
|
|
|
|
}
|
|
|
|
q.writerLocalOffset += bufLen
|
|
|
|
q.writerOffset += bufLen
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (q *Queue) readFull(buf []byte) error {
|
|
|
|
bufLen := uint64(len(buf))
|
|
|
|
if q.readerOffset+bufLen > q.writerFlushedOffset {
|
|
|
|
q.writer.MustFlush()
|
|
|
|
q.writerFlushedOffset = q.writerOffset
|
|
|
|
}
|
|
|
|
n, err := io.ReadFull(q.reader, buf)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if uint64(n) != bufLen {
|
|
|
|
return fmt.Errorf("unexpected number of bytes read; got %d bytes; want %d bytes", n, bufLen)
|
|
|
|
}
|
|
|
|
q.readerLocalOffset += bufLen
|
|
|
|
q.readerOffset += bufLen
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-09-16 16:30:04 +02:00
|
|
|
func (q *Queue) flushMetainfoIfNeededLocked() error {
|
|
|
|
t := fasttime.UnixTimestamp()
|
|
|
|
if t == q.lastMetainfoFlushTime {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if err := q.flushMetainfoLocked(); err != nil {
|
|
|
|
return fmt.Errorf("cannot flush metainfo: %w", err)
|
|
|
|
}
|
|
|
|
q.lastMetainfoFlushTime = t
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (q *Queue) flushMetainfoLocked() error {
|
2020-02-23 12:35:47 +01:00
|
|
|
mi := &metainfo{
|
|
|
|
Name: q.name,
|
|
|
|
ReaderOffset: q.readerOffset,
|
|
|
|
WriterOffset: q.writerOffset,
|
|
|
|
}
|
|
|
|
metainfoPath := q.metainfoPath()
|
|
|
|
if err := mi.WriteToFile(metainfoPath); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot write metainfo to %q: %w", metainfoPath, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
var headerBufPool bytesutil.ByteBufferPool
|
|
|
|
|
|
|
|
type metainfo struct {
|
|
|
|
Name string
|
|
|
|
ReaderOffset uint64
|
|
|
|
WriterOffset uint64
|
|
|
|
}
|
|
|
|
|
|
|
|
func (mi *metainfo) Reset() {
|
|
|
|
mi.ReaderOffset = 0
|
|
|
|
mi.WriterOffset = 0
|
|
|
|
}
|
|
|
|
|
|
|
|
func (mi *metainfo) WriteToFile(path string) error {
|
|
|
|
data, err := json.Marshal(mi)
|
|
|
|
if err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot marshal persistent queue metainfo %#v: %w", mi, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
if err := ioutil.WriteFile(path, data, 0600); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot write persistent queue metainfo to %q: %w", path, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
2020-09-16 16:30:04 +02:00
|
|
|
fs.MustSyncPath(path)
|
2020-02-23 12:35:47 +01:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (mi *metainfo) ReadFromFile(path string) error {
|
|
|
|
mi.Reset()
|
|
|
|
data, err := ioutil.ReadFile(path)
|
|
|
|
if err != nil {
|
|
|
|
if os.IsNotExist(err) {
|
|
|
|
return err
|
|
|
|
}
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot read %q: %w", path, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
if err := json.Unmarshal(data, mi); err != nil {
|
2020-06-30 21:58:18 +02:00
|
|
|
return fmt.Errorf("cannot unmarshal persistent queue metainfo from %q: %w", path, err)
|
2020-02-23 12:35:47 +01:00
|
|
|
}
|
|
|
|
if mi.ReaderOffset > mi.WriterOffset {
|
|
|
|
return fmt.Errorf("invalid data read from %q: readerOffset=%d cannot exceed writerOffset=%d", path, mi.ReaderOffset, mi.WriterOffset)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|