VictoriaMetrics/lib/mergeset/part_header.go
Aliaksandr Valialkin 43b24164ef
all: add Windows build for VictoriaMetrics
This commit changes background merge algorithm, so it becomes compatible with Windows file semantics.

The previous algorithm for background merge:

1. Merge source parts into a destination part inside tmp directory.
2. Create a file in txn directory with instructions on how to atomically
   swap source parts with the destination part.
3. Perform instructions from the file.
4. Delete the file with instructions.

This algorithm guarantees that either source parts or destination part
is visible in the partition after unclean shutdown at any step above,
since the remaining files with instructions is replayed on the next restart,
after that the remaining contents of the tmp directory is deleted.

Unfortunately this algorithm doesn't work under Windows because
it disallows removing and moving files, which are in use.

So the new algorithm for background merge has been implemented:

1. Merge source parts into a destination part inside the partition directory itself.
   E.g. now the partition directory may contain both complete and incomplete parts.
2. Atomically update the parts.json file with the new list of parts after the merge,
   e.g. remove the source parts from the list and add the destination part to the list
   before storing it to parts.json file.
3. Remove the source parts from disk when they are no longer used.

This algorithm guarantees that either source parts or destination part
is visible in the partition after unclean shutdown at any step above,
since incomplete partitions from step 1 or old source parts from step 3 are removed
on the next startup by inspecting parts.json file.

This algorithm should work under Windows, since it doesn't remove or move files in use.
This algorithm has also the following benefits:

- It should work better for NFS.
- It fits object storage semantics.

The new algorithm changes data storage format, so it is impossible to downgrade
to the previous versions of VictoriaMetrics after upgrading to this algorithm.

Updates https://github.com/VictoriaMetrics/VictoriaMetrics/issues/3236
Updates https://github.com/VictoriaMetrics/VictoriaMetrics/issues/3821
Updates https://github.com/VictoriaMetrics/VictoriaMetrics/issues/70
2023-03-19 01:36:51 -07:00

132 lines
3.4 KiB
Go

package mergeset
import (
"encoding/hex"
"encoding/json"
"fmt"
"os"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fs"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
)
type partHeader struct {
// The number of items the part contains.
itemsCount uint64
// The number of blocks the part contains.
blocksCount uint64
// The first item in the part.
firstItem []byte
// The last item in the part.
lastItem []byte
}
type partHeaderJSON struct {
ItemsCount uint64
BlocksCount uint64
FirstItem hexString
LastItem hexString
}
type hexString []byte
func (hs hexString) MarshalJSON() ([]byte, error) {
h := hex.EncodeToString(hs)
b := make([]byte, 0, len(h)+2)
b = append(b, '"')
b = append(b, h...)
b = append(b, '"')
return b, nil
}
func (hs *hexString) UnmarshalJSON(data []byte) error {
if len(data) < 2 {
return fmt.Errorf("too small data string: got %q; must be at least 2 bytes", data)
}
if data[0] != '"' || data[len(data)-1] != '"' {
return fmt.Errorf("missing heading and/or tailing quotes in the data string %q", data)
}
data = data[1 : len(data)-1]
b, err := hex.DecodeString(string(data))
if err != nil {
return fmt.Errorf("cannot hex-decode %q: %w", data, err)
}
*hs = b
return nil
}
func (ph *partHeader) Reset() {
ph.itemsCount = 0
ph.blocksCount = 0
ph.firstItem = ph.firstItem[:0]
ph.lastItem = ph.lastItem[:0]
}
func (ph *partHeader) String() string {
return fmt.Sprintf("partHeader{itemsCount: %d, blocksCount: %d, firstItem: %X, lastItem: %X}",
ph.itemsCount, ph.blocksCount, ph.firstItem, ph.lastItem)
}
func (ph *partHeader) CopyFrom(src *partHeader) {
ph.itemsCount = src.itemsCount
ph.blocksCount = src.blocksCount
ph.firstItem = append(ph.firstItem[:0], src.firstItem...)
ph.lastItem = append(ph.lastItem[:0], src.lastItem...)
}
func (ph *partHeader) ReadMetadata(partPath string) error {
ph.Reset()
// Read ph fields from metadata.
metadataPath := partPath + "/metadata.json"
metadata, err := os.ReadFile(metadataPath)
if err != nil {
return fmt.Errorf("cannot read %q: %w", metadataPath, err)
}
var phj partHeaderJSON
if err := json.Unmarshal(metadata, &phj); err != nil {
return fmt.Errorf("cannot parse %q: %w", metadataPath, err)
}
if phj.ItemsCount <= 0 {
return fmt.Errorf("part %q cannot contain zero items", partPath)
}
ph.itemsCount = phj.ItemsCount
if phj.BlocksCount <= 0 {
return fmt.Errorf("part %q cannot contain zero blocks", partPath)
}
if phj.BlocksCount > phj.ItemsCount {
return fmt.Errorf("the number of blocks cannot exceed the number of items in the part %q; got blocksCount=%d, itemsCount=%d",
partPath, phj.BlocksCount, phj.ItemsCount)
}
ph.blocksCount = phj.BlocksCount
ph.firstItem = append(ph.firstItem[:0], phj.FirstItem...)
ph.lastItem = append(ph.lastItem[:0], phj.LastItem...)
return nil
}
func (ph *partHeader) WriteMetadata(partPath string) error {
phj := &partHeaderJSON{
ItemsCount: ph.itemsCount,
BlocksCount: ph.blocksCount,
FirstItem: append([]byte{}, ph.firstItem...),
LastItem: append([]byte{}, ph.lastItem...),
}
metadata, err := json.Marshal(&phj)
if err != nil {
logger.Panicf("BUG: cannot marshal partHeader metadata: %s", err)
}
metadataPath := partPath + "/metadata.json"
if err := fs.WriteFileAtomically(metadataPath, metadata, false); err != nil {
return fmt.Errorf("cannot create %q: %w", metadataPath, err)
}
return nil
}