2023-06-20 05:55:12 +00:00
package logstorage
import (
"path/filepath"
"sync"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/filestream"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/fs"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
)
type readerWithStats struct {
r filestream . ReadCloser
bytesRead uint64
}
func ( r * readerWithStats ) reset ( ) {
r . r = nil
r . bytesRead = 0
}
func ( r * readerWithStats ) init ( rc filestream . ReadCloser ) {
r . reset ( )
r . r = rc
}
// Path returns the path to r file
func ( r * readerWithStats ) Path ( ) string {
return r . r . Path ( )
}
// MustReadFull reads len(data) to r.
func ( r * readerWithStats ) MustReadFull ( data [ ] byte ) {
fs . MustReadData ( r . r , data )
r . bytesRead += uint64 ( len ( data ) )
}
func ( r * readerWithStats ) Read ( p [ ] byte ) ( int , error ) {
n , err := r . r . Read ( p )
r . bytesRead += uint64 ( n )
return n , err
}
func ( r * readerWithStats ) MustClose ( ) {
r . r . MustClose ( )
r . r = nil
}
// streamReaders contains readers for blockStreamReader
type streamReaders struct {
metaindexReader readerWithStats
indexReader readerWithStats
columnsHeaderReader readerWithStats
timestampsReader readerWithStats
fieldValuesReader readerWithStats
fieldBloomFilterReader readerWithStats
messageValuesReader readerWithStats
messageBloomFilterReader readerWithStats
}
func ( sr * streamReaders ) reset ( ) {
sr . metaindexReader . reset ( )
sr . indexReader . reset ( )
sr . columnsHeaderReader . reset ( )
sr . timestampsReader . reset ( )
sr . fieldValuesReader . reset ( )
sr . fieldBloomFilterReader . reset ( )
sr . messageValuesReader . reset ( )
sr . messageBloomFilterReader . reset ( )
}
func ( sr * streamReaders ) init ( metaindexReader , indexReader , columnsHeaderReader , timestampsReader , fieldValuesReader , fieldBloomFilterReader ,
messageValuesReader , messageBloomFilterReader filestream . ReadCloser ,
) {
sr . metaindexReader . init ( metaindexReader )
sr . indexReader . init ( indexReader )
sr . columnsHeaderReader . init ( columnsHeaderReader )
sr . timestampsReader . init ( timestampsReader )
sr . fieldValuesReader . init ( fieldValuesReader )
sr . fieldBloomFilterReader . init ( fieldBloomFilterReader )
sr . messageValuesReader . init ( messageValuesReader )
sr . messageBloomFilterReader . init ( messageBloomFilterReader )
}
func ( sr * streamReaders ) totalBytesRead ( ) uint64 {
n := uint64 ( 0 )
n += sr . metaindexReader . bytesRead
n += sr . indexReader . bytesRead
n += sr . columnsHeaderReader . bytesRead
n += sr . timestampsReader . bytesRead
n += sr . fieldValuesReader . bytesRead
n += sr . fieldBloomFilterReader . bytesRead
n += sr . messageValuesReader . bytesRead
n += sr . messageBloomFilterReader . bytesRead
return n
}
func ( sr * streamReaders ) MustClose ( ) {
sr . metaindexReader . MustClose ( )
sr . indexReader . MustClose ( )
sr . columnsHeaderReader . MustClose ( )
sr . timestampsReader . MustClose ( )
sr . fieldValuesReader . MustClose ( )
sr . fieldBloomFilterReader . MustClose ( )
sr . messageValuesReader . MustClose ( )
sr . messageBloomFilterReader . MustClose ( )
}
// blockStreamReader is used for reading blocks in streaming manner from a part.
type blockStreamReader struct {
// blockData contains the data for the last read block
blockData blockData
2024-05-12 14:33:29 +00:00
// a contains data for blockData
a arena
2023-06-20 05:55:12 +00:00
// ph is the header for the part
ph partHeader
// streamReaders contains data readers in stream mode
streamReaders streamReaders
// indexBlockHeaders contains the list of all the indexBlockHeader entries for the part
indexBlockHeaders [ ] indexBlockHeader
// blockHeaders contains the list of blockHeader entries for the current indexBlockHeader pointed by nextIndexBlockIdx
blockHeaders [ ] blockHeader
// nextIndexBlockIdx is the index of the next item to read from indexBlockHeaders
nextIndexBlockIdx int
// nextBlockIdx is the index of the next item to read from blockHeaders
nextBlockIdx int
// globalUncompressedSizeBytes is the total size of log entries seen in the part
globalUncompressedSizeBytes uint64
// globalRowsCount is the number of log entries seen in the part
globalRowsCount uint64
// globalBlocksCount is the number of blocks seen in the part
globalBlocksCount uint64
// sidLast is the stream id for the previously read block
sidLast streamID
// minTimestampLast is the minimum timestamp for the previously read block
minTimestampLast int64
}
// reset resets bsr, so it can be re-used
func ( bsr * blockStreamReader ) reset ( ) {
bsr . blockData . reset ( )
2024-05-12 14:33:29 +00:00
bsr . a . reset ( )
2023-06-20 05:55:12 +00:00
bsr . ph . reset ( )
bsr . streamReaders . reset ( )
ihs := bsr . indexBlockHeaders
if len ( ihs ) > 10e3 {
// The ihs len is unbound, so it is better to drop too long indexBlockHeaders in order to reduce memory usage
ihs = nil
}
for i := range ihs {
ihs [ i ] . reset ( )
}
bsr . indexBlockHeaders = ihs [ : 0 ]
bhs := bsr . blockHeaders
for i := range bhs {
bhs [ i ] . reset ( )
}
bsr . blockHeaders = bhs [ : 0 ]
bsr . nextIndexBlockIdx = 0
bsr . nextBlockIdx = 0
bsr . globalUncompressedSizeBytes = 0
bsr . globalRowsCount = 0
bsr . globalBlocksCount = 0
bsr . sidLast . reset ( )
bsr . minTimestampLast = 0
}
// Path returns part path for bsr (e.g. file path, url or in-memory reference)
func ( bsr * blockStreamReader ) Path ( ) string {
path := bsr . streamReaders . metaindexReader . Path ( )
return filepath . Dir ( path )
}
// MustInitFromInmemoryPart initializes bsr from mp.
func ( bsr * blockStreamReader ) MustInitFromInmemoryPart ( mp * inmemoryPart ) {
bsr . reset ( )
bsr . ph = mp . ph
// Initialize streamReaders
metaindexReader := mp . metaindex . NewReader ( )
indexReader := mp . index . NewReader ( )
columnsHeaderReader := mp . columnsHeader . NewReader ( )
timestampsReader := mp . timestamps . NewReader ( )
fieldValuesReader := mp . fieldValues . NewReader ( )
fieldBloomFilterReader := mp . fieldBloomFilter . NewReader ( )
messageValuesReader := mp . messageValues . NewReader ( )
messageBloomFilterReader := mp . messageBloomFilter . NewReader ( )
bsr . streamReaders . init ( metaindexReader , indexReader , columnsHeaderReader , timestampsReader ,
fieldValuesReader , fieldBloomFilterReader , messageValuesReader , messageBloomFilterReader )
// Read metaindex data
bsr . indexBlockHeaders = mustReadIndexBlockHeaders ( bsr . indexBlockHeaders [ : 0 ] , & bsr . streamReaders . metaindexReader )
}
// MustInitFromFilePart initializes bsr from file part at the given path.
func ( bsr * blockStreamReader ) MustInitFromFilePart ( path string ) {
bsr . reset ( )
// Files in the part are always read without OS cache pollution,
// since they are usually deleted after the merge.
const nocache = true
metaindexPath := filepath . Join ( path , metaindexFilename )
indexPath := filepath . Join ( path , indexFilename )
columnsHeaderPath := filepath . Join ( path , columnsHeaderFilename )
timestampsPath := filepath . Join ( path , timestampsFilename )
fieldValuesPath := filepath . Join ( path , fieldValuesFilename )
fieldBloomFilterPath := filepath . Join ( path , fieldBloomFilename )
messageValuesPath := filepath . Join ( path , messageValuesFilename )
messageBloomFilterPath := filepath . Join ( path , messageBloomFilename )
bsr . ph . mustReadMetadata ( path )
// Open data readers
metaindexReader := filestream . MustOpen ( metaindexPath , nocache )
indexReader := filestream . MustOpen ( indexPath , nocache )
columnsHeaderReader := filestream . MustOpen ( columnsHeaderPath , nocache )
timestampsReader := filestream . MustOpen ( timestampsPath , nocache )
fieldValuesReader := filestream . MustOpen ( fieldValuesPath , nocache )
fieldBloomFilterReader := filestream . MustOpen ( fieldBloomFilterPath , nocache )
messageValuesReader := filestream . MustOpen ( messageValuesPath , nocache )
messageBloomFilterReader := filestream . MustOpen ( messageBloomFilterPath , nocache )
// Initialize streamReaders
bsr . streamReaders . init ( metaindexReader , indexReader , columnsHeaderReader , timestampsReader ,
fieldValuesReader , fieldBloomFilterReader , messageValuesReader , messageBloomFilterReader )
// Read metaindex data
bsr . indexBlockHeaders = mustReadIndexBlockHeaders ( bsr . indexBlockHeaders [ : 0 ] , & bsr . streamReaders . metaindexReader )
}
// NextBlock reads the next block from bsr and puts it into bsr.blockData.
//
// false is returned if there are no other blocks.
2024-05-12 14:33:29 +00:00
//
// bsr.blockData is valid until the next call to NextBlock().
2023-06-20 05:55:12 +00:00
func ( bsr * blockStreamReader ) NextBlock ( ) bool {
for bsr . nextBlockIdx >= len ( bsr . blockHeaders ) {
if ! bsr . nextIndexBlock ( ) {
return false
}
}
ih := & bsr . indexBlockHeaders [ bsr . nextIndexBlockIdx - 1 ]
bh := & bsr . blockHeaders [ bsr . nextBlockIdx ]
th := & bh . timestampsHeader
// Validate bh
if bh . streamID . less ( & bsr . sidLast ) {
logger . Panicf ( "FATAL: %s: blockHeader.streamID=%s cannot be smaller than the streamID from the previously read block: %s" , bsr . Path ( ) , & bh . streamID , & bsr . sidLast )
}
if bh . streamID . equal ( & bsr . sidLast ) && th . minTimestamp < bsr . minTimestampLast {
logger . Panicf ( "FATAL: %s: timestamps.minTimestamp=%d cannot be smaller than the minTimestamp for the previously read block for the same streamID: %d" ,
bsr . Path ( ) , th . minTimestamp , bsr . minTimestampLast )
}
bsr . minTimestampLast = th . minTimestamp
bsr . sidLast = bh . streamID
if th . minTimestamp < ih . minTimestamp {
logger . Panicf ( "FATAL: %s: timestampsHeader.minTimestamp=%d cannot be smaller than indexBlockHeader.minTimestamp=%d" , bsr . Path ( ) , th . minTimestamp , ih . minTimestamp )
}
if th . maxTimestamp > ih . maxTimestamp {
logger . Panicf ( "FATAL: %s: timestampsHeader.maxTimestamp=%d cannot be bigger than indexBlockHeader.maxTimestamp=%d" , bsr . Path ( ) , th . maxTimestamp , ih . minTimestamp )
}
// Read bsr.blockData
2024-05-12 14:33:29 +00:00
bsr . a . reset ( )
bsr . blockData . mustReadFrom ( & bsr . a , bh , & bsr . streamReaders )
2023-06-20 05:55:12 +00:00
bsr . globalUncompressedSizeBytes += bh . uncompressedSizeBytes
bsr . globalRowsCount += bh . rowsCount
bsr . globalBlocksCount ++
if bsr . globalUncompressedSizeBytes > bsr . ph . UncompressedSizeBytes {
logger . Panicf ( "FATAL: %s: too big size of entries read: %d; mustn't exceed partHeader.UncompressedSizeBytes=%d" ,
bsr . Path ( ) , bsr . globalUncompressedSizeBytes , bsr . ph . UncompressedSizeBytes )
}
if bsr . globalRowsCount > bsr . ph . RowsCount {
logger . Panicf ( "FATAL: %s: too many log entries read so far: %d; mustn't exceed partHeader.RowsCount=%d" , bsr . Path ( ) , bsr . globalRowsCount , bsr . ph . RowsCount )
}
if bsr . globalBlocksCount > bsr . ph . BlocksCount {
logger . Panicf ( "FATAL: %s: too many blocks read so far: %d; mustn't exceed partHeader.BlocksCount=%d" , bsr . Path ( ) , bsr . globalBlocksCount , bsr . ph . BlocksCount )
}
// The block has been sucessfully read
bsr . nextBlockIdx ++
return true
}
func ( bsr * blockStreamReader ) nextIndexBlock ( ) bool {
// Advance to the next indexBlockHeader
if bsr . nextIndexBlockIdx >= len ( bsr . indexBlockHeaders ) {
// No more blocks left
// Validate bsr.ph
totalBytesRead := bsr . streamReaders . totalBytesRead ( )
if bsr . ph . CompressedSizeBytes != totalBytesRead {
logger . Panicf ( "FATAL: %s: partHeader.CompressedSizeBytes=%d must match the size of data read: %d" , bsr . Path ( ) , bsr . ph . CompressedSizeBytes , totalBytesRead )
}
if bsr . ph . UncompressedSizeBytes != bsr . globalUncompressedSizeBytes {
logger . Panicf ( "FATAL: %s: partHeader.UncompressedSizeBytes=%d must match the size of entries read: %d" ,
bsr . Path ( ) , bsr . ph . UncompressedSizeBytes , bsr . globalUncompressedSizeBytes )
}
if bsr . ph . RowsCount != bsr . globalRowsCount {
logger . Panicf ( "FATAL: %s: partHeader.RowsCount=%d must match the number of log entries read: %d" , bsr . Path ( ) , bsr . ph . RowsCount , bsr . globalRowsCount )
}
if bsr . ph . BlocksCount != bsr . globalBlocksCount {
logger . Panicf ( "FATAL: %s: partHeader.BlocksCount=%d must match the number of blocks read: %d" , bsr . Path ( ) , bsr . ph . BlocksCount , bsr . globalBlocksCount )
}
return false
}
ih := & bsr . indexBlockHeaders [ bsr . nextIndexBlockIdx ]
// Validate ih
metaindexReader := & bsr . streamReaders . metaindexReader
if ih . minTimestamp < bsr . ph . MinTimestamp {
logger . Panicf ( "FATAL: %s: indexBlockHeader.minTimestamp=%d cannot be smaller than partHeader.MinTimestamp=%d" ,
metaindexReader . Path ( ) , ih . minTimestamp , bsr . ph . MinTimestamp )
}
if ih . maxTimestamp > bsr . ph . MaxTimestamp {
logger . Panicf ( "FATAL: %s: indexBlockHeader.maxTimestamp=%d cannot be bigger than partHeader.MaxTimestamp=%d" ,
metaindexReader . Path ( ) , ih . maxTimestamp , bsr . ph . MaxTimestamp )
}
// Read indexBlock for the given ih
bb := longTermBufPool . Get ( )
bb . B = ih . mustReadNextIndexBlock ( bb . B [ : 0 ] , & bsr . streamReaders )
bsr . blockHeaders = resetBlockHeaders ( bsr . blockHeaders )
var err error
bsr . blockHeaders , err = unmarshalBlockHeaders ( bsr . blockHeaders [ : 0 ] , bb . B )
longTermBufPool . Put ( bb )
if err != nil {
logger . Panicf ( "FATAL: %s: cannot unmarshal blockHeader entries: %s" , bsr . streamReaders . indexReader . Path ( ) , err )
}
bsr . nextIndexBlockIdx ++
bsr . nextBlockIdx = 0
return true
}
// MustClose closes bsr.
func ( bsr * blockStreamReader ) MustClose ( ) {
bsr . streamReaders . MustClose ( )
bsr . reset ( )
}
// getBlockStreamReader returns blockStreamReader.
//
// The returned blockStreamReader must be initialized with MustInit().
// call putBlockStreamReader() when the retruend blockStreamReader is no longer needed.
func getBlockStreamReader ( ) * blockStreamReader {
v := blockStreamReaderPool . Get ( )
if v == nil {
v = & blockStreamReader { }
}
bsr := v . ( * blockStreamReader )
return bsr
}
// putBlockStreamReader returns bsr to the pool.
//
// bsr cannot be used after returning to the pool.
func putBlockStreamReader ( bsr * blockStreamReader ) {
bsr . reset ( )
blockStreamReaderPool . Put ( bsr )
}
var blockStreamReaderPool sync . Pool
// mustCloseBlockStreamReaders calls MustClose() on the given bsrs.
func mustCloseBlockStreamReaders ( bsrs [ ] * blockStreamReader ) {
for _ , bsr := range bsrs {
bsr . MustClose ( )
}
}