package logstorage import ( "fmt" "math" "sync" "github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil" "github.com/VictoriaMetrics/VictoriaMetrics/lib/encoding" "github.com/VictoriaMetrics/VictoriaMetrics/lib/logger" "github.com/VictoriaMetrics/VictoriaMetrics/lib/slicesutil" ) // blockHeader contains information about a single block. // // blockHeader is stored in the indexFilename file. type blockHeader struct { // streamID is a stream id for entries in the block streamID streamID // uncompressedSizeBytes is the original (uncompressed) size of log entries stored in the block uncompressedSizeBytes uint64 // rowsCount is the number of log entries stored in the block rowsCount uint64 // timestampsHeader contains information about timestamps for log entries in the block timestampsHeader timestampsHeader // columnsHeaderOffset is the offset of columnsHeader at columnsHeaderFilename columnsHeaderOffset uint64 // columnsHeaderSize is the size of columnsHeader at columnsHeaderFilename columnsHeaderSize uint64 } // reset resets bh, so it can be re-used. func (bh *blockHeader) reset() { bh.streamID.reset() bh.uncompressedSizeBytes = 0 bh.rowsCount = 0 bh.timestampsHeader.reset() bh.columnsHeaderOffset = 0 bh.columnsHeaderSize = 0 } func (bh *blockHeader) copyFrom(src *blockHeader) { bh.reset() bh.streamID = src.streamID bh.uncompressedSizeBytes = src.uncompressedSizeBytes bh.rowsCount = src.rowsCount bh.timestampsHeader.copyFrom(&src.timestampsHeader) bh.columnsHeaderOffset = src.columnsHeaderOffset bh.columnsHeaderSize = src.columnsHeaderSize } // marshal appends the marshaled bh to dst and returns the result. func (bh *blockHeader) marshal(dst []byte) []byte { // Do not store the version used for encoding directly in the block header, since: // - all the block headers in the same part use the same encoding // - the block header encoding version can be put in metadata file for the part (aka metadataFilename) dst = bh.streamID.marshal(dst) dst = encoding.MarshalVarUint64(dst, bh.uncompressedSizeBytes) dst = encoding.MarshalVarUint64(dst, bh.rowsCount) dst = bh.timestampsHeader.marshal(dst) dst = encoding.MarshalVarUint64(dst, bh.columnsHeaderOffset) dst = encoding.MarshalVarUint64(dst, bh.columnsHeaderSize) return dst } // unmarshal unmarshals bh from src and returns the remaining tail. func (bh *blockHeader) unmarshal(src []byte) ([]byte, error) { bh.reset() srcOrig := src // unmarshal bh.streamID tail, err := bh.streamID.unmarshal(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal streamID: %w", err) } src = tail // unmarshal bh.uncompressedSizeBytes n, nSize := encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal uncompressedSizeBytes") } src = src[nSize:] bh.uncompressedSizeBytes = n // unmarshal bh.rowsCount n, nSize = encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal rowsCount") } src = src[nSize:] if n > maxRowsPerBlock { return srcOrig, fmt.Errorf("too big value for rowsCount: %d; mustn't exceed %d", n, maxRowsPerBlock) } bh.rowsCount = n // unmarshal bh.timestampsHeader tail, err = bh.timestampsHeader.unmarshal(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal timestampsHeader: %w", err) } src = tail // unmarshal columnsHeaderOffset n, nSize = encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal columnsHeaderOffset") } src = src[nSize:] bh.columnsHeaderOffset = n // unmarshal columnsHeaderSize n, nSize = encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal columnsHeaderSize") } src = src[nSize:] if n > maxColumnsHeaderSize { return srcOrig, fmt.Errorf("too big value for columnsHeaderSize: %d; mustn't exceed %d", n, maxColumnsHeaderSize) } bh.columnsHeaderSize = n return src, nil } func getBlockHeader() *blockHeader { v := blockHeaderPool.Get() if v == nil { return &blockHeader{} } return v.(*blockHeader) } func putBlockHeader(bh *blockHeader) { bh.reset() blockHeaderPool.Put(bh) } var blockHeaderPool sync.Pool // unmarshalBlockHeaders appends unmarshaled from src blockHeader entries to dst and returns the result. func unmarshalBlockHeaders(dst []blockHeader, src []byte) ([]blockHeader, error) { dstOrig := dst for len(src) > 0 { if len(dst) < cap(dst) { dst = dst[:len(dst)+1] } else { dst = append(dst, blockHeader{}) } bh := &dst[len(dst)-1] tail, err := bh.unmarshal(src) if err != nil { return dstOrig, fmt.Errorf("cannot unmarshal blockHeader entries: %w", err) } src = tail } if err := validateBlockHeaders(dst[len(dstOrig):]); err != nil { return dstOrig, err } return dst, nil } func validateBlockHeaders(bhs []blockHeader) error { for i := 1; i < len(bhs); i++ { bhCurr := &bhs[i] bhPrev := &bhs[i-1] if bhCurr.streamID.less(&bhPrev.streamID) { return fmt.Errorf("unexpected blockHeader with smaller streamID=%s after bigger streamID=%s at position %d", &bhCurr.streamID, &bhPrev.streamID, i) } if !bhCurr.streamID.equal(&bhPrev.streamID) { continue } thCurr := bhCurr.timestampsHeader thPrev := bhPrev.timestampsHeader if thCurr.minTimestamp < thPrev.minTimestamp { return fmt.Errorf("unexpected blockHeader with smaller timestamp=%d after bigger timestamp=%d at position %d", thCurr.minTimestamp, thPrev.minTimestamp, i) } } return nil } func resetBlockHeaders(bhs []blockHeader) []blockHeader { for i := range bhs { bhs[i].reset() } return bhs[:0] } func getColumnsHeader() *columnsHeader { v := columnsHeaderPool.Get() if v == nil { return &columnsHeader{} } return v.(*columnsHeader) } func putColumnsHeader(csh *columnsHeader) { csh.reset() columnsHeaderPool.Put(csh) } var columnsHeaderPool sync.Pool // columnsHeader contains information about columns in a single block. // // columnsHeader is stored in the columnsHeaderFilename file. type columnsHeader struct { // columnHeaders contains the information about every column seen in the block. columnHeaders []columnHeader // constColumns contain fields with constant values across all the block entries. constColumns []Field } func (csh *columnsHeader) reset() { chs := csh.columnHeaders for i := range chs { chs[i].reset() } csh.columnHeaders = chs[:0] ccs := csh.constColumns for i := range ccs { ccs[i].Reset() } csh.constColumns = ccs[:0] } func (csh *columnsHeader) resizeConstColumns(columnsLen int) []Field { csh.constColumns = slicesutil.SetLength(csh.constColumns, columnsLen) return csh.constColumns } func (csh *columnsHeader) resizeColumnHeaders(columnHeadersLen int) []columnHeader { csh.columnHeaders = slicesutil.SetLength(csh.columnHeaders, columnHeadersLen) return csh.columnHeaders } func (csh *columnsHeader) marshal(dst []byte) []byte { chs := csh.columnHeaders dst = encoding.MarshalVarUint64(dst, uint64(len(chs))) for i := range chs { dst = chs[i].marshal(dst) } ccs := csh.constColumns dst = encoding.MarshalVarUint64(dst, uint64(len(ccs))) for i := range ccs { dst = ccs[i].marshal(dst) } return dst } // unmarshalNoArena unmarshals csh from src. // // csh is valid until src is changed. func (csh *columnsHeader) unmarshalNoArena(src []byte) error { csh.reset() // unmarshal columnHeaders n, nSize := encoding.UnmarshalVarUint64(src) if nSize <= 0 { return fmt.Errorf("cannot unmarshal columnHeaders len") } src = src[nSize:] if n > maxColumnsPerBlock { return fmt.Errorf("too many column headers: %d; mustn't exceed %d", n, maxColumnsPerBlock) } chs := csh.resizeColumnHeaders(int(n)) for i := range chs { tail, err := chs[i].unmarshalNoArena(src) if err != nil { return fmt.Errorf("cannot unmarshal columnHeader %d out of %d columnHeaders: %w", i, len(chs), err) } src = tail } csh.columnHeaders = chs // unmarshal constColumns n, nSize = encoding.UnmarshalVarUint64(src) if nSize <= 0 { return fmt.Errorf("cannot unmarshal constColumns len") } src = src[nSize:] if n+uint64(len(csh.columnHeaders)) > maxColumnsPerBlock { return fmt.Errorf("too many columns: %d; mustn't exceed %d", n+uint64(len(csh.columnHeaders)), maxColumnsPerBlock) } ccs := csh.resizeConstColumns(int(n)) for i := range ccs { tail, err := ccs[i].unmarshalNoArena(src) if err != nil { return fmt.Errorf("cannot unmarshal constColumn %d out of %d columns: %w", i, len(ccs), err) } src = tail } // Verify that the src is empty if len(src) > 0 { return fmt.Errorf("unexpected non-empty tail left after unmarshaling columnsHeader: len(tail)=%d", len(src)) } return nil } // columnHeaders contains information for values, which belong to a single label in a single block. // // The main column with an empty name is stored in messageValuesFilename, // while the rest of columns are stored in fieldValuesFilename. // This allows minimizing disk read IO when filtering by non-message columns. // // Every block column contains also a bloom filter for all the tokens stored in the column. // This bloom filter is used for fast determining whether the given block may contain the given tokens. // // Tokens in bloom filter depend on valueType: // // - valueTypeString stores tokens seen in all the values // - valueTypeDict doesn't store anything in the bloom filter, since all the encoded values // are available directly in the valuesDict field // - valueTypeUint8, valueTypeUint16, valueTypeUint32 and valueTypeUint64 stores encoded uint values // - valueTypeFloat64 stores encoded float64 values // - valueTypeIPv4 stores encoded into uint32 ips // - valueTypeTimestampISO8601 stores encoded into uint64 timestamps // // Bloom filters for main column with an empty name is stored in messageBloomFilename, // while the rest of columns are stored in fieldBloomFilename. type columnHeader struct { // name contains column name aka label name name string // valueType is the type of values stored in the block valueType valueType // minValue is the minimum encoded value for uint*, ipv4, timestamp and float64 value in the columnHeader // // It is used for fast detection of whether the given columnHeader contains values in the given range minValue uint64 // maxValue is the maximum encoded value for uint*, ipv4, timestamp and float64 value in the columnHeader // // It is used for fast detection of whether the given columnHeader contains values in the given range maxValue uint64 // valuesDict contains unique values for valueType = valueTypeDict valuesDict valuesDict // valuesOffset contains the offset of the block in either messageValuesFilename or fieldValuesFilename valuesOffset uint64 // valuesSize contains the size of the block in either messageValuesFilename or fieldValuesFilename valuesSize uint64 // bloomFilterOffset contains the offset of the bloom filter in either messageBloomFilename or fieldBloomFilename bloomFilterOffset uint64 // bloomFilterSize contains the size of the bloom filter in either messageBloomFilename or fieldBloomFilename bloomFilterSize uint64 } // reset resets ch func (ch *columnHeader) reset() { ch.name = "" ch.valueType = 0 ch.minValue = 0 ch.maxValue = 0 ch.valuesDict.reset() ch.valuesOffset = 0 ch.valuesSize = 0 ch.bloomFilterOffset = 0 ch.bloomFilterSize = 0 } // marshal appends marshaled ch to dst and returns the result. func (ch *columnHeader) marshal(dst []byte) []byte { // check minValue/maxValue if ch.valueType == valueTypeFloat64 { minValue := math.Float64frombits(ch.minValue) maxValue := math.Float64frombits(ch.maxValue) if minValue > maxValue { logger.Panicf("BUG: minValue=%g must be smaller than maxValue=%g for valueTypeFloat64", minValue, maxValue) } } else if ch.valueType == valueTypeTimestampISO8601 { minValue := int64(ch.minValue) maxValue := int64(ch.maxValue) if minValue > maxValue { logger.Panicf("BUG: minValue=%g must be smaller than maxValue=%g for valueTypeTimestampISO8601", minValue, maxValue) } } else if ch.minValue > ch.maxValue { logger.Panicf("BUG: minValue=%d must be smaller than maxValue=%d for valueType=%d", ch.minValue, ch.maxValue, ch.valueType) } // Encode common fields - ch.name and ch.valueType dst = encoding.MarshalBytes(dst, bytesutil.ToUnsafeBytes(ch.name)) dst = append(dst, byte(ch.valueType)) // Encode other fields depending on ch.valueType switch ch.valueType { case valueTypeString: dst = ch.marshalValuesAndBloomFilters(dst) case valueTypeDict: dst = ch.valuesDict.marshal(dst) dst = ch.marshalValues(dst) case valueTypeUint8: dst = append(dst, byte(ch.minValue)) dst = append(dst, byte(ch.maxValue)) dst = ch.marshalValuesAndBloomFilters(dst) case valueTypeUint16: dst = encoding.MarshalUint16(dst, uint16(ch.minValue)) dst = encoding.MarshalUint16(dst, uint16(ch.maxValue)) dst = ch.marshalValuesAndBloomFilters(dst) case valueTypeUint32: dst = encoding.MarshalUint32(dst, uint32(ch.minValue)) dst = encoding.MarshalUint32(dst, uint32(ch.maxValue)) dst = ch.marshalValuesAndBloomFilters(dst) case valueTypeUint64: dst = encoding.MarshalUint64(dst, ch.minValue) dst = encoding.MarshalUint64(dst, ch.maxValue) dst = ch.marshalValuesAndBloomFilters(dst) case valueTypeFloat64: // float64 values are encoded as uint64 via math.Float64bits() dst = encoding.MarshalUint64(dst, ch.minValue) dst = encoding.MarshalUint64(dst, ch.maxValue) dst = ch.marshalValuesAndBloomFilters(dst) case valueTypeIPv4: dst = encoding.MarshalUint32(dst, uint32(ch.minValue)) dst = encoding.MarshalUint32(dst, uint32(ch.maxValue)) dst = ch.marshalValuesAndBloomFilters(dst) case valueTypeTimestampISO8601: // timestamps are encoded in nanoseconds dst = encoding.MarshalUint64(dst, ch.minValue) dst = encoding.MarshalUint64(dst, ch.maxValue) dst = ch.marshalValuesAndBloomFilters(dst) default: logger.Panicf("BUG: unknown valueType=%d", ch.valueType) } return dst } func (ch *columnHeader) marshalValuesAndBloomFilters(dst []byte) []byte { dst = ch.marshalValues(dst) dst = ch.marshalBloomFilters(dst) return dst } func (ch *columnHeader) marshalValues(dst []byte) []byte { dst = encoding.MarshalVarUint64(dst, ch.valuesOffset) dst = encoding.MarshalVarUint64(dst, ch.valuesSize) return dst } func (ch *columnHeader) marshalBloomFilters(dst []byte) []byte { dst = encoding.MarshalVarUint64(dst, ch.bloomFilterOffset) dst = encoding.MarshalVarUint64(dst, ch.bloomFilterSize) return dst } // unmarshalNoArena unmarshals ch from src and returns the tail left after unmarshaling. // // ch is valid until src is changed. func (ch *columnHeader) unmarshalNoArena(src []byte) ([]byte, error) { ch.reset() srcOrig := src // Unmarshal column name data, nSize := encoding.UnmarshalBytes(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal column name") } src = src[nSize:] ch.name = bytesutil.ToUnsafeString(data) // Unmarshal value type if len(src) < 1 { return srcOrig, fmt.Errorf("cannot unmarshal valueType from 0 bytes for column %q; need at least 1 byte", ch.name) } ch.valueType = valueType(src[0]) src = src[1:] // Unmarshal the rest of data depending on valueType switch ch.valueType { case valueTypeString: tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeString for column %q: %w", ch.name, err) } src = tail case valueTypeDict: tail, err := ch.valuesDict.unmarshalNoArena(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal dict at valueTypeDict for column %q: %w", ch.name, err) } src = tail tail, err = ch.unmarshalValues(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values at valueTypeDict for column %q: %w", ch.name, err) } src = tail case valueTypeUint8: if len(src) < 2 { return srcOrig, fmt.Errorf("cannot unmarshal min/max values at valueTypeUint8 from %d bytes for column %q; need at least 2 bytes", len(src), ch.name) } ch.minValue = uint64(src[0]) ch.maxValue = uint64(src[1]) src = src[2:] tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeUint8 for column %q: %w", ch.name, err) } src = tail case valueTypeUint16: if len(src) < 4 { return srcOrig, fmt.Errorf("cannot unmarshal min/max values at valueTypeUint16 from %d bytes for column %q; need at least 4 bytes", len(src), ch.name) } ch.minValue = uint64(encoding.UnmarshalUint16(src)) ch.maxValue = uint64(encoding.UnmarshalUint16(src[2:])) src = src[4:] tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeUint16 for column %q: %w", ch.name, err) } src = tail case valueTypeUint32: if len(src) < 8 { return srcOrig, fmt.Errorf("cannot unmarshal min/max values at valueTypeUint32 from %d bytes for column %q; need at least 8 bytes", len(src), ch.name) } ch.minValue = uint64(encoding.UnmarshalUint32(src)) ch.maxValue = uint64(encoding.UnmarshalUint32(src[4:])) src = src[8:] tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeUint32 for column %q: %w", ch.name, err) } src = tail case valueTypeUint64: if len(src) < 16 { return srcOrig, fmt.Errorf("cannot unmarshal min/max values at valueTypeUint64 from %d bytes for column %q; need at least 16 bytes", len(src), ch.name) } ch.minValue = encoding.UnmarshalUint64(src) ch.maxValue = encoding.UnmarshalUint64(src[8:]) src = src[16:] tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeUint64 for column %q: %w", ch.name, err) } src = tail case valueTypeFloat64: if len(src) < 16 { return srcOrig, fmt.Errorf("cannot unmarshal min/max values at valueTypeFloat64 from %d bytes for column %q; need at least 16 bytes", len(src), ch.name) } // min and max values must be converted to real values with math.Float64frombits() during querying. ch.minValue = encoding.UnmarshalUint64(src) ch.maxValue = encoding.UnmarshalUint64(src[8:]) src = src[16:] tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeFloat64 for column %q: %w", ch.name, err) } src = tail case valueTypeIPv4: if len(src) < 8 { return srcOrig, fmt.Errorf("cannot unmarshal min/max values at valueTypeIPv4 from %d bytes for column %q; need at least 8 bytes", len(src), ch.name) } ch.minValue = uint64(encoding.UnmarshalUint32(src)) ch.maxValue = uint64(encoding.UnmarshalUint32(src[4:])) src = src[8:] tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeIPv4 for column %q: %w", ch.name, err) } src = tail case valueTypeTimestampISO8601: if len(src) < 16 { return srcOrig, fmt.Errorf("cannot unmarshal min/max values at valueTypeTimestampISO8601 from %d bytes for column %q; need at least 16 bytes", len(src), ch.name) } ch.minValue = encoding.UnmarshalUint64(src) ch.maxValue = encoding.UnmarshalUint64(src[8:]) src = src[16:] tail, err := ch.unmarshalValuesAndBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values and bloom filters at valueTypeTimestampISO8601 for column %q: %w", ch.name, err) } src = tail default: return srcOrig, fmt.Errorf("unexpected valueType=%d for column %q", ch.valueType, ch.name) } return src, nil } func (ch *columnHeader) unmarshalValuesAndBloomFilters(src []byte) ([]byte, error) { srcOrig := src tail, err := ch.unmarshalValues(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal values: %w", err) } src = tail tail, err = ch.unmarshalBloomFilters(src) if err != nil { return srcOrig, fmt.Errorf("cannot unmarshal bloom filters: %w", err) } src = tail return src, nil } func (ch *columnHeader) unmarshalValues(src []byte) ([]byte, error) { srcOrig := src n, nSize := encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal valuesOffset") } src = src[nSize:] ch.valuesOffset = n n, nSize = encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal valuesSize") } src = src[nSize:] if n > maxValuesBlockSize { return srcOrig, fmt.Errorf("too big valuesSize: %d bytes; mustn't exceed %d bytes", n, maxValuesBlockSize) } ch.valuesSize = n return src, nil } func (ch *columnHeader) unmarshalBloomFilters(src []byte) ([]byte, error) { srcOrig := src n, nSize := encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal bloomFilterOffset") } src = src[nSize:] ch.bloomFilterOffset = n n, nSize = encoding.UnmarshalVarUint64(src) if nSize <= 0 { return srcOrig, fmt.Errorf("cannot unmarshal bloomFilterSize") } src = src[nSize:] if n > maxBloomFilterBlockSize { return srcOrig, fmt.Errorf("too big bloomFilterSize: %d bytes; mustn't exceed %d bytes", n, maxBloomFilterBlockSize) } ch.bloomFilterSize = n return src, nil } // timestampsHeader contains the information about timestamps block. type timestampsHeader struct { // blockOffset is an offset of timestamps block inside timestampsFilename file blockOffset uint64 // blockSize is the size of the timestamps block inside timestampsFilename file blockSize uint64 // minTimestamp is the mimumum timestamp seen in the block in nanoseconds minTimestamp int64 // maxTimestamp is the maximum timestamp seen in the block in nanoseconds maxTimestamp int64 // marshalType is the type used for encoding the timestamps block marshalType encoding.MarshalType } // reset resets th, so it can be reused func (th *timestampsHeader) reset() { th.blockOffset = 0 th.blockSize = 0 th.minTimestamp = 0 th.maxTimestamp = 0 th.marshalType = 0 } func (th *timestampsHeader) copyFrom(src *timestampsHeader) { th.blockOffset = src.blockOffset th.blockSize = src.blockSize th.minTimestamp = src.minTimestamp th.maxTimestamp = src.maxTimestamp th.marshalType = src.marshalType } // marshal appends marshaled th to dst and returns the result. func (th *timestampsHeader) marshal(dst []byte) []byte { dst = encoding.MarshalUint64(dst, th.blockOffset) dst = encoding.MarshalUint64(dst, th.blockSize) dst = encoding.MarshalUint64(dst, uint64(th.minTimestamp)) dst = encoding.MarshalUint64(dst, uint64(th.maxTimestamp)) dst = append(dst, byte(th.marshalType)) return dst } // unmarshal unmarshals th from src and returns the tail left after the unmarshaling. func (th *timestampsHeader) unmarshal(src []byte) ([]byte, error) { th.reset() if len(src) < 33 { return src, fmt.Errorf("cannot unmarshal timestampsHeader from %d bytes; need at least 33 bytes", len(src)) } th.blockOffset = encoding.UnmarshalUint64(src) th.blockSize = encoding.UnmarshalUint64(src[8:]) th.minTimestamp = int64(encoding.UnmarshalUint64(src[16:])) th.maxTimestamp = int64(encoding.UnmarshalUint64(src[24:])) th.marshalType = encoding.MarshalType(src[32]) return src[33:], nil }