lib/mergeset: remove inmemoryBlock pooling, since it wasn't effecitve

This should reduce memory usage a bit when new time series are ingested at high rate (aka high churn rate)
This commit is contained in:
Aliaksandr Valialkin 2024-01-26 21:33:08 +01:00
parent 7de19c3748
commit e84c877503
No known key found for this signature in database
GPG key ID: 52C003EE2BCDB9EB
3 changed files with 3 additions and 29 deletions

View file

@ -11,7 +11,6 @@ import (
"unsafe" "unsafe"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil" "github.com/VictoriaMetrics/VictoriaMetrics/lib/bytesutil"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/cgroup"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/encoding" "github.com/VictoriaMetrics/VictoriaMetrics/lib/encoding"
"github.com/VictoriaMetrics/VictoriaMetrics/lib/logger" "github.com/VictoriaMetrics/VictoriaMetrics/lib/logger"
) )
@ -571,26 +570,3 @@ func getLensBuffer(n int) *lensBuffer {
func putLensBuffer(lb *lensBuffer) { func putLensBuffer(lb *lensBuffer) {
lensBufferPool.Put(lb) lensBufferPool.Put(lb)
} }
func getInmemoryBlock() *inmemoryBlock {
select {
case ib := <-ibPoolCh:
return ib
default:
return &inmemoryBlock{}
}
}
func putInmemoryBlock(ib *inmemoryBlock) {
ib.Reset()
select {
case ibPoolCh <- ib:
default:
// drop ib in order to reduce memory usage on systems with big number of CPU cores
}
}
// Every inmemoryBlock struct occupies at least 64KB of memory, e.g. quite big amounts of memory.
// Use a chan instead of sync.Pool in order to reduce memory usage on systems
// with big number of CPU cores.
var ibPoolCh = make(chan *inmemoryBlock, 100*cgroup.AvailableCPUs())

View file

@ -325,7 +325,7 @@ func (ps *partSearch) readInmemoryBlock(bh *blockHeader) (*inmemoryBlock, error)
ps.sb.lensData = bytesutil.ResizeNoCopyMayOverallocate(ps.sb.lensData, int(bh.lensBlockSize)) ps.sb.lensData = bytesutil.ResizeNoCopyMayOverallocate(ps.sb.lensData, int(bh.lensBlockSize))
ps.p.lensFile.MustReadAt(ps.sb.lensData, int64(bh.lensBlockOffset)) ps.p.lensFile.MustReadAt(ps.sb.lensData, int64(bh.lensBlockOffset))
ib := getInmemoryBlock() ib := &inmemoryBlock{}
if err := ib.UnmarshalData(&ps.sb, bh.firstItem, bh.commonPrefix, bh.itemsCount, bh.marshalType); err != nil { if err := ib.UnmarshalData(&ps.sb, bh.firstItem, bh.commonPrefix, bh.itemsCount, bh.marshalType); err != nil {
return nil, fmt.Errorf("cannot unmarshal storage block with %d items: %w", bh.itemsCount, err) return nil, fmt.Errorf("cannot unmarshal storage block with %d items: %w", bh.itemsCount, err)
} }

View file

@ -233,7 +233,7 @@ func (ris *rawItemsShard) addItems(tb *Table, items [][]byte) [][]byte {
ris.mu.Lock() ris.mu.Lock()
ibs := ris.ibs ibs := ris.ibs
if len(ibs) == 0 { if len(ibs) == 0 {
ib := getInmemoryBlock() ib := &nmemoryBlock{}
ibs = append(ibs, ib) ibs = append(ibs, ib)
ris.ibs = ibs ris.ibs = ibs
} }
@ -249,12 +249,11 @@ func (ris *rawItemsShard) addItems(tb *Table, items [][]byte) [][]byte {
atomic.StoreUint64(&ris.lastFlushTime, fasttime.UnixTimestamp()) atomic.StoreUint64(&ris.lastFlushTime, fasttime.UnixTimestamp())
break break
} }
ib = getInmemoryBlock() ib = &nmemoryBlock{}
if ib.Add(item) { if ib.Add(item) {
ibs = append(ibs, ib) ibs = append(ibs, ib)
continue continue
} }
putInmemoryBlock(ib)
logger.Panicf("BUG: cannot insert too big item into an empty inmemoryBlock len(item)=%d; the caller should be responsible for avoiding too big items", len(item)) logger.Panicf("BUG: cannot insert too big item into an empty inmemoryBlock len(item)=%d; the caller should be responsible for avoiding too big items", len(item))
} }
ris.ibs = ibs ris.ibs = ibs
@ -882,7 +881,6 @@ func (tb *Table) createInmemoryPart(ibs []*inmemoryBlock) *partWrapper {
} }
bsr := getBlockStreamReader() bsr := getBlockStreamReader()
bsr.MustInitFromInmemoryBlock(ib) bsr.MustInitFromInmemoryBlock(ib)
putInmemoryBlock(ib)
bsrs = append(bsrs, bsr) bsrs = append(bsrs, bsr)
} }
if len(bsrs) == 0 { if len(bsrs) == 0 {