mirror of
https://github.com/VictoriaMetrics/VictoriaMetrics.git
synced 2025-03-21 15:45:01 +00:00
all: fix misspellings
This commit is contained in:
parent
d6523ffe90
commit
54fb8b21f9
9 changed files with 12 additions and 9 deletions
|
@ -19,7 +19,7 @@ var (
|
||||||
maxPointsPerTimeseries = flag.Int("search.maxPointsPerTimeseries", 10e3, "The maximum points per a single timeseries returned from the search")
|
maxPointsPerTimeseries = flag.Int("search.maxPointsPerTimeseries", 10e3, "The maximum points per a single timeseries returned from the search")
|
||||||
)
|
)
|
||||||
|
|
||||||
// The minumum number of points per timeseries for enabling time rounding.
|
// The minimum number of points per timeseries for enabling time rounding.
|
||||||
// This improves cache hit ratio for frequently requested queries over
|
// This improves cache hit ratio for frequently requested queries over
|
||||||
// big time ranges.
|
// big time ranges.
|
||||||
const minTimeseriesPointsForTimeRounding = 50
|
const minTimeseriesPointsForTimeRounding = 50
|
||||||
|
|
|
@ -44,7 +44,7 @@ func (bb *ByteBuffer) ReadAt(p []byte, offset int64) {
|
||||||
logger.Panicf("BUG: too big offset=%d; cannot exceed len(bb.B)=%d", offset, len(bb.B))
|
logger.Panicf("BUG: too big offset=%d; cannot exceed len(bb.B)=%d", offset, len(bb.B))
|
||||||
}
|
}
|
||||||
if n := copy(p, bb.B[offset:]); n < len(p) {
|
if n := copy(p, bb.B[offset:]); n < len(p) {
|
||||||
logger.Panicf("BUG: EOF occured after reading %d bytes out of %d bytes at offset %d", n, len(p), offset)
|
logger.Panicf("BUG: EOF occurred after reading %d bytes out of %d bytes at offset %d", n, len(p), offset)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -117,7 +117,7 @@ func marshalInt64Array(dst []byte, a []int64, precisionBits uint8) (result []byt
|
||||||
|
|
||||||
bb := bbPool.Get()
|
bb := bbPool.Get()
|
||||||
if isGauge(a) {
|
if isGauge(a) {
|
||||||
// Guage values are better compressed with delta encoding.
|
// Gauge values are better compressed with delta encoding.
|
||||||
mt = MarshalTypeZSTDNearestDelta
|
mt = MarshalTypeZSTDNearestDelta
|
||||||
pb := precisionBits
|
pb := precisionBits
|
||||||
if pb < 6 {
|
if pb < 6 {
|
||||||
|
|
|
@ -208,7 +208,7 @@ func (tb *Table) MustClose() {
|
||||||
logger.Infof("%d inmemory parts have been flushed to files in %s on %q", len(pws), time.Since(startTime), tb.path)
|
logger.Infof("%d inmemory parts have been flushed to files in %s on %q", len(pws), time.Since(startTime), tb.path)
|
||||||
|
|
||||||
// Remove references to parts from the tb, so they may be eventually closed
|
// Remove references to parts from the tb, so they may be eventually closed
|
||||||
// after all the seraches are done.
|
// after all the searches are done.
|
||||||
tb.partsLock.Lock()
|
tb.partsLock.Lock()
|
||||||
parts := tb.parts
|
parts := tb.parts
|
||||||
tb.parts = nil
|
tb.parts = nil
|
||||||
|
|
|
@ -33,6 +33,9 @@ func benchmarkTableSearch(b *testing.B, itemsCount int) {
|
||||||
// Force finishing pending merges
|
// Force finishing pending merges
|
||||||
tb.MustClose()
|
tb.MustClose()
|
||||||
tb, err = OpenTable(path)
|
tb, err = OpenTable(path)
|
||||||
|
if err != nil {
|
||||||
|
b.Fatalf("unexpected error when re-opening table %q: %s", path, err)
|
||||||
|
}
|
||||||
defer tb.MustClose()
|
defer tb.MustClose()
|
||||||
|
|
||||||
keys := make([][]byte, len(items))
|
keys := make([][]byte, len(items))
|
||||||
|
|
|
@ -945,7 +945,7 @@ func (is *indexSearch) getTSIDByMetricName(dst *TSID, metricName []byte) error {
|
||||||
if len(dmis) > 0 {
|
if len(dmis) > 0 {
|
||||||
// Verify whether the dst is marked as deleted.
|
// Verify whether the dst is marked as deleted.
|
||||||
if _, deleted := dmis[dst.MetricID]; deleted {
|
if _, deleted := dmis[dst.MetricID]; deleted {
|
||||||
// The dst is deleted. Continue seraching.
|
// The dst is deleted. Continue searching.
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -599,7 +599,7 @@ func (pt *partition) MustClose() {
|
||||||
logger.Infof("%d inmemory parts have been flushed to files in %s on %q", len(pws), time.Since(startTime), pt.smallPartsPath)
|
logger.Infof("%d inmemory parts have been flushed to files in %s on %q", len(pws), time.Since(startTime), pt.smallPartsPath)
|
||||||
|
|
||||||
// Remove references to smallParts from the pt, so they may be eventually closed
|
// Remove references to smallParts from the pt, so they may be eventually closed
|
||||||
// after all the seraches are done.
|
// after all the searches are done.
|
||||||
pt.partsLock.Lock()
|
pt.partsLock.Lock()
|
||||||
smallParts := pt.smallParts
|
smallParts := pt.smallParts
|
||||||
pt.smallParts = nil
|
pt.smallParts = nil
|
||||||
|
|
|
@ -19,7 +19,7 @@ type rawRow struct {
|
||||||
// Value is time series value for the given timestamp.
|
// Value is time series value for the given timestamp.
|
||||||
Value float64
|
Value float64
|
||||||
|
|
||||||
// PrecisionBits is the number of the siginificant bits in the Value
|
// PrecisionBits is the number of the significant bits in the Value
|
||||||
// to store. Possible values are [1..64].
|
// to store. Possible values are [1..64].
|
||||||
// 1 means max. 50% error, 2 - 25%, 3 - 12.5%, 64 means no error, i.e.
|
// 1 means max. 50% error, 2 - 25%, 3 - 12.5%, 64 means no error, i.e.
|
||||||
// Value stored without information loss.
|
// Value stored without information loss.
|
||||||
|
|
|
@ -30,7 +30,7 @@ type Storage struct {
|
||||||
cachePath string
|
cachePath string
|
||||||
retentionMonths int
|
retentionMonths int
|
||||||
|
|
||||||
// lock file for excluse access to the storage on the given path.
|
// lock file for exclusive access to the storage on the given path.
|
||||||
flockF *os.File
|
flockF *os.File
|
||||||
|
|
||||||
idbCurr atomic.Value
|
idbCurr atomic.Value
|
||||||
|
@ -656,7 +656,7 @@ func (s *Storage) add(rows []rawRow, mrs []MetricRow, precisionBits uint8) ([]ra
|
||||||
errors = s.updateDateMetricIDCache(rows, errors)
|
errors = s.updateDateMetricIDCache(rows, errors)
|
||||||
if len(errors) > 0 {
|
if len(errors) > 0 {
|
||||||
// Return only the first error, since it has no sense in returning all errors.
|
// Return only the first error, since it has no sense in returning all errors.
|
||||||
return rows, fmt.Errorf("errors occured during rows addition: %s", errors[0])
|
return rows, fmt.Errorf("errors occurred during rows addition: %s", errors[0])
|
||||||
}
|
}
|
||||||
return rows, nil
|
return rows, nil
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue