From 95b8cf76f851b58586d1d0257ca63ce9d72b43b1 Mon Sep 17 00:00:00 2001 From: yumeiyin <155420652+yumeiyin@users.noreply.github.com> Date: Wed, 29 May 2024 20:08:38 +0800 Subject: [PATCH] chore: remove redundant words (#6348) (cherry picked from commit 9289c7512dab757c1401a81955ed99237b2537e7) --- app/vmagent/remotewrite/pendingseries_test.go | 2 +- app/vmctl/opentsdb/opentsdb.go | 2 +- docs/stream-aggregation.md | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/app/vmagent/remotewrite/pendingseries_test.go b/app/vmagent/remotewrite/pendingseries_test.go index 487fdab491..6e949a94f4 100644 --- a/app/vmagent/remotewrite/pendingseries_test.go +++ b/app/vmagent/remotewrite/pendingseries_test.go @@ -34,7 +34,7 @@ func testPushWriteRequest(t *testing.T, rowsCount, expectedBlockLenProm, expecte return true } if !tryPushWriteRequest(wr, pushBlock, isVMRemoteWrite) { - t.Fatalf("cannot push data to to remote storage") + t.Fatalf("cannot push data to remote storage") } if math.Abs(float64(pushBlockLen-expectedBlockLen)/float64(expectedBlockLen)*100) > tolerancePrc { t.Fatalf("unexpected block len for rowsCount=%d, isVMRemoteWrite=%v; got %d bytes; expecting %d bytes +- %.0f%%", diff --git a/app/vmctl/opentsdb/opentsdb.go b/app/vmctl/opentsdb/opentsdb.go index 4efc18c6c2..b75e20b5a2 100644 --- a/app/vmctl/opentsdb/opentsdb.go +++ b/app/vmctl/opentsdb/opentsdb.go @@ -26,7 +26,7 @@ type Retention struct { FirstOrder string SecondOrder string AggTime string - // The actual ranges will will attempt to query (as offsets from now) + // The actual ranges will attempt to query (as offsets from now) QueryRanges []TimeRange } diff --git a/docs/stream-aggregation.md b/docs/stream-aggregation.md index f94b42e496..daaa59e4bf 100644 --- a/docs/stream-aggregation.md +++ b/docs/stream-aggregation.md @@ -1073,7 +1073,7 @@ This may lead to the following issues: since they ignore the first sample in a new time series. - Unexpected spikes for [total](#total) and [increase](#increase) outputs, since they assume that new time series start from 0. -These issues can be be fixed in the following ways: +These issues can be fixed in the following ways: - By increasing the `interval` option at [stream aggregation config](#stream-aggregation-config), so it covers the expected delays in data ingestion pipelines.