Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix for index block appearing as a flushed index block, now explicit check has on disk segment #3622

Merged
Merged
Show file tree
Hide file tree
Changes from 5 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
291 changes: 291 additions & 0 deletions src/dbnode/integration/index_active_block_rotate_test.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,291 @@
// +build integration
//
// Copyright (c) 2021 Uber Technologies, Inc.
//
// Permission is hereby granted, free of charge, to any person obtaining a copy
// of this software and associated documentation files (the "Software"), to deal
// in the Software without restriction, including without limitation the rights
// to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
// copies of the Software, and to permit persons to whom the Software is
// furnished to do so, subject to the following conditions:
//
// The above copyright notice and this permission notice shall be included in
// all copies or substantial portions of the Software.
//
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
// THE SOFTWARE.

package integration

import (
"fmt"
"sync"
"testing"
"time"

"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/require"
"github.com/uber-go/tally"
"go.uber.org/atomic"
"go.uber.org/zap"
"go.uber.org/zap/zapcore"
"go.uber.org/zap/zaptest/observer"

"github.com/m3db/m3/src/dbnode/namespace"
"github.com/m3db/m3/src/dbnode/retention"
"github.com/m3db/m3/src/dbnode/storage/index/compaction"
xclock "github.com/m3db/m3/src/x/clock"
"github.com/m3db/m3/src/x/ident"
xtime "github.com/m3db/m3/src/x/time"
)

func TestIndexActiveBlockRotate(t *testing.T) {
var (
testNsID = ident.StringID("testns")
numWrites = 50
numTags = 10
blockSize = 2 * time.Hour
indexBlockSize = blockSize
bufferPast = 10 * time.Minute
rOpts = retention.NewOptions().
SetRetentionPeriod(blockSize).
SetBlockSize(blockSize).
SetBufferPast(bufferPast)

idxOpts = namespace.NewIndexOptions().SetEnabled(true).SetBlockSize(indexBlockSize)
nsOpts = namespace.NewOptions().
SetRetentionOptions(rOpts).
SetIndexOptions(idxOpts).
SetColdWritesEnabled(true)

defaultTimeout = time.Minute
// verifyTimeout = time.Minute
)
ns, err := namespace.NewMetadata(testNsID, nsOpts)
require.NoError(t, err)

// Set time to next warm flushable block transition
// (i.e. align by block + bufferPast - time.Second)
currTime := time.Now().UTC()
progressTime := false
progressTimeDelta := time.Duration(0)
lockTime := sync.RWMutex{}
setTime := func(t time.Time) {
lockTime.Lock()
defer lockTime.Unlock()
progressTime = false
currTime = t.UTC()
}
setProgressTime := func() {
lockTime.Lock()
defer lockTime.Unlock()
progressTime = true
actualNow := time.Now().UTC()
progressTimeDelta = currTime.Sub(actualNow)
}
nowFn := func() time.Time {
lockTime.RLock()
at := currTime
progress := progressTime
progressDelta := progressTimeDelta
lockTime.RUnlock()
if progress {
return time.Now().UTC().Add(progressDelta)
}
return at
}

testOpts := NewTestOptions(t).
SetNamespaces([]namespace.Metadata{ns}).
SetWriteNewSeriesAsync(true).
SetNowFn(nowFn)

testSetup, err := NewTestSetup(t, testOpts, nil)
require.NoError(t, err)
defer testSetup.Close()

// Set foreground compaction planner options to force index compaction.
minCompactSize := 10
foregroundCompactionOpts := compaction.DefaultOptions
foregroundCompactionOpts.Levels = []compaction.Level{
{
MinSizeInclusive: 0,
MaxSizeExclusive: int64(minCompactSize),
},
}
indexOpts := testSetup.StorageOpts().IndexOptions().
SetForegroundCompactionPlannerOptions(foregroundCompactionOpts)
testSetup.SetStorageOpts(testSetup.StorageOpts().SetIndexOptions(indexOpts))

// Configure log capture
log := testSetup.StorageOpts().InstrumentOptions().Logger()
captureCore, logs := observer.New(zapcore.ErrorLevel)
zapOpt := zap.WrapCore(func(existingCore zapcore.Core) zapcore.Core {
return zapcore.NewTee(existingCore, captureCore)
})
log = log.WithOptions(zapOpt)

// Wire up logger.
instrumentOpts := testSetup.StorageOpts().InstrumentOptions().
SetLogger(log)
testSetup.SetStorageOpts(testSetup.StorageOpts().SetInstrumentOptions(instrumentOpts))
scope := testSetup.Scope()

// Start the server.
require.NoError(t, testSetup.StartServer())

// Stop the server.
defer func() {
require.NoError(t, testSetup.StopServer())
log.Debug("server is now down")
}()

// Write test data.
session, err := testSetup.M3DBClient().DefaultSession()
require.NoError(t, err)

var (
metricGCSeries = "index.block.active-block.gc-series+namespace=" + testNsID.String()
metricFlushIndex = "database.flushIndex.success+namespace=" + testNsID.String()
)
prevWarmFlushes := counterValue(t, scope, metricFlushIndex)
prevNumGCSeries := 0
numGCSeries := counterValue(t, scope, metricGCSeries)
require.Equal(t, 0, numGCSeries)

prevLog := log
for i := 0; i < 4; i++ {
log = prevLog.With(zap.Int("checkIteration", i))

// Progress to next time just before a flush and freeze (using setTime).
prevTime := nowFn()
newTime := prevTime.
Truncate(indexBlockSize).
Add(2 * indexBlockSize)
setTime(newTime)
log.Info("progressing time to before next block edge",
zap.Stringer("prevTime", prevTime),
zap.Stringer("newTime", newTime))

start := time.Now()
log.Info("writing test data")

t0 := xtime.ToUnixNano(newTime.Add(-1 * (bufferPast / 2)))
t1 := xtime.ToUnixNano(newTime)
writesPeriodIter := GenerateTestIndexWrite(i, numWrites, numTags, t0, t1)
writesPeriodIter.Write(t, testNsID, session)
log.Info("test data written", zap.Duration("took", time.Since(start)))

log.Info("waiting till data is indexed")
indexed := xclock.WaitUntil(func() bool {
indexedPeriod := writesPeriodIter.NumIndexed(t, testNsID, session)
return indexedPeriod == len(writesPeriodIter)
}, 15*time.Second)
require.True(t, indexed,
fmt.Sprintf("unexpected data indexed: actual=%d, expected=%d",
writesPeriodIter.NumIndexedWithOptions(t, testNsID, session, NumIndexedOptions{Logger: log}),
len(writesPeriodIter)))
log.Info("verified data is indexed", zap.Duration("took", time.Since(start)))

newTime = prevTime.
Truncate(indexBlockSize).
Add(2 * indexBlockSize).
Add(bufferPast).
Add(-100 * time.Millisecond)
setTime(newTime)
log.Info("progressing time to before next flush",
zap.Stringer("prevTime", prevTime),
zap.Stringer("newTime", newTime))

log.Info("waiting till warm flush occurs")

// Resume time progressing by wall clock.
setProgressTime()

// Start checks to ensure metrics are visible the whole time.
checkFailed := atomic.NewUint64(0)
checkIndexable := func() {
numGCSeriesBefore := counterValue(t, scope, metricGCSeries)
indexedPeriod := writesPeriodIter.NumIndexed(t, testNsID, session)
numGCSeriesAfter := counterValue(t, scope, metricGCSeries)
if len(writesPeriodIter) != indexedPeriod {
assert.Equal(t, len(writesPeriodIter), indexedPeriod,
fmt.Sprintf("some metrics not indexed/visible: actual=%d, expected=%d, numGCBefore=%d, numGCAfter=%d",
writesPeriodIter.NumIndexedWithOptions(t, testNsID, session, NumIndexedOptions{Logger: log}),
len(writesPeriodIter),
numGCSeriesBefore,
numGCSeriesAfter))
checkFailed.Inc()
}
}

ticker := time.NewTicker(10 * time.Millisecond)
stopTickCh := make(chan struct{})
closedTickCh := make(chan struct{})
go func() {
defer func() {
ticker.Stop()
close(closedTickCh)
}()

for {
select {
case <-ticker.C:
checkIndexable()
case <-stopTickCh:
return
}
}
}()

start = time.Now()
warmFlushed := xclock.WaitUntil(func() bool {
return counterValue(t, scope, metricFlushIndex)-prevWarmFlushes > 0
}, defaultTimeout)
counter := counterValue(t, scope, metricFlushIndex)
require.True(t, warmFlushed,
fmt.Sprintf("warm flush stats: current=%d, previous=%d", counter, prevWarmFlushes))
log.Info("verified data has been warm flushed", zap.Duration("took", time.Since(start)))
prevWarmFlushes = counter

start = time.Now()
log.Info("waiting for GC of series")

expectedNumGCSeries := prevNumGCSeries + numWrites - minCompactSize
gcSeries := xclock.WaitUntil(func() bool {
numGCSeries := counterValue(t, scope, metricGCSeries)
return numGCSeries >= expectedNumGCSeries
}, defaultTimeout)
numGCSeries := counterValue(t, scope, metricGCSeries)
require.True(t, gcSeries,
fmt.Sprintf("unexpected num gc series: actual=%d, expected=%d",
numGCSeries, expectedNumGCSeries))
require.True(t, numGCSeries >= expectedNumGCSeries)
log.Info("verified series have been GC'd", zap.Duration("took", time.Since(start)))
prevNumGCSeries = numGCSeries

require.Equal(t, 0, logs.Len(), "errors found in logs during flush/indexing")

// Keep running indexable check for a few seconds, then progress next iter.
time.Sleep(5 * time.Second)
close(stopTickCh)
<-closedTickCh

// Ensure check did not fail.
require.True(t, checkFailed.Load() == 0,
fmt.Sprintf("check indexable errors: %d", checkFailed.Load()))
}

log.Info("checks passed")
}

func counterValue(t *testing.T, r tally.TestScope, key string) int {
v, ok := r.Snapshot().Counters()[key]
require.True(t, ok)
return int(v.Value())
}
44 changes: 42 additions & 2 deletions src/dbnode/integration/index_helpers.go
Original file line number Diff line number Diff line change
Expand Up @@ -29,14 +29,16 @@ import (
"testing"
"time"

"github.com/stretchr/testify/require"
"go.uber.org/zap"

"github.com/m3db/m3/src/dbnode/client"
"github.com/m3db/m3/src/dbnode/encoding"
"github.com/m3db/m3/src/dbnode/storage/index"
"github.com/m3db/m3/src/m3ninx/idx"
"github.com/m3db/m3/src/query/storage/m3/consolidators"
"github.com/m3db/m3/src/x/ident"
xtime "github.com/m3db/m3/src/x/time"

"github.com/stretchr/testify/require"
)

// TestIndexWrites holds index writes for testing.
Expand Down Expand Up @@ -161,6 +163,21 @@ func (w TestIndexWrites) Write(t *testing.T, ns ident.ID, s client.Session) {

// NumIndexed gets number of indexed series.
func (w TestIndexWrites) NumIndexed(t *testing.T, ns ident.ID, s client.Session) int {
return w.NumIndexedWithOptions(t, ns, s, NumIndexedOptions{})
}

// NumIndexedOptions is options when performing num indexed check.
type NumIndexedOptions struct {
Logger *zap.Logger
}

// NumIndexedWithOptions gets number of indexed series with a set of options.
func (w TestIndexWrites) NumIndexedWithOptions(
t *testing.T,
ns ident.ID,
s client.Session,
opts NumIndexedOptions,
) int {
numFound := 0
for i := 0; i < len(w); i++ {
wi := w[i]
Expand All @@ -173,19 +190,42 @@ func (w TestIndexWrites) NumIndexed(t *testing.T, ns ident.ID, s client.Session)
SeriesLimit: 10,
})
if err != nil {
if l := opts.Logger; l != nil {
l.Error("fetch tagged IDs error", zap.Error(err))
}
continue
}
if !iter.Next() {
if l := opts.Logger; l != nil {
l.Warn("missing result",
zap.String("queryID", wi.ID.String()),
zap.ByteString("queryTags", consolidators.MustIdentTagIteratorToTags(wi.Tags, nil).ID()))
}
continue
}
cuNs, cuID, cuTag := iter.Current()
if ns.String() != cuNs.String() {
if l := opts.Logger; l != nil {
l.Warn("namespace mismatch",
zap.String("queryNamespace", ns.String()),
zap.String("resultNamespace", cuNs.String()))
}
continue
}
if wi.ID.String() != cuID.String() {
if l := opts.Logger; l != nil {
l.Warn("id mismatch",
zap.String("queryID", wi.ID.String()),
zap.String("resultID", cuID.String()))
}
continue
}
if !ident.NewTagIterMatcher(wi.Tags).Matches(cuTag) {
if l := opts.Logger; l != nil {
l.Warn("tag mismatch",
zap.ByteString("queryTags", consolidators.MustIdentTagIteratorToTags(wi.Tags, nil).ID()),
zap.ByteString("resultTags", consolidators.MustIdentTagIteratorToTags(cuTag, nil).ID()))
}
continue
}
numFound++
Expand Down
Loading