Change LogRecordsCount to LogRecordCount everywhere (#3551)

Signed-off-by: Bogdan Drutu <bogdandrutu@gmail.com>
This commit is contained in:
Bogdan Drutu 2021-07-01 17:07:14 -07:00 committed by GitHub
parent 9a7bccea3e
commit 32c2d0f131
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
6 changed files with 21 additions and 21 deletions

View File

@ -122,9 +122,9 @@ func (sme *MetricsSink) Reset() {
// stores all logs and allows querying them for testing.
type LogsSink struct {
nonMutatingConsumer
mu sync.Mutex
logs []pdata.Logs
logRecordsCount int
mu sync.Mutex
logs []pdata.Logs
logRecordCount int
}
var _ consumer.Logs = (*LogsSink)(nil)
@ -135,7 +135,7 @@ func (sle *LogsSink) ConsumeLogs(_ context.Context, ld pdata.Logs) error {
defer sle.mu.Unlock()
sle.logs = append(sle.logs, ld)
sle.logRecordsCount += ld.LogRecordCount()
sle.logRecordCount += ld.LogRecordCount()
return nil
}
@ -150,11 +150,11 @@ func (sle *LogsSink) AllLogs() []pdata.Logs {
return copyLogs
}
// LogRecordsCount returns the number of log records stored by this sink since last Reset.
func (sle *LogsSink) LogRecordsCount() int {
// LogRecordCount returns the number of log records stored by this sink since last Reset.
func (sle *LogsSink) LogRecordCount() int {
sle.mu.Lock()
defer sle.mu.Unlock()
return sle.logRecordsCount
return sle.logRecordCount
}
// Reset deletes any stored data.
@ -163,5 +163,5 @@ func (sle *LogsSink) Reset() {
defer sle.mu.Unlock()
sle.logs = nil
sle.logRecordsCount = 0
sle.logRecordCount = 0
}

View File

@ -64,8 +64,8 @@ func TestLogsSink(t *testing.T) {
want = append(want, md)
}
assert.Equal(t, want, sink.AllLogs())
assert.Equal(t, len(want), sink.LogRecordsCount())
assert.Equal(t, len(want), sink.LogRecordCount())
sink.Reset()
assert.Equal(t, 0, len(sink.AllLogs()))
assert.Equal(t, 0, sink.LogRecordsCount())
assert.Equal(t, 0, sink.LogRecordCount())
}

View File

@ -283,7 +283,7 @@ func TestLogsRoundTrip(t *testing.T) {
md := testdata.GenerateLogsOneLogRecord()
assert.NoError(t, exp.ConsumeLogs(context.Background(), md))
require.Eventually(t, func() bool {
return sink.LogRecordsCount() > 0
return sink.LogRecordCount() > 0
}, 1*time.Second, 10*time.Millisecond)
allLogs := sink.AllLogs()
require.Len(t, allLogs, 1)

View File

@ -606,7 +606,7 @@ func TestBatchLogProcessor_ReceivingData(t *testing.T) {
require.NoError(t, batcher.Shutdown(context.Background()))
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordsCount())
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordCount())
receivedMds := sink.AllLogs()
logsReceivedByName := logsReceivedByName(receivedMds)
for requestNum := 0; requestNum < requestCount; requestNum++ {
@ -656,7 +656,7 @@ func TestBatchLogProcessor_BatchSize(t *testing.T) {
expectedBatchesNum := requestCount * logsPerRequest / int(cfg.SendBatchSize)
expectedBatchingFactor := int(cfg.SendBatchSize) / logsPerRequest
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordsCount())
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordCount())
receivedMds := sink.AllLogs()
require.Equal(t, expectedBatchesNum, len(receivedMds))
for _, ld := range receivedMds {
@ -671,7 +671,7 @@ func TestBatchLogProcessor_BatchSize(t *testing.T) {
assert.Equal(t, 1, len(viewData))
distData := viewData[0].Data.(*view.DistributionData)
assert.Equal(t, int64(expectedBatchesNum), distData.Count)
assert.Equal(t, sink.LogRecordsCount(), int(distData.Sum()))
assert.Equal(t, sink.LogRecordCount(), int(distData.Sum()))
assert.Equal(t, cfg.SendBatchSize, uint32(distData.Min))
assert.Equal(t, cfg.SendBatchSize, uint32(distData.Max))
@ -706,7 +706,7 @@ func TestBatchLogsProcessor_Timeout(t *testing.T) {
// Wait for at least one batch to be sent.
for {
if sink.LogRecordsCount() != 0 {
if sink.LogRecordCount() != 0 {
break
}
<-time.After(cfg.Timeout)
@ -721,7 +721,7 @@ func TestBatchLogsProcessor_Timeout(t *testing.T) {
expectedBatchesNum := 1
expectedBatchingFactor := 5
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordsCount())
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordCount())
receivedMds := sink.AllLogs()
require.Equal(t, expectedBatchesNum, len(receivedMds))
for _, ld := range receivedMds {
@ -754,7 +754,7 @@ func TestBatchLogProcessor_Shutdown(t *testing.T) {
require.NoError(t, batcher.Shutdown(context.Background()))
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordsCount())
require.Equal(t, requestCount*logsPerRequest, sink.LogRecordCount())
require.Equal(t, 1, len(sink.AllLogs()))
}

View File

@ -138,7 +138,7 @@ func TestLogsProcessorCloningMultiplexing(t *testing.T) {
for i, p := range processors {
m := p.(*consumertest.LogsSink)
assert.Equal(t, wantMetricsCount, m.LogRecordsCount())
assert.Equal(t, wantMetricsCount, m.LogRecordCount())
metricOrig := ld.ResourceLogs().At(0).InstrumentationLibraryLogs().At(0).Logs().At(0)
allLogs := m.AllLogs()
metricClone := allLogs[0].ResourceLogs().At(0).InstrumentationLibraryLogs().At(0).Logs().At(0)

View File

@ -161,7 +161,7 @@ func TestLogsProcessorMultiplexing(t *testing.T) {
for _, p := range processors {
m := p.(*consumertest.LogsSink)
assert.Equal(t, wantMetricsCount, m.LogRecordsCount())
assert.Equal(t, wantMetricsCount, m.LogRecordCount())
assert.EqualValues(t, ld, m.AllLogs()[0])
}
}
@ -184,6 +184,6 @@ func TestLogsProcessorWhenOneErrors(t *testing.T) {
assert.Error(t, lfc.ConsumeLogs(context.Background(), ld))
}
assert.Equal(t, wantMetricsCount, processors[0].(*consumertest.LogsSink).LogRecordsCount())
assert.Equal(t, wantMetricsCount, processors[2].(*consumertest.LogsSink).LogRecordsCount())
assert.Equal(t, wantMetricsCount, processors[0].(*consumertest.LogsSink).LogRecordCount())
assert.Equal(t, wantMetricsCount, processors[2].(*consumertest.LogsSink).LogRecordCount())
}