opentelemetry-collector/scraper/scraperhelper/controller_test.go

813 lines
23 KiB
Go

// Copyright The OpenTelemetry Authors
// SPDX-License-Identifier: Apache-2.0
package scraperhelper
import (
"context"
"errors"
"testing"
"time"
"github.com/stretchr/testify/assert"
"github.com/stretchr/testify/require"
"go.opentelemetry.io/otel/attribute"
"go.opentelemetry.io/otel/codes"
"go.opentelemetry.io/otel/sdk/metric/metricdata"
"go.opentelemetry.io/otel/sdk/metric/metricdata/metricdatatest"
sdktrace "go.opentelemetry.io/otel/sdk/trace"
"go.uber.org/multierr"
"go.opentelemetry.io/collector/component"
"go.opentelemetry.io/collector/component/componenttest"
"go.opentelemetry.io/collector/consumer/consumertest"
"go.opentelemetry.io/collector/pdata/plog"
"go.opentelemetry.io/collector/pdata/pmetric"
"go.opentelemetry.io/collector/receiver"
"go.opentelemetry.io/collector/receiver/receivertest"
"go.opentelemetry.io/collector/scraper"
"go.opentelemetry.io/collector/scraper/scrapererror"
"go.opentelemetry.io/collector/scraper/scraperhelper/internal/metadatatest"
)
type testInitialize struct {
ch chan bool
err error
}
func (ts *testInitialize) start(context.Context, component.Host) error {
ts.ch <- true
return ts.err
}
type testClose struct {
ch chan bool
err error
}
func (ts *testClose) shutdown(context.Context) error {
ts.ch <- true
return ts.err
}
type testScrape struct {
ch chan int
timesScrapeCalled int
err error
}
func (ts *testScrape) scrapeLogs(context.Context) (plog.Logs, error) {
ts.timesScrapeCalled++
ts.ch <- ts.timesScrapeCalled
if ts.err != nil {
return plog.Logs{}, ts.err
}
md := plog.NewLogs()
md.ResourceLogs().AppendEmpty().ScopeLogs().AppendEmpty().LogRecords().AppendEmpty().Body().SetStr("")
return md, nil
}
func (ts *testScrape) scrapeMetrics(context.Context) (pmetric.Metrics, error) {
ts.timesScrapeCalled++
ts.ch <- ts.timesScrapeCalled
if ts.err != nil {
return pmetric.Metrics{}, ts.err
}
md := pmetric.NewMetrics()
md.ResourceMetrics().AppendEmpty().ScopeMetrics().AppendEmpty().Metrics().AppendEmpty().SetEmptyGauge().DataPoints().AppendEmpty()
return md, nil
}
func newTestNoDelaySettings() *ControllerConfig {
return &ControllerConfig{
CollectionInterval: time.Second,
InitialDelay: 0,
}
}
type scraperTestCase struct {
name string
scrapers int
scraperControllerSettings *ControllerConfig
scrapeErr error
expectScraped bool
initialize bool
close bool
initializeErr error
closeErr error
}
func TestLogsScrapeController(t *testing.T) {
testCases := []scraperTestCase{
{
name: "NoScrapers",
},
{
name: "AddLogsScrapersWithCollectionInterval",
scrapers: 2,
expectScraped: true,
},
{
name: "AddLogsScrapers_ScrapeError",
scrapers: 2,
scrapeErr: errors.New("err1"),
},
{
name: "AddLogsScrapersWithInitializeAndClose",
scrapers: 2,
initialize: true,
expectScraped: true,
close: true,
},
{
name: "AddLogsScrapersWithInitializeAndCloseErrors",
scrapers: 2,
initialize: true,
close: true,
initializeErr: errors.New("err1"),
closeErr: errors.New("err2"),
},
}
for _, test := range testCases {
t.Run(test.name, func(t *testing.T) {
receiverID := component.MustNewID("receiver")
tel := componenttest.NewTelemetry()
t.Cleanup(func() { require.NoError(t, tel.Shutdown(context.Background())) })
set := tel.NewTelemetrySettings()
_, parentSpan := set.TracerProvider.Tracer("test").Start(context.Background(), t.Name())
defer parentSpan.End()
initializeChs := make([]chan bool, test.scrapers)
scrapeLogsChs := make([]chan int, test.scrapers)
closeChs := make([]chan bool, test.scrapers)
options := configureLogOptions(t, test, initializeChs, scrapeLogsChs, closeChs)
tickerCh := make(chan time.Time)
options = append(options, WithTickerChannel(tickerCh))
sink := new(consumertest.LogsSink)
cfg := newTestNoDelaySettings()
if test.scraperControllerSettings != nil {
cfg = test.scraperControllerSettings
}
mr, err := NewLogsController(cfg, receiver.Settings{ID: receiverID, TelemetrySettings: set, BuildInfo: component.NewDefaultBuildInfo()}, sink, options...)
require.NoError(t, err)
err = mr.Start(context.Background(), componenttest.NewNopHost())
expectedStartErr := getExpectedStartErr(test)
if expectedStartErr != nil {
assert.Equal(t, expectedStartErr, err)
} else if test.initialize {
assertChannelsCalled(t, initializeChs, "start was not called")
}
const iterations = 5
if test.expectScraped || test.scrapeErr != nil {
// validate that scrape is called at least N times for each configured scraper
for _, ch := range scrapeLogsChs {
<-ch
}
// Consume the initial scrapes on start
for i := 0; i < iterations; i++ {
tickerCh <- time.Now()
for _, ch := range scrapeLogsChs {
<-ch
}
}
// wait until all calls to scrape have completed
if test.scrapeErr == nil {
require.Eventually(t, func() bool {
return sink.LogRecordCount() == (1+iterations)*(test.scrapers)
}, time.Second, time.Millisecond)
}
if test.expectScraped {
assert.GreaterOrEqual(t, sink.LogRecordCount(), iterations)
}
spans := tel.SpanRecorder.Ended()
assertReceiverSpan(t, spans)
assertScraperSpan(t, test.scrapeErr, spans, "scraper/scraper/ScrapeLogs")
assertLogsScraperObsMetrics(t, tel, receiverID, component.MustNewID("scraper"), test.scrapeErr, sink)
}
err = mr.Shutdown(context.Background())
expectedShutdownErr := getExpectedShutdownErr(test)
if expectedShutdownErr != nil {
assert.EqualError(t, err, expectedShutdownErr.Error())
} else if test.close {
assertChannelsCalled(t, closeChs, "shutdown was not called")
}
})
}
}
func TestMetricsScrapeController(t *testing.T) {
testCases := []scraperTestCase{
{
name: "NoScrapers",
},
{
name: "AddMetricsScrapersWithCollectionInterval",
scrapers: 2,
expectScraped: true,
},
{
name: "AddMetricsScrapers_ScrapeError",
scrapers: 2,
scrapeErr: errors.New("err1"),
},
{
name: "AddMetricsScrapersWithInitializeAndClose",
scrapers: 2,
initialize: true,
expectScraped: true,
close: true,
},
{
name: "AddMetricsScrapersWithInitializeAndCloseErrors",
scrapers: 2,
initialize: true,
close: true,
initializeErr: errors.New("err1"),
closeErr: errors.New("err2"),
},
}
for _, test := range testCases {
t.Run(test.name, func(t *testing.T) {
receiverID := component.MustNewID("receiver")
tel := componenttest.NewTelemetry()
t.Cleanup(func() { require.NoError(t, tel.Shutdown(context.Background())) })
set := tel.NewTelemetrySettings()
_, parentSpan := set.TracerProvider.Tracer("test").Start(context.Background(), t.Name())
defer parentSpan.End()
initializeChs := make([]chan bool, test.scrapers)
scrapeMetricsChs := make([]chan int, test.scrapers)
closeChs := make([]chan bool, test.scrapers)
options := configureMetricOptions(t, test, initializeChs, scrapeMetricsChs, closeChs)
tickerCh := make(chan time.Time)
options = append(options, WithTickerChannel(tickerCh))
sink := new(consumertest.MetricsSink)
cfg := newTestNoDelaySettings()
if test.scraperControllerSettings != nil {
cfg = test.scraperControllerSettings
}
mr, err := NewMetricsController(cfg, receiver.Settings{ID: receiverID, TelemetrySettings: set, BuildInfo: component.NewDefaultBuildInfo()}, sink, options...)
require.NoError(t, err)
err = mr.Start(context.Background(), componenttest.NewNopHost())
expectedStartErr := getExpectedStartErr(test)
if expectedStartErr != nil {
assert.Equal(t, expectedStartErr, err)
} else if test.initialize {
assertChannelsCalled(t, initializeChs, "start was not called")
}
const iterations = 5
if test.expectScraped || test.scrapeErr != nil {
// validate that scrape is called at least N times for each configured scraper
for _, ch := range scrapeMetricsChs {
<-ch
}
// Consume the initial scrapes on start
for i := 0; i < iterations; i++ {
tickerCh <- time.Now()
for _, ch := range scrapeMetricsChs {
<-ch
}
}
// wait until all calls to scrape have completed
if test.scrapeErr == nil {
require.Eventually(t, func() bool {
return sink.DataPointCount() == (1+iterations)*(test.scrapers)
}, time.Second, time.Millisecond)
}
if test.expectScraped {
assert.GreaterOrEqual(t, sink.DataPointCount(), iterations)
}
spans := tel.SpanRecorder.Ended()
assertReceiverSpan(t, spans)
assertScraperSpan(t, test.scrapeErr, spans, "scraper/scraper/ScrapeMetrics")
assertMetricsScraperObsMetrics(t, tel, receiverID, component.MustNewID("scraper"), test.scrapeErr, sink)
}
err = mr.Shutdown(context.Background())
expectedShutdownErr := getExpectedShutdownErr(test)
if expectedShutdownErr != nil {
assert.EqualError(t, err, expectedShutdownErr.Error())
} else if test.close {
assertChannelsCalled(t, closeChs, "shutdown was not called")
}
})
}
}
func configureLogOptions(t *testing.T, test scraperTestCase, initializeChs []chan bool, scrapeLogsChs []chan int, closeChs []chan bool) []ControllerOption {
var logsOptions []ControllerOption
for i := 0; i < test.scrapers; i++ {
var scraperOptions []scraper.Option
if test.initialize {
initializeChs[i] = make(chan bool, 1)
ti := &testInitialize{ch: initializeChs[i], err: test.initializeErr}
scraperOptions = append(scraperOptions, scraper.WithStart(ti.start))
}
if test.close {
closeChs[i] = make(chan bool, 1)
tc := &testClose{ch: closeChs[i], err: test.closeErr}
scraperOptions = append(scraperOptions, scraper.WithShutdown(tc.shutdown))
}
scrapeLogsChs[i] = make(chan int)
ts := &testScrape{ch: scrapeLogsChs[i], err: test.scrapeErr}
scp, err := scraper.NewLogs(ts.scrapeLogs, scraperOptions...)
require.NoError(t, err)
logsOptions = append(logsOptions, addLogsScraper(component.MustNewType("scraper"), scp))
}
return logsOptions
}
func configureMetricOptions(t *testing.T, test scraperTestCase, initializeChs []chan bool, scrapeMetricsChs []chan int, closeChs []chan bool) []ControllerOption {
var metricOptions []ControllerOption
for i := 0; i < test.scrapers; i++ {
var scraperOptions []scraper.Option
if test.initialize {
initializeChs[i] = make(chan bool, 1)
ti := &testInitialize{ch: initializeChs[i], err: test.initializeErr}
scraperOptions = append(scraperOptions, scraper.WithStart(ti.start))
}
if test.close {
closeChs[i] = make(chan bool, 1)
tc := &testClose{ch: closeChs[i], err: test.closeErr}
scraperOptions = append(scraperOptions, scraper.WithShutdown(tc.shutdown))
}
scrapeMetricsChs[i] = make(chan int)
ts := &testScrape{ch: scrapeMetricsChs[i], err: test.scrapeErr}
scp, err := scraper.NewMetrics(ts.scrapeMetrics, scraperOptions...)
require.NoError(t, err)
metricOptions = append(metricOptions, AddScraper(component.MustNewType("scraper"), scp))
}
return metricOptions
}
func getExpectedStartErr(test scraperTestCase) error {
return test.initializeErr
}
func getExpectedShutdownErr(test scraperTestCase) error {
var errs error
if test.closeErr != nil {
for i := 0; i < test.scrapers; i++ {
errs = multierr.Append(errs, test.closeErr)
}
}
return errs
}
func assertChannelsCalled(t *testing.T, chs []chan bool, message string) {
for _, ic := range chs {
assertChannelCalled(t, ic, message)
}
}
func assertChannelCalled(t *testing.T, ch chan bool, message string) {
select {
case <-ch:
default:
assert.Fail(t, message)
}
}
func assertReceiverSpan(t *testing.T, spans []sdktrace.ReadOnlySpan) {
receiverSpan := false
for _, span := range spans {
if span.Name() == "receiver/receiver/MetricsReceived" {
receiverSpan = true
break
}
}
assert.True(t, receiverSpan)
}
func assertScraperSpan(t *testing.T, expectedErr error, spans []sdktrace.ReadOnlySpan, expectedSpanName string) {
expectedStatusCode := codes.Unset
expectedStatusMessage := ""
if expectedErr != nil {
expectedStatusCode = codes.Error
expectedStatusMessage = expectedErr.Error()
}
scraperSpan := false
for _, span := range spans {
if span.Name() == expectedSpanName {
scraperSpan = true
assert.Equal(t, expectedStatusCode, span.Status().Code)
assert.Equal(t, expectedStatusMessage, span.Status().Description)
break
}
}
assert.True(t, scraperSpan)
}
func assertLogsScraperObsMetrics(t *testing.T, tel *componenttest.Telemetry, receiver component.ID, scraper component.ID, expectedErr error, sink *consumertest.LogsSink) {
logRecordCounts := 0
for _, md := range sink.AllLogs() {
logRecordCounts += md.LogRecordCount()
}
expectedScraped := int64(sink.LogRecordCount())
expectedErrored := int64(0)
if expectedErr != nil {
var partialError scrapererror.PartialScrapeError
if errors.As(expectedErr, &partialError) {
expectedErrored = int64(partialError.Failed)
} else {
expectedScraped = int64(0)
expectedErrored = int64(sink.LogRecordCount())
}
}
metadatatest.AssertEqualScraperScrapedLogRecords(t, tel,
[]metricdata.DataPoint[int64]{
{
Attributes: attribute.NewSet(
attribute.String(receiverKey, receiver.String()),
attribute.String(scraperKey, scraper.String())),
Value: expectedScraped,
},
}, metricdatatest.IgnoreTimestamp(), metricdatatest.IgnoreExemplars())
metadatatest.AssertEqualScraperErroredLogRecords(t, tel,
[]metricdata.DataPoint[int64]{
{
Attributes: attribute.NewSet(
attribute.String(receiverKey, receiver.String()),
attribute.String(scraperKey, scraper.String())),
Value: expectedErrored,
},
}, metricdatatest.IgnoreTimestamp(), metricdatatest.IgnoreExemplars())
}
func assertMetricsScraperObsMetrics(t *testing.T, tel *componenttest.Telemetry, receiver component.ID, scraper component.ID, expectedErr error, sink *consumertest.MetricsSink) {
dataPointCounts := 0
for _, md := range sink.AllMetrics() {
dataPointCounts += md.DataPointCount()
}
expectedScraped := int64(sink.DataPointCount())
expectedErrored := int64(0)
if expectedErr != nil {
var partialError scrapererror.PartialScrapeError
if errors.As(expectedErr, &partialError) {
expectedErrored = int64(partialError.Failed)
} else {
expectedScraped = int64(0)
expectedErrored = int64(sink.DataPointCount())
}
}
metadatatest.AssertEqualScraperScrapedMetricPoints(t, tel,
[]metricdata.DataPoint[int64]{
{
Attributes: attribute.NewSet(
attribute.String(receiverKey, receiver.String()),
attribute.String(scraperKey, scraper.String())),
Value: expectedScraped,
},
}, metricdatatest.IgnoreTimestamp(), metricdatatest.IgnoreExemplars())
metadatatest.AssertEqualScraperErroredMetricPoints(t, tel,
[]metricdata.DataPoint[int64]{
{
Attributes: attribute.NewSet(
attribute.String(receiverKey, receiver.String()),
attribute.String(scraperKey, scraper.String())),
Value: expectedErrored,
},
}, metricdatatest.IgnoreTimestamp(), metricdatatest.IgnoreExemplars())
}
func TestSingleLogsScraperPerInterval(t *testing.T) {
scrapeCh := make(chan int, 10)
ts := &testScrape{ch: scrapeCh}
cfg := newTestNoDelaySettings()
tickerCh := make(chan time.Time)
scp, err := scraper.NewLogs(ts.scrapeLogs)
require.NoError(t, err)
recv, err := NewLogsController(
cfg,
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.LogsSink),
addLogsScraper(component.MustNewType("scraper"), scp),
WithTickerChannel(tickerCh),
)
require.NoError(t, err)
require.NoError(t, recv.Start(context.Background(), componenttest.NewNopHost()))
defer func() { require.NoError(t, recv.Shutdown(context.Background())) }()
tickerCh <- time.Now()
assert.Eventually(
t,
func() bool {
return <-scrapeCh == 2
},
300*time.Millisecond,
100*time.Millisecond,
"Make sure the scraper channel is called twice",
)
select {
case <-scrapeCh:
assert.Fail(t, "Scrape was called more than twice")
case <-time.After(100 * time.Millisecond):
return
}
}
func TestSingleMetricsScraperPerInterval(t *testing.T) {
scrapeCh := make(chan int, 10)
ts := &testScrape{ch: scrapeCh}
cfg := newTestNoDelaySettings()
tickerCh := make(chan time.Time)
scp, err := scraper.NewMetrics(ts.scrapeMetrics)
require.NoError(t, err)
recv, err := NewMetricsController(
cfg,
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.MetricsSink),
AddScraper(component.MustNewType("scraper"), scp),
WithTickerChannel(tickerCh),
)
require.NoError(t, err)
require.NoError(t, recv.Start(context.Background(), componenttest.NewNopHost()))
defer func() { require.NoError(t, recv.Shutdown(context.Background())) }()
tickerCh <- time.Now()
assert.Eventually(
t,
func() bool {
return <-scrapeCh == 2
},
300*time.Millisecond,
100*time.Millisecond,
"Make sure the scraper channel is called twice",
)
select {
case <-scrapeCh:
assert.Fail(t, "Scrape was called more than twice")
case <-time.After(100 * time.Millisecond):
return
}
}
func TestLogsScraperControllerStartsOnInit(t *testing.T) {
t.Parallel()
ts := &testScrape{
ch: make(chan int, 1),
}
scp, err := scraper.NewLogs(ts.scrapeLogs)
require.NoError(t, err, "Must not error when creating scraper")
r, err := NewLogsController(
&ControllerConfig{
CollectionInterval: time.Hour,
InitialDelay: 0,
},
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.LogsSink),
addLogsScraper(component.MustNewType("scraper"), scp),
)
require.NoError(t, err, "Must not error when creating scrape controller")
assert.NoError(t, r.Start(context.Background(), componenttest.NewNopHost()), "Must not error on start")
<-time.After(500 * time.Nanosecond)
require.NoError(t, r.Shutdown(context.Background()), "Must not have errored on shutdown")
assert.Equal(t, 1, ts.timesScrapeCalled, "Must have been called as soon as the controller started")
}
func TestMetricsScraperControllerStartsOnInit(t *testing.T) {
t.Parallel()
ts := &testScrape{
ch: make(chan int, 1),
}
scp, err := scraper.NewMetrics(ts.scrapeMetrics)
require.NoError(t, err, "Must not error when creating scraper")
r, err := NewMetricsController(
&ControllerConfig{
CollectionInterval: time.Hour,
InitialDelay: 0,
},
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.MetricsSink),
AddScraper(component.MustNewType("scraper"), scp),
)
require.NoError(t, err, "Must not error when creating scrape controller")
assert.NoError(t, r.Start(context.Background(), componenttest.NewNopHost()), "Must not error on start")
<-time.After(500 * time.Nanosecond)
require.NoError(t, r.Shutdown(context.Background()), "Must not have errored on shutdown")
assert.Equal(t, 1, ts.timesScrapeCalled, "Must have been called as soon as the controller started")
}
func TestLogsScraperControllerInitialDelay(t *testing.T) {
if testing.Short() {
t.Skip("This requires real time to pass, skipping")
return
}
t.Parallel()
var (
elapsed = make(chan time.Time, 1)
cfg = ControllerConfig{
CollectionInterval: time.Second,
InitialDelay: 300 * time.Millisecond,
}
)
scp, err := scraper.NewLogs(func(context.Context) (plog.Logs, error) {
elapsed <- time.Now()
return plog.NewLogs(), nil
})
require.NoError(t, err, "Must not error when creating scraper")
r, err := NewLogsController(
&cfg,
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.LogsSink),
addLogsScraper(component.MustNewType("scraper"), scp),
)
require.NoError(t, err, "Must not error when creating receiver")
t0 := time.Now()
require.NoError(t, r.Start(context.Background(), componenttest.NewNopHost()), "Must not error when starting")
t1 := <-elapsed
assert.GreaterOrEqual(t, t1.Sub(t0), 300*time.Millisecond, "Must have had 300ms pass as defined by initial delay")
assert.NoError(t, r.Shutdown(context.Background()), "Must not error closing down")
}
func TestMetricsScraperControllerInitialDelay(t *testing.T) {
if testing.Short() {
t.Skip("This requires real time to pass, skipping")
return
}
t.Parallel()
var (
elapsed = make(chan time.Time, 1)
cfg = ControllerConfig{
CollectionInterval: time.Second,
InitialDelay: 300 * time.Millisecond,
}
)
scp, err := scraper.NewMetrics(func(context.Context) (pmetric.Metrics, error) {
elapsed <- time.Now()
return pmetric.NewMetrics(), nil
})
require.NoError(t, err, "Must not error when creating scraper")
r, err := NewMetricsController(
&cfg,
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.MetricsSink),
AddScraper(component.MustNewType("scraper"), scp),
)
require.NoError(t, err, "Must not error when creating receiver")
t0 := time.Now()
require.NoError(t, r.Start(context.Background(), componenttest.NewNopHost()), "Must not error when starting")
t1 := <-elapsed
assert.GreaterOrEqual(t, t1.Sub(t0), 300*time.Millisecond, "Must have had 300ms pass as defined by initial delay")
assert.NoError(t, r.Shutdown(context.Background()), "Must not error closing down")
}
func TestLogsScraperShutdownBeforeScrapeCanStart(t *testing.T) {
cfg := ControllerConfig{
CollectionInterval: time.Second,
InitialDelay: 5 * time.Second,
}
scp, err := scraper.NewLogs(func(context.Context) (plog.Logs, error) {
// make the scraper wait for long enough it would disrupt a shutdown.
time.Sleep(30 * time.Second)
return plog.NewLogs(), nil
})
require.NoError(t, err, "Must not error when creating scraper")
r, err := NewLogsController(
&cfg,
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.LogsSink),
addLogsScraper(component.MustNewType("scraper"), scp),
)
require.NoError(t, err, "Must not error when creating receiver")
require.NoError(t, r.Start(context.Background(), componenttest.NewNopHost()))
shutdown := make(chan struct{}, 1)
go func() {
assert.NoError(t, r.Shutdown(context.Background()))
close(shutdown)
}()
timer := time.NewTicker(10 * time.Second)
select {
case <-timer.C:
require.Fail(t, "shutdown should not wait for scraping")
case <-shutdown:
}
}
func TestMetricsScraperShutdownBeforeScrapeCanStart(t *testing.T) {
cfg := ControllerConfig{
CollectionInterval: time.Second,
InitialDelay: 5 * time.Second,
}
scp, err := scraper.NewMetrics(func(context.Context) (pmetric.Metrics, error) {
// make the scraper wait for long enough it would disrupt a shutdown.
time.Sleep(30 * time.Second)
return pmetric.NewMetrics(), nil
})
require.NoError(t, err, "Must not error when creating scraper")
r, err := NewMetricsController(
&cfg,
receivertest.NewNopSettings(receivertest.NopType),
new(consumertest.MetricsSink),
AddScraper(component.MustNewType("scraper"), scp),
)
require.NoError(t, err, "Must not error when creating receiver")
require.NoError(t, r.Start(context.Background(), componenttest.NewNopHost()))
shutdown := make(chan struct{}, 1)
go func() {
assert.NoError(t, r.Shutdown(context.Background()))
close(shutdown)
}()
timer := time.NewTicker(10 * time.Second)
select {
case <-timer.C:
require.Fail(t, "shutdown should not wait for scraping")
case <-shutdown:
}
}
func addLogsScraper(t component.Type, sc scraper.Logs) ControllerOption {
f := scraper.NewFactory(t, nil,
scraper.WithLogs(func(context.Context, scraper.Settings, component.Config) (scraper.Logs, error) {
return sc, nil
}, component.StabilityLevelAlpha))
return AddFactoryWithConfig(f, nil)
}