603 lines
19 KiB
Go
603 lines
19 KiB
Go
// Copyright The OpenTelemetry Authors
|
|
// SPDX-License-Identifier: Apache-2.0
|
|
|
|
package exporterhelper
|
|
|
|
import (
|
|
"context"
|
|
"testing"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"go.opentelemetry.io/collector/exporter/exporterbatcher"
|
|
"go.opentelemetry.io/collector/exporter/exporterhelper/internal/sizer"
|
|
"go.opentelemetry.io/collector/pdata/pmetric"
|
|
"go.opentelemetry.io/collector/pdata/testdata"
|
|
)
|
|
|
|
var metricsBytesSizer = &sizer.MetricsBytesSizer{}
|
|
|
|
func TestMergeMetrics(t *testing.T) {
|
|
mr1 := newMetricsRequest(testdata.GenerateMetrics(2), nil)
|
|
mr2 := newMetricsRequest(testdata.GenerateMetrics(3), nil)
|
|
res, err := mr1.MergeSplit(context.Background(), exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems}, mr2)
|
|
require.NoError(t, err)
|
|
// Every metric has 2 data points.
|
|
assert.Equal(t, 2*5, res[0].ItemsCount())
|
|
}
|
|
|
|
func TestMergeMetricsInvalidInput(t *testing.T) {
|
|
mr1 := newTracesRequest(testdata.GenerateTraces(2), nil)
|
|
mr2 := newMetricsRequest(testdata.GenerateMetrics(3), nil)
|
|
_, err := mr1.MergeSplit(context.Background(), exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems}, mr2)
|
|
require.Error(t, err)
|
|
}
|
|
|
|
func TestMergeSplitMetrics(t *testing.T) {
|
|
s := sizer.MetricsCountSizer{}
|
|
tests := []struct {
|
|
name string
|
|
cfg exporterbatcher.SizeConfig
|
|
mr1 Request
|
|
mr2 Request
|
|
expected []Request
|
|
}{
|
|
{
|
|
name: "both_requests_empty",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 10},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
expected: []Request{newMetricsRequest(pmetric.NewMetrics(), nil)},
|
|
},
|
|
{
|
|
name: "first_request_empty",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 10},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(5), nil),
|
|
expected: []Request{newMetricsRequest(testdata.GenerateMetrics(5), nil)},
|
|
},
|
|
{
|
|
name: "first_empty_second_nil",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 10},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: nil,
|
|
expected: []Request{newMetricsRequest(pmetric.NewMetrics(), nil)},
|
|
},
|
|
{
|
|
name: "merge_only",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 60},
|
|
mr1: newMetricsRequest(testdata.GenerateMetrics(10), nil),
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(14), nil),
|
|
expected: []Request{newMetricsRequest(func() pmetric.Metrics {
|
|
metrics := testdata.GenerateMetrics(10)
|
|
testdata.GenerateMetrics(14).ResourceMetrics().MoveAndAppendTo(metrics.ResourceMetrics())
|
|
return metrics
|
|
}(), nil)},
|
|
},
|
|
{
|
|
name: "split_only",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 14},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(15), nil), // 15 metrics, 30 data points
|
|
expected: []Request{
|
|
newMetricsRequest(testdata.GenerateMetrics(7), nil), // 7 metrics, 14 data points
|
|
newMetricsRequest(testdata.GenerateMetrics(7), nil), // 7 metrics, 14 data points
|
|
newMetricsRequest(testdata.GenerateMetrics(1), nil), // 1 metric, 2 data points
|
|
},
|
|
},
|
|
{
|
|
name: "split_and_merge",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 28},
|
|
mr1: newMetricsRequest(testdata.GenerateMetrics(7), nil), // 7 metrics, 14 data points
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(25), nil), // 25 metrics, 50 data points
|
|
expected: []Request{
|
|
newMetricsRequest(func() pmetric.Metrics {
|
|
metrics := testdata.GenerateMetrics(7)
|
|
testdata.GenerateMetrics(7).ResourceMetrics().MoveAndAppendTo(metrics.ResourceMetrics())
|
|
return metrics
|
|
}(), nil),
|
|
newMetricsRequest(testdata.GenerateMetrics(14), nil), // 14 metrics, 28 data points
|
|
newMetricsRequest(testdata.GenerateMetrics(4), nil), // 4 metrics, 8 data points
|
|
},
|
|
},
|
|
{
|
|
name: "scope_metrics_split",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 8},
|
|
mr1: newMetricsRequest(func() pmetric.Metrics {
|
|
md := testdata.GenerateMetrics(4)
|
|
extraScopeMetrics := md.ResourceMetrics().At(0).ScopeMetrics().AppendEmpty()
|
|
testdata.GenerateMetrics(4).ResourceMetrics().At(0).ScopeMetrics().At(0).MoveTo(extraScopeMetrics)
|
|
extraScopeMetrics.Scope().SetName("extra scope")
|
|
return md
|
|
}(), nil),
|
|
mr2: nil,
|
|
expected: []Request{
|
|
newMetricsRequest(testdata.GenerateMetrics(4), nil),
|
|
newMetricsRequest(func() pmetric.Metrics {
|
|
md := testdata.GenerateMetrics(4)
|
|
md.ResourceMetrics().At(0).ScopeMetrics().At(0).Scope().SetName("extra scope")
|
|
return md
|
|
}(), nil),
|
|
},
|
|
},
|
|
}
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
res, err := tt.mr1.MergeSplit(context.Background(), tt.cfg, tt.mr2)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, len(tt.expected), len(res))
|
|
for i := range res {
|
|
expected := tt.expected[i].(*metricsRequest)
|
|
actual := res[i].(*metricsRequest)
|
|
assert.Equal(t, expected.size(&s), actual.size(&s))
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestMergeSplitMetricsInputNotModifiedIfErrorReturned(t *testing.T) {
|
|
r1 := newMetricsRequest(testdata.GenerateMetrics(18), nil) // 18 metrics, 36 data points
|
|
r2 := newLogsRequest(testdata.GenerateLogs(3), nil)
|
|
_, err := r1.MergeSplit(context.Background(), exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 10}, r2)
|
|
require.Error(t, err)
|
|
assert.Equal(t, 36, r1.ItemsCount())
|
|
}
|
|
|
|
func TestMergeSplitMetricsInvalidInput(t *testing.T) {
|
|
r1 := newTracesRequest(testdata.GenerateTraces(2), nil)
|
|
r2 := newMetricsRequest(testdata.GenerateMetrics(3), nil)
|
|
_, err := r1.MergeSplit(context.Background(), exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 10}, r2)
|
|
require.Error(t, err)
|
|
}
|
|
|
|
func TestExtractMetrics(t *testing.T) {
|
|
for i := 0; i < 20; i++ {
|
|
md := testdata.GenerateMetrics(10)
|
|
extractedMetrics, _ := extractMetrics(md, i, &sizer.MetricsCountSizer{})
|
|
assert.Equal(t, i, extractedMetrics.DataPointCount())
|
|
assert.Equal(t, 20-i, md.DataPointCount())
|
|
}
|
|
}
|
|
|
|
func TestExtractMetricsInvalidMetric(t *testing.T) {
|
|
md := testdata.GenerateMetricsMetricTypeInvalid()
|
|
extractedMetrics, _ := extractMetrics(md, 10, &sizer.MetricsCountSizer{})
|
|
assert.Equal(t, testdata.GenerateMetricsMetricTypeInvalid(), extractedMetrics)
|
|
assert.Equal(t, 0, md.ResourceMetrics().Len())
|
|
}
|
|
|
|
func TestMergeSplitManySmallMetrics(t *testing.T) {
|
|
// All requests merge into a single batch.
|
|
cfg := exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 20000}
|
|
merged := []Request{newMetricsRequest(testdata.GenerateMetrics(1), nil)}
|
|
for j := 0; j < 1000; j++ {
|
|
lr2 := newMetricsRequest(testdata.GenerateMetrics(10), nil)
|
|
res, _ := merged[len(merged)-1].MergeSplit(context.Background(), cfg, lr2)
|
|
merged = append(merged[0:len(merged)-1], res...)
|
|
}
|
|
assert.Len(t, merged, 2)
|
|
}
|
|
|
|
func BenchmarkSplittingBasedOnItemCountManySmallMetrics(b *testing.B) {
|
|
// All requests merge into a single batch.
|
|
cfg := exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 20020}
|
|
b.ReportAllocs()
|
|
for i := 0; i < b.N; i++ {
|
|
merged := []Request{newMetricsRequest(testdata.GenerateMetrics(10), nil)}
|
|
for j := 0; j < 1000; j++ {
|
|
lr2 := newMetricsRequest(testdata.GenerateMetrics(10), nil)
|
|
res, _ := merged[len(merged)-1].MergeSplit(context.Background(), cfg, lr2)
|
|
merged = append(merged[0:len(merged)-1], res...)
|
|
}
|
|
assert.Len(b, merged, 1)
|
|
}
|
|
}
|
|
|
|
func BenchmarkSplittingBasedOnItemCountManyMetricsSlightlyAboveLimit(b *testing.B) {
|
|
// Every incoming request results in a split.
|
|
cfg := exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 20000}
|
|
b.ReportAllocs()
|
|
for i := 0; i < b.N; i++ {
|
|
merged := []Request{newMetricsRequest(testdata.GenerateMetrics(0), nil)}
|
|
for j := 0; j < 10; j++ {
|
|
lr2 := newMetricsRequest(testdata.GenerateMetrics(10001), nil)
|
|
res, _ := merged[len(merged)-1].MergeSplit(context.Background(), cfg, lr2)
|
|
merged = append(merged[0:len(merged)-1], res...)
|
|
}
|
|
assert.Len(b, merged, 11)
|
|
}
|
|
}
|
|
|
|
func BenchmarkSplittingBasedOnItemCountHugeMetrics(b *testing.B) {
|
|
// One request splits into many batches.
|
|
cfg := exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeItems, MaxSize: 20000}
|
|
b.ReportAllocs()
|
|
for i := 0; i < b.N; i++ {
|
|
merged := []Request{newMetricsRequest(testdata.GenerateMetrics(0), nil)}
|
|
lr2 := newMetricsRequest(testdata.GenerateMetrics(100000), nil)
|
|
res, _ := merged[len(merged)-1].MergeSplit(context.Background(), cfg, lr2)
|
|
merged = append(merged[0:len(merged)-1], res...)
|
|
assert.Len(b, merged, 10)
|
|
}
|
|
}
|
|
|
|
func TestMergeSplitMetricsBasedOnByteSize(t *testing.T) {
|
|
s := sizer.MetricsBytesSizer{}
|
|
tests := []struct {
|
|
name string
|
|
cfg exporterbatcher.SizeConfig
|
|
mr1 Request
|
|
mr2 Request
|
|
expectedSizes []int
|
|
}{
|
|
{
|
|
name: "both_requests_empty",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeBytes, MaxSize: metricsBytesSizer.MetricsSize(testdata.GenerateMetrics(10))},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
expectedSizes: []int{0},
|
|
},
|
|
{
|
|
name: "first_request_empty",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeBytes, MaxSize: metricsBytesSizer.MetricsSize(testdata.GenerateMetrics(10))},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(5), nil),
|
|
expectedSizes: []int{1035},
|
|
},
|
|
{
|
|
name: "first_empty_second_nil",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeBytes, MaxSize: metricsBytesSizer.MetricsSize(testdata.GenerateMetrics(10))},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: nil,
|
|
expectedSizes: []int{0},
|
|
},
|
|
{
|
|
name: "merge_only",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeBytes, MaxSize: metricsBytesSizer.MetricsSize(testdata.GenerateMetrics(11))},
|
|
mr1: newMetricsRequest(testdata.GenerateMetrics(4), nil),
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(6), nil),
|
|
expectedSizes: []int{2102},
|
|
},
|
|
{
|
|
name: "split_only",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeBytes, MaxSize: s.MetricsSize(testdata.GenerateMetrics(4))},
|
|
mr1: newMetricsRequest(pmetric.NewMetrics(), nil),
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(10), nil),
|
|
expectedSizes: []int{706, 504, 625, 378},
|
|
},
|
|
{
|
|
name: "merge_and_split",
|
|
cfg: exporterbatcher.SizeConfig{
|
|
Sizer: exporterbatcher.SizerTypeBytes,
|
|
MaxSize: metricsBytesSizer.MetricsSize(testdata.GenerateMetrics(10))/2 + metricsBytesSizer.MetricsSize(testdata.GenerateMetrics(11))/2,
|
|
},
|
|
mr1: newMetricsRequest(testdata.GenerateMetrics(8), nil),
|
|
mr2: newMetricsRequest(testdata.GenerateMetrics(20), nil),
|
|
expectedSizes: []int{2107, 2022, 1954, 290},
|
|
},
|
|
{
|
|
name: "scope_metrics_split",
|
|
cfg: exporterbatcher.SizeConfig{Sizer: exporterbatcher.SizerTypeBytes, MaxSize: metricsBytesSizer.MetricsSize(testdata.GenerateMetrics(4))},
|
|
mr1: newMetricsRequest(func() pmetric.Metrics {
|
|
md := testdata.GenerateMetrics(4)
|
|
extraScopeMetrics := md.ResourceMetrics().At(0).ScopeMetrics().AppendEmpty()
|
|
testdata.GenerateMetrics(4).ResourceMetrics().At(0).ScopeMetrics().At(0).MoveTo(extraScopeMetrics)
|
|
extraScopeMetrics.Scope().SetName("extra scope")
|
|
return md
|
|
}(), nil),
|
|
mr2: nil,
|
|
expectedSizes: []int{706, 700, 85},
|
|
},
|
|
}
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
res, err := tt.mr1.MergeSplit(context.Background(), tt.cfg, tt.mr2)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, len(tt.expectedSizes), len(res))
|
|
for i := range res {
|
|
assert.Equal(t, tt.expectedSizes[i], res[i].(*metricsRequest).size(&s))
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestMetricsRequest_MergeSplit_UnknownSizerType(t *testing.T) {
|
|
// Create a logs request
|
|
req := newMetricsRequest(pmetric.NewMetrics(), nil)
|
|
|
|
// Create config with invalid sizer type by using zero value
|
|
cfg := exporterbatcher.SizeConfig{
|
|
Sizer: exporterbatcher.SizerType{}, // Empty struct will have empty string as val
|
|
}
|
|
|
|
// Call MergeSplit with invalid sizer
|
|
result, err := req.MergeSplit(context.Background(), cfg, nil)
|
|
|
|
// Verify results
|
|
assert.Nil(t, result)
|
|
assert.EqualError(t, err, "unknown sizer type")
|
|
}
|
|
|
|
func TestExtractGaugeDataPoints(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
capacity int
|
|
numDataPoints int
|
|
expectedPoints int
|
|
}{
|
|
{
|
|
name: "extract_all_points",
|
|
capacity: 100,
|
|
numDataPoints: 2,
|
|
expectedPoints: 2,
|
|
},
|
|
{
|
|
name: "extract_partial_points",
|
|
capacity: 1,
|
|
numDataPoints: 2,
|
|
expectedPoints: 1,
|
|
},
|
|
{
|
|
name: "no_capacity",
|
|
capacity: 0,
|
|
numDataPoints: 2,
|
|
expectedPoints: 0,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
srcMetric := pmetric.NewMetric()
|
|
gauge := srcMetric.SetEmptyGauge()
|
|
for i := 0; i < tt.numDataPoints; i++ {
|
|
dp := gauge.DataPoints().AppendEmpty()
|
|
dp.SetIntValue(int64(i))
|
|
}
|
|
|
|
sz := &mockMetricsSizer{dpSize: 1}
|
|
|
|
destMetric, removedSize := extractGaugeDataPoints(gauge, tt.capacity, sz)
|
|
|
|
assert.Equal(t, tt.expectedPoints, destMetric.Gauge().DataPoints().Len())
|
|
if tt.expectedPoints > 0 {
|
|
assert.Equal(t, tt.expectedPoints, removedSize)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestExtractSumDataPoints(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
capacity int
|
|
numDataPoints int
|
|
expectedPoints int
|
|
}{
|
|
{
|
|
name: "extract_all_points",
|
|
capacity: 100,
|
|
numDataPoints: 2,
|
|
expectedPoints: 2,
|
|
},
|
|
{
|
|
name: "extract_partial_points",
|
|
capacity: 1,
|
|
numDataPoints: 2,
|
|
expectedPoints: 1,
|
|
},
|
|
{
|
|
name: "no_capacity",
|
|
capacity: 0,
|
|
numDataPoints: 2,
|
|
expectedPoints: 0,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
srcMetric := pmetric.NewMetric()
|
|
sum := srcMetric.SetEmptySum()
|
|
for i := 0; i < tt.numDataPoints; i++ {
|
|
dp := sum.DataPoints().AppendEmpty()
|
|
dp.SetIntValue(int64(i))
|
|
}
|
|
|
|
sz := &mockMetricsSizer{dpSize: 1}
|
|
|
|
destMetric, removedSize := extractSumDataPoints(sum, tt.capacity, sz)
|
|
|
|
assert.Equal(t, tt.expectedPoints, destMetric.Sum().DataPoints().Len())
|
|
if tt.expectedPoints > 0 {
|
|
assert.Equal(t, tt.expectedPoints, removedSize)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestExtractHistogramDataPoints(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
capacity int
|
|
numDataPoints int
|
|
expectedPoints int
|
|
}{
|
|
{
|
|
name: "extract_all_points",
|
|
capacity: 100,
|
|
numDataPoints: 2,
|
|
expectedPoints: 2,
|
|
},
|
|
{
|
|
name: "extract_partial_points",
|
|
capacity: 1,
|
|
numDataPoints: 2,
|
|
expectedPoints: 1,
|
|
},
|
|
{
|
|
name: "no_capacity",
|
|
capacity: 0,
|
|
numDataPoints: 2,
|
|
expectedPoints: 0,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
srcMetric := pmetric.NewMetric()
|
|
histogram := srcMetric.SetEmptyHistogram()
|
|
|
|
for i := 0; i < tt.numDataPoints; i++ {
|
|
dp := histogram.DataPoints().AppendEmpty()
|
|
dp.SetCount(uint64(i)) //nolint:gosec // disable G115
|
|
}
|
|
|
|
sz := &mockMetricsSizer{dpSize: 1}
|
|
|
|
destMetric, removedSize := extractHistogramDataPoints(histogram, tt.capacity, sz)
|
|
|
|
assert.Equal(t, tt.expectedPoints, destMetric.Histogram().DataPoints().Len())
|
|
if tt.expectedPoints > 0 {
|
|
assert.Equal(t, tt.expectedPoints, removedSize)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestExtractExponentialHistogramDataPoints(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
capacity int
|
|
numDataPoints int
|
|
expectedPoints int
|
|
}{
|
|
{
|
|
name: "extract_all_points",
|
|
capacity: 100,
|
|
numDataPoints: 2,
|
|
expectedPoints: 2,
|
|
},
|
|
{
|
|
name: "extract_partial_points",
|
|
capacity: 1,
|
|
numDataPoints: 2,
|
|
expectedPoints: 1,
|
|
},
|
|
{
|
|
name: "no_capacity",
|
|
capacity: 0,
|
|
numDataPoints: 2,
|
|
expectedPoints: 0,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
srcMetric := pmetric.NewMetric()
|
|
expHistogram := srcMetric.SetEmptyExponentialHistogram()
|
|
for i := 0; i < tt.numDataPoints; i++ {
|
|
dp := expHistogram.DataPoints().AppendEmpty()
|
|
dp.SetCount(uint64(i)) //nolint:gosec // disable G115
|
|
}
|
|
|
|
sz := &mockMetricsSizer{dpSize: 1}
|
|
|
|
destMetric, removedSize := extractExponentialHistogramDataPoints(expHistogram, tt.capacity, sz)
|
|
|
|
assert.Equal(t, tt.expectedPoints, destMetric.ExponentialHistogram().DataPoints().Len())
|
|
if tt.expectedPoints > 0 {
|
|
assert.Equal(t, tt.expectedPoints, removedSize)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestExtractSummaryDataPoints(t *testing.T) {
|
|
tests := []struct {
|
|
name string
|
|
capacity int
|
|
numDataPoints int
|
|
expectedPoints int
|
|
}{
|
|
{
|
|
name: "extract_all_points",
|
|
capacity: 100,
|
|
numDataPoints: 2,
|
|
expectedPoints: 2,
|
|
},
|
|
{
|
|
name: "extract_partial_points",
|
|
capacity: 1,
|
|
numDataPoints: 2,
|
|
expectedPoints: 1,
|
|
},
|
|
{
|
|
name: "no_capacity",
|
|
capacity: 0,
|
|
numDataPoints: 2,
|
|
expectedPoints: 0,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
srcMetric := pmetric.NewMetric()
|
|
summary := srcMetric.SetEmptySummary()
|
|
for i := 0; i < tt.numDataPoints; i++ {
|
|
dp := summary.DataPoints().AppendEmpty()
|
|
dp.SetCount(uint64(i)) //nolint:gosec // disable G115
|
|
}
|
|
|
|
sz := &mockMetricsSizer{dpSize: 1}
|
|
|
|
destMetric, removedSize := extractSummaryDataPoints(summary, tt.capacity, sz)
|
|
|
|
assert.Equal(t, tt.expectedPoints, destMetric.Summary().DataPoints().Len())
|
|
if tt.expectedPoints > 0 {
|
|
assert.Equal(t, tt.expectedPoints, removedSize)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
// mockMetricsSizer implements sizer.MetricsSizer interface for testing
|
|
type mockMetricsSizer struct {
|
|
dpSize int
|
|
}
|
|
|
|
func (m *mockMetricsSizer) MetricsSize(_ pmetric.Metrics) int {
|
|
return 0
|
|
}
|
|
|
|
func (m *mockMetricsSizer) MetricSize(_ pmetric.Metric) int {
|
|
return 0
|
|
}
|
|
|
|
func (m *mockMetricsSizer) NumberDataPointSize(_ pmetric.NumberDataPoint) int {
|
|
return m.dpSize
|
|
}
|
|
|
|
func (m *mockMetricsSizer) HistogramDataPointSize(_ pmetric.HistogramDataPoint) int {
|
|
return m.dpSize
|
|
}
|
|
|
|
func (m *mockMetricsSizer) ExponentialHistogramDataPointSize(_ pmetric.ExponentialHistogramDataPoint) int {
|
|
return m.dpSize
|
|
}
|
|
|
|
func (m *mockMetricsSizer) SummaryDataPointSize(_ pmetric.SummaryDataPoint) int {
|
|
return m.dpSize
|
|
}
|
|
|
|
func (m *mockMetricsSizer) ResourceMetricsSize(_ pmetric.ResourceMetrics) int {
|
|
return 0
|
|
}
|
|
|
|
func (m *mockMetricsSizer) ScopeMetricsSize(_ pmetric.ScopeMetrics) int {
|
|
return 0
|
|
}
|
|
|
|
func (m *mockMetricsSizer) DeltaSize(size int) int {
|
|
return size
|
|
}
|