rename DoubleSum -> Sum (#3583)

* rename DoubleSum -> Sum

* update changelog

* revert pict changes
This commit is contained in:
alrex 2021-07-12 15:21:51 -07:00 committed by GitHub
parent e69cebab8e
commit aa60edff16
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
48 changed files with 293 additions and 289 deletions

View File

@ -2,6 +2,10 @@
## Unreleased
## 🛑 Breaking changes 🛑
- Rename `pdata.DoubleSum` to `pdata.Sum` (#3583)
## 🧰 Bug fixes 🧰
- `scraperhelper`: Include the scraper name in log messages (#3487)

View File

@ -40,7 +40,7 @@ metrics:
description: Total CPU seconds broken down by different states.
unit: s
data:
type: double sum
type: sum
monotonic: true
aggregation: cumulative
labels: [freeFormLabel, freeFormLabelWithValue, enumLabel]
@ -53,7 +53,7 @@ metrics:
description: Total CPU seconds broken down by different states.
unit: s
data:
type: double sum
type: sum
monotonic: true
aggregation: cumulative
labels: [missing]
@ -97,7 +97,7 @@ func Test_loadMetadata(t *testing.T) {
"system.cpu.time": {
Description: "Total CPU seconds broken down by different states.",
Unit: "s",
Data: &doubleSum{
Data: &sum{
Aggregated: Aggregated{Aggregation: "cumulative"},
Mono: Mono{Monotonic: true},
},

View File

@ -31,7 +31,7 @@ metrics:
description: Total CPU seconds broken down by different states.
unit: s
data:
type: double sum
type: sum
aggregation: cumulative
labels: []
`

View File

@ -23,7 +23,7 @@ var (
_ MetricData = &intSum{}
_ MetricData = &intHistogram{}
_ MetricData = &doubleGauge{}
_ MetricData = &doubleSum{}
_ MetricData = &sum{}
_ MetricData = &histogram{}
)
@ -52,8 +52,8 @@ func (e *ymlMetricData) UnmarshalYAML(unmarshal func(interface{}) error) error {
md = &intHistogram{}
case "double gauge":
md = &doubleGauge{}
case "double sum":
md = &doubleSum{}
case "sum":
md = &sum{}
case "histogram":
md = &histogram{}
default:
@ -148,20 +148,20 @@ func (i intSum) HasAggregated() bool {
return true
}
type doubleSum struct {
type sum struct {
Aggregated `yaml:",inline"`
Mono `yaml:",inline"`
}
func (d doubleSum) Type() string {
return "DoubleSum"
func (d sum) Type() string {
return "Sum"
}
func (d doubleSum) HasMonotonic() bool {
func (d sum) HasMonotonic() bool {
return true
}
func (d doubleSum) HasAggregated() bool {
func (d sum) HasAggregated() bool {
return true
}

View File

@ -31,7 +31,7 @@ func TestMetricData(t *testing.T) {
{&intSum{}, "IntSum", true, true},
{&intHistogram{}, "IntHistogram", true, false},
{&doubleGauge{}, "DoubleGauge", false, false},
{&doubleSum{}, "DoubleSum", true, true},
{&sum{}, "Sum", true, true},
{&histogram{}, "Histogram", true, false},
} {
assert.Equal(t, arg.typ, arg.metricData.Type())

View File

@ -169,8 +169,8 @@ var intSum = &messageValueStruct{
}
var doubleSum = &messageValueStruct{
structName: "DoubleSum",
description: "// DoubleSum represents the type of a numeric double scalar metric that is calculated as a sum of all reported measurements over a time interval.",
structName: "Sum",
description: "// Sum represents the type of a numeric double scalar metric that is calculated as a sum of all reported measurements over a time interval.",
originFullName: "otlpmetrics.Sum",
fields: []baseField{
aggregationTemporalityField,

View File

@ -23,11 +23,11 @@ metrics:
unit:
# Required
data:
# Required: one of int gauge, int sum, int histogram, double gauge, double sum, or histogram.
# Required: one of int gauge, int sum, int histogram, double gauge, sum, or histogram.
type:
# Required for int sum and double sum.
# Required for int sum and sum.
monotonic: # true | false
# Required for int sum, int histogram, double sum, and histogram.
# Required for int sum, int histogram, sum, and histogram.
aggregation: # delta | cumulative
# Optional: array of labels that were defined in the labels section that are emitted by this metric.
labels:

View File

@ -77,8 +77,8 @@ func addLabelsToMetric(metric *pdata.Metric, labelMap pdata.StringMap) {
addLabelsToDoubleDataPoints(metric.DoubleGauge().DataPoints(), labelMap)
case pdata.MetricDataTypeIntSum:
addLabelsToIntDataPoints(metric.IntSum().DataPoints(), labelMap)
case pdata.MetricDataTypeDoubleSum:
addLabelsToDoubleDataPoints(metric.DoubleSum().DataPoints(), labelMap)
case pdata.MetricDataTypeSum:
addLabelsToDoubleDataPoints(metric.Sum().DataPoints(), labelMap)
case pdata.MetricDataTypeIntHistogram:
addLabelsToIntHistogramDataPoints(metric.IntHistogram().DataPoints(), labelMap)
case pdata.MetricDataTypeHistogram:

View File

@ -48,7 +48,7 @@ func TestConvertResourceToLabelsAllDataTypesEmptyDataPoint(t *testing.T) {
assert.Equal(t, 1, md.ResourceMetrics().At(0).Resource().Attributes().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(0).DoubleGauge().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(1).IntGauge().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(2).DoubleSum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(2).Sum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(3).IntSum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(4).Histogram().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(5).IntHistogram().DataPoints().At(0).LabelsMap().Len())
@ -59,7 +59,7 @@ func TestConvertResourceToLabelsAllDataTypesEmptyDataPoint(t *testing.T) {
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).Resource().Attributes().Len())
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(0).DoubleGauge().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(1).IntGauge().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(2).DoubleSum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(2).Sum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(3).IntSum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(4).Histogram().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 1, cloneMd.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(5).IntHistogram().DataPoints().At(0).LabelsMap().Len())
@ -67,7 +67,7 @@ func TestConvertResourceToLabelsAllDataTypesEmptyDataPoint(t *testing.T) {
assert.Equal(t, 1, md.ResourceMetrics().At(0).Resource().Attributes().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(0).DoubleGauge().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(1).IntGauge().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(2).DoubleSum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(2).Sum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(3).IntSum().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(4).Histogram().DataPoints().At(0).LabelsMap().Len())
assert.Equal(t, 0, md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(5).IntHistogram().DataPoints().At(0).LabelsMap().Len())

View File

@ -88,8 +88,8 @@ func (a *lastValueAccumulator) addMetric(metric pdata.Metric, il pdata.Instrumen
return a.accumulateIntSum(metric, il, now)
case pdata.MetricDataTypeDoubleGauge:
return a.accumulateDoubleGauge(metric, il, now)
case pdata.MetricDataTypeDoubleSum:
return a.accumulateDoubleSum(metric, il, now)
case pdata.MetricDataTypeSum:
return a.accumulateSum(metric, il, now)
case pdata.MetricDataTypeIntHistogram:
return a.accumulateIntHistogram(metric, il, now)
case pdata.MetricDataTypeHistogram:
@ -232,8 +232,8 @@ func (a *lastValueAccumulator) accumulateIntSum(metric pdata.Metric, il pdata.In
return
}
func (a *lastValueAccumulator) accumulateDoubleSum(metric pdata.Metric, il pdata.InstrumentationLibrary, now time.Time) (n int) {
doubleSum := metric.DoubleSum()
func (a *lastValueAccumulator) accumulateSum(metric pdata.Metric, il pdata.InstrumentationLibrary, now time.Time) (n int) {
doubleSum := metric.Sum()
// Drop metrics with non-cumulative aggregations
if doubleSum.AggregationTemporality() != pdata.AggregationTemporalityCumulative {
@ -249,24 +249,24 @@ func (a *lastValueAccumulator) accumulateDoubleSum(metric pdata.Metric, il pdata
v, ok := a.registeredMetrics.Load(signature)
if !ok {
m := createMetric(metric)
m.DoubleSum().SetIsMonotonic(metric.DoubleSum().IsMonotonic())
m.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
ip.CopyTo(m.DoubleSum().DataPoints().AppendEmpty())
m.Sum().SetIsMonotonic(metric.Sum().IsMonotonic())
m.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
ip.CopyTo(m.Sum().DataPoints().AppendEmpty())
a.registeredMetrics.Store(signature, &accumulatedValue{value: m, instrumentationLibrary: il, updated: now})
n++
continue
}
mv := v.(*accumulatedValue)
if ip.Timestamp().AsTime().Before(mv.value.DoubleSum().DataPoints().At(0).Timestamp().AsTime()) {
if ip.Timestamp().AsTime().Before(mv.value.Sum().DataPoints().At(0).Timestamp().AsTime()) {
// only keep datapoint with latest timestamp
continue
}
m := createMetric(metric)
m.DoubleSum().SetIsMonotonic(metric.DoubleSum().IsMonotonic())
m.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
ip.CopyTo(m.DoubleSum().DataPoints().AppendEmpty())
m.Sum().SetIsMonotonic(metric.Sum().IsMonotonic())
m.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
ip.CopyTo(m.Sum().DataPoints().AppendEmpty())
a.registeredMetrics.Store(signature, &accumulatedValue{value: m, instrumentationLibrary: il, updated: now})
n++
}

View File

@ -52,12 +52,12 @@ func TestAccumulateDeltaAggregation(t *testing.T) {
},
},
{
name: "DoubleSum",
name: "Sum",
fillMetric: func(ts time.Time, metric pdata.Metric) {
metric.SetName("test_metric")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityDelta)
dp := metric.DoubleSum().DataPoints().AppendEmpty()
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityDelta)
dp := metric.Sum().DataPoints().AppendEmpty()
dp.SetValue(42.42)
dp.LabelsMap().Insert("label_1", "1")
dp.LabelsMap().Insert("label_2", "2")
@ -168,15 +168,15 @@ func TestAccumulateMetrics(t *testing.T) {
},
},
{
name: "DoubleSum",
name: "Sum",
metric: func(ts time.Time, v float64, metrics pdata.MetricSlice) {
metric := metrics.AppendEmpty()
metric.SetName("test_metric")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(false)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(false)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDescription("test description")
dp := metric.DoubleSum().DataPoints().AppendEmpty()
dp := metric.Sum().DataPoints().AppendEmpty()
dp.SetValue(v)
dp.LabelsMap().Insert("label_1", "1")
dp.LabelsMap().Insert("label_2", "2")
@ -200,15 +200,15 @@ func TestAccumulateMetrics(t *testing.T) {
},
},
{
name: "MonotonicDoubleSum",
name: "MonotonicSum",
metric: func(ts time.Time, v float64, metrics pdata.MetricSlice) {
metric := metrics.AppendEmpty()
metric.SetName("test_metric")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(true)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(true)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDescription("test description")
dp := metric.DoubleSum().DataPoints().AppendEmpty()
dp := metric.Sum().DataPoints().AppendEmpty()
dp.SetValue(v)
dp.LabelsMap().Insert("label_1", "1")
dp.LabelsMap().Insert("label_2", "2")
@ -345,12 +345,12 @@ func getMerticProperties(metric pdata.Metric) (
value = metric.DoubleGauge().DataPoints().At(0).Value()
temporality = pdata.AggregationTemporalityUnspecified
isMonotonic = false
case pdata.MetricDataTypeDoubleSum:
labels = metric.DoubleSum().DataPoints().At(0).LabelsMap()
ts = metric.DoubleSum().DataPoints().At(0).Timestamp().AsTime()
value = metric.DoubleSum().DataPoints().At(0).Value()
temporality = metric.DoubleSum().AggregationTemporality()
isMonotonic = metric.DoubleSum().IsMonotonic()
case pdata.MetricDataTypeSum:
labels = metric.Sum().DataPoints().At(0).LabelsMap()
ts = metric.Sum().DataPoints().At(0).Timestamp().AsTime()
value = metric.Sum().DataPoints().At(0).Value()
temporality = metric.Sum().AggregationTemporality()
isMonotonic = metric.Sum().IsMonotonic()
case pdata.MetricDataTypeIntHistogram:
labels = metric.IntHistogram().DataPoints().At(0).LabelsMap()
ts = metric.IntHistogram().DataPoints().At(0).Timestamp().AsTime()

View File

@ -64,8 +64,8 @@ func (c *collector) convertMetric(metric pdata.Metric) (prometheus.Metric, error
return c.convertIntSum(metric)
case pdata.MetricDataTypeDoubleGauge:
return c.convertDoubleGauge(metric)
case pdata.MetricDataTypeDoubleSum:
return c.convertDoubleSum(metric)
case pdata.MetricDataTypeSum:
return c.convertSum(metric)
case pdata.MetricDataTypeIntHistogram:
return c.convertIntHistogram(metric)
case pdata.MetricDataTypeHistogram:
@ -152,11 +152,11 @@ func (c *collector) convertIntSum(metric pdata.Metric) (prometheus.Metric, error
return m, nil
}
func (c *collector) convertDoubleSum(metric pdata.Metric) (prometheus.Metric, error) {
ip := metric.DoubleSum().DataPoints().At(0)
func (c *collector) convertSum(metric pdata.Metric) (prometheus.Metric, error) {
ip := metric.Sum().DataPoints().At(0)
metricType := prometheus.GaugeValue
if metric.DoubleSum().IsMonotonic() {
if metric.Sum().IsMonotonic() {
metricType = prometheus.CounterValue
}

View File

@ -69,7 +69,7 @@ func TestConvertInvalidMetric(t *testing.T) {
for _, mType := range []pdata.MetricDataType{
pdata.MetricDataTypeHistogram,
pdata.MetricDataTypeIntHistogram,
pdata.MetricDataTypeDoubleSum,
pdata.MetricDataTypeSum,
pdata.MetricDataTypeIntSum,
pdata.MetricDataTypeDoubleGauge,
pdata.MetricDataTypeIntGauge,
@ -83,8 +83,8 @@ func TestConvertInvalidMetric(t *testing.T) {
metric.IntSum().DataPoints().AppendEmpty()
case pdata.MetricDataTypeDoubleGauge:
metric.DoubleGauge().DataPoints().AppendEmpty()
case pdata.MetricDataTypeDoubleSum:
metric.DoubleSum().DataPoints().AppendEmpty()
case pdata.MetricDataTypeSum:
metric.Sum().DataPoints().AppendEmpty()
case pdata.MetricDataTypeIntHistogram:
metric.IntHistogram().DataPoints().AppendEmpty()
case pdata.MetricDataTypeHistogram:
@ -225,17 +225,17 @@ func TestCollectMetrics(t *testing.T) {
},
},
{
name: "DoubleSum",
name: "Sum",
metricType: prometheus.GaugeValue,
value: 42.42,
metric: func(ts time.Time) (metric pdata.Metric) {
metric = pdata.NewMetric()
metric.SetName("test_metric")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(false)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(false)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDescription("test description")
dp := metric.DoubleSum().DataPoints().AppendEmpty()
dp := metric.Sum().DataPoints().AppendEmpty()
dp.SetValue(42.42)
dp.LabelsMap().Insert("label_1", "1")
dp.LabelsMap().Insert("label_2", "2")
@ -265,17 +265,17 @@ func TestCollectMetrics(t *testing.T) {
},
},
{
name: "MonotonicDoubleSum",
name: "MonotonicSum",
metricType: prometheus.CounterValue,
value: 42.42,
metric: func(ts time.Time) (metric pdata.Metric) {
metric = pdata.NewMetric()
metric.SetName("test_metric")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(true)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(true)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDescription("test description")
dp := metric.DoubleSum().DataPoints().AppendEmpty()
dp := metric.Sum().DataPoints().AppendEmpty()
dp.SetValue(42.42)
dp.LabelsMap().Insert("label_1", "1")
dp.LabelsMap().Insert("label_2", "2")

View File

@ -142,8 +142,8 @@ func (prwe *PRWExporter) PushMetrics(ctx context.Context, md pdata.Metrics) erro
dropped++
errs = append(errs, err)
}
case pdata.MetricDataTypeDoubleSum:
dataPoints := metric.DoubleSum().DataPoints()
case pdata.MetricDataTypeSum:
dataPoints := metric.Sum().DataPoints()
if err := prwe.addDoubleDataPointSlice(dataPoints, tsMap, resource, metric); err != nil {
dropped++
errs = append(errs, err)

View File

@ -357,7 +357,7 @@ func Test_PushMetrics(t *testing.T) {
// success cases
intSumBatch := testdata.GenerateMetricsManyMetricsSameResource(10)
doubleSumBatch := getMetricsFromMetricList(validMetrics1[validDoubleSum], validMetrics2[validDoubleSum])
sumBatch := getMetricsFromMetricList(validMetrics1[validSum], validMetrics2[validSum])
intGaugeBatch := getMetricsFromMetricList(validMetrics1[validIntGauge], validMetrics2[validIntGauge])
@ -381,7 +381,7 @@ func Test_PushMetrics(t *testing.T) {
emptyCumulativeIntSumBatch := getMetricsFromMetricList(invalidMetrics[emptyCumulativeIntSum])
emptyCumulativeDoubleSumBatch := getMetricsFromMetricList(invalidMetrics[emptyCumulativeDoubleSum])
emptyCumulativeSumBatch := getMetricsFromMetricList(invalidMetrics[emptyCumulativeSum])
emptyCumulativeIntHistogramBatch := getMetricsFromMetricList(invalidMetrics[emptyCumulativeIntHistogram])
@ -434,7 +434,7 @@ func Test_PushMetrics(t *testing.T) {
},
{
"doubleSum_case",
&doubleSumBatch,
&sumBatch,
checkFunc,
2,
http.StatusAccepted,
@ -521,8 +521,8 @@ func Test_PushMetrics(t *testing.T) {
true,
},
{
"emptyCumulativeDoubleSum_case",
&emptyCumulativeDoubleSumBatch,
"emptyCumulativeSum_case",
&emptyCumulativeSumBatch,
checkFunc,
0,
http.StatusAccepted,

View File

@ -55,8 +55,8 @@ func validateMetrics(metric pdata.Metric) bool {
return metric.DoubleGauge().DataPoints().Len() != 0
case pdata.MetricDataTypeIntGauge:
return metric.IntGauge().DataPoints().Len() != 0
case pdata.MetricDataTypeDoubleSum:
return metric.DoubleSum().DataPoints().Len() != 0 && metric.DoubleSum().AggregationTemporality() == pdata.AggregationTemporalityCumulative
case pdata.MetricDataTypeSum:
return metric.Sum().DataPoints().Len() != 0 && metric.Sum().AggregationTemporality() == pdata.AggregationTemporalityCumulative
case pdata.MetricDataTypeIntSum:
return metric.IntSum().DataPoints().Len() != 0 && metric.IntSum().AggregationTemporality() == pdata.AggregationTemporalityCumulative
case pdata.MetricDataTypeHistogram:

View File

@ -89,7 +89,7 @@ var (
validIntGauge = "valid_IntGauge"
validDoubleGauge = "valid_DoubleGauge"
validIntSum = "valid_IntSum"
validDoubleSum = "valid_DoubleSum"
validSum = "valid_Sum"
validIntHistogram = "valid_IntHistogram"
validHistogram = "valid_Histogram"
validSummary = "valid_Summary"
@ -106,7 +106,7 @@ var (
validDoubleGauge: getDoubleGaugeMetric(validDoubleGauge, lbs1, floatVal1, time1),
validIntSum: getIntSumMetric(validIntSum, lbs1, intVal1, time1),
suffixedCounter: getIntSumMetric(suffixedCounter, lbs1, intVal1, time1),
validDoubleSum: getDoubleSumMetric(validDoubleSum, lbs1, floatVal1, time1),
validSum: getSumMetric(validSum, lbs1, floatVal1, time1),
validIntHistogram: getIntHistogramMetric(validIntHistogram, lbs1, time1, floatVal1, uint64(intVal1), bounds, buckets),
validHistogram: getHistogramMetric(validHistogram, lbs1, time1, floatVal1, uint64(intVal1), bounds, buckets),
validSummary: getSummaryMetric(validSummary, lbs1, time1, floatVal1, uint64(intVal1), quantiles),
@ -115,7 +115,7 @@ var (
validIntGauge: getIntGaugeMetric(validIntGauge, lbs2, intVal2, time2),
validDoubleGauge: getDoubleGaugeMetric(validDoubleGauge, lbs2, floatVal2, time2),
validIntSum: getIntSumMetric(validIntSum, lbs2, intVal2, time2),
validDoubleSum: getDoubleSumMetric(validDoubleSum, lbs2, floatVal2, time2),
validSum: getSumMetric(validSum, lbs2, floatVal2, time2),
validIntHistogram: getIntHistogramMetric(validIntHistogram, lbs2, time2, floatVal2, uint64(intVal2), bounds, buckets),
validHistogram: getHistogramMetric(validHistogram, lbs2, time2, floatVal2, uint64(intVal2), bounds, buckets),
validSummary: getSummaryMetric(validSummary, lbs2, time2, floatVal2, uint64(intVal2), quantiles),
@ -130,14 +130,14 @@ var (
emptyIntGauge = "emptyIntGauge"
emptyDoubleGauge = "emptyDoubleGauge"
emptyIntSum = "emptyIntSum"
emptyDoubleSum = "emptyDoubleSum"
emptySum = "emptySum"
emptyIntHistogram = "emptyIntHistogram"
emptyHistogram = "emptyHistogram"
emptySummary = "emptySummary"
// Category 2: invalid type and temporality combination
emptyCumulativeIntSum = "emptyCumulativeIntSum"
emptyCumulativeDoubleSum = "emptyCumulativeDoubleSum"
emptyCumulativeSum = "emptyCumulativeSum"
emptyCumulativeIntHistogram = "emptyCumulativeIntHistogram"
emptyCumulativeHistogram = "emptyCumulativeHistogram"
@ -147,12 +147,12 @@ var (
emptyIntGauge: getEmptyIntGaugeMetric(emptyIntGauge),
emptyDoubleGauge: getEmptyDoubleGaugeMetric(emptyDoubleGauge),
emptyIntSum: getEmptyIntSumMetric(emptyIntSum),
emptyDoubleSum: getEmptyDoubleSumMetric(emptyDoubleSum),
emptySum: getEmptySumMetric(emptySum),
emptyIntHistogram: getEmptyIntHistogramMetric(emptyIntHistogram),
emptyHistogram: getEmptyHistogramMetric(emptyHistogram),
emptySummary: getEmptySummaryMetric(emptySummary),
emptyCumulativeIntSum: getEmptyCumulativeIntSumMetric(emptyCumulativeIntSum),
emptyCumulativeDoubleSum: getEmptyCumulativeDoubleSumMetric(emptyCumulativeDoubleSum),
emptyCumulativeSum: getEmptyCumulativeSumMetric(emptyCumulativeSum),
emptyCumulativeIntHistogram: getEmptyCumulativeIntHistogramMetric(emptyCumulativeIntHistogram),
emptyCumulativeHistogram: getEmptyCumulativeHistogramMetric(emptyCumulativeHistogram),
}
@ -302,27 +302,27 @@ func getIntSumMetric(name string, labels pdata.StringMap, value int64, ts uint64
return metric
}
func getEmptyDoubleSumMetric(name string) pdata.Metric {
func getEmptySumMetric(name string) pdata.Metric {
metric := pdata.NewMetric()
metric.SetName(name)
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.SetDataType(pdata.MetricDataTypeSum)
return metric
}
func getEmptyCumulativeDoubleSumMetric(name string) pdata.Metric {
func getEmptyCumulativeSumMetric(name string) pdata.Metric {
metric := pdata.NewMetric()
metric.SetName(name)
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
return metric
}
func getDoubleSumMetric(name string, labels pdata.StringMap, value float64, ts uint64) pdata.Metric {
func getSumMetric(name string, labels pdata.StringMap, value float64, ts uint64) pdata.Metric {
metric := pdata.NewMetric()
metric.SetName(name)
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
dp := metric.DoubleSum().DataPoints().AppendEmpty()
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
dp := metric.Sum().DataPoints().AppendEmpty()
dp.SetValue(value)
labels.Range(func(k string, v string) bool {

View File

@ -129,9 +129,9 @@ func (g *metricGenerator) populateMetrics(cfg MetricsCfg, ilm pdata.Instrumentat
sum.SetIsMonotonic(cfg.IsMonotonicSum)
sum.SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
populateIntPoints(cfg, sum.DataPoints())
case pdata.MetricDataTypeDoubleSum:
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
sum := metric.DoubleSum()
case pdata.MetricDataTypeSum:
metric.SetDataType(pdata.MetricDataTypeSum)
sum := metric.Sum()
sum.SetIsMonotonic(cfg.IsMonotonicSum)
sum.SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
populateDoublePoints(cfg, sum.DataPoints())

View File

@ -75,10 +75,10 @@ func pictToCfg(inputs PICTMetricInputs) MetricsCfg {
case MetricTypeDoubleGauge:
cfg.MetricDescriptorType = pdata.MetricDataTypeDoubleGauge
case MetricTypeMonotonicDoubleSum:
cfg.MetricDescriptorType = pdata.MetricDataTypeDoubleSum
cfg.MetricDescriptorType = pdata.MetricDataTypeSum
cfg.IsMonotonicSum = true
case MetricTypeNonMonotonicDoubleSum:
cfg.MetricDescriptorType = pdata.MetricDataTypeDoubleSum
cfg.MetricDescriptorType = pdata.MetricDataTypeSum
cfg.IsMonotonicSum = false
case MetricTypeIntHistogram:
cfg.MetricDescriptorType = pdata.MetricDataTypeIntHistogram

View File

@ -89,8 +89,8 @@ func (b *dataBuffer) logMetricDataPoints(m pdata.Metric) {
b.logEntry(" -> IsMonotonic: %t", data.IsMonotonic())
b.logEntry(" -> AggregationTemporality: %s", data.AggregationTemporality().String())
b.logIntDataPoints(data.DataPoints())
case pdata.MetricDataTypeDoubleSum:
data := m.DoubleSum()
case pdata.MetricDataTypeSum:
data := m.Sum()
b.logEntry(" -> IsMonotonic: %t", data.IsMonotonic())
b.logEntry(" -> AggregationTemporality: %s", data.AggregationTemporality().String())
b.logDoubleDataPoints(data.DataPoints())

View File

@ -50,8 +50,8 @@ func (m *Matcher) MatchMetric(metric pdata.Metric) (bool, error) {
return m.matchDoubleGauge(metricName, metric.DoubleGauge())
case pdata.MetricDataTypeIntSum:
return m.matchIntSum(metricName, metric.IntSum())
case pdata.MetricDataTypeDoubleSum:
return m.matchDoubleSum(metricName, metric.DoubleSum())
case pdata.MetricDataTypeSum:
return m.matchSum(metricName, metric.Sum())
case pdata.MetricDataTypeIntHistogram:
return m.matchIntHistogram(metricName, metric.IntHistogram())
case pdata.MetricDataTypeHistogram:
@ -89,7 +89,7 @@ func (m *Matcher) matchDoubleGauge(metricName string, gauge pdata.DoubleGauge) (
return false, nil
}
func (m *Matcher) matchDoubleSum(metricName string, sum pdata.DoubleSum) (bool, error) {
func (m *Matcher) matchSum(metricName string, sum pdata.Sum) (bool, error) {
pts := sum.DataPoints()
for i := 0; i < pts.Len(); i++ {
matched, err := m.matchEnv(metricName, pts.At(i).LabelsMap())

View File

@ -56,8 +56,8 @@ func TestNilDoubleGauge(t *testing.T) {
testNilValue(t, dataType)
}
func TestNilDoubleSum(t *testing.T) {
dataType := pdata.MetricDataTypeDoubleSum
func TestNilSum(t *testing.T) {
dataType := pdata.MetricDataTypeSum
testNilValue(t, dataType)
}
@ -111,13 +111,13 @@ func TestDoubleGaugeEmptyDataPoint(t *testing.T) {
assert.True(t, matched)
}
func TestDoubleSumEmptyDataPoint(t *testing.T) {
func TestSumEmptyDataPoint(t *testing.T) {
matcher, err := NewMatcher(`MetricName == 'my.metric'`)
require.NoError(t, err)
m := pdata.NewMetric()
m.SetName("my.metric")
m.SetDataType(pdata.MetricDataTypeDoubleSum)
m.DoubleSum().DataPoints().AppendEmpty()
m.SetDataType(pdata.MetricDataTypeSum)
m.Sum().DataPoints().AppendEmpty()
matched, err := matcher.MatchMetric(m)
assert.NoError(t, err)
assert.True(t, matched)
@ -248,21 +248,21 @@ func testMatchDoubleGauge(t *testing.T, metricName string) bool {
return match
}
func TestMatchDoubleSumByMetricName(t *testing.T) {
assert.True(t, matchDoubleSum(t, "my.metric"))
func TestMatchSumByMetricName(t *testing.T) {
assert.True(t, matchSum(t, "my.metric"))
}
func TestNonMatchDoubleSumByMetricName(t *testing.T) {
assert.False(t, matchDoubleSum(t, "foo.metric"))
func TestNonMatchSumByMetricName(t *testing.T) {
assert.False(t, matchSum(t, "foo.metric"))
}
func matchDoubleSum(t *testing.T, metricName string) bool {
func matchSum(t *testing.T, metricName string) bool {
matcher, err := NewMatcher(`MetricName == 'my.metric'`)
require.NoError(t, err)
m := pdata.NewMetric()
m.SetName(metricName)
m.SetDataType(pdata.MetricDataTypeDoubleSum)
dps := m.DoubleSum().DataPoints()
m.SetDataType(pdata.MetricDataTypeSum)
dps := m.Sum().DataPoints()
dps.AppendEmpty()
matched, err := matcher.MatchMetric(m)
assert.NoError(t, err)

View File

@ -112,7 +112,7 @@ func GenerateMetricsAllTypesNoDataPoints() pdata.Metrics {
ms := ilm0.Metrics()
initMetric(ms.AppendEmpty(), TestGaugeDoubleMetricName, pdata.MetricDataTypeDoubleGauge)
initMetric(ms.AppendEmpty(), TestGaugeIntMetricName, pdata.MetricDataTypeIntGauge)
initMetric(ms.AppendEmpty(), TestCounterDoubleMetricName, pdata.MetricDataTypeDoubleSum)
initMetric(ms.AppendEmpty(), TestCounterDoubleMetricName, pdata.MetricDataTypeSum)
initMetric(ms.AppendEmpty(), TestCounterIntMetricName, pdata.MetricDataTypeIntSum)
initMetric(ms.AppendEmpty(), TestDoubleHistogramMetricName, pdata.MetricDataTypeHistogram)
initMetric(ms.AppendEmpty(), TestIntHistogramMetricName, pdata.MetricDataTypeIntHistogram)
@ -132,8 +132,8 @@ func GenerateMetricsAllTypesEmptyDataPoint() pdata.Metrics {
initMetric(intGauge, TestGaugeIntMetricName, pdata.MetricDataTypeIntGauge)
intGauge.IntGauge().DataPoints().AppendEmpty()
doubleSum := ms.AppendEmpty()
initMetric(doubleSum, TestCounterDoubleMetricName, pdata.MetricDataTypeDoubleSum)
doubleSum.DoubleSum().DataPoints().AppendEmpty()
initMetric(doubleSum, TestCounterDoubleMetricName, pdata.MetricDataTypeSum)
doubleSum.Sum().DataPoints().AppendEmpty()
intSum := ms.AppendEmpty()
initMetric(intSum, TestCounterIntMetricName, pdata.MetricDataTypeIntSum)
intSum.IntSum().DataPoints().AppendEmpty()
@ -199,9 +199,9 @@ func initGaugeIntMetricOneDataPoint(im pdata.Metric) {
}
func initSumDoubleMetric(dm pdata.Metric) {
initMetric(dm, TestCounterDoubleMetricName, pdata.MetricDataTypeDoubleSum)
initMetric(dm, TestCounterDoubleMetricName, pdata.MetricDataTypeSum)
ddps := dm.DoubleSum().DataPoints()
ddps := dm.Sum().DataPoints()
ddp0 := ddps.AppendEmpty()
initMetricLabels12(ddp0.LabelsMap())
ddp0.SetStartTimestamp(TestMetricStartTimestamp)
@ -295,8 +295,8 @@ func initMetric(m pdata.Metric, name string, ty pdata.MetricDataType) {
sum := m.IntSum()
sum.SetIsMonotonic(true)
sum.SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
case pdata.MetricDataTypeDoubleSum:
sum := m.DoubleSum()
case pdata.MetricDataTypeSum:
sum := m.Sum()
sum.SetIsMonotonic(true)
sum.SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
case pdata.MetricDataTypeIntHistogram:

View File

@ -675,55 +675,55 @@ func (ms IntSum) CopyTo(dest IntSum) {
ms.DataPoints().CopyTo(dest.DataPoints())
}
// DoubleSum represents the type of a numeric double scalar metric that is calculated as a sum of all reported measurements over a time interval.
// Sum represents the type of a numeric double scalar metric that is calculated as a sum of all reported measurements over a time interval.
//
// This is a reference type, if passed by value and callee modifies it the
// caller will see the modification.
//
// Must use NewDoubleSum function to create new instances.
// Must use NewSum function to create new instances.
// Important: zero-initialized instance is not valid for use.
type DoubleSum struct {
type Sum struct {
orig *otlpmetrics.Sum
}
func newDoubleSum(orig *otlpmetrics.Sum) DoubleSum {
return DoubleSum{orig: orig}
func newSum(orig *otlpmetrics.Sum) Sum {
return Sum{orig: orig}
}
// NewDoubleSum creates a new empty DoubleSum.
// NewSum creates a new empty Sum.
//
// This must be used only in testing code since no "Set" method available.
func NewDoubleSum() DoubleSum {
return newDoubleSum(&otlpmetrics.Sum{})
func NewSum() Sum {
return newSum(&otlpmetrics.Sum{})
}
// AggregationTemporality returns the aggregationtemporality associated with this DoubleSum.
func (ms DoubleSum) AggregationTemporality() AggregationTemporality {
// AggregationTemporality returns the aggregationtemporality associated with this Sum.
func (ms Sum) AggregationTemporality() AggregationTemporality {
return AggregationTemporality((*ms.orig).AggregationTemporality)
}
// SetAggregationTemporality replaces the aggregationtemporality associated with this DoubleSum.
func (ms DoubleSum) SetAggregationTemporality(v AggregationTemporality) {
// SetAggregationTemporality replaces the aggregationtemporality associated with this Sum.
func (ms Sum) SetAggregationTemporality(v AggregationTemporality) {
(*ms.orig).AggregationTemporality = otlpmetrics.AggregationTemporality(v)
}
// IsMonotonic returns the ismonotonic associated with this DoubleSum.
func (ms DoubleSum) IsMonotonic() bool {
// IsMonotonic returns the ismonotonic associated with this Sum.
func (ms Sum) IsMonotonic() bool {
return (*ms.orig).IsMonotonic
}
// SetIsMonotonic replaces the ismonotonic associated with this DoubleSum.
func (ms DoubleSum) SetIsMonotonic(v bool) {
// SetIsMonotonic replaces the ismonotonic associated with this Sum.
func (ms Sum) SetIsMonotonic(v bool) {
(*ms.orig).IsMonotonic = v
}
// DataPoints returns the DataPoints associated with this DoubleSum.
func (ms DoubleSum) DataPoints() DoubleDataPointSlice {
// DataPoints returns the DataPoints associated with this Sum.
func (ms Sum) DataPoints() DoubleDataPointSlice {
return newDoubleDataPointSlice(&(*ms.orig).DataPoints)
}
// CopyTo copies all properties from the current struct to the dest.
func (ms DoubleSum) CopyTo(dest DoubleSum) {
func (ms Sum) CopyTo(dest Sum) {
dest.SetAggregationTemporality(ms.AggregationTemporality())
dest.SetIsMonotonic(ms.IsMonotonic())
ms.DataPoints().CopyTo(dest.DataPoints())

View File

@ -504,30 +504,30 @@ func TestIntSum_DataPoints(t *testing.T) {
assert.EqualValues(t, testValDataPoints, ms.DataPoints())
}
func TestDoubleSum_CopyTo(t *testing.T) {
ms := NewDoubleSum()
generateTestDoubleSum().CopyTo(ms)
assert.EqualValues(t, generateTestDoubleSum(), ms)
func TestSum_CopyTo(t *testing.T) {
ms := NewSum()
generateTestSum().CopyTo(ms)
assert.EqualValues(t, generateTestSum(), ms)
}
func TestDoubleSum_AggregationTemporality(t *testing.T) {
ms := NewDoubleSum()
func TestSum_AggregationTemporality(t *testing.T) {
ms := NewSum()
assert.EqualValues(t, AggregationTemporalityUnspecified, ms.AggregationTemporality())
testValAggregationTemporality := AggregationTemporalityCumulative
ms.SetAggregationTemporality(testValAggregationTemporality)
assert.EqualValues(t, testValAggregationTemporality, ms.AggregationTemporality())
}
func TestDoubleSum_IsMonotonic(t *testing.T) {
ms := NewDoubleSum()
func TestSum_IsMonotonic(t *testing.T) {
ms := NewSum()
assert.EqualValues(t, false, ms.IsMonotonic())
testValIsMonotonic := true
ms.SetIsMonotonic(testValIsMonotonic)
assert.EqualValues(t, testValIsMonotonic, ms.IsMonotonic())
}
func TestDoubleSum_DataPoints(t *testing.T) {
ms := NewDoubleSum()
func TestSum_DataPoints(t *testing.T) {
ms := NewSum()
assert.EqualValues(t, NewDoubleDataPointSlice(), ms.DataPoints())
fillTestDoubleDataPointSlice(ms.DataPoints())
testValDataPoints := generateTestDoubleDataPointSlice()
@ -2003,13 +2003,13 @@ func fillTestIntSum(tv IntSum) {
fillTestIntDataPointSlice(tv.DataPoints())
}
func generateTestDoubleSum() DoubleSum {
tv := NewDoubleSum()
fillTestDoubleSum(tv)
func generateTestSum() Sum {
tv := NewSum()
fillTestSum(tv)
return tv
}
func fillTestDoubleSum(tv DoubleSum) {
func fillTestSum(tv Sum) {
tv.SetAggregationTemporality(AggregationTemporalityCumulative)
tv.SetIsMonotonic(true)
fillTestDoubleDataPointSlice(tv.DataPoints())

View File

@ -111,8 +111,8 @@ func (md Metrics) DataPointCount() (dataPointCount int) {
dataPointCount += m.DoubleGauge().DataPoints().Len()
case MetricDataTypeIntSum:
dataPointCount += m.IntSum().DataPoints().Len()
case MetricDataTypeDoubleSum:
dataPointCount += m.DoubleSum().DataPoints().Len()
case MetricDataTypeSum:
dataPointCount += m.Sum().DataPoints().Len()
case MetricDataTypeIntHistogram:
dataPointCount += m.IntHistogram().DataPoints().Len()
case MetricDataTypeHistogram:
@ -134,7 +134,7 @@ const (
MetricDataTypeIntGauge
MetricDataTypeDoubleGauge
MetricDataTypeIntSum
MetricDataTypeDoubleSum
MetricDataTypeSum
MetricDataTypeIntHistogram
MetricDataTypeHistogram
MetricDataTypeSummary
@ -151,8 +151,8 @@ func (mdt MetricDataType) String() string {
return "DoubleGauge"
case MetricDataTypeIntSum:
return "IntSum"
case MetricDataTypeDoubleSum:
return "DoubleSum"
case MetricDataTypeSum:
return "Sum"
case MetricDataTypeIntHistogram:
return "IntHistogram"
case MetricDataTypeHistogram:
@ -174,7 +174,7 @@ func (ms Metric) DataType() MetricDataType {
case *otlpmetrics.Metric_IntSum:
return MetricDataTypeIntSum
case *otlpmetrics.Metric_Sum:
return MetricDataTypeDoubleSum
return MetricDataTypeSum
case *otlpmetrics.Metric_IntHistogram:
return MetricDataTypeIntHistogram
case *otlpmetrics.Metric_Histogram:
@ -195,7 +195,7 @@ func (ms Metric) SetDataType(ty MetricDataType) {
ms.orig.Data = &otlpmetrics.Metric_Gauge{Gauge: &otlpmetrics.Gauge{}}
case MetricDataTypeIntSum:
ms.orig.Data = &otlpmetrics.Metric_IntSum{IntSum: &otlpmetrics.IntSum{}}
case MetricDataTypeDoubleSum:
case MetricDataTypeSum:
ms.orig.Data = &otlpmetrics.Metric_Sum{Sum: &otlpmetrics.Sum{}}
case MetricDataTypeIntHistogram:
ms.orig.Data = &otlpmetrics.Metric_IntHistogram{IntHistogram: &otlpmetrics.IntHistogram{}}
@ -227,11 +227,11 @@ func (ms Metric) IntSum() IntSum {
return newIntSum(ms.orig.Data.(*otlpmetrics.Metric_IntSum).IntSum)
}
// DoubleSum returns the data as DoubleSum.
// Calling this function when DataType() != MetricDataTypeDoubleSum will cause a panic.
// Sum returns the data as Sum.
// Calling this function when DataType() != MetricDataTypeSum will cause a panic.
// Calling this function on zero-initialized Metric will cause a panic.
func (ms Metric) DoubleSum() DoubleSum {
return newDoubleSum(ms.orig.Data.(*otlpmetrics.Metric_Sum).Sum)
func (ms Metric) Sum() Sum {
return newSum(ms.orig.Data.(*otlpmetrics.Metric_Sum).Sum)
}
// IntHistogram returns the data as IntHistogram.
@ -271,7 +271,7 @@ func copyData(src, dest *otlpmetrics.Metric) {
dest.Data = data
case *otlpmetrics.Metric_Sum:
data := &otlpmetrics.Metric_Sum{Sum: &otlpmetrics.Sum{}}
newDoubleSum(srcData.Sum).CopyTo(newDoubleSum(data.Sum))
newSum(srcData.Sum).CopyTo(newSum(data.Sum))
dest.Data = data
case *otlpmetrics.Metric_IntHistogram:
data := &otlpmetrics.Metric_IntHistogram{IntHistogram: &otlpmetrics.IntHistogram{}}

View File

@ -65,7 +65,7 @@ func TestCopyData(t *testing.T) {
},
},
{
name: "DoubleSum",
name: "Sum",
src: &otlpmetrics.Metric{
Data: &otlpmetrics.Metric_Sum{
Sum: &otlpmetrics.Sum{},
@ -110,8 +110,8 @@ func TestDataType(t *testing.T) {
assert.Equal(t, MetricDataTypeDoubleGauge, m.DataType())
m.SetDataType(MetricDataTypeIntSum)
assert.Equal(t, MetricDataTypeIntSum, m.DataType())
m.SetDataType(MetricDataTypeDoubleSum)
assert.Equal(t, MetricDataTypeDoubleSum, m.DataType())
m.SetDataType(MetricDataTypeSum)
assert.Equal(t, MetricDataTypeSum, m.DataType())
m.SetDataType(MetricDataTypeIntHistogram)
assert.Equal(t, MetricDataTypeIntHistogram, m.DataType())
m.SetDataType(MetricDataTypeHistogram)
@ -260,7 +260,7 @@ func TestDataPointCountWithNilDataPoints(t *testing.T) {
intSum := ilm.Metrics().AppendEmpty()
intSum.SetDataType(MetricDataTypeIntSum)
doubleSum := ilm.Metrics().AppendEmpty()
doubleSum.SetDataType(MetricDataTypeDoubleSum)
doubleSum.SetDataType(MetricDataTypeSum)
assert.EqualValues(t, 0, metrics.DataPointCount())
}
@ -272,7 +272,7 @@ func TestOtlpToInternalReadOnly(t *testing.T) {
InstrumentationLibraryMetrics: []*otlpmetrics.InstrumentationLibraryMetrics{
{
InstrumentationLibrary: generateTestProtoInstrumentationLibrary(),
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoDoubleSumMetric(), generateTestProtoDoubleHistogramMetric()},
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoSumMetric(), generateTestProtoDoubleHistogramMetric()},
},
},
},
@ -312,8 +312,8 @@ func TestOtlpToInternalReadOnly(t *testing.T) {
assert.EqualValues(t, "my_metric_double", metricDouble.Name())
assert.EqualValues(t, "My metric", metricDouble.Description())
assert.EqualValues(t, "ms", metricDouble.Unit())
assert.EqualValues(t, MetricDataTypeDoubleSum, metricDouble.DataType())
dsd := metricDouble.DoubleSum()
assert.EqualValues(t, MetricDataTypeSum, metricDouble.DataType())
dsd := metricDouble.Sum()
assert.EqualValues(t, AggregationTemporalityCumulative, dsd.AggregationTemporality())
doubleDataPoints := dsd.DataPoints()
assert.EqualValues(t, 2, doubleDataPoints.Len())
@ -360,7 +360,7 @@ func TestOtlpToFromInternalReadOnly(t *testing.T) {
InstrumentationLibraryMetrics: []*otlpmetrics.InstrumentationLibraryMetrics{
{
InstrumentationLibrary: generateTestProtoInstrumentationLibrary(),
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoDoubleSumMetric(), generateTestProtoDoubleHistogramMetric()},
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoSumMetric(), generateTestProtoDoubleHistogramMetric()},
},
},
},
@ -374,7 +374,7 @@ func TestOtlpToFromInternalReadOnly(t *testing.T) {
InstrumentationLibraryMetrics: []*otlpmetrics.InstrumentationLibraryMetrics{
{
InstrumentationLibrary: generateTestProtoInstrumentationLibrary(),
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoDoubleSumMetric(), generateTestProtoDoubleHistogramMetric()},
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoSumMetric(), generateTestProtoDoubleHistogramMetric()},
},
},
},
@ -462,7 +462,7 @@ func TestOtlpToFromInternalIntGaugeMutating(t *testing.T) {
}, internal.MetricsToOtlp(md.InternalRep()))
}
func TestOtlpToFromInternalDoubleSumMutating(t *testing.T) {
func TestOtlpToFromInternalSumMutating(t *testing.T) {
newLabels := NewStringMap().InitFromMap(map[string]string{"k": "v"})
md := MetricsFromInternalRep(internal.MetricsFromOtlp(&otlpcollectormetrics.ExportMetricsServiceRequest{
@ -472,7 +472,7 @@ func TestOtlpToFromInternalDoubleSumMutating(t *testing.T) {
InstrumentationLibraryMetrics: []*otlpmetrics.InstrumentationLibraryMetrics{
{
InstrumentationLibrary: generateTestProtoInstrumentationLibrary(),
Metrics: []*otlpmetrics.Metric{generateTestProtoDoubleSumMetric()},
Metrics: []*otlpmetrics.Metric{generateTestProtoSumMetric()},
},
},
},
@ -488,7 +488,7 @@ func TestOtlpToFromInternalDoubleSumMutating(t *testing.T) {
metric.SetUnit("1")
assert.EqualValues(t, "1", metric.Unit())
// Mutate DataPoints
dsd := metric.DoubleSum()
dsd := metric.Sum()
assert.EqualValues(t, 2, dsd.DataPoints().Len())
dsd.DataPoints().Resize(1)
assert.EqualValues(t, 1, dsd.DataPoints().Len())
@ -653,7 +653,7 @@ func BenchmarkOtlpToFromInternal_PassThrough(b *testing.B) {
InstrumentationLibraryMetrics: []*otlpmetrics.InstrumentationLibraryMetrics{
{
InstrumentationLibrary: generateTestProtoInstrumentationLibrary(),
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoDoubleSumMetric(), generateTestProtoDoubleHistogramMetric()},
Metrics: []*otlpmetrics.Metric{generateTestProtoIntGaugeMetric(), generateTestProtoSumMetric(), generateTestProtoDoubleHistogramMetric()},
},
},
},
@ -696,7 +696,7 @@ func BenchmarkOtlpToFromInternal_IntGauge_MutateOneLabel(b *testing.B) {
}
}
func BenchmarkOtlpToFromInternal_DoubleSum_MutateOneLabel(b *testing.B) {
func BenchmarkOtlpToFromInternal_Sum_MutateOneLabel(b *testing.B) {
req := &otlpcollectormetrics.ExportMetricsServiceRequest{
ResourceMetrics: []*otlpmetrics.ResourceMetrics{
{
@ -704,7 +704,7 @@ func BenchmarkOtlpToFromInternal_DoubleSum_MutateOneLabel(b *testing.B) {
InstrumentationLibraryMetrics: []*otlpmetrics.InstrumentationLibraryMetrics{
{
InstrumentationLibrary: generateTestProtoInstrumentationLibrary(),
Metrics: []*otlpmetrics.Metric{generateTestProtoDoubleSumMetric()},
Metrics: []*otlpmetrics.Metric{generateTestProtoSumMetric()},
},
},
},
@ -714,7 +714,7 @@ func BenchmarkOtlpToFromInternal_DoubleSum_MutateOneLabel(b *testing.B) {
b.ResetTimer()
for n := 0; n < b.N; n++ {
md := MetricsFromInternalRep(internal.MetricsFromOtlp(req))
md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(0).DoubleSum().DataPoints().At(0).LabelsMap().Upsert("key0", "value2")
md.ResourceMetrics().At(0).InstrumentationLibraryMetrics().At(0).Metrics().At(0).Sum().DataPoints().At(0).LabelsMap().Upsert("key0", "value2")
newReq := internal.MetricsToOtlp(md.InternalRep())
if len(req.ResourceMetrics) != len(newReq.ResourceMetrics) {
b.Fail()
@ -801,7 +801,7 @@ func generateTestProtoIntGaugeMetric() *otlpmetrics.Metric {
},
}
}
func generateTestProtoDoubleSumMetric() *otlpmetrics.Metric {
func generateTestProtoSumMetric() *otlpmetrics.Metric {
return &otlpmetrics.Metric{
Name: "my_metric_double",
Description: "My metric",

View File

@ -88,8 +88,8 @@ func metricDataPointCount(ms pdata.Metric) (dataPointCount int) {
dataPointCount = ms.DoubleGauge().DataPoints().Len()
case pdata.MetricDataTypeIntSum:
dataPointCount = ms.IntSum().DataPoints().Len()
case pdata.MetricDataTypeDoubleSum:
dataPointCount = ms.DoubleSum().DataPoints().Len()
case pdata.MetricDataTypeSum:
dataPointCount = ms.Sum().DataPoints().Len()
case pdata.MetricDataTypeIntHistogram:
dataPointCount = ms.IntHistogram().DataPoints().Len()
case pdata.MetricDataTypeHistogram:
@ -126,9 +126,9 @@ func splitMetric(ms, dest pdata.Metric, size int) (int, bool) {
ms.IntSum().DataPoints().RemoveIf(func(_ pdata.IntDataPoint) bool {
return filterDataPoints()
})
case pdata.MetricDataTypeDoubleSum:
dest.DoubleSum().DataPoints().Resize(size)
ms.DoubleSum().DataPoints().RemoveIf(func(_ pdata.DoubleDataPoint) bool {
case pdata.MetricDataTypeSum:
dest.Sum().DataPoints().Resize(size)
ms.Sum().DataPoints().RemoveIf(func(_ pdata.DoubleDataPoint) bool {
return filterDataPoints()
})
case pdata.MetricDataTypeIntHistogram:

View File

@ -59,7 +59,7 @@ func TestExprProcessor(t *testing.T) {
testFilter(t, pdata.MetricDataTypeIntGauge)
testFilter(t, pdata.MetricDataTypeDoubleGauge)
testFilter(t, pdata.MetricDataTypeIntSum)
testFilter(t, pdata.MetricDataTypeDoubleSum)
testFilter(t, pdata.MetricDataTypeSum)
testFilter(t, pdata.MetricDataTypeIntHistogram)
testFilter(t, pdata.MetricDataTypeHistogram)
}
@ -105,8 +105,8 @@ func testFilter(t *testing.T, mdType pdata.MetricDataType) {
for l := 0; l < pts.Len(); l++ {
assertFiltered(t, pts.At(l).LabelsMap())
}
case pdata.MetricDataTypeDoubleSum:
pts := metric.DoubleSum().DataPoints()
case pdata.MetricDataTypeSum:
pts := metric.Sum().DataPoints()
for l := 0; l < pts.Len(); l++ {
assertFiltered(t, pts.At(l).LabelsMap())
}

View File

@ -196,9 +196,9 @@ var Metrics = &metricStruct{
metric.SetName("process.cpu.time")
metric.SetDescription("Total CPU seconds broken down by different states.")
metric.SetUnit("s")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(true)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(true)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
},
},
&metricImpl{
@ -267,9 +267,9 @@ var Metrics = &metricStruct{
metric.SetName("system.cpu.time")
metric.SetDescription("Total CPU seconds broken down by different states.")
metric.SetUnit("s")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(true)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(true)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
},
},
&metricImpl{
@ -289,9 +289,9 @@ var Metrics = &metricStruct{
metric.SetName("system.disk.io_time")
metric.SetDescription("Time disk spent activated. On Windows, this is calculated as the inverse of disk idle time.")
metric.SetUnit("s")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(true)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(true)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
},
},
&metricImpl{
@ -311,9 +311,9 @@ var Metrics = &metricStruct{
metric.SetName("system.disk.operation_time")
metric.SetDescription("Time spent in disk operations.")
metric.SetUnit("s")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(true)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(true)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
},
},
&metricImpl{
@ -344,9 +344,9 @@ var Metrics = &metricStruct{
metric.SetName("system.disk.weighted_io_time")
metric.SetDescription("Time disk spent activated multiplied by the queue length.")
metric.SetUnit("s")
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
metric.DoubleSum().SetIsMonotonic(true)
metric.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
metric.SetDataType(pdata.MetricDataTypeSum)
metric.Sum().SetIsMonotonic(true)
metric.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
},
},
&metricImpl{

View File

@ -71,7 +71,7 @@ func (s *scraper) scrape(_ context.Context) (pdata.MetricSlice, error) {
func initializeCPUTimeMetric(metric pdata.Metric, startTime, now pdata.Timestamp, cpuTimes []cpu.TimesStat) {
metadata.Metrics.SystemCPUTime.Init(metric)
ddps := metric.DoubleSum().DataPoints()
ddps := metric.Sum().DataPoints()
ddps.Resize(len(cpuTimes) * cpuStatesLen)
for i, cpuTime := range cpuTimes {
appendCPUTimeStateDataPoints(ddps, i*cpuStatesLen, startTime, now, cpuTime)

View File

@ -109,19 +109,19 @@ func TestScrape(t *testing.T) {
func assertCPUMetricValid(t *testing.T, metric pdata.Metric, descriptor pdata.Metric, startTime pdata.Timestamp) {
internal.AssertDescriptorEqual(t, descriptor, metric)
if startTime != 0 {
internal.AssertDoubleSumMetricStartTimeEquals(t, metric, startTime)
internal.AssertSumMetricStartTimeEquals(t, metric, startTime)
}
assert.GreaterOrEqual(t, metric.DoubleSum().DataPoints().Len(), 4*runtime.NumCPU())
internal.AssertDoubleSumMetricLabelExists(t, metric, 0, metadata.Labels.Cpu)
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 0, metadata.Labels.CPUState, metadata.LabelCPUState.User)
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 1, metadata.Labels.CPUState, metadata.LabelCPUState.System)
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 2, metadata.Labels.CPUState, metadata.LabelCPUState.Idle)
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 3, metadata.Labels.CPUState, metadata.LabelCPUState.Interrupt)
assert.GreaterOrEqual(t, metric.Sum().DataPoints().Len(), 4*runtime.NumCPU())
internal.AssertSumMetricLabelExists(t, metric, 0, metadata.Labels.Cpu)
internal.AssertSumMetricLabelHasValue(t, metric, 0, metadata.Labels.CPUState, metadata.LabelCPUState.User)
internal.AssertSumMetricLabelHasValue(t, metric, 1, metadata.Labels.CPUState, metadata.LabelCPUState.System)
internal.AssertSumMetricLabelHasValue(t, metric, 2, metadata.Labels.CPUState, metadata.LabelCPUState.Idle)
internal.AssertSumMetricLabelHasValue(t, metric, 3, metadata.Labels.CPUState, metadata.LabelCPUState.Interrupt)
}
func assertCPUMetricHasLinuxSpecificStateLabels(t *testing.T, metric pdata.Metric) {
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 4, metadata.Labels.CPUState, metadata.LabelCPUState.Nice)
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 5, metadata.Labels.CPUState, metadata.LabelCPUState.Softirq)
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 6, metadata.Labels.CPUState, metadata.LabelCPUState.Steal)
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 7, metadata.Labels.CPUState, metadata.LabelCPUState.Wait)
internal.AssertSumMetricLabelHasValue(t, metric, 4, metadata.Labels.CPUState, metadata.LabelCPUState.Nice)
internal.AssertSumMetricLabelHasValue(t, metric, 5, metadata.Labels.CPUState, metadata.LabelCPUState.Softirq)
internal.AssertSumMetricLabelHasValue(t, metric, 6, metadata.Labels.CPUState, metadata.LabelCPUState.Steal)
internal.AssertSumMetricLabelHasValue(t, metric, 7, metadata.Labels.CPUState, metadata.LabelCPUState.Wait)
}

View File

@ -137,7 +137,7 @@ func initializeDiskOperationsMetric(metric pdata.Metric, startTime, now pdata.Ti
func initializeDiskIOTimeMetric(metric pdata.Metric, startTime, now pdata.Timestamp, ioCounters map[string]disk.IOCountersStat) {
metadata.Metrics.SystemDiskIoTime.Init(metric)
ddps := metric.DoubleSum().DataPoints()
ddps := metric.Sum().DataPoints()
ddps.Resize(len(ioCounters))
idx := 0
@ -150,7 +150,7 @@ func initializeDiskIOTimeMetric(metric pdata.Metric, startTime, now pdata.Timest
func initializeDiskOperationTimeMetric(metric pdata.Metric, startTime, now pdata.Timestamp, ioCounters map[string]disk.IOCountersStat) {
metadata.Metrics.SystemDiskOperationTime.Init(metric)
ddps := metric.DoubleSum().DataPoints()
ddps := metric.Sum().DataPoints()
ddps.Resize(2 * len(ioCounters))
idx := 0

View File

@ -33,7 +33,7 @@ func appendSystemSpecificMetrics(metrics pdata.MetricSlice, startIdx int, startT
func initializeDiskWeightedIOTimeMetric(metric pdata.Metric, startTime, now pdata.Timestamp, ioCounters map[string]disk.IOCountersStat) {
metadata.Metrics.SystemDiskWeightedIoTime.Init(metric)
ddps := metric.DoubleSum().DataPoints()
ddps := metric.Sum().DataPoints()
ddps.Resize(len(ioCounters))
idx := 0

View File

@ -137,19 +137,19 @@ func assertInt64DiskMetricValid(t *testing.T, metric pdata.Metric, expectedDescr
func assertDoubleDiskMetricValid(t *testing.T, metric pdata.Metric, expectedDescriptor pdata.Metric, expectDirectionLabels bool, startTime pdata.Timestamp) {
internal.AssertDescriptorEqual(t, expectedDescriptor, metric)
if startTime != 0 {
internal.AssertDoubleSumMetricStartTimeEquals(t, metric, startTime)
internal.AssertSumMetricStartTimeEquals(t, metric, startTime)
}
minExpectedPoints := 1
if expectDirectionLabels {
minExpectedPoints = 2
}
assert.GreaterOrEqual(t, metric.DoubleSum().DataPoints().Len(), minExpectedPoints)
assert.GreaterOrEqual(t, metric.Sum().DataPoints().Len(), minExpectedPoints)
internal.AssertDoubleSumMetricLabelExists(t, metric, 0, "device")
internal.AssertSumMetricLabelExists(t, metric, 0, "device")
if expectDirectionLabels {
internal.AssertDoubleSumMetricLabelHasValue(t, metric, 0, "direction", "read")
internal.AssertDoubleSumMetricLabelHasValue(t, metric, metric.DoubleSum().DataPoints().Len()-1, "direction", "write")
internal.AssertSumMetricLabelHasValue(t, metric, 0, "direction", "read")
internal.AssertSumMetricLabelHasValue(t, metric, metric.Sum().DataPoints().Len()-1, "direction", "write")
}
}

View File

@ -155,7 +155,7 @@ func initializeDiskOperationsMetric(metric pdata.Metric, startTime, now pdata.Ti
func initializeDiskIOTimeMetric(metric pdata.Metric, startTime, now pdata.Timestamp, logicalDiskCounterValues []*perfcounters.CounterValues) {
metadata.Metrics.SystemDiskIoTime.Init(metric)
ddps := metric.DoubleSum().DataPoints()
ddps := metric.Sum().DataPoints()
ddps.Resize(len(logicalDiskCounterValues))
for idx, logicalDiskCounter := range logicalDiskCounterValues {
// disk active time = system boot time - disk idle time
@ -166,7 +166,7 @@ func initializeDiskIOTimeMetric(metric pdata.Metric, startTime, now pdata.Timest
func initializeDiskOperationTimeMetric(metric pdata.Metric, startTime, now pdata.Timestamp, logicalDiskCounterValues []*perfcounters.CounterValues) {
metadata.Metrics.SystemDiskOperationTime.Init(metric)
ddps := metric.DoubleSum().DataPoints()
ddps := metric.Sum().DataPoints()
ddps.Resize(2 * len(logicalDiskCounterValues))
for idx, logicalDiskCounter := range logicalDiskCounterValues {
initializeDoubleDataPoint(ddps.At(2*idx+0), startTime, now, logicalDiskCounter.InstanceName, metadata.LabelDiskDirection.Read, float64(logicalDiskCounter.Values[avgDiskSecsPerRead])/1e7)

View File

@ -188,7 +188,7 @@ func scrapeAndAppendCPUTimeMetric(metrics pdata.MetricSlice, startTime, now pdat
func initializeCPUTimeMetric(metric pdata.Metric, startTime, now pdata.Timestamp, times *cpu.TimesStat) {
metadata.Metrics.ProcessCPUTime.Init(metric)
ddps := metric.DoubleSum().DataPoints()
ddps := metric.Sum().DataPoints()
ddps.Resize(cpuStatesLen)
appendCPUTimeStateDataPoints(ddps, startTime, now, times)
}

View File

@ -94,12 +94,12 @@ func assertCPUTimeMetricValid(t *testing.T, resourceMetrics pdata.ResourceMetric
cpuTimeMetric := getMetric(t, metadata.Metrics.ProcessCPUTime.New(), resourceMetrics)
internal.AssertDescriptorEqual(t, metadata.Metrics.ProcessCPUTime.New(), cpuTimeMetric)
if startTime != 0 {
internal.AssertDoubleSumMetricStartTimeEquals(t, cpuTimeMetric, startTime)
internal.AssertSumMetricStartTimeEquals(t, cpuTimeMetric, startTime)
}
internal.AssertDoubleSumMetricLabelHasValue(t, cpuTimeMetric, 0, "state", "user")
internal.AssertDoubleSumMetricLabelHasValue(t, cpuTimeMetric, 1, "state", "system")
internal.AssertSumMetricLabelHasValue(t, cpuTimeMetric, 0, "state", "user")
internal.AssertSumMetricLabelHasValue(t, cpuTimeMetric, 1, "state", "system")
if runtime.GOOS == "linux" {
internal.AssertDoubleSumMetricLabelHasValue(t, cpuTimeMetric, 2, "state", "wait")
internal.AssertSumMetricLabelHasValue(t, cpuTimeMetric, 2, "state", "wait")
}
}

View File

@ -47,8 +47,8 @@ func AssertIntGaugeMetricLabelHasValue(t *testing.T, metric pdata.Metric, index
assert.Equal(t, expectedVal, val)
}
func AssertDoubleSumMetricLabelHasValue(t *testing.T, metric pdata.Metric, index int, labelName string, expectedVal string) {
val, ok := metric.DoubleSum().DataPoints().At(index).LabelsMap().Get(labelName)
func AssertSumMetricLabelHasValue(t *testing.T, metric pdata.Metric, index int, labelName string, expectedVal string) {
val, ok := metric.Sum().DataPoints().At(index).LabelsMap().Get(labelName)
assert.Truef(t, ok, "Missing label %q in metric %q", labelName, metric.Name())
assert.Equal(t, expectedVal, val)
}
@ -58,8 +58,8 @@ func AssertIntSumMetricLabelExists(t *testing.T, metric pdata.Metric, index int,
assert.Truef(t, ok, "Missing label %q in metric %q", labelName, metric.Name())
}
func AssertDoubleSumMetricLabelExists(t *testing.T, metric pdata.Metric, index int, labelName string) {
_, ok := metric.DoubleSum().DataPoints().At(index).LabelsMap().Get(labelName)
func AssertSumMetricLabelExists(t *testing.T, metric pdata.Metric, index int, labelName string) {
_, ok := metric.Sum().DataPoints().At(index).LabelsMap().Get(labelName)
assert.Truef(t, ok, "Missing label %q in metric %q", labelName, metric.Name())
}
@ -70,8 +70,8 @@ func AssertIntSumMetricStartTimeEquals(t *testing.T, metric pdata.Metric, startT
}
}
func AssertDoubleSumMetricStartTimeEquals(t *testing.T, metric pdata.Metric, startTime pdata.Timestamp) {
ddps := metric.DoubleSum().DataPoints()
func AssertSumMetricStartTimeEquals(t *testing.T, metric pdata.Metric, startTime pdata.Timestamp) {
ddps := metric.Sum().DataPoints()
for i := 0; i < ddps.Len(); i++ {
require.Equal(t, startTime, ddps.At(i).StartTimestamp())
}
@ -97,8 +97,8 @@ func AssertSameTimeStampForMetrics(t *testing.T, metrics pdata.MetricSlice, star
}
}
if dt == pdata.MetricDataTypeDoubleSum {
ddps := metric.DoubleSum().DataPoints()
if dt == pdata.MetricDataTypeSum {
ddps := metric.Sum().DataPoints()
for j := 0; j < ddps.Len(); j++ {
if ts == 0 {
ts = ddps.At(j).Timestamp()

View File

@ -101,7 +101,7 @@ metrics:
description: Total CPU seconds broken down by different states.
unit: s
data:
type: double sum
type: sum
aggregation: cumulative
monotonic: true
labels: [process.state]
@ -135,7 +135,7 @@ metrics:
description: Total CPU seconds broken down by different states.
unit: s
data:
type: double sum
type: sum
aggregation: cumulative
monotonic: true
labels: [cpu.state]
@ -189,7 +189,7 @@ metrics:
description: Time disk spent activated. On Windows, this is calculated as the inverse of disk idle time.
unit: s
data:
type: double sum
type: sum
aggregation: cumulative
monotonic: true
labels: [disk.device]
@ -198,7 +198,7 @@ metrics:
description: Time spent in disk operations.
unit: s
data:
type: double sum
type: sum
aggregation: cumulative
monotonic: true
labels: [disk.device, disk.direction]
@ -207,7 +207,7 @@ metrics:
description: Time disk spent activated multiplied by the queue length.
unit: s
data:
type: double sum
type: sum
aggregation: cumulative
monotonic: true
labels: [disk.device]

View File

@ -100,7 +100,7 @@ func (mf *metricFamilyPdata) updateLabelKeys(ls labels.Labels) {
var _ = (*metricFamilyPdata)(nil).updateLabelKeys
func (mf *metricFamilyPdata) isCumulativeTypePdata() bool {
return mf.mtype == pdata.MetricDataTypeDoubleSum ||
return mf.mtype == pdata.MetricDataTypeSum ||
mf.mtype == pdata.MetricDataTypeIntSum ||
mf.mtype == pdata.MetricDataTypeHistogram ||
mf.mtype == pdata.MetricDataTypeSummary

View File

@ -60,7 +60,7 @@ func convToPdataMetricType(metricType textparse.MetricType) pdata.MetricDataType
switch metricType {
case textparse.MetricTypeCounter:
// always use float64, as it's the internal data type used in prometheus
return pdata.MetricDataTypeDoubleSum
return pdata.MetricDataTypeSum
// textparse.MetricTypeUnknown is converted to gauge by default to fix Prometheus untyped metrics from being dropped
case textparse.MetricTypeGauge, textparse.MetricTypeUnknown:
return pdata.MetricDataTypeDoubleGauge

View File

@ -194,12 +194,12 @@ func TestConvToPdataMetricType(t *testing.T) {
{
name: "textparse.counter",
mtype: textparse.MetricTypeCounter,
want: pdata.MetricDataTypeDoubleSum,
want: pdata.MetricDataTypeSum,
},
{
name: "textparse.gauge",
mtype: textparse.MetricTypeCounter,
want: pdata.MetricDataTypeDoubleSum,
want: pdata.MetricDataTypeSum,
},
{
name: "textparse.unknown",
@ -250,7 +250,7 @@ func TestIsusefulLabelPdata(t *testing.T) {
model.MetricNameLabel, model.InstanceLabel, model.SchemeLabel, model.MetricsPathLabel, model.JobLabel,
},
mtypes: []pdata.MetricDataType{
pdata.MetricDataTypeDoubleSum,
pdata.MetricDataTypeSum,
pdata.MetricDataTypeDoubleGauge,
pdata.MetricDataTypeIntHistogram,
pdata.MetricDataTypeHistogram,
@ -271,7 +271,7 @@ func TestIsusefulLabelPdata(t *testing.T) {
{
name: `bucket label with non "int_histogram", "histogram":: useful`,
mtypes: []pdata.MetricDataType{
pdata.MetricDataTypeDoubleSum,
pdata.MetricDataTypeSum,
pdata.MetricDataTypeDoubleGauge,
pdata.MetricDataTypeSummary,
pdata.MetricDataTypeIntSum,
@ -294,7 +294,7 @@ func TestIsusefulLabelPdata(t *testing.T) {
name: `quantile label with non-"summary": useful`,
labelKeys: []string{model.QuantileLabel},
mtypes: []pdata.MetricDataType{
pdata.MetricDataTypeDoubleSum,
pdata.MetricDataTypeSum,
pdata.MetricDataTypeDoubleGauge,
pdata.MetricDataTypeIntHistogram,
pdata.MetricDataTypeHistogram,
@ -309,7 +309,7 @@ func TestIsusefulLabelPdata(t *testing.T) {
name: `any other label with any type:: useful`,
labelKeys: []string{"any_label", "foo.bar"},
mtypes: []pdata.MetricDataType{
pdata.MetricDataTypeDoubleSum,
pdata.MetricDataTypeSum,
pdata.MetricDataTypeDoubleGauge,
pdata.MetricDataTypeIntHistogram,
pdata.MetricDataTypeHistogram,

View File

@ -112,10 +112,10 @@ func DiffMetric(diffs []*MetricDiff, expected pdata.Metric, actual pdata.Metric)
diffs = diff(diffs, expected.IntSum().IsMonotonic(), actual.IntSum().IsMonotonic(), "IntSum IsMonotonic")
diffs = diff(diffs, expected.IntSum().AggregationTemporality(), actual.IntSum().AggregationTemporality(), "IntSum AggregationTemporality")
diffs = diffIntPts(diffs, expected.IntSum().DataPoints(), actual.IntSum().DataPoints())
case pdata.MetricDataTypeDoubleSum:
diffs = diff(diffs, expected.DoubleSum().IsMonotonic(), actual.DoubleSum().IsMonotonic(), "DoubleSum IsMonotonic")
diffs = diff(diffs, expected.DoubleSum().AggregationTemporality(), actual.DoubleSum().AggregationTemporality(), "DoubleSum AggregationTemporality")
diffs = diffDoublePts(diffs, expected.DoubleSum().DataPoints(), actual.DoubleSum().DataPoints())
case pdata.MetricDataTypeSum:
diffs = diff(diffs, expected.Sum().IsMonotonic(), actual.Sum().IsMonotonic(), "Sum IsMonotonic")
diffs = diff(diffs, expected.Sum().AggregationTemporality(), actual.Sum().AggregationTemporality(), "Sum AggregationTemporality")
diffs = diffDoublePts(diffs, expected.Sum().DataPoints(), actual.Sum().DataPoints())
case pdata.MetricDataTypeIntHistogram:
diffs = diff(diffs, expected.IntHistogram().AggregationTemporality(), actual.IntHistogram().AggregationTemporality(), "IntHistogram AggregationTemporality")
diffs = diffIntHistogramPts(diffs, expected.IntHistogram().DataPoints(), actual.IntHistogram().DataPoints())

View File

@ -176,9 +176,9 @@ func SortedMetrics(metrics pdata.Metrics) pdata.Metrics {
for l := 0; l < m.DoubleGauge().DataPoints().Len(); l++ {
m.DoubleGauge().DataPoints().At(l).LabelsMap().Sort()
}
case pdata.MetricDataTypeDoubleSum:
for l := 0; l < m.DoubleSum().DataPoints().Len(); l++ {
m.DoubleSum().DataPoints().At(l).LabelsMap().Sort()
case pdata.MetricDataTypeSum:
for l := 0; l < m.Sum().DataPoints().Len(); l++ {
m.Sum().DataPoints().At(l).LabelsMap().Sort()
}
case pdata.MetricDataTypeIntHistogram:
for l := 0; l < m.IntHistogram().DataPoints().Len(); l++ {

View File

@ -92,8 +92,8 @@ func collectLabelKeys(metric pdata.Metric) *labelKeys {
collectLabelKeysDoubleDataPoints(metric.DoubleGauge().DataPoints(), keySet)
case pdata.MetricDataTypeIntSum:
collectLabelKeysIntDataPoints(metric.IntSum().DataPoints(), keySet)
case pdata.MetricDataTypeDoubleSum:
collectLabelKeysDoubleDataPoints(metric.DoubleSum().DataPoints(), keySet)
case pdata.MetricDataTypeSum:
collectLabelKeysDoubleDataPoints(metric.Sum().DataPoints(), keySet)
case pdata.MetricDataTypeIntHistogram:
collectLabelKeysIntHistogramDataPoints(metric.IntHistogram().DataPoints(), keySet)
case pdata.MetricDataTypeHistogram:
@ -193,8 +193,8 @@ func descriptorTypeToOC(metric pdata.Metric) ocmetrics.MetricDescriptor_Type {
return ocmetrics.MetricDescriptor_CUMULATIVE_INT64
}
return ocmetrics.MetricDescriptor_GAUGE_INT64
case pdata.MetricDataTypeDoubleSum:
sd := metric.DoubleSum()
case pdata.MetricDataTypeSum:
sd := metric.Sum()
if sd.IsMonotonic() && sd.AggregationTemporality() == pdata.AggregationTemporalityCumulative {
return ocmetrics.MetricDescriptor_CUMULATIVE_DOUBLE
}
@ -225,8 +225,8 @@ func dataPointsToTimeseries(metric pdata.Metric, labelKeys *labelKeys) []*ocmetr
return doublePointToOC(metric.DoubleGauge().DataPoints(), labelKeys)
case pdata.MetricDataTypeIntSum:
return intPointsToOC(metric.IntSum().DataPoints(), labelKeys)
case pdata.MetricDataTypeDoubleSum:
return doublePointToOC(metric.DoubleSum().DataPoints(), labelKeys)
case pdata.MetricDataTypeSum:
return doublePointToOC(metric.Sum().DataPoints(), labelKeys)
case pdata.MetricDataTypeIntHistogram:
return intHistogramPointToOC(metric.IntHistogram().DataPoints(), labelKeys)
case pdata.MetricDataTypeHistogram:

View File

@ -236,9 +236,9 @@ func TestMetricsType(t *testing.T) {
name: "double-non-monotonic-delta-sum",
internal: func() pdata.Metric {
m := pdata.NewMetric()
m.SetDataType(pdata.MetricDataTypeDoubleSum)
m.DoubleSum().SetIsMonotonic(false)
m.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityDelta)
m.SetDataType(pdata.MetricDataTypeSum)
m.Sum().SetIsMonotonic(false)
m.Sum().SetAggregationTemporality(pdata.AggregationTemporalityDelta)
return m
},
descType: ocmetrics.MetricDescriptor_GAUGE_DOUBLE,
@ -247,9 +247,9 @@ func TestMetricsType(t *testing.T) {
name: "double-non-monotonic-cumulative-sum",
internal: func() pdata.Metric {
m := pdata.NewMetric()
m.SetDataType(pdata.MetricDataTypeDoubleSum)
m.DoubleSum().SetIsMonotonic(false)
m.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
m.SetDataType(pdata.MetricDataTypeSum)
m.Sum().SetIsMonotonic(false)
m.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
return m
},
descType: ocmetrics.MetricDescriptor_GAUGE_DOUBLE,
@ -258,9 +258,9 @@ func TestMetricsType(t *testing.T) {
name: "double-monotonic-delta-sum",
internal: func() pdata.Metric {
m := pdata.NewMetric()
m.SetDataType(pdata.MetricDataTypeDoubleSum)
m.DoubleSum().SetIsMonotonic(true)
m.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityDelta)
m.SetDataType(pdata.MetricDataTypeSum)
m.Sum().SetIsMonotonic(true)
m.Sum().SetAggregationTemporality(pdata.AggregationTemporalityDelta)
return m
},
descType: ocmetrics.MetricDescriptor_GAUGE_DOUBLE,
@ -269,9 +269,9 @@ func TestMetricsType(t *testing.T) {
name: "double-monotonic-cumulative-sum",
internal: func() pdata.Metric {
m := pdata.NewMetric()
m.SetDataType(pdata.MetricDataTypeDoubleSum)
m.DoubleSum().SetIsMonotonic(true)
m.DoubleSum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
m.SetDataType(pdata.MetricDataTypeSum)
m.Sum().SetIsMonotonic(true)
m.Sum().SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
return m
},
descType: ocmetrics.MetricDescriptor_CUMULATIVE_DOUBLE,

View File

@ -182,11 +182,11 @@ func descriptorTypeToMetrics(t ocmetrics.MetricDescriptor_Type, metric pdata.Met
sum.SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
return pdata.MetricDataTypeIntSum
case ocmetrics.MetricDescriptor_CUMULATIVE_DOUBLE:
metric.SetDataType(pdata.MetricDataTypeDoubleSum)
sum := metric.DoubleSum()
metric.SetDataType(pdata.MetricDataTypeSum)
sum := metric.Sum()
sum.SetIsMonotonic(true)
sum.SetAggregationTemporality(pdata.AggregationTemporalityCumulative)
return pdata.MetricDataTypeDoubleSum
return pdata.MetricDataTypeSum
case ocmetrics.MetricDescriptor_CUMULATIVE_DISTRIBUTION:
metric.SetDataType(pdata.MetricDataTypeHistogram)
histo := metric.Histogram()
@ -209,8 +209,8 @@ func setDataPoints(ocMetric *ocmetrics.Metric, metric pdata.Metric) {
fillDoubleDataPoint(ocMetric, metric.DoubleGauge().DataPoints())
case pdata.MetricDataTypeIntSum:
fillIntDataPoint(ocMetric, metric.IntSum().DataPoints())
case pdata.MetricDataTypeDoubleSum:
fillDoubleDataPoint(ocMetric, metric.DoubleSum().DataPoints())
case pdata.MetricDataTypeSum:
fillDoubleDataPoint(ocMetric, metric.Sum().DataPoints())
case pdata.MetricDataTypeHistogram:
fillDoubleHistogramDataPoint(ocMetric, metric.Histogram().DataPoints())
case pdata.MetricDataTypeSummary: