diff --git a/.golangci.yml b/.golangci.yml index 6ff58e516..6eb0a04eb 100644 --- a/.golangci.yml +++ b/.golangci.yml @@ -36,6 +36,7 @@ issues: - ^pkg.*client.*clientset.*versioned.* - ^pkg.*client.*informers.*externalversions.* - ^pkg.*proto.* + - pkg/proto # output configuration options output: diff --git a/Makefile b/Makefile index 5eea51ffd..c6ee56bd1 100644 --- a/Makefile +++ b/Makefile @@ -402,7 +402,7 @@ test-integration-parallel: test-deps # You can download version v1.64.6 at https://github.com/golangci/golangci-lint/releases/tag/v1.64.6 .PHONY: lint lint: check-linter - $(GOLANGCI_LINT) run --build-tags=$(GOLANGCI_LINT_TAGS) --timeout=20m + $(GOLANGCI_LINT) run --build-tags=$(GOLANGCI_LINT_TAGS) --timeout=20m --max-same-issues 0 --max-issues-per-linter 0 ################################################################################ diff --git a/pkg/actors/api/period.go b/pkg/actors/api/period.go index 7a5760d47..72f553bfc 100644 --- a/pkg/actors/api/period.go +++ b/pkg/actors/api/period.go @@ -23,6 +23,8 @@ import ( ) // ReminderPeriod contains the parsed period for a reminder. +// +//nolint:recvcheck type ReminderPeriod struct { value string // Raw value as received from the user diff --git a/pkg/actors/api/period_test.go b/pkg/actors/api/period_test.go index de6cdd73d..b6f119b97 100644 --- a/pkg/actors/api/period_test.go +++ b/pkg/actors/api/period_test.go @@ -105,7 +105,7 @@ func TestReminderPeriodJSON(t *testing.T) { out := &bytes.Buffer{} err = json.Compact(out, got) require.NoError(t, err) - assert.Equal(t, wantJSON, out.String()) + assert.JSONEq(t, wantJSON, out.String()) // Unmarshal dec := ReminderPeriod{} diff --git a/pkg/actors/api/reminder.go b/pkg/actors/api/reminder.go index 5f162b6a4..b1ae0029e 100644 --- a/pkg/actors/api/reminder.go +++ b/pkg/actors/api/reminder.go @@ -28,6 +28,8 @@ import ( ) // Reminder represents a reminder or timer for a unique actor. +// +//nolint:recvcheck type Reminder struct { ActorID string `json:"actorID,omitempty"` ActorType string `json:"actorType,omitempty"` diff --git a/pkg/actors/api/reminder_test.go b/pkg/actors/api/reminder_test.go index 352c57c1c..e060a9fc8 100644 --- a/pkg/actors/api/reminder_test.go +++ b/pkg/actors/api/reminder_test.go @@ -118,7 +118,7 @@ func TestReminderProperties(t *testing.T) { require.Equal(t, -1, r.RepeatsLeft()) require.Equal(t, -1, r.Period.repeats) - for i := 0; i <= 3; i++ { + for i := range 4 { nextTick, active := r.NextTick() require.Equal(t, time1.Add((2*time.Second)*time.Duration(i)), nextTick) @@ -207,7 +207,7 @@ func TestReminderJSON(t *testing.T) { // Marshal enc, err := json.Marshal(dec) require.NoError(t, err) - require.Equal(t, payload, string(enc)) + require.JSONEq(t, payload, string(enc)) }) t.Run("failed to unmarshal", func(t *testing.T) { diff --git a/pkg/actors/api/reminders.go b/pkg/actors/api/reminders.go index 206719db3..1df000c18 100644 --- a/pkg/actors/api/reminders.go +++ b/pkg/actors/api/reminders.go @@ -30,6 +30,8 @@ type GetReminderRequest struct { } // CreateReminderRequest is the request object to create a new reminder. +// +//nolint:recvcheck type CreateReminderRequest struct { Name string ActorType string @@ -96,6 +98,8 @@ func (req *CreateReminderRequest) UnmarshalJSON(data []byte) error { } // CreateTimerRequest is the request object to create a new timer. +// +//nolint:recvcheck type CreateTimerRequest struct { Name string ActorType string diff --git a/pkg/api/errors/pubsub.go b/pkg/api/errors/pubsub.go index 6e8d5b5ad..d3782b41b 100644 --- a/pkg/api/errors/pubsub.go +++ b/pkg/api/errors/pubsub.go @@ -75,7 +75,7 @@ func (p *PubSubError) withTopicError(topic string, err error) *PubSubMetadataErr } } -func (p PubSubError) PublishMessage(topic string, err error) error { +func (p *PubSubError) PublishMessage(topic string, err error) error { return p.withTopicError(topic, err).build( codes.Internal, http.StatusInternalServerError, @@ -96,7 +96,7 @@ func (p *PubSubError) PublishForbidden(topic, appID string, err error) error { // TestNotFound is specifically for the error we are expecting for the api_tests. The not found // expected error codes are different than the existing ones for PubSubNotFound, hence // why this one is needed -func (p PubSubError) TestNotFound(topic string, err error) error { +func (p *PubSubError) TestNotFound(topic string, err error) error { return p.withTopicError(topic, err).build( codes.NotFound, http.StatusBadRequest, diff --git a/pkg/api/grpc/actor_test.go b/pkg/api/grpc/actor_test.go index 352290eee..f17743fad 100644 --- a/pkg/api/grpc/actor_test.go +++ b/pkg/api/grpc/actor_test.go @@ -59,7 +59,7 @@ func TestGetActorState(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) // act - res, err := client.GetActorState(context.Background(), &runtimev1pb.GetActorStateRequest{ + res, err := client.GetActorState(t.Context(), &runtimev1pb.GetActorStateRequest{ ActorId: "fakeActorID", ActorType: "fakeActorType", Key: "key1", @@ -101,7 +101,7 @@ func TestExecuteActorStateTransaction(t *testing.T) { // act res, err := client.ExecuteActorStateTransaction( - context.Background(), + t.Context(), &runtimev1pb.ExecuteActorStateTransactionRequest{ ActorId: "fakeActorID", ActorType: "fakeActorType", diff --git a/pkg/api/grpc/crypto_test.go b/pkg/api/grpc/crypto_test.go index 4cfae3443..42490c46c 100644 --- a/pkg/api/grpc/crypto_test.go +++ b/pkg/api/grpc/crypto_test.go @@ -15,7 +15,6 @@ package grpc import ( "bytes" - "context" "crypto/rand" "errors" "fmt" @@ -58,7 +57,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("data and options in single chunk", func(t *testing.T) { var enc []byte t.Run("encrypt", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -80,7 +79,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("decrypt", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -103,7 +102,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("one data chunk", func(t *testing.T) { var enc []byte t.Run("encrypt", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -127,7 +126,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("decrypt", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -152,7 +151,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("multiple data chunks", func(t *testing.T) { var enc []byte t.Run("encrypt", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -182,7 +181,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("decrypt - whole header in first chunk", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -211,7 +210,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("decrypt - header split in multiple chunks", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() defer stream.CloseSend() @@ -253,7 +252,7 @@ func TestCryptoAlpha1(t *testing.T) { _, err := io.ReadFull(rand.Reader, largeData) require.NoError(t, err) - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -276,7 +275,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("decrypt without header", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -297,7 +296,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("invalid sequence number", func(t *testing.T) { t.Run("encrypt", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -319,7 +318,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("decrypt", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -341,7 +340,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("options in non-leading message", func(t *testing.T) { t.Run("encrypt", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -370,7 +369,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("decrypt", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -397,7 +396,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("encrypt without required options", func(t *testing.T) { t.Run("missing options", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -411,7 +410,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("missing component name", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -429,7 +428,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("missing key name", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -447,7 +446,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("missing algorithm", func(t *testing.T) { - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -467,7 +466,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("decrypt without required options", func(t *testing.T) { t.Run("missing options", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -481,7 +480,7 @@ func TestCryptoAlpha1(t *testing.T) { }) t.Run("missing component name", func(t *testing.T) { - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() send := []runtimev1pb.CryptoRequests{ @@ -500,7 +499,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("time out while waiting for first chunk", func(t *testing.T) { t.Run("encrypt", func(t *testing.T) { start := time.Now() - stream, err := client.EncryptAlpha1(context.Background()) + stream, err := client.EncryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() @@ -512,7 +511,7 @@ func TestCryptoAlpha1(t *testing.T) { t.Run("decrypt", func(t *testing.T) { start := time.Now() - stream, err := client.DecryptAlpha1(context.Background()) + stream, err := client.DecryptAlpha1(t.Context()) require.NoError(t, err) defer stream.CloseSend() diff --git a/pkg/api/grpc/daprinternal_test.go b/pkg/api/grpc/daprinternal_test.go index b3678aa37..a95e6c4b5 100644 --- a/pkg/api/grpc/daprinternal_test.go +++ b/pkg/api/grpc/daprinternal_test.go @@ -14,7 +14,6 @@ limitations under the License. package grpc import ( - "context" "errors" "io" "testing" @@ -50,7 +49,7 @@ func TestCallLocal(t *testing.T) { request := invokev1.NewInvokeMethodRequest("method") defer request.Close() - _, err := client.CallLocal(context.Background(), request.Proto()) + _, err := client.CallLocal(t.Context(), request.Proto()) assert.Equal(t, codes.Internal, status.Code(err)) }) @@ -72,7 +71,7 @@ func TestCallLocal(t *testing.T) { Message: nil, } - _, err := client.CallLocal(context.Background(), request) + _, err := client.CallLocal(t.Context(), request) assert.Equal(t, codes.InvalidArgument, status.Code(err)) }) @@ -97,7 +96,7 @@ func TestCallLocal(t *testing.T) { request := invokev1.NewInvokeMethodRequest("method") defer request.Close() - _, err := client.CallLocal(context.Background(), request.Proto()) + _, err := client.CallLocal(t.Context(), request.Proto()) assert.Equal(t, codes.Internal, status.Code(err)) }) } @@ -116,7 +115,7 @@ func TestCallLocalStream(t *testing.T) { defer clientConn.Close() client := internalv1pb.NewServiceInvocationClient(clientConn) - st, err := client.CallLocalStream(context.Background()) + st, err := client.CallLocalStream(t.Context()) require.NoError(t, err) request := invokev1.NewInvokeMethodRequest("method") @@ -146,7 +145,7 @@ func TestCallLocalStream(t *testing.T) { defer clientConn.Close() client := internalv1pb.NewServiceInvocationClient(clientConn) - st, err := client.CallLocalStream(context.Background()) + st, err := client.CallLocalStream(t.Context()) require.NoError(t, err) err = st.Send(&internalv1pb.InternalInvokeRequestStream{ @@ -183,7 +182,7 @@ func TestCallLocalStream(t *testing.T) { defer clientConn.Close() client := internalv1pb.NewServiceInvocationClient(clientConn) - st, err := client.CallLocalStream(context.Background()) + st, err := client.CallLocalStream(t.Context()) require.NoError(t, err) request := invokev1.NewInvokeMethodRequest("method"). @@ -221,7 +220,7 @@ func TestCallRemoteAppWithTracing(t *testing.T) { request := invokev1.NewInvokeMethodRequest("method") defer request.Close() - resp, err := client.CallLocal(context.Background(), request.Proto()) + resp, err := client.CallLocal(t.Context(), request.Proto()) require.NoError(t, err) assert.NotEmpty(t, resp.GetMessage(), "failed to generate trace context with app call") } @@ -239,7 +238,7 @@ func TestCallActorWithTracing(t *testing.T) { WithActor("test-actor", "actor-1") defer request.Close() - resp, err := client.CallActor(context.Background(), request.Proto()) + resp, err := client.CallActor(t.Context(), request.Proto()) require.NoError(t, err) assert.NotEmpty(t, resp.GetMessage(), "failed to generate trace context with actor call") } diff --git a/pkg/api/grpc/grpc_test.go b/pkg/api/grpc/grpc_test.go index 1884cdd68..d4632cb37 100644 --- a/pkg/api/grpc/grpc_test.go +++ b/pkg/api/grpc/grpc_test.go @@ -429,7 +429,7 @@ func TestAPIToken(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) md := grpcMetadata.Pairs("dapr-api-token", token) - ctx := grpcMetadata.NewOutgoingContext(context.Background(), md) + ctx := grpcMetadata.NewOutgoingContext(t.Context(), md) t.Run("unary", func(t *testing.T) { // act @@ -496,7 +496,7 @@ func TestAPIToken(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) md := grpcMetadata.Pairs("dapr-api-token", "bad, bad token") - ctx := grpcMetadata.NewOutgoingContext(context.Background(), md) + ctx := grpcMetadata.NewOutgoingContext(t.Context(), md) t.Run("unary", func(t *testing.T) { // act @@ -559,7 +559,7 @@ func TestAPIToken(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - ctx := context.Background() + ctx := t.Context() t.Run("unary", func(t *testing.T) { // act @@ -687,7 +687,7 @@ func TestInvokeServiceFromHTTPResponse(t *testing.T) { }, } var header grpcMetadata.MD - _, err := client.InvokeService(context.Background(), req, grpc.Header(&header)) + _, err := client.InvokeService(t.Context(), req, grpc.Header(&header)) // assert mockDirectMessaging.AssertNumberOfCalls(t, "Invoke", 1) @@ -759,7 +759,7 @@ func TestInvokeServiceFromGRPCResponse(t *testing.T) { Data: &anypb.Any{Value: []byte("testData")}, }, } - _, err := client.InvokeService(context.Background(), req) + _, err := client.InvokeService(t.Context(), req) // assert mockDirectMessaging.AssertNumberOfCalls(t, "Invoke", 1) @@ -790,7 +790,7 @@ func TestSecretStoreNotConfigured(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - _, err := client.GetSecret(context.Background(), &runtimev1pb.GetSecretRequest{}) + _, err := client.GetSecret(t.Context(), &runtimev1pb.GetSecretRequest{}) assert.Equal(t, codes.FailedPrecondition, status.Code(err)) } @@ -928,7 +928,7 @@ func TestGetSecret(t *testing.T) { StoreName: tt.storeName, Key: tt.key, } - resp, err := client.GetSecret(context.Background(), req) + resp, err := client.GetSecret(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err) @@ -1004,7 +1004,7 @@ func TestGetBulkSecret(t *testing.T) { req := &runtimev1pb.GetBulkSecretRequest{ StoreName: tt.storeName, } - resp, err := client.GetBulkSecret(context.Background(), req) + resp, err := client.GetBulkSecret(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err) @@ -1032,7 +1032,7 @@ func TestGetStateWhenStoreNotConfigured(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - _, err := client.GetState(context.Background(), &runtimev1pb.GetStateRequest{}) + _, err := client.GetState(t.Context(), &runtimev1pb.GetStateRequest{}) assert.Equal(t, codes.FailedPrecondition, status.Code(err)) } @@ -1227,7 +1227,7 @@ func TestSaveState(t *testing.T) { // test and assert for _, tt := range testCases { t.Run(tt.testName, func(t *testing.T) { - _, err := client.SaveState(context.Background(), &runtimev1pb.SaveStateRequest{ + _, err := client.SaveState(t.Context(), &runtimev1pb.SaveStateRequest{ StoreName: tt.storeName, States: tt.states, }) @@ -1321,7 +1321,7 @@ func TestGetState(t *testing.T) { Key: tt.key, } - resp, err := client.GetState(context.Background(), req) + resp, err := client.GetState(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err) assert.Equal(t, resp.GetData(), tt.expectedResponse.GetData(), "Expected response Data to be same") @@ -1459,7 +1459,7 @@ func TestGetConfiguration(t *testing.T) { Keys: tt.keys, } - resp, err := client.GetConfigurationAlpha1(context.Background(), req) + resp, err := client.GetConfigurationAlpha1(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err) assert.Equal(t, resp.GetItems(), tt.expectedResponse.GetItems(), "Expected response items to be same") @@ -1475,7 +1475,7 @@ func TestGetConfiguration(t *testing.T) { Keys: tt.keys, } - resp, err := client.GetConfiguration(context.Background(), req) + resp, err := client.GetConfiguration(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err) assert.Equal(t, resp.GetItems(), tt.expectedResponse.GetItems(), "Expected response items to be same") @@ -1499,7 +1499,7 @@ func TestSubscribeConfiguration(t *testing.T) { }), mock.MatchedBy(func(f configuration.UpdateHandler) bool { if len(tempReq.Keys) == 1 && tempReq.Keys[0] == goodKey { - go f(context.Background(), &configuration.UpdateEvent{ + go f(t.Context(), &configuration.UpdateEvent{ Items: map[string]*configuration.Item{ goodKey: { Value: "test-data", @@ -1525,7 +1525,7 @@ func TestSubscribeConfiguration(t *testing.T) { }), mock.MatchedBy(func(f configuration.UpdateHandler) bool { if len(tempReq.Keys) == 2 && tempReq.Keys[0] == goodKey && tempReq.Keys[1] == goodKey2 { - go f(context.Background(), &configuration.UpdateEvent{ + go f(t.Context(), &configuration.UpdateEvent{ Items: map[string]*configuration.Item{ goodKey: { Value: "test-data", @@ -1634,7 +1634,7 @@ func TestSubscribeConfiguration(t *testing.T) { Keys: tt.keys, } - resp, _ := subscribeFn(context.Background(), req) + resp, _ := subscribeFn(t.Context(), req) if !tt.errorExcepted { // First message should contain the ID only @@ -1714,7 +1714,7 @@ func TestUnSubscribeConfiguration(t *testing.T) { return default: } - if err := f(context.Background(), &configuration.UpdateEvent{ + if err := f(t.Context(), &configuration.UpdateEvent{ Items: map[string]*configuration.Item{ goodKey: { Value: "test-data", @@ -1748,7 +1748,7 @@ func TestUnSubscribeConfiguration(t *testing.T) { return default: } - if err := f(context.Background(), &configuration.UpdateEvent{ + if err := f(t.Context(), &configuration.UpdateEvent{ Items: map[string]*configuration.Item{ goodKey: { Value: "test-data", @@ -1830,7 +1830,7 @@ func TestUnSubscribeConfiguration(t *testing.T) { Keys: tt.keys, } - resp, err := client.SubscribeConfigurationAlpha1(context.Background(), req) + resp, err := client.SubscribeConfigurationAlpha1(t.Context(), req) require.NoError(t, err, "Error should be nil") const retry = 3 count := 0 @@ -1852,7 +1852,7 @@ func TestUnSubscribeConfiguration(t *testing.T) { subscribeID = rsp.GetId() } require.NoError(t, err, "Error should be nil") - _, err = client.UnsubscribeConfigurationAlpha1(context.Background(), &runtimev1pb.UnsubscribeConfigurationRequest{ + _, err = client.UnsubscribeConfigurationAlpha1(t.Context(), &runtimev1pb.UnsubscribeConfigurationRequest{ StoreName: tt.storeName, Id: subscribeID, }) @@ -1879,7 +1879,7 @@ func TestUnSubscribeConfiguration(t *testing.T) { Keys: tt.keys, } - resp, err := client.SubscribeConfiguration(context.Background(), req) + resp, err := client.SubscribeConfiguration(t.Context(), req) require.NoError(t, err, "Error should be nil") const retry = 3 count := 0 @@ -1901,7 +1901,7 @@ func TestUnSubscribeConfiguration(t *testing.T) { subscribeID = rsp.GetId() } require.NoError(t, err, "Error should be nil") - _, err = client.UnsubscribeConfiguration(context.Background(), &runtimev1pb.UnsubscribeConfigurationRequest{ + _, err = client.UnsubscribeConfiguration(t.Context(), &runtimev1pb.UnsubscribeConfigurationRequest{ StoreName: tt.storeName, Id: subscribeID, }) @@ -1975,7 +1975,7 @@ func TestUnsubscribeConfigurationErrScenario(t *testing.T) { Id: tt.id, } - resp, err := client.UnsubscribeConfigurationAlpha1(context.Background(), req) + resp, err := client.UnsubscribeConfigurationAlpha1(t.Context(), req) assert.Equal(t, tt.expectedResponse, resp != nil) assert.Equal(t, tt.expectedError, err != nil) }) @@ -1985,7 +1985,7 @@ func TestUnsubscribeConfigurationErrScenario(t *testing.T) { Id: tt.id, } - resp, err := client.UnsubscribeConfiguration(context.Background(), req) + resp, err := client.UnsubscribeConfiguration(t.Context(), req) assert.Equal(t, tt.expectedResponse, resp != nil) assert.Equal(t, tt.expectedError, err != nil) }) @@ -2096,7 +2096,7 @@ func TestGetBulkState(t *testing.T) { Keys: tt.keys, } - resp, err := client.GetBulkState(context.Background(), req) + resp, err := client.GetBulkState(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err) @@ -2210,7 +2210,7 @@ func TestDeleteState(t *testing.T) { for _, tt := range testCases { t.Run(tt.testName, func(t *testing.T) { - _, err := client.DeleteState(context.Background(), &runtimev1pb.DeleteStateRequest{ + _, err := client.DeleteState(t.Context(), &runtimev1pb.DeleteStateRequest{ StoreName: tt.storeName, Key: tt.key, }) @@ -2338,7 +2338,7 @@ func TestDeleteBulkState(t *testing.T) { for _, tt := range testCases { t.Run(tt.testName, func(t *testing.T) { - _, err := client.DeleteBulkState(context.Background(), &runtimev1pb.DeleteBulkStateRequest{ + _, err := client.DeleteBulkState(t.Context(), &runtimev1pb.DeleteBulkStateRequest{ StoreName: tt.storeName, States: tt.states, }) @@ -2403,19 +2403,19 @@ func TestPublishTopic(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) t.Run("err: empty publish event request", func(t *testing.T) { - _, err := client.PublishEvent(context.Background(), &runtimev1pb.PublishEventRequest{}) + _, err := client.PublishEvent(t.Context(), &runtimev1pb.PublishEventRequest{}) assert.Equal(t, codes.InvalidArgument, status.Code(err)) }) t.Run("err: publish event request with empty topic", func(t *testing.T) { - _, err := client.PublishEvent(context.Background(), &runtimev1pb.PublishEventRequest{ + _, err := client.PublishEvent(t.Context(), &runtimev1pb.PublishEventRequest{ PubsubName: "pubsub", }) assert.Equal(t, codes.InvalidArgument, status.Code(err)) }) t.Run("no err: publish event request with topic and pubsub alone", func(t *testing.T) { - _, err := client.PublishEvent(context.Background(), &runtimev1pb.PublishEventRequest{ + _, err := client.PublishEvent(t.Context(), &runtimev1pb.PublishEventRequest{ PubsubName: "pubsub", Topic: "topic", }) @@ -2423,7 +2423,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("no err: publish event request with topic, pubsub and ce metadata override", func(t *testing.T) { - _, err := client.PublishEvent(context.Background(), &runtimev1pb.PublishEventRequest{ + _, err := client.PublishEvent(t.Context(), &runtimev1pb.PublishEventRequest{ PubsubName: "pubsub", Topic: "topic", Metadata: map[string]string{ @@ -2436,7 +2436,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: publish event request with error-topic and pubsub", func(t *testing.T) { - _, err := client.PublishEvent(context.Background(), &runtimev1pb.PublishEventRequest{ + _, err := client.PublishEvent(t.Context(), &runtimev1pb.PublishEventRequest{ PubsubName: "pubsub", Topic: "error-topic", }) @@ -2444,7 +2444,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: publish event request with err-not-found topic and pubsub", func(t *testing.T) { - _, err := client.PublishEvent(context.Background(), &runtimev1pb.PublishEventRequest{ + _, err := client.PublishEvent(t.Context(), &runtimev1pb.PublishEventRequest{ PubsubName: "pubsub", Topic: "err-not-found", }) @@ -2452,7 +2452,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: publish event request with err-not-allowed topic and pubsub", func(t *testing.T) { - _, err := client.PublishEvent(context.Background(), &runtimev1pb.PublishEventRequest{ + _, err := client.PublishEvent(t.Context(), &runtimev1pb.PublishEventRequest{ PubsubName: "pubsub", Topic: "err-not-allowed", }) @@ -2460,12 +2460,12 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: empty bulk publish event request", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{}) + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{}) assert.Equal(t, codes.InvalidArgument, status.Code(err)) }) t.Run("err: bulk publish event request with duplicate entry Ids", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "topic", Entries: []*runtimev1pb.BulkPublishRequestEntry{ @@ -2489,7 +2489,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: bulk publish event request with missing entry Ids", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "topic", Entries: []*runtimev1pb.BulkPublishRequestEntry{ @@ -2511,14 +2511,14 @@ func TestPublishTopic(t *testing.T) { assert.Contains(t, err.Error(), "not present for entry") }) t.Run("err: bulk publish event request with pubsub and empty topic", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", }) assert.Equal(t, codes.InvalidArgument, status.Code(err)) }) t.Run("no err: bulk publish event request with pubsub, topic and empty entries", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "topic", }) @@ -2526,7 +2526,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: bulk publish event request with error-topic and pubsub", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "error-topic", }) @@ -2534,7 +2534,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: bulk publish event request with err-not-found topic and pubsub", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "err-not-found", }) @@ -2542,7 +2542,7 @@ func TestPublishTopic(t *testing.T) { }) t.Run("err: bulk publish event request with err-not-allowed topic and pubsub", func(t *testing.T) { - _, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + _, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "err-not-allowed", }) @@ -2605,7 +2605,7 @@ func TestBulkPublish(t *testing.T) { } t.Run("no failures", func(t *testing.T) { - res, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + res, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "topic", Entries: sampleEntries, @@ -2615,7 +2615,7 @@ func TestBulkPublish(t *testing.T) { }) t.Run("no failures with ce metadata override", func(t *testing.T) { - res, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + res, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "topic", Entries: sampleEntries, @@ -2630,7 +2630,7 @@ func TestBulkPublish(t *testing.T) { }) t.Run("all failures from component", func(t *testing.T) { - res, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + res, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "error-topic", Entries: sampleEntries, @@ -2643,7 +2643,7 @@ func TestBulkPublish(t *testing.T) { }) t.Run("partial failures from component", func(t *testing.T) { - res, err := client.BulkPublishEventAlpha1(context.Background(), &runtimev1pb.BulkPublishRequest{ + res, err := client.BulkPublishEventAlpha1(t.Context(), &runtimev1pb.BulkPublishRequest{ PubsubName: "pubsub", Topic: "even-error-topic", Entries: sampleEntries, @@ -2671,12 +2671,12 @@ func TestInvokeBinding(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - _, err := client.InvokeBinding(context.Background(), &runtimev1pb.InvokeBindingRequest{}) + _, err := client.InvokeBinding(t.Context(), &runtimev1pb.InvokeBindingRequest{}) require.NoError(t, err) - _, err = client.InvokeBinding(context.Background(), &runtimev1pb.InvokeBindingRequest{Name: "error-binding"}) + _, err = client.InvokeBinding(t.Context(), &runtimev1pb.InvokeBindingRequest{Name: "error-binding"}) assert.Equal(t, codes.Internal, status.Code(err)) - ctx := grpcMetadata.AppendToOutgoingContext(context.Background(), "traceparent", "Test") + ctx := grpcMetadata.AppendToOutgoingContext(t.Context(), "traceparent", "Test") resp, err := client.InvokeBinding(ctx, &runtimev1pb.InvokeBindingRequest{Metadata: map[string]string{"userMetadata": "val1"}}) require.NoError(t, err) assert.NotNil(t, resp) @@ -2699,7 +2699,7 @@ func TestTransactionStateStoreNotConfigured(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - _, err := client.ExecuteStateTransaction(context.Background(), &runtimev1pb.ExecuteStateTransactionRequest{}) + _, err := client.ExecuteStateTransaction(t.Context(), &runtimev1pb.ExecuteStateTransactionRequest{}) assert.Equal(t, codes.FailedPrecondition, status.Code(err)) } @@ -2718,7 +2718,7 @@ func TestTransactionStateStoreNotImplemented(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - _, err := client.ExecuteStateTransaction(context.Background(), &runtimev1pb.ExecuteStateTransactionRequest{ + _, err := client.ExecuteStateTransaction(t.Context(), &runtimev1pb.ExecuteStateTransactionRequest{ StoreName: "store1", }) assert.Equal(t, codes.Unimplemented, status.Code(err)) @@ -2742,12 +2742,12 @@ func TestExecuteStateTransaction(t *testing.T) { fakeStore.On("Multi", mock.MatchedBy(matchContextInterface), mock.MatchedBy(func(req *state.TransactionalStateRequest) bool { - return matchKeyFn(context.Background(), req, goodKey) + return matchKeyFn(t.Context(), req, goodKey) })).Return(nil) fakeStore.On("Multi", mock.MatchedBy(matchContextInterface), mock.MatchedBy(func(req *state.TransactionalStateRequest) bool { - return matchKeyFn(context.Background(), req, "error-key") + return matchKeyFn(t.Context(), req, "error-key") })).Return(errors.New("error to execute with key2")) compStore := compstore.New() @@ -2855,7 +2855,7 @@ func TestExecuteStateTransaction(t *testing.T) { } } - _, err := client.ExecuteStateTransaction(context.Background(), req) + _, err := client.ExecuteStateTransaction(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err) } else { @@ -3037,7 +3037,7 @@ func TestQueryState(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) - resp, err := client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + resp, err := client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: "store1", Query: queryTestRequestOK, }) @@ -3047,20 +3047,20 @@ func TestQueryState(t *testing.T) { assert.NotNil(t, resp.GetResults()[0].GetData()) } - resp, err = client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + resp, err = client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: "store1", Query: queryTestRequestNoRes, }) assert.Empty(t, resp.GetResults()) assert.Equal(t, codes.OK, status.Code(err)) - _, err = client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + _, err = client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: "store1", Query: queryTestRequestErr, }) assert.Equal(t, codes.Internal, status.Code(err)) - _, err = client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + _, err = client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: "store1", Query: queryTestRequestSyntaxErr, }) @@ -3083,7 +3083,7 @@ func TestStateStoreQuerierNotImplemented(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - _, err := client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + _, err := client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: "store1", }) assert.Equal(t, codes.Internal, status.Code(err)) @@ -3107,7 +3107,7 @@ func TestStateStoreQuerierEncrypted(t *testing.T) { defer clientConn.Close() client := runtimev1pb.NewDaprClient(clientConn) - _, err := client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + _, err := client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: storeName, }) assert.Equal(t, codes.Internal, status.Code(err)) @@ -3139,7 +3139,7 @@ func TestGetConfigurationAPI(t *testing.T) { testFn := func(getFn getConfigurationFn) func(t *testing.T) { return func(t *testing.T) { - r, err := getFn(context.Background(), &runtimev1pb.GetConfigurationRequest{ + r, err := getFn(t.Context(), &runtimev1pb.GetConfigurationRequest{ StoreName: "store1", Keys: []string{ "key1", @@ -3178,7 +3178,7 @@ func TestSubscribeConfigurationAPI(t *testing.T) { getConfigurationItemTest := func(subscribeFn subscribeConfigurationFn) func(t *testing.T) { return func(t *testing.T) { - s, err := subscribeFn(context.Background(), &runtimev1pb.SubscribeConfigurationRequest{ + s, err := subscribeFn(t.Context(), &runtimev1pb.SubscribeConfigurationRequest{ StoreName: "store1", Keys: []string{ "key1", @@ -3222,7 +3222,7 @@ func TestSubscribeConfigurationAPI(t *testing.T) { getAllConfigurationItemTest := func(subscribeFn subscribeConfigurationFn) func(t *testing.T) { return func(t *testing.T) { - s, err := subscribeFn(context.Background(), &runtimev1pb.SubscribeConfigurationRequest{ + s, err := subscribeFn(t.Context(), &runtimev1pb.SubscribeConfigurationRequest{ StoreName: "store1", Keys: []string{}, }) @@ -3317,7 +3317,7 @@ func TestStateAPIWithResiliency(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) t.Run("get state request retries with resiliency", func(t *testing.T) { - _, err := client.GetState(context.Background(), &runtimev1pb.GetStateRequest{ + _, err := client.GetState(t.Context(), &runtimev1pb.GetStateRequest{ StoreName: "failStore", Key: "failingGetKey", }) @@ -3327,7 +3327,7 @@ func TestStateAPIWithResiliency(t *testing.T) { t.Run("get state request times out with resiliency", func(t *testing.T) { start := time.Now() - _, err := client.GetState(context.Background(), &runtimev1pb.GetStateRequest{ + _, err := client.GetState(t.Context(), &runtimev1pb.GetStateRequest{ StoreName: "failStore", Key: "timeoutGetKey", }) @@ -3339,7 +3339,7 @@ func TestStateAPIWithResiliency(t *testing.T) { }) t.Run("set state request retries with resiliency", func(t *testing.T) { - _, err := client.SaveState(context.Background(), &runtimev1pb.SaveStateRequest{ + _, err := client.SaveState(t.Context(), &runtimev1pb.SaveStateRequest{ StoreName: "failStore", States: []*commonv1pb.StateItem{ { @@ -3354,7 +3354,7 @@ func TestStateAPIWithResiliency(t *testing.T) { t.Run("set state request times out with resiliency", func(t *testing.T) { start := time.Now() - _, err := client.SaveState(context.Background(), &runtimev1pb.SaveStateRequest{ + _, err := client.SaveState(t.Context(), &runtimev1pb.SaveStateRequest{ StoreName: "failStore", States: []*commonv1pb.StateItem{ { @@ -3371,7 +3371,7 @@ func TestStateAPIWithResiliency(t *testing.T) { }) t.Run("delete state request retries with resiliency", func(t *testing.T) { - _, err := client.DeleteState(context.Background(), &runtimev1pb.DeleteStateRequest{ + _, err := client.DeleteState(t.Context(), &runtimev1pb.DeleteStateRequest{ StoreName: "failStore", Key: "failingDeleteKey", }) @@ -3381,7 +3381,7 @@ func TestStateAPIWithResiliency(t *testing.T) { t.Run("delete state request times out with resiliency", func(t *testing.T) { start := time.Now() - _, err := client.DeleteState(context.Background(), &runtimev1pb.DeleteStateRequest{ + _, err := client.DeleteState(t.Context(), &runtimev1pb.DeleteStateRequest{ StoreName: "failStore", Key: "timeoutDeleteKey", }) @@ -3399,7 +3399,7 @@ func TestStateAPIWithResiliency(t *testing.T) { failingStore.BulkFailKey.Store(ptr.Of("")) }) - _, err := client.GetBulkState(context.Background(), &runtimev1pb.GetBulkStateRequest{ + _, err := client.GetBulkState(t.Context(), &runtimev1pb.GetBulkStateRequest{ StoreName: "failStore", Keys: []string{"failingBulkGetKey", "goodBulkGetKey"}, }) @@ -3408,7 +3408,7 @@ func TestStateAPIWithResiliency(t *testing.T) { }) t.Run("bulk state set recovers from single key failure with resiliency", func(t *testing.T) { - _, err := client.SaveState(context.Background(), &runtimev1pb.SaveStateRequest{ + _, err := client.SaveState(t.Context(), &runtimev1pb.SaveStateRequest{ StoreName: "failStore", States: []*commonv1pb.StateItem{ { @@ -3429,7 +3429,7 @@ func TestStateAPIWithResiliency(t *testing.T) { t.Run("bulk state set times out with resiliency", func(t *testing.T) { start := time.Now() - _, err := client.SaveState(context.Background(), &runtimev1pb.SaveStateRequest{ + _, err := client.SaveState(t.Context(), &runtimev1pb.SaveStateRequest{ StoreName: "failStore", States: []*commonv1pb.StateItem{ { @@ -3451,7 +3451,7 @@ func TestStateAPIWithResiliency(t *testing.T) { }) t.Run("state transaction passes after retries with resiliency", func(t *testing.T) { - _, err := client.ExecuteStateTransaction(context.Background(), &runtimev1pb.ExecuteStateTransactionRequest{ + _, err := client.ExecuteStateTransaction(t.Context(), &runtimev1pb.ExecuteStateTransactionRequest{ StoreName: "failStore", Operations: []*runtimev1pb.TransactionalStateOperation{ { @@ -3468,7 +3468,7 @@ func TestStateAPIWithResiliency(t *testing.T) { }) t.Run("state transaction times out with resiliency", func(t *testing.T) { - _, err := client.ExecuteStateTransaction(context.Background(), &runtimev1pb.ExecuteStateTransactionRequest{ + _, err := client.ExecuteStateTransaction(t.Context(), &runtimev1pb.ExecuteStateTransactionRequest{ StoreName: "failStore", Operations: []*runtimev1pb.TransactionalStateOperation{ { @@ -3485,7 +3485,7 @@ func TestStateAPIWithResiliency(t *testing.T) { }) t.Run("state query retries with resiliency", func(t *testing.T) { - _, err := client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + _, err := client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: "failStore", Query: queryTestRequestOK, Metadata: map[string]string{"key": "failingQueryKey"}, @@ -3496,7 +3496,7 @@ func TestStateAPIWithResiliency(t *testing.T) { }) t.Run("state query times out with resiliency", func(t *testing.T) { - _, err := client.QueryStateAlpha1(context.Background(), &runtimev1pb.QueryStateRequest{ + _, err := client.QueryStateAlpha1(t.Context(), &runtimev1pb.QueryStateRequest{ StoreName: "failStore", Query: queryTestRequestOK, Metadata: map[string]string{"key": "timeoutQueryKey"}, @@ -3544,7 +3544,7 @@ func TestConfigurationAPIWithResiliency(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) t.Run("test get configuration retries with resiliency", func(t *testing.T) { - _, err := client.GetConfiguration(context.Background(), &runtimev1pb.GetConfigurationRequest{ + _, err := client.GetConfiguration(t.Context(), &runtimev1pb.GetConfigurationRequest{ StoreName: "failConfig", Keys: []string{}, Metadata: map[string]string{"key": "failingGetKey"}, @@ -3555,7 +3555,7 @@ func TestConfigurationAPIWithResiliency(t *testing.T) { }) t.Run("test get configuration fails due to timeout with resiliency", func(t *testing.T) { - _, err := client.GetConfiguration(context.Background(), &runtimev1pb.GetConfigurationRequest{ + _, err := client.GetConfiguration(t.Context(), &runtimev1pb.GetConfigurationRequest{ StoreName: "failConfig", Keys: []string{}, Metadata: map[string]string{"key": "timeoutGetKey"}, @@ -3566,7 +3566,7 @@ func TestConfigurationAPIWithResiliency(t *testing.T) { }) t.Run("test subscribe configuration retries with resiliency", func(t *testing.T) { - resp, err := client.SubscribeConfiguration(context.Background(), &runtimev1pb.SubscribeConfigurationRequest{ + resp, err := client.SubscribeConfiguration(t.Context(), &runtimev1pb.SubscribeConfigurationRequest{ StoreName: "failConfig", Keys: []string{}, Metadata: map[string]string{"key": "failingSubscribeKey"}, @@ -3580,7 +3580,7 @@ func TestConfigurationAPIWithResiliency(t *testing.T) { }) t.Run("test subscribe configuration fails due to timeout with resiliency", func(t *testing.T) { - resp, err := client.SubscribeConfiguration(context.Background(), &runtimev1pb.SubscribeConfigurationRequest{ + resp, err := client.SubscribeConfiguration(t.Context(), &runtimev1pb.SubscribeConfigurationRequest{ StoreName: "failConfig", Keys: []string{}, Metadata: map[string]string{"key": "timeoutSubscribeKey"}, @@ -3626,7 +3626,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) t.Run("Get secret - retries on initial failure with resiliency", func(t *testing.T) { - _, err := client.GetSecret(context.Background(), &runtimev1pb.GetSecretRequest{ + _, err := client.GetSecret(t.Context(), &runtimev1pb.GetSecretRequest{ StoreName: "failSecret", Key: "key", }) @@ -3638,7 +3638,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { t.Run("Get secret - timeout before request ends", func(t *testing.T) { // Store sleeps for 30 seconds, let's make sure our timeout takes less time than that. start := time.Now() - _, err := client.GetSecret(context.Background(), &runtimev1pb.GetSecretRequest{ + _, err := client.GetSecret(t.Context(), &runtimev1pb.GetSecretRequest{ StoreName: "failSecret", Key: "timeout", }) @@ -3650,7 +3650,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { }) t.Run("Get bulk secret - retries on initial failure with resiliency", func(t *testing.T) { - _, err := client.GetBulkSecret(context.Background(), &runtimev1pb.GetBulkSecretRequest{ + _, err := client.GetBulkSecret(t.Context(), &runtimev1pb.GetBulkSecretRequest{ StoreName: "failSecret", Metadata: map[string]string{"key": "bulk"}, }) @@ -3661,7 +3661,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { t.Run("Get bulk secret - timeout before request ends", func(t *testing.T) { start := time.Now() - _, err := client.GetBulkSecret(context.Background(), &runtimev1pb.GetBulkSecretRequest{ + _, err := client.GetBulkSecret(t.Context(), &runtimev1pb.GetBulkSecretRequest{ StoreName: "failSecret", Metadata: map[string]string{"key": "bulkTimeout"}, }) @@ -3710,7 +3710,7 @@ func TestServiceInvocationWithResiliency(t *testing.T) { t.Run("Test invoke direct message retries with resiliency", func(t *testing.T) { val := []byte("failingKey") - res, err := client.InvokeService(context.Background(), &runtimev1pb.InvokeServiceRequest{ + res, err := client.InvokeService(t.Context(), &runtimev1pb.InvokeServiceRequest{ Id: "failingApp", Message: &commonv1pb.InvokeRequest{ Method: "test", @@ -3727,7 +3727,7 @@ func TestServiceInvocationWithResiliency(t *testing.T) { t.Run("Test invoke direct message fails with timeout", func(t *testing.T) { start := time.Now() - _, err := client.InvokeService(context.Background(), &runtimev1pb.InvokeServiceRequest{ + _, err := client.InvokeService(t.Context(), &runtimev1pb.InvokeServiceRequest{ Id: "failingApp", Message: &commonv1pb.InvokeRequest{ Method: "test", @@ -3742,7 +3742,7 @@ func TestServiceInvocationWithResiliency(t *testing.T) { }) t.Run("Test invoke direct messages fails after exhausting retries", func(t *testing.T) { - _, err := client.InvokeService(context.Background(), &runtimev1pb.InvokeServiceRequest{ + _, err := client.InvokeService(t.Context(), &runtimev1pb.InvokeServiceRequest{ Id: "failingApp", Message: &commonv1pb.InvokeRequest{ Method: "test", @@ -3756,7 +3756,7 @@ func TestServiceInvocationWithResiliency(t *testing.T) { t.Run("Test invoke direct messages opens circuit breaker after consecutive failures", func(t *testing.T) { // Circuit breaker trips on the 5th request, ending the retries. - _, err := client.InvokeService(context.Background(), &runtimev1pb.InvokeServiceRequest{ + _, err := client.InvokeService(t.Context(), &runtimev1pb.InvokeServiceRequest{ Id: "circuitBreakerApp", Message: &commonv1pb.InvokeRequest{ Method: "test", @@ -3767,7 +3767,7 @@ func TestServiceInvocationWithResiliency(t *testing.T) { assert.Equal(t, 5, failingDirectMessaging.Failure.CallCount("circuitBreakerKey")) // Additional requests should fail due to the circuit breaker. - _, err = client.InvokeService(context.Background(), &runtimev1pb.InvokeServiceRequest{ + _, err = client.InvokeService(t.Context(), &runtimev1pb.InvokeServiceRequest{ Id: "circuitBreakerApp", Message: &commonv1pb.InvokeRequest{ Method: "test", @@ -3858,7 +3858,7 @@ func TestTryLock(t *testing.T) { StoreName: "abc", ExpiryInSeconds: 10, } - _, err := api.TryLockAlpha1(context.Background(), req) + _, err := api.TryLockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = FailedPrecondition desc = lock store is not configured", err.Error()) }) @@ -3879,7 +3879,7 @@ func TestTryLock(t *testing.T) { StoreName: "mock", ExpiryInSeconds: 10, } - _, err := api.TryLockAlpha1(context.Background(), req) + _, err := api.TryLockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = InvalidArgument desc = ResourceId is empty in lock store mock", err.Error()) }) @@ -3902,7 +3902,7 @@ func TestTryLock(t *testing.T) { ResourceId: "resource", ExpiryInSeconds: 10, } - _, err := api.TryLockAlpha1(context.Background(), req) + _, err := api.TryLockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = InvalidArgument desc = LockOwner is empty in lock store mock", err.Error()) }) @@ -3927,7 +3927,7 @@ func TestTryLock(t *testing.T) { LockOwner: "owner", ExpiryInSeconds: 0, } - _, err := api.TryLockAlpha1(context.Background(), req) + _, err := api.TryLockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = InvalidArgument desc = ExpiryInSeconds is not positive in lock store mock", err.Error()) }) @@ -3952,7 +3952,7 @@ func TestTryLock(t *testing.T) { LockOwner: "owner", ExpiryInSeconds: 1, } - _, err := api.TryLockAlpha1(context.Background(), req) + _, err := api.TryLockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = InvalidArgument desc = lock store abc not found", err.Error()) }) @@ -3962,7 +3962,7 @@ func TestTryLock(t *testing.T) { mockLockStore := daprt.NewMockStore(ctl) - mockLockStore.EXPECT().TryLock(context.Background(), gomock.Any()).DoAndReturn(func(ctx context.Context, req *lock.TryLockRequest) (*lock.TryLockResponse, error) { + mockLockStore.EXPECT().TryLock(t.Context(), gomock.Any()).DoAndReturn(func(ctx context.Context, req *lock.TryLockRequest) (*lock.TryLockResponse, error) { assert.Equal(t, "lock||resource", req.ResourceID) assert.Equal(t, "owner", req.LockOwner) assert.Equal(t, int32(1), req.ExpiryInSeconds) @@ -3986,7 +3986,7 @@ func TestTryLock(t *testing.T) { LockOwner: "owner", ExpiryInSeconds: 1, } - resp, err := api.TryLockAlpha1(context.Background(), req) + resp, err := api.TryLockAlpha1(t.Context(), req) require.NoError(t, err) assert.True(t, resp.GetSuccess()) }) @@ -4009,7 +4009,7 @@ func TestUnlock(t *testing.T) { req := &runtimev1pb.UnlockRequest{ StoreName: "abc", } - _, err := api.UnlockAlpha1(context.Background(), req) + _, err := api.UnlockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = FailedPrecondition desc = lock store is not configured", err.Error()) }) @@ -4030,7 +4030,7 @@ func TestUnlock(t *testing.T) { req := &runtimev1pb.UnlockRequest{ StoreName: "abc", } - _, err := api.UnlockAlpha1(context.Background(), req) + _, err := api.UnlockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = InvalidArgument desc = ResourceId is empty in lock store abc", err.Error()) }) @@ -4051,7 +4051,7 @@ func TestUnlock(t *testing.T) { StoreName: "abc", ResourceId: "resource", } - _, err := api.UnlockAlpha1(context.Background(), req) + _, err := api.UnlockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = InvalidArgument desc = LockOwner is empty in lock store abc", err.Error()) }) @@ -4074,7 +4074,7 @@ func TestUnlock(t *testing.T) { ResourceId: "resource", LockOwner: "owner", } - _, err := api.UnlockAlpha1(context.Background(), req) + _, err := api.UnlockAlpha1(t.Context(), req) assert.Equal(t, "api error: code = InvalidArgument desc = lock store abc not found", err.Error()) }) @@ -4084,7 +4084,7 @@ func TestUnlock(t *testing.T) { mockLockStore := daprt.NewMockStore(ctl) - mockLockStore.EXPECT().Unlock(context.Background(), gomock.Any()).DoAndReturn(func(ctx context.Context, req *lock.UnlockRequest) (*lock.UnlockResponse, error) { + mockLockStore.EXPECT().Unlock(t.Context(), gomock.Any()).DoAndReturn(func(ctx context.Context, req *lock.UnlockRequest) (*lock.UnlockResponse, error) { assert.Equal(t, "lock||resource", req.ResourceID) assert.Equal(t, "owner", req.LockOwner) return &lock.UnlockResponse{ @@ -4105,7 +4105,7 @@ func TestUnlock(t *testing.T) { ResourceId: "resource", LockOwner: "owner", } - resp, err := api.UnlockAlpha1(context.Background(), req) + resp, err := api.UnlockAlpha1(t.Context(), req) require.NoError(t, err) assert.Equal(t, runtimev1pb.UnlockResponse_SUCCESS, resp.GetStatus()) //nolint:nosnakecase }) @@ -4228,7 +4228,7 @@ func TestMetadata(t *testing.T) { client := runtimev1pb.NewDaprClient(clientConn) t.Run("Set Metadata", func(t *testing.T) { - _, err := client.SetMetadata(context.Background(), &runtimev1pb.SetMetadataRequest{ + _, err := client.SetMetadata(t.Context(), &runtimev1pb.SetMetadataRequest{ Key: "foo", Value: "bar", }) @@ -4236,7 +4236,7 @@ func TestMetadata(t *testing.T) { }) t.Run("Get Metadata", func(t *testing.T) { - res, err := client.GetMetadata(context.Background(), &runtimev1pb.GetMetadataRequest{}) + res, err := client.GetMetadata(t.Context(), &runtimev1pb.GetMetadataRequest{}) require.NoError(t, err) assert.Equal(t, "fakeAPI", res.GetId()) diff --git a/pkg/api/grpc/proxy/codec/codec.go b/pkg/api/grpc/proxy/codec/codec.go index 5af597dc7..14c69ecf8 100644 --- a/pkg/api/grpc/proxy/codec/codec.go +++ b/pkg/api/grpc/proxy/codec/codec.go @@ -55,12 +55,6 @@ func (f *Frame) ProtoMessage() { // nop } -// Raw returns the raw message. -// This is primarily useful for debugging. -func (f Frame) Raw() []byte { - return f.payload -} - // Marshal implements the encoding.Codec interface method. func (p *Proxy) Marshal(v any) ([]byte, error) { out, ok := v.(*Frame) diff --git a/pkg/api/grpc/proxy/handler_test.go b/pkg/api/grpc/proxy/handler_test.go index 868cb7043..bd55e837a 100644 --- a/pkg/api/grpc/proxy/handler_test.go +++ b/pkg/api/grpc/proxy/handler_test.go @@ -476,7 +476,7 @@ func (s *proxyTestSuite) TestResiliencyUnary() { setupMetrics(s) - ctx := metadata.NewOutgoingContext(context.Background(), metadata.Pairs(diag.GRPCProxyAppIDKey, testAppID)) + ctx := metadata.NewOutgoingContext(t.Context(), metadata.Pairs(diag.GRPCProxyAppIDKey, testAppID)) _, err := s.testClient.Ping(ctx, &pb.PingRequest{Value: message}) require.Error(t, err, "Ping should fail due to timeouts") @@ -515,7 +515,7 @@ func (s *proxyTestSuite) TestResiliencyUnary() { go func(i int) { for j := range numOperations { pingMsg := fmt.Sprintf("%d:%d", i, j) - ctx := metadata.NewOutgoingContext(context.Background(), metadata.Pairs(diag.GRPCProxyAppIDKey, testAppID)) + ctx := metadata.NewOutgoingContext(t.Context(), metadata.Pairs(diag.GRPCProxyAppIDKey, testAppID)) res, err := s.testClient.Ping(ctx, &pb.PingRequest{Value: pingMsg}) require.NoErrorf(t, err, "Ping should succeed for operation %d:%d", i, j) require.NotNilf(t, res, "Response should not be nil for operation %d:%d", i, j) @@ -581,7 +581,7 @@ func (s *proxyTestSuite) TestResiliencyStreaming() { s.T().Run("retries are not allowed", func(t *testing.T) { // We're purposely not setting dapr-stream=true in this context because we want to simulate the failure when the RPC is not marked as streaming - ctx, cancel := context.WithTimeout(context.Background(), time.Minute) + ctx, cancel := context.WithTimeout(t.Context(), time.Minute) defer cancel() ctx = metadata.NewOutgoingContext(ctx, metadata.Pairs( diag.GRPCProxyAppIDKey, "test", @@ -616,7 +616,7 @@ func (s *proxyTestSuite) TestResiliencyStreaming() { }) s.T().Run("timeouts do not apply after initial handshake", func(t *testing.T) { - ctx, cancel := context.WithTimeout(context.Background(), time.Minute) + ctx, cancel := context.WithTimeout(t.Context(), time.Minute) defer cancel() setupMetrics(s) @@ -667,7 +667,7 @@ func (s *proxyTestSuite) TestResiliencyStreaming() { s.service.simulateConnectionFailures.Store(0) }() - ctx, cancel := context.WithTimeout(context.Background(), time.Minute) + ctx, cancel := context.WithTimeout(t.Context(), time.Minute) defer cancel() setupMetrics(s) diff --git a/pkg/api/grpc/server_test.go b/pkg/api/grpc/server_test.go index 716daad18..abb596c67 100644 --- a/pkg/api/grpc/server_test.go +++ b/pkg/api/grpc/server_test.go @@ -168,7 +168,7 @@ func TestGrpcAPILoggingMiddlewares(t *testing.T) { if userAgent != "" { md["user-agent"] = []string{userAgent} } - ctx := grpcMetadata.NewIncomingContext(context.Background(), md) + ctx := grpcMetadata.NewIncomingContext(t.Context(), md) info := &grpcGo.UnaryServerInfo{ FullMethod: "/dapr.proto.runtime.v1.Dapr/GetState", diff --git a/pkg/api/http/http_test.go b/pkg/api/http/http_test.go index 48c34ae40..96b959d15 100644 --- a/pkg/api/http/http_test.go +++ b/pkg/api/http/http_test.go @@ -3907,11 +3907,11 @@ type fakeConfigurationStore struct { counter int } -func (c fakeConfigurationStore) Ping() error { +func (c *fakeConfigurationStore) Ping() error { return nil } -func (c fakeConfigurationStore) Get(ctx context.Context, req *configuration.GetRequest) (*configuration.GetResponse, error) { +func (c *fakeConfigurationStore) Get(ctx context.Context, req *configuration.GetRequest) (*configuration.GetResponse, error) { if len(req.Keys) == 0 { return &configuration.GetResponse{ Items: map[string]*configuration.Item{ @@ -3974,8 +3974,8 @@ func (c fakeConfigurationStore) Get(ctx context.Context, req *configuration.GetR return nil, errors.New("get key error: value not found") } -func (c fakeConfigurationStore) Init(ctx context.Context, metadata configuration.Metadata) error { - c.counter = 0 //nolint:staticcheck +func (c *fakeConfigurationStore) Init(ctx context.Context, metadata configuration.Metadata) error { + c.counter = 0 return nil } @@ -3992,7 +3992,7 @@ func (c *fakeConfigurationStore) Unsubscribe(ctx context.Context, req *configura type fakeLockStore struct{} -func (l fakeLockStore) Ping() error { +func (l *fakeLockStore) Ping() error { return nil } @@ -4427,11 +4427,11 @@ func matchContextInterface(v any) bool { return ok } -func (c fakeConfigurationStore) Close() error { +func (c *fakeConfigurationStore) Close() error { return nil } -func (l fakeLockStore) Close() error { +func (l *fakeLockStore) Close() error { return nil } diff --git a/pkg/api/http/server_test.go b/pkg/api/http/server_test.go index 775f1afea..2eda016d0 100644 --- a/pkg/api/http/server_test.go +++ b/pkg/api/http/server_test.go @@ -93,7 +93,7 @@ func TestUnescapeRequestParametersHandler(t *testing.T) { newCtx := func() context.Context { chiCtx := chi.NewRouteContext() chiCtx.URLParams.Add("testparam", "foo%20bar") - return context.WithValue(context.Background(), chi.RouteCtxKey, chiCtx) + return context.WithValue(t.Context(), chi.RouteCtxKey, chiCtx) } t.Run("unescapeRequestParametersHandler is added as middleware if the endpoint includes Parameters in its path", func(t *testing.T) { diff --git a/pkg/api/http/universal_test.go b/pkg/api/http/universal_test.go index 82e348209..ed4d82d78 100644 --- a/pkg/api/http/universal_test.go +++ b/pkg/api/http/universal_test.go @@ -158,7 +158,7 @@ func TestUniversalHTTPHandler(t *testing.T) { respBody, err := io.ReadAll(resp.Body) require.NoError(t, err) - assert.Equal(t, `{"errorCode":"ERR_BAD_REQUEST","message":"invalid request: unexpected message"}`, string(respBody)) + assert.JSONEq(t, `{"errorCode":"ERR_BAD_REQUEST","message":"invalid request: unexpected message"}`, string(respBody)) }) t.Run("Handler returns nil", func(t *testing.T) { diff --git a/pkg/api/universal/metadata_test.go b/pkg/api/universal/metadata_test.go index 62b9dc3f7..3623c7a71 100644 --- a/pkg/api/universal/metadata_test.go +++ b/pkg/api/universal/metadata_test.go @@ -14,7 +14,6 @@ limitations under the License. package universal import ( - "context" "encoding/json" "testing" "time" @@ -110,7 +109,7 @@ func TestGetMetadata(t *testing.T) { globalConfig: &config.Configuration{}, } - response, err := fakeAPI.GetMetadata(context.Background(), &runtimev1pb.GetMetadataRequest{}) + response, err := fakeAPI.GetMetadata(t.Context(), &runtimev1pb.GetMetadataRequest{}) require.NoError(t, err, "Expected no error") bytes, err := json.Marshal(response) @@ -141,7 +140,7 @@ func TestSetMetadata(t *testing.T) { appID: "fakeAPI", } - _, err := fakeAPI.SetMetadata(context.Background(), &runtimev1pb.SetMetadataRequest{ + _, err := fakeAPI.SetMetadata(t.Context(), &runtimev1pb.SetMetadataRequest{ Key: "testKey", Value: "testValue", }) diff --git a/pkg/api/universal/secrets_test.go b/pkg/api/universal/secrets_test.go index fec531fa8..7d3cee2be 100644 --- a/pkg/api/universal/secrets_test.go +++ b/pkg/api/universal/secrets_test.go @@ -14,7 +14,6 @@ limitations under the License. package universal import ( - "context" "testing" "time" @@ -41,13 +40,13 @@ func TestSecretStoreNotConfigured(t *testing.T) { // act t.Run("GetSecret", func(t *testing.T) { - _, err := fakeAPI.GetSecret(context.Background(), &runtimev1pb.GetSecretRequest{}) + _, err := fakeAPI.GetSecret(t.Context(), &runtimev1pb.GetSecretRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrSecretStoreNotConfigured) }) t.Run("GetBulkSecret", func(t *testing.T) { - _, err := fakeAPI.GetBulkSecret(context.Background(), &runtimev1pb.GetBulkSecretRequest{}) + _, err := fakeAPI.GetBulkSecret(t.Context(), &runtimev1pb.GetBulkSecretRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrSecretStoreNotConfigured) }) @@ -175,7 +174,7 @@ func TestGetSecret(t *testing.T) { StoreName: tt.storeName, Key: tt.key, } - resp, err := fakeAPI.GetSecret(context.Background(), req) + resp, err := fakeAPI.GetSecret(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err, "Expected no error") @@ -239,7 +238,7 @@ func TestGetBulkSecret(t *testing.T) { req := &runtimev1pb.GetBulkSecretRequest{ StoreName: tt.storeName, } - resp, err := fakeAPI.GetBulkSecret(context.Background(), req) + resp, err := fakeAPI.GetBulkSecret(t.Context(), req) if !tt.errorExcepted { require.NoError(t, err, "Expected no error") @@ -273,7 +272,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { // act t.Run("Get secret - retries on initial failure with resiliency", func(t *testing.T) { - _, err := fakeAPI.GetSecret(context.Background(), &runtimev1pb.GetSecretRequest{ + _, err := fakeAPI.GetSecret(t.Context(), &runtimev1pb.GetSecretRequest{ StoreName: "failSecret", Key: "key", }) @@ -285,7 +284,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { t.Run("Get secret - timeout before request ends", func(t *testing.T) { // Store sleeps for 30 seconds, let's make sure our timeout takes less time than that. start := time.Now() - _, err := fakeAPI.GetSecret(context.Background(), &runtimev1pb.GetSecretRequest{ + _, err := fakeAPI.GetSecret(t.Context(), &runtimev1pb.GetSecretRequest{ StoreName: "failSecret", Key: "timeout", }) @@ -297,7 +296,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { }) t.Run("Get bulk secret - retries on initial failure with resiliency", func(t *testing.T) { - _, err := fakeAPI.GetBulkSecret(context.Background(), &runtimev1pb.GetBulkSecretRequest{ + _, err := fakeAPI.GetBulkSecret(t.Context(), &runtimev1pb.GetBulkSecretRequest{ StoreName: "failSecret", Metadata: map[string]string{"key": "bulk"}, }) @@ -308,7 +307,7 @@ func TestSecretAPIWithResiliency(t *testing.T) { t.Run("Get bulk secret - timeout before request ends", func(t *testing.T) { start := time.Now() - _, err := fakeAPI.GetBulkSecret(context.Background(), &runtimev1pb.GetBulkSecretRequest{ + _, err := fakeAPI.GetBulkSecret(t.Context(), &runtimev1pb.GetBulkSecretRequest{ StoreName: "failSecret", Metadata: map[string]string{"key": "bulkTimeout"}, }) diff --git a/pkg/api/universal/shutdown_test.go b/pkg/api/universal/shutdown_test.go index c35b7793e..3badb4904 100644 --- a/pkg/api/universal/shutdown_test.go +++ b/pkg/api/universal/shutdown_test.go @@ -34,7 +34,7 @@ func TestShutdownEndpoint(t *testing.T) { } t.Run("Shutdown successfully", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) _, err := fakeAPI.Shutdown(ctx, &runtimev1pb.ShutdownRequest{}) cancel() require.NoError(t, err, "Expected no error") diff --git a/pkg/api/universal/subtlecrypto_subtlecrypto_test.go b/pkg/api/universal/subtlecrypto_subtlecrypto_test.go index 88948f952..00ab5a095 100644 --- a/pkg/api/universal/subtlecrypto_subtlecrypto_test.go +++ b/pkg/api/universal/subtlecrypto_subtlecrypto_test.go @@ -18,7 +18,6 @@ limitations under the License. package universal import ( - "context" "crypto/sha256" "encoding/hex" "testing" @@ -46,7 +45,7 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { } t.Run("return key in PEM format", func(t *testing.T) { - res, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + res, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "myvault", Name: "good-key", Format: runtimev1pb.SubtleGetKeyRequest_PEM, @@ -58,7 +57,7 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { }) t.Run("return key in JSON format", func(t *testing.T) { - res, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + res, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "myvault", Name: "good-key", Format: runtimev1pb.SubtleGetKeyRequest_JSON, @@ -70,7 +69,7 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { }) t.Run("default to PEM format", func(t *testing.T) { - res, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + res, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "myvault", Name: "good-key", }) @@ -81,7 +80,7 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { }) t.Run("key not found", func(t *testing.T) { - res, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + res, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "myvault", Name: "not-found", }) @@ -92,7 +91,7 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { }) t.Run("key has key ID", func(t *testing.T) { - res, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + res, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "myvault", Name: "with-name", }) @@ -108,13 +107,13 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { compStore.AddCryptoProvider("myvault", fakeCryptoProvider) }() - _, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{}) + _, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrCryptoProvidersNotConfigured) }) t.Run("provider not found", func(t *testing.T) { - _, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + _, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "notfound", }) require.Error(t, err) @@ -122,7 +121,7 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { }) t.Run("invalid format", func(t *testing.T) { - _, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + _, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "myvault", Format: runtimev1pb.SubtleGetKeyRequest_KeyFormat(-9000), }) @@ -132,7 +131,7 @@ func TestSubtleGetKeyAlpha1(t *testing.T) { }) t.Run("failed to get key", func(t *testing.T) { - _, err := fakeAPI.SubtleGetKeyAlpha1(context.Background(), &runtimev1pb.SubtleGetKeyRequest{ + _, err := fakeAPI.SubtleGetKeyAlpha1(t.Context(), &runtimev1pb.SubtleGetKeyRequest{ ComponentName: "myvault", Name: "error-key", }) @@ -153,7 +152,7 @@ func TestSubtleEncryptAlpha1(t *testing.T) { } t.Run("encrypt message", func(t *testing.T) { - res, err := fakeAPI.SubtleEncryptAlpha1(context.Background(), &runtimev1pb.SubtleEncryptRequest{ + res, err := fakeAPI.SubtleEncryptAlpha1(t.Context(), &runtimev1pb.SubtleEncryptRequest{ ComponentName: "myvault", Plaintext: []byte("hello world"), KeyName: "good-tag", @@ -171,13 +170,13 @@ func TestSubtleEncryptAlpha1(t *testing.T) { compStore.AddCryptoProvider("myvault", fakeCryptoProvider) }() - _, err := fakeAPI.SubtleEncryptAlpha1(context.Background(), &runtimev1pb.SubtleEncryptRequest{}) + _, err := fakeAPI.SubtleEncryptAlpha1(t.Context(), &runtimev1pb.SubtleEncryptRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrCryptoProvidersNotConfigured) }) t.Run("provider not found", func(t *testing.T) { - _, err := fakeAPI.SubtleEncryptAlpha1(context.Background(), &runtimev1pb.SubtleEncryptRequest{ + _, err := fakeAPI.SubtleEncryptAlpha1(t.Context(), &runtimev1pb.SubtleEncryptRequest{ ComponentName: "notfound", }) require.Error(t, err) @@ -185,7 +184,7 @@ func TestSubtleEncryptAlpha1(t *testing.T) { }) t.Run("failed to encrypt", func(t *testing.T) { - _, err := fakeAPI.SubtleEncryptAlpha1(context.Background(), &runtimev1pb.SubtleEncryptRequest{ + _, err := fakeAPI.SubtleEncryptAlpha1(t.Context(), &runtimev1pb.SubtleEncryptRequest{ ComponentName: "myvault", KeyName: "error", }) @@ -207,7 +206,7 @@ func TestSubtleDecryptAlpha1(t *testing.T) { } t.Run("decrypt message", func(t *testing.T) { - res, err := fakeAPI.SubtleDecryptAlpha1(context.Background(), &runtimev1pb.SubtleDecryptRequest{ + res, err := fakeAPI.SubtleDecryptAlpha1(t.Context(), &runtimev1pb.SubtleDecryptRequest{ ComponentName: "myvault", Ciphertext: []byte("hello world"), KeyName: "good", @@ -224,13 +223,13 @@ func TestSubtleDecryptAlpha1(t *testing.T) { compStore.AddCryptoProvider("myvault", fakeCryptoProvider) }() - _, err := fakeAPI.SubtleDecryptAlpha1(context.Background(), &runtimev1pb.SubtleDecryptRequest{}) + _, err := fakeAPI.SubtleDecryptAlpha1(t.Context(), &runtimev1pb.SubtleDecryptRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrCryptoProvidersNotConfigured) }) t.Run("provider not found", func(t *testing.T) { - _, err := fakeAPI.SubtleDecryptAlpha1(context.Background(), &runtimev1pb.SubtleDecryptRequest{ + _, err := fakeAPI.SubtleDecryptAlpha1(t.Context(), &runtimev1pb.SubtleDecryptRequest{ ComponentName: "notfound", }) require.Error(t, err) @@ -238,7 +237,7 @@ func TestSubtleDecryptAlpha1(t *testing.T) { }) t.Run("failed to decrypt", func(t *testing.T) { - _, err := fakeAPI.SubtleDecryptAlpha1(context.Background(), &runtimev1pb.SubtleDecryptRequest{ + _, err := fakeAPI.SubtleDecryptAlpha1(t.Context(), &runtimev1pb.SubtleDecryptRequest{ ComponentName: "myvault", KeyName: "error", }) @@ -260,7 +259,7 @@ func TestSubtleWrapKeyAlpha1(t *testing.T) { } t.Run("wrap key", func(t *testing.T) { - res, err := fakeAPI.SubtleWrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleWrapKeyRequest{ + res, err := fakeAPI.SubtleWrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleWrapKeyRequest{ ComponentName: "myvault", PlaintextKey: []byte("hello world"), KeyName: "good-tag", @@ -277,13 +276,13 @@ func TestSubtleWrapKeyAlpha1(t *testing.T) { compStore.AddCryptoProvider("myvault", fakeCryptoProvider) }() - _, err := fakeAPI.SubtleWrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleWrapKeyRequest{}) + _, err := fakeAPI.SubtleWrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleWrapKeyRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrCryptoProvidersNotConfigured) }) t.Run("provider not found", func(t *testing.T) { - _, err := fakeAPI.SubtleWrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleWrapKeyRequest{ + _, err := fakeAPI.SubtleWrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleWrapKeyRequest{ ComponentName: "notfound", }) require.Error(t, err) @@ -291,7 +290,7 @@ func TestSubtleWrapKeyAlpha1(t *testing.T) { }) t.Run("key is empty", func(t *testing.T) { - _, err := fakeAPI.SubtleWrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleWrapKeyRequest{ + _, err := fakeAPI.SubtleWrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleWrapKeyRequest{ ComponentName: "myvault", KeyName: "error", }) @@ -301,7 +300,7 @@ func TestSubtleWrapKeyAlpha1(t *testing.T) { }) t.Run("failed to wrap key", func(t *testing.T) { - _, err := fakeAPI.SubtleWrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleWrapKeyRequest{ + _, err := fakeAPI.SubtleWrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleWrapKeyRequest{ ComponentName: "myvault", KeyName: "error", PlaintextKey: oneHundredTwentyEightBits, @@ -324,7 +323,7 @@ func TestSubtleUnwrapKeyAlpha1(t *testing.T) { } t.Run("unwrap key", func(t *testing.T) { - res, err := fakeAPI.SubtleUnwrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleUnwrapKeyRequest{ + res, err := fakeAPI.SubtleUnwrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleUnwrapKeyRequest{ ComponentName: "myvault", WrappedKey: []byte("hello world"), KeyName: "good", @@ -341,13 +340,13 @@ func TestSubtleUnwrapKeyAlpha1(t *testing.T) { compStore.AddCryptoProvider("myvault", fakeCryptoProvider) }() - _, err := fakeAPI.SubtleUnwrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleUnwrapKeyRequest{}) + _, err := fakeAPI.SubtleUnwrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleUnwrapKeyRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrCryptoProvidersNotConfigured) }) t.Run("provider not found", func(t *testing.T) { - _, err := fakeAPI.SubtleUnwrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleUnwrapKeyRequest{ + _, err := fakeAPI.SubtleUnwrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleUnwrapKeyRequest{ ComponentName: "notfound", }) require.Error(t, err) @@ -355,7 +354,7 @@ func TestSubtleUnwrapKeyAlpha1(t *testing.T) { }) t.Run("failed to unwrap key", func(t *testing.T) { - _, err := fakeAPI.SubtleUnwrapKeyAlpha1(context.Background(), &runtimev1pb.SubtleUnwrapKeyRequest{ + _, err := fakeAPI.SubtleUnwrapKeyAlpha1(t.Context(), &runtimev1pb.SubtleUnwrapKeyRequest{ ComponentName: "myvault", KeyName: "error", WrappedKey: oneHundredTwentyEightBits, @@ -378,7 +377,7 @@ func TestSubtleSignAlpha1(t *testing.T) { } t.Run("sign message", func(t *testing.T) { - res, err := fakeAPI.SubtleSignAlpha1(context.Background(), &runtimev1pb.SubtleSignRequest{ + res, err := fakeAPI.SubtleSignAlpha1(t.Context(), &runtimev1pb.SubtleSignRequest{ ComponentName: "myvault", Digest: []byte("hello world"), KeyName: "good", @@ -395,13 +394,13 @@ func TestSubtleSignAlpha1(t *testing.T) { compStore.AddCryptoProvider("myvault", fakeCryptoProvider) }() - _, err := fakeAPI.SubtleSignAlpha1(context.Background(), &runtimev1pb.SubtleSignRequest{}) + _, err := fakeAPI.SubtleSignAlpha1(t.Context(), &runtimev1pb.SubtleSignRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrCryptoProvidersNotConfigured) }) t.Run("provider not found", func(t *testing.T) { - _, err := fakeAPI.SubtleSignAlpha1(context.Background(), &runtimev1pb.SubtleSignRequest{ + _, err := fakeAPI.SubtleSignAlpha1(t.Context(), &runtimev1pb.SubtleSignRequest{ ComponentName: "notfound", }) require.Error(t, err) @@ -409,7 +408,7 @@ func TestSubtleSignAlpha1(t *testing.T) { }) t.Run("failed to sign", func(t *testing.T) { - _, err := fakeAPI.SubtleSignAlpha1(context.Background(), &runtimev1pb.SubtleSignRequest{ + _, err := fakeAPI.SubtleSignAlpha1(t.Context(), &runtimev1pb.SubtleSignRequest{ ComponentName: "myvault", KeyName: "error", Digest: oneHundredTwentyEightBits, @@ -432,7 +431,7 @@ func TestSubtleVerifyAlpha1(t *testing.T) { } t.Run("signature is valid", func(t *testing.T) { - res, err := fakeAPI.SubtleVerifyAlpha1(context.Background(), &runtimev1pb.SubtleVerifyRequest{ + res, err := fakeAPI.SubtleVerifyAlpha1(t.Context(), &runtimev1pb.SubtleVerifyRequest{ ComponentName: "myvault", Digest: oneHundredTwentyEightBits, Signature: oneHundredTwentyEightBits, @@ -444,7 +443,7 @@ func TestSubtleVerifyAlpha1(t *testing.T) { }) t.Run("signature is invalid", func(t *testing.T) { - res, err := fakeAPI.SubtleVerifyAlpha1(context.Background(), &runtimev1pb.SubtleVerifyRequest{ + res, err := fakeAPI.SubtleVerifyAlpha1(t.Context(), &runtimev1pb.SubtleVerifyRequest{ ComponentName: "myvault", Digest: oneHundredTwentyEightBits, Signature: oneHundredTwentyEightBits, @@ -461,13 +460,13 @@ func TestSubtleVerifyAlpha1(t *testing.T) { compStore.AddCryptoProvider("myvault", fakeCryptoProvider) }() - _, err := fakeAPI.SubtleVerifyAlpha1(context.Background(), &runtimev1pb.SubtleVerifyRequest{}) + _, err := fakeAPI.SubtleVerifyAlpha1(t.Context(), &runtimev1pb.SubtleVerifyRequest{}) require.Error(t, err) require.ErrorIs(t, err, messages.ErrCryptoProvidersNotConfigured) }) t.Run("provider not found", func(t *testing.T) { - _, err := fakeAPI.SubtleVerifyAlpha1(context.Background(), &runtimev1pb.SubtleVerifyRequest{ + _, err := fakeAPI.SubtleVerifyAlpha1(t.Context(), &runtimev1pb.SubtleVerifyRequest{ ComponentName: "notfound", }) require.Error(t, err) @@ -475,7 +474,7 @@ func TestSubtleVerifyAlpha1(t *testing.T) { }) t.Run("failed to verify", func(t *testing.T) { - _, err := fakeAPI.SubtleVerifyAlpha1(context.Background(), &runtimev1pb.SubtleVerifyRequest{ + _, err := fakeAPI.SubtleVerifyAlpha1(t.Context(), &runtimev1pb.SubtleVerifyRequest{ ComponentName: "myvault", KeyName: "error", }) diff --git a/pkg/api/universal/workflow_test.go b/pkg/api/universal/workflow_test.go index 42ff6e587..18588be20 100644 --- a/pkg/api/universal/workflow_test.go +++ b/pkg/api/universal/workflow_test.go @@ -101,7 +101,7 @@ func TestStartWorkflowAPI(t *testing.T) { InstanceId: tt.instanceID, WorkflowName: tt.workflowName, } - _, err := fakeAPI.StartWorkflow(context.Background(), req) + _, err := fakeAPI.StartWorkflow(t.Context(), req) if tt.expectedError == nil { require.NoError(t, err) @@ -146,7 +146,7 @@ func TestGetWorkflowAPI(t *testing.T) { WorkflowComponent: tt.workflowComponent, InstanceId: tt.instanceID, } - _, err := fakeAPI.GetWorkflow(context.Background(), req) + _, err := fakeAPI.GetWorkflow(t.Context(), req) if tt.expectedError == nil { require.NoError(t, err) @@ -199,7 +199,7 @@ func TestTerminateWorkflowAPI(t *testing.T) { WorkflowComponent: tt.workflowComponent, InstanceId: tt.instanceID, } - _, err := fakeAPI.TerminateWorkflow(context.Background(), req) + _, err := fakeAPI.TerminateWorkflow(t.Context(), req) if tt.expectedError == nil { require.NoError(t, err) @@ -258,7 +258,7 @@ func TestRaiseEventWorkflowApi(t *testing.T) { EventName: tt.eventName, EventData: []byte("fake_input"), } - _, err := fakeAPI.RaiseEventWorkflow(context.Background(), req) + _, err := fakeAPI.RaiseEventWorkflow(t.Context(), req) if tt.expectedError == nil { require.NoError(t, err) @@ -311,7 +311,7 @@ func TestPauseWorkflowApi(t *testing.T) { WorkflowComponent: tt.workflowComponent, InstanceId: tt.instanceID, } - _, err := fakeAPI.PauseWorkflow(context.Background(), req) + _, err := fakeAPI.PauseWorkflow(t.Context(), req) if tt.expectedError == nil { require.NoError(t, err) @@ -356,7 +356,7 @@ func TestResumeWorkflowApi(t *testing.T) { WorkflowComponent: tt.workflowComponent, InstanceId: tt.instanceID, } - _, err := fakeAPI.ResumeWorkflow(context.Background(), req) + _, err := fakeAPI.ResumeWorkflow(t.Context(), req) if tt.expectedError == nil { require.NoError(t, err) diff --git a/pkg/apis/common/namevalue.go b/pkg/apis/common/namevalue.go index 129ad4fe4..7c6623db8 100644 --- a/pkg/apis/common/namevalue.go +++ b/pkg/apis/common/namevalue.go @@ -37,7 +37,7 @@ type NameValuePair struct { } // HasValue returns true if the NameValuePair has a non-empty value. -func (nvp NameValuePair) HasValue() bool { +func (nvp *NameValuePair) HasValue() bool { return len(nvp.Value.JSON.Raw) > 0 } diff --git a/pkg/apis/common/scoped.go b/pkg/apis/common/scoped.go index 164d29f22..b28fe9e48 100644 --- a/pkg/apis/common/scoped.go +++ b/pkg/apis/common/scoped.go @@ -22,7 +22,7 @@ type Scoped struct { } // IsAppScoped returns true if the appID is allowed in the scopes for the resource. -func (s Scoped) IsAppScoped(appID string) bool { +func (s *Scoped) IsAppScoped(appID string) bool { if len(s.Scopes) == 0 { // If there are no scopes, then every app is allowed return true diff --git a/pkg/apis/common/zz_generated.deepcopy.go b/pkg/apis/common/zz_generated.deepcopy.go index e04771dc8..f150ff076 100644 --- a/pkg/apis/common/zz_generated.deepcopy.go +++ b/pkg/apis/common/zz_generated.deepcopy.go @@ -20,8 +20,6 @@ limitations under the License. package common -import () - // DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil. func (in *DynamicValue) DeepCopyInto(out *DynamicValue) { *out = *in diff --git a/pkg/apis/components/v1alpha1/types.go b/pkg/apis/components/v1alpha1/types.go index b7b6b0779..cace49d2b 100644 --- a/pkg/apis/components/v1alpha1/types.go +++ b/pkg/apis/components/v1alpha1/types.go @@ -32,6 +32,8 @@ const ( //+kubebuilder:object:root=true // Component describes an Dapr component type. +// +//nolint:recvcheck type Component struct { metav1.TypeMeta `json:",inline"` //+optional diff --git a/pkg/apis/httpEndpoint/v1alpha1/types.go b/pkg/apis/httpEndpoint/v1alpha1/types.go index 855e6a184..a20f7a70b 100644 --- a/pkg/apis/httpEndpoint/v1alpha1/types.go +++ b/pkg/apis/httpEndpoint/v1alpha1/types.go @@ -32,6 +32,8 @@ const ( // HTTPEndpoint describes a Dapr HTTPEndpoint type for external service invocation. // This endpoint can be external to Dapr, or external to the environment. +// +//nolint:recvcheck type HTTPEndpoint struct { metav1.TypeMeta `json:",inline"` //+optional diff --git a/pkg/apis/resiliency/v1alpha1/types.go b/pkg/apis/resiliency/v1alpha1/types.go index 2c83e3817..7b71f42fd 100644 --- a/pkg/apis/resiliency/v1alpha1/types.go +++ b/pkg/apis/resiliency/v1alpha1/types.go @@ -34,7 +34,7 @@ type Resiliency struct { } // String implements fmt.Stringer and is used for debugging. It returns the policy object encoded as JSON. -func (r Resiliency) String() string { +func (r *Resiliency) String() string { b, _ := json.Marshal(r) return string(b) } diff --git a/pkg/apis/subscriptions/v1alpha1/types.go b/pkg/apis/subscriptions/v1alpha1/types.go index d83ab23e2..7df7d791d 100644 --- a/pkg/apis/subscriptions/v1alpha1/types.go +++ b/pkg/apis/subscriptions/v1alpha1/types.go @@ -31,6 +31,8 @@ const ( // +kubebuilder:object:root=true // Subscription describes an pub/sub event subscription. +// +//nolint:recvcheck type Subscription struct { metav1.TypeMeta `json:",inline"` // +optional diff --git a/pkg/apis/subscriptions/v2alpha1/types.go b/pkg/apis/subscriptions/v2alpha1/types.go index 7e7ab0924..ba917c8d1 100644 --- a/pkg/apis/subscriptions/v2alpha1/types.go +++ b/pkg/apis/subscriptions/v2alpha1/types.go @@ -32,6 +32,8 @@ const ( // +kubebuilder:storageversion // Subscription describes an pub/sub event subscription. +// +//nolint:recvcheck type Subscription struct { metav1.TypeMeta `json:",inline"` // +optional diff --git a/pkg/apphealth/health_test.go b/pkg/apphealth/health_test.go index 2c74ed539..5691152d7 100644 --- a/pkg/apphealth/health_test.go +++ b/pkg/apphealth/health_test.go @@ -35,7 +35,7 @@ func TestAppHealth_setResult(t *testing.T) { }, nil) // Set the initial state to healthy - h.setResult(context.Background(), true) + h.setResult(t.Context(), true) statusChange := make(chan uint8, 1) unexpectedStatusChanges := atomic.Int32{} @@ -55,7 +55,7 @@ func TestAppHealth_setResult(t *testing.T) { if i == threshold-1 { <-statusChange // Allow the channel to be written into } - h.setResult(context.Background(), false) + h.setResult(t.Context(), false) if i == threshold-1 { select { case v := <-statusChange: @@ -77,7 +77,7 @@ func TestAppHealth_setResult(t *testing.T) { // First success should bring the app back to healthy <-statusChange // Allow the channel to be written into - h.setResult(context.Background(), true) + h.setResult(t.Context(), true) select { case v := <-statusChange: assert.Equal(t, AppStatusHealthy, v) @@ -93,7 +93,7 @@ func TestAppHealth_setResult(t *testing.T) { wg.Add(1) go func() { for range threshold + 5 { - h.setResult(context.Background(), false) + h.setResult(t.Context(), false) } wg.Done() }() @@ -114,7 +114,7 @@ func TestAppHealth_setResult(t *testing.T) { h.failureCount.Store(int32(math.MaxInt32 - 2)) statusChange <- 255 // Fill the channel again for range 5 { - h.setResult(context.Background(), false) + h.setResult(t.Context(), false) } assert.Empty(t, unexpectedStatusChanges.Load()) assert.Equal(t, threshold+3, h.failureCount.Load()) @@ -172,7 +172,7 @@ func TestAppHealth_ratelimitReports(t *testing.T) { func Test_StartProbes(t *testing.T) { t.Run("closing context should return", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) done := make(chan struct{}) @@ -203,7 +203,7 @@ func Test_StartProbes(t *testing.T) { }) t.Run("calling StartProbes after it has already closed should error", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) h := New(config.AppHealthConfig{ @@ -229,7 +229,7 @@ func Test_StartProbes(t *testing.T) { }) t.Run("should return after closed", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) h := New(config.AppHealthConfig{ @@ -260,7 +260,7 @@ func Test_StartProbes(t *testing.T) { }) t.Run("should call app probe function after interval", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) var probeCalls atomic.Int64 diff --git a/pkg/channel/grpc/grpc_channel_test.go b/pkg/channel/grpc/grpc_channel_test.go index 2c5a45dd9..a6fb048f2 100644 --- a/pkg/channel/grpc/grpc_channel_test.go +++ b/pkg/channel/grpc/grpc_channel_test.go @@ -74,7 +74,7 @@ func TestMain(m *testing.M) { } func createConnection(t *testing.T) *grpc.ClientConn { - ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second) + ctx, cancel := context.WithTimeout(t.Context(), 2*time.Second) conn, err := grpc.DialContext(ctx, "localhost:9998", //nolint:staticcheck grpc.WithTransportCredentials(insecure.NewCredentials()), grpc.WithBlock(), //nolint:staticcheck @@ -99,7 +99,7 @@ func TestInvokeMethod(t *testing.T) { appMetadataToken: "token1", maxRequestBodySize: 4 << 20, } - ctx := context.Background() + ctx := t.Context() t.Run("successful request", func(t *testing.T) { req := invokev1.NewInvokeMethodRequest("method"). @@ -145,7 +145,7 @@ func TestHealthProbe(t *testing.T) { appMetadataToken: "token1", maxRequestBodySize: 4 << 20, } - ctx := context.Background() + ctx := t.Context() var ( success bool diff --git a/pkg/channel/http/http_channel_test.go b/pkg/channel/http/http_channel_test.go index 6abda157c..a820f9c48 100644 --- a/pkg/channel/http/http_channel_test.go +++ b/pkg/channel/http/http_channel_test.go @@ -14,7 +14,6 @@ limitations under the License. package http import ( - "context" "encoding/json" "io" "net/http" @@ -143,7 +142,7 @@ func (t *testUppercaseHandler) ServeHTTP(w http.ResponseWriter, r *http.Request) func TestInvokeMethodMiddlewaresPipeline(t *testing.T) { var th http.Handler = &testStatusCodeHandler{Code: http.StatusOK} server := httptest.NewServer(th) - ctx := context.Background() + ctx := t.Context() t.Run("pipeline should be called when handlers are not empty", func(t *testing.T) { called := 0 @@ -403,7 +402,7 @@ func TestInvokeMethodMiddlewaresPipeline(t *testing.T) { func TestInvokeMethodHeaders(t *testing.T) { th := &testHeadersHandler{} - ctx := context.Background() + ctx := t.Context() server := httptest.NewServer(th) defer server.Close() @@ -467,7 +466,7 @@ func TestInvokeMethodHeaders(t *testing.T) { func TestInvokeMethod(t *testing.T) { th := &testQueryStringHandler{t: t, serverURL: ""} - ctx := context.Background() + ctx := t.Context() server := httptest.NewServer(th) defer server.Close() @@ -523,7 +522,7 @@ func TestInvokeMethod(t *testing.T) { } func TestInvokeMethodMaxConcurrency(t *testing.T) { - ctx := context.Background() + ctx := t.Context() t.Run("single concurrency", func(t *testing.T) { handler := testConcurrencyHandler{ maxCalls: 1, @@ -638,7 +637,7 @@ func TestInvokeMethodMaxConcurrency(t *testing.T) { } func TestInvokeWithHeaders(t *testing.T) { - ctx := context.Background() + ctx := t.Context() testServer := httptest.NewServer(&testHandlerHeaders{}) c := Channel{ baseAddress: testServer.URL, @@ -673,7 +672,7 @@ func TestInvokeWithHeaders(t *testing.T) { } func TestContentType(t *testing.T) { - ctx := context.Background() + ctx := t.Context() t.Run("no default content type", func(t *testing.T) { handler := &testContentTypeHandler{} @@ -754,7 +753,7 @@ func TestContentType(t *testing.T) { } func TestContentLength(t *testing.T) { - ctx := context.Background() + ctx := t.Context() handler := &testHandlerHeaders{} testServer := httptest.NewServer(handler) @@ -788,7 +787,7 @@ func TestContentLength(t *testing.T) { func TestAppToken(t *testing.T) { t.Run("token present", func(t *testing.T) { - ctx := context.Background() + ctx := t.Context() testServer := httptest.NewServer(&testHandlerHeaders{}) c := Channel{ baseAddress: testServer.URL, @@ -820,7 +819,7 @@ func TestAppToken(t *testing.T) { }) t.Run("token not present", func(t *testing.T) { - ctx := context.Background() + ctx := t.Context() testServer := httptest.NewServer(&testHandlerHeaders{}) c := Channel{ baseAddress: testServer.URL, @@ -852,7 +851,7 @@ func TestAppToken(t *testing.T) { } func TestHealthProbe(t *testing.T) { - ctx := context.Background() + ctx := t.Context() h := &testStatusCodeHandler{} testServer := httptest.NewServer(h) c := Channel{ @@ -887,7 +886,7 @@ func TestHealthProbe(t *testing.T) { } func TestNoInvalidTraceContext(t *testing.T) { - ctx := context.Background() + ctx := t.Context() handler := &testHandlerHeaders{} testServer := httptest.NewServer(handler) diff --git a/pkg/channel/http/rwrecorder.go b/pkg/channel/http/rwrecorder.go index 4c750f108..72a6841fb 100644 --- a/pkg/channel/http/rwrecorder.go +++ b/pkg/channel/http/rwrecorder.go @@ -107,6 +107,5 @@ func parseContentLength(cl string) int64 { if err != nil { return -1 } - //nolint:gosec return int64(n) } diff --git a/pkg/components/bindings/input_pluggable_test.go b/pkg/components/bindings/input_pluggable_test.go index 4ab76d3ff..b8e81f1a6 100644 --- a/pkg/components/bindings/input_pluggable_test.go +++ b/pkg/components/bindings/input_pluggable_test.go @@ -125,7 +125,7 @@ func TestInputBindingCalls(t *testing.T) { }() conn := inputFromConnector(testLogger, connector) - err = conn.Init(context.Background(), bindings.Metadata{ + err = conn.Init(t.Context(), bindings.Metadata{ Base: contribMetadata.Base{}, }) @@ -180,7 +180,7 @@ func TestInputBindingCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = binding.Read(context.Background(), func(_ context.Context, resp *bindings.ReadResponse) ([]byte, error) { + err = binding.Read(t.Context(), func(_ context.Context, resp *bindings.ReadResponse) ([]byte, error) { handleCalled.Add(1) messagesProcessed.Done() assert.Contains(t, messagesData, resp.Data) diff --git a/pkg/components/bindings/output_pluggable_test.go b/pkg/components/bindings/output_pluggable_test.go index 6b8f3ae3d..be0a56e17 100644 --- a/pkg/components/bindings/output_pluggable_test.go +++ b/pkg/components/bindings/output_pluggable_test.go @@ -127,7 +127,7 @@ func TestOutputBindingCalls(t *testing.T) { }() conn := outputFromConnector(testLogger, connector) - err = conn.Init(context.Background(), bindings.Metadata{ + err = conn.Init(t.Context(), bindings.Metadata{ Base: contribMetadata.Base{}, }) @@ -171,7 +171,7 @@ func TestOutputBindingCalls(t *testing.T) { defer cleanup() require.NoError(t, err) - resp, err := outputSvc.Invoke(context.Background(), &bindings.InvokeRequest{ + resp, err := outputSvc.Invoke(t.Context(), &bindings.InvokeRequest{ Data: fakeDataReq, Metadata: fakeMetadata, Operation: fakeOp, @@ -194,7 +194,7 @@ func TestOutputBindingCalls(t *testing.T) { defer cleanup() require.NoError(t, err) - _, err = outputSvc.Invoke(context.Background(), &bindings.InvokeRequest{}) + _, err = outputSvc.Invoke(t.Context(), &bindings.InvokeRequest{}) require.Error(t, err) assert.Equal(t, int64(1), srv.invokeCalled.Load()) diff --git a/pkg/components/loader/localloader_test.go b/pkg/components/loader/localloader_test.go index 15f1c4818..6e46873d0 100644 --- a/pkg/components/loader/localloader_test.go +++ b/pkg/components/loader/localloader_test.go @@ -14,7 +14,6 @@ limitations under the License. package loader import ( - "context" "io/fs" "os" "path/filepath" @@ -36,7 +35,7 @@ spec: `), fs.FileMode(0o600))) loader := NewLocalLoader("", []string{tmp}) - components, err := loader.Load(context.Background()) + components, err := loader.Load(t.Context()) require.NoError(t, err) require.Len(t, components, 1) require.Equal(t, "statestore", components[0].Name) @@ -55,14 +54,14 @@ spec: `), fs.FileMode(0o600))) loader := NewLocalLoader("", []string{tmp}) - components, err := loader.Load(context.Background()) + components, err := loader.Load(t.Context()) require.NoError(t, err) require.Empty(t, components) }) t.Run("Test Non Existent Directory", func(t *testing.T) { loader := NewLocalLoader("", []string{"/non-existent-directory"}) - _, err := loader.Load(context.Background()) + _, err := loader.Load(t.Context()) require.Error(t, err) }) } @@ -80,7 +79,7 @@ spec: `), fs.FileMode(0o600))) loader := NewLocalLoader("", []string{tmp}) - err := loader.Validate(context.Background()) + err := loader.Validate(t.Context()) require.NoError(t, err) }) @@ -97,13 +96,13 @@ spec: `), fs.FileMode(0o600))) loader := NewLocalLoader("", []string{tmp}) - err := loader.Validate(context.Background()) + err := loader.Validate(t.Context()) require.NoError(t, err) }) t.Run("Test Validate Non Existent Directory", func(t *testing.T) { loader := NewLocalLoader("", []string{"/non-existent-directory"}) - err := loader.Validate(context.Background()) + err := loader.Validate(t.Context()) require.Error(t, err) }) } diff --git a/pkg/components/pluggable/grpc_test.go b/pkg/components/pluggable/grpc_test.go index 3a3dbd0d6..3eb132986 100644 --- a/pkg/components/pluggable/grpc_test.go +++ b/pkg/components/pluggable/grpc_test.go @@ -115,7 +115,7 @@ func TestGRPCConnector(t *testing.T) { assert.Contains(t, acceptedStatus, connector.conn.GetState()) assert.Equal(t, 1, fakeFactoryCalled) - require.NoError(t, connector.conn.Invoke(context.Background(), fmt.Sprintf("/%s/%s", fakeSvcName, fakeMethodName), structpb.NewNullValue(), structpb.NewNullValue())) + require.NoError(t, connector.conn.Invoke(t.Context(), fmt.Sprintf("/%s/%s", fakeSvcName, fakeMethodName), structpb.NewNullValue(), structpb.NewNullValue())) assert.Equal(t, 1, handlerCalled) }) @@ -173,7 +173,7 @@ func TestGRPCConnector(t *testing.T) { require.NoError(t, connector.Dial("")) defer connector.Close() - ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second) + ctx, cancel := context.WithTimeout(t.Context(), 5*time.Second) defer cancel() assert.True(t, connector.conn.WaitForStateChange(ctx, connectivity.Idle)) diff --git a/pkg/components/pubsub/pluggable_test.go b/pkg/components/pubsub/pluggable_test.go index 79d43d407..3563668d6 100644 --- a/pkg/components/pubsub/pluggable_test.go +++ b/pkg/components/pubsub/pluggable_test.go @@ -149,7 +149,7 @@ func TestPubSubPluggableCalls(t *testing.T) { }() ps := fromConnector(testLogger, connector) - err = ps.Init(context.Background(), pubsub.Metadata{ + err = ps.Init(t.Context(), pubsub.Metadata{ Base: contribMetadata.Base{}, }) @@ -181,7 +181,7 @@ func TestPubSubPluggableCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = ps.Publish(context.Background(), &pubsub.PublishRequest{ + err = ps.Publish(t.Context(), &pubsub.PublishRequest{ Topic: fakeTopic, }) @@ -202,7 +202,7 @@ func TestPubSubPluggableCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = ps.Publish(context.Background(), &pubsub.PublishRequest{ + err = ps.Publish(t.Context(), &pubsub.PublishRequest{ Topic: fakeTopic, }) @@ -265,7 +265,7 @@ func TestPubSubPluggableCalls(t *testing.T) { handleErrors <- errors.New("fake-error") close(handleErrors) - err = ps.Subscribe(context.Background(), pubsub.SubscribeRequest{ + err = ps.Subscribe(t.Context(), pubsub.SubscribeRequest{ Topic: fakeTopic, }, func(_ context.Context, m *pubsub.NewMessage) error { handleCalled.Add(1) diff --git a/pkg/components/secretstores/pluggable_test.go b/pkg/components/secretstores/pluggable_test.go index d285162e6..9db61bb43 100644 --- a/pkg/components/secretstores/pluggable_test.go +++ b/pkg/components/secretstores/pluggable_test.go @@ -130,7 +130,7 @@ func TestComponentCalls(t *testing.T) { }() secretStore := fromConnector(testLogger, connector) - err = secretStore.Init(context.Background(), secretstores.Metadata{ + err = secretStore.Init(t.Context(), secretstores.Metadata{ Base: contribMetadata.Base{}, }) require.NoError(t, err) @@ -162,7 +162,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := secretStore.GetSecret(context.Background(), secretstores.GetSecretRequest{ + resp, err := secretStore.GetSecret(t.Context(), secretstores.GetSecretRequest{ Name: key, }) assert.Equal(t, int64(1), svc.getSecretCalled.Load()) @@ -183,7 +183,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := gSecretStores.BulkGetSecret(context.Background(), secretstores.BulkGetSecretRequest{}) + resp, err := gSecretStores.BulkGetSecret(t.Context(), secretstores.BulkGetSecretRequest{}) assert.Equal(t, int64(1), svc.bulkGetSecretCalled.Load()) str := err.Error() assert.Equal(t, err.Error(), str) diff --git a/pkg/components/state/bulk_test.go b/pkg/components/state/bulk_test.go index 41b90444d..3ff644e4e 100644 --- a/pkg/components/state/bulk_test.go +++ b/pkg/components/state/bulk_test.go @@ -69,7 +69,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("no error", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, func(ctx context.Context, req *state.SetRequest) error { count.Add(1) return nil @@ -82,7 +82,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("does not retry on etag error", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, func(ctx context.Context, req *state.SetRequest) error { count.Add(1) return etagInvalidErr @@ -97,7 +97,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("retries on other errors", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, func(ctx context.Context, req *state.SetRequest) error { count.Add(1) return simulatedErr @@ -110,7 +110,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("success on second attempt", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, func(ctx context.Context, req *state.SetRequest) error { if count.Add(1) == 1 { return simulatedErr @@ -132,7 +132,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("all successful", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, nil, // The single method should not be invoked, so this will panic if it happens func(ctx context.Context, req []state.SetRequest, opts state.BulkStoreOpts) error { count.Add(1) @@ -145,7 +145,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("key1 successful, key2 etag mismatch", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, nil, // The single method should not be invoked, so this will panic if it happens func(ctx context.Context, req []state.SetRequest, opts state.BulkStoreOpts) error { count.Add(1) @@ -163,7 +163,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("key1 etag invalid, key2 etag mismatch", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, nil, // The single method should not be invoked, so this will panic if it happens func(ctx context.Context, req []state.SetRequest, opts state.BulkStoreOpts) error { count.Add(1) @@ -183,7 +183,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("key1 successful, key2 fails and is retried", func(t *testing.T) { count := atomic.Uint32{} // This should retry, but the second time only key2 should be requested - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, func(ctx context.Context, req *state.SetRequest) error { require.Equal(t, "key2", req.Key) count.Add(1) @@ -204,7 +204,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("key1 fails and is retried, key2 has etag error", func(t *testing.T) { count := atomic.Uint32{} // This should retry, but the second time only key1 should be requested - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, func(ctx context.Context, req *state.SetRequest) error { require.Equal(t, "key1", req.Key) count.Add(1) @@ -232,7 +232,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { count := atomic.Uint32{} // This should retry, but the second time only key1 should be requested - err := PerformBulkStoreOperation(context.Background(), reqs2, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs2, policyDef, state.BulkStoreOpts{}, func(ctx context.Context, req *state.SetRequest) error { require.Equal(t, "key1", req.Key) count.Add(1) @@ -259,7 +259,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { } count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs2, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs2, policyDef, state.BulkStoreOpts{}, nil, // The single method should not be invoked, so this will panic if it happens func(ctx context.Context, req []state.SetRequest, opts state.BulkStoreOpts) error { if count.Add(1) == 1 { @@ -294,7 +294,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("retries when error is not a multierror", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, nil, // The single method should not be invoked, so this will panic if it happens func(ctx context.Context, req []state.SetRequest, opts state.BulkStoreOpts) error { count.Add(1) @@ -308,7 +308,7 @@ func TestPerformBulkStoreOperation(t *testing.T) { t.Run("retries when multierror contains a non-BulkStoreError error", func(t *testing.T) { count := atomic.Uint32{} - err := PerformBulkStoreOperation(context.Background(), reqs, policyDef, state.BulkStoreOpts{}, + err := PerformBulkStoreOperation(t.Context(), reqs, policyDef, state.BulkStoreOpts{}, nil, // The single method should not be invoked, so this will panic if it happens func(ctx context.Context, req []state.SetRequest, opts state.BulkStoreOpts) error { count.Add(1) diff --git a/pkg/components/state/pluggable_test.go b/pkg/components/state/pluggable_test.go index 082711d58..149924bd4 100644 --- a/pkg/components/state/pluggable_test.go +++ b/pkg/components/state/pluggable_test.go @@ -174,7 +174,7 @@ func TestComponentCalls(t *testing.T) { return dialer(ctx, opts...) }) client := clientFactory(testLogger).(*grpcStateStore) - require.NoError(t, client.Init(context.Background(), state.Metadata{})) + require.NoError(t, client.Init(t.Context(), state.Metadata{})) return client, cleanup, err } @@ -208,7 +208,7 @@ func TestComponentCalls(t *testing.T) { }() ps := fromConnector(testLogger, connector) - err = ps.Init(context.Background(), state.Metadata{ + err = ps.Init(t.Context(), state.Metadata{ Base: contribMetadata.Base{}, }) @@ -241,7 +241,7 @@ func TestComponentCalls(t *testing.T) { stStore, cleanup, err := getStateStore(svc) require.NoError(t, err) defer cleanup() - err = stStore.Delete(context.Background(), &state.DeleteRequest{ + err = stStore.Delete(t.Context(), &state.DeleteRequest{ Key: fakeKey, }) @@ -262,7 +262,7 @@ func TestComponentCalls(t *testing.T) { stStore, cleanup, err := getStateStore(svc) require.NoError(t, err) defer cleanup() - err = stStore.Delete(context.Background(), &state.DeleteRequest{ + err = stStore.Delete(t.Context(), &state.DeleteRequest{ Key: fakeKey, }) @@ -292,7 +292,7 @@ func TestComponentCalls(t *testing.T) { stStore, cleanup, err := getStateStore(svc) require.NoError(t, err) defer cleanup() - err = stStore.Delete(context.Background(), &state.DeleteRequest{ + err = stStore.Delete(t.Context(), &state.DeleteRequest{ Key: fakeKey, }) @@ -325,7 +325,7 @@ func TestComponentCalls(t *testing.T) { stStore, cleanup, err := getStateStore(svc) require.NoError(t, err) defer cleanup() - err = stStore.Delete(context.Background(), &state.DeleteRequest{ + err = stStore.Delete(t.Context(), &state.DeleteRequest{ Key: fakeKey, }) @@ -349,7 +349,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := stStore.Get(context.Background(), &state.GetRequest{ + resp, err := stStore.Get(t.Context(), &state.GetRequest{ Key: fakeKey, }) @@ -370,7 +370,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := stStore.Get(context.Background(), &state.GetRequest{ + resp, err := stStore.Get(t.Context(), &state.GetRequest{ Key: fakeKey, }) @@ -400,7 +400,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := stStore.Get(context.Background(), &state.GetRequest{ + resp, err := stStore.Get(t.Context(), &state.GetRequest{ Key: fakeKey, }) @@ -423,7 +423,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.Set(context.Background(), &state.SetRequest{ + err = stStore.Set(t.Context(), &state.SetRequest{ Key: fakeKey, Value: fakeData, }) @@ -445,7 +445,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.Set(context.Background(), &state.SetRequest{ + err = stStore.Set(t.Context(), &state.SetRequest{ Key: fakeKey, Value: fakeData, }) @@ -488,7 +488,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.BulkSet(context.Background(), []state.SetRequest{}, state.BulkStoreOpts{}) + err = stStore.BulkSet(t.Context(), []state.SetRequest{}, state.BulkStoreOpts{}) require.Error(t, err) assert.Equal(t, int64(1), svc.bulkSetCalled.Load()) @@ -509,7 +509,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.BulkSet(context.Background(), requests, state.BulkStoreOpts{}) + err = stStore.BulkSet(t.Context(), requests, state.BulkStoreOpts{}) require.ErrorIs(t, ErrNilSetValue, err) assert.Equal(t, int64(0), svc.bulkSetCalled.Load()) @@ -536,7 +536,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.BulkSet(context.Background(), requests, state.BulkStoreOpts{}) + err = stStore.BulkSet(t.Context(), requests, state.BulkStoreOpts{}) require.NoError(t, err) assert.Equal(t, int64(1), svc.bulkSetCalled.Load()) @@ -561,7 +561,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.BulkDelete(context.Background(), requests, state.BulkStoreOpts{}) + err = stStore.BulkDelete(t.Context(), requests, state.BulkStoreOpts{}) require.NoError(t, err) assert.Equal(t, int64(1), svc.bulkDeleteCalled.Load()) @@ -583,7 +583,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.BulkDelete(context.Background(), requests, state.BulkStoreOpts{}) + err = stStore.BulkDelete(t.Context(), requests, state.BulkStoreOpts{}) require.Error(t, err) assert.Equal(t, int64(1), svc.bulkDeleteCalled.Load()) @@ -615,7 +615,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.BulkDelete(context.Background(), requests, state.BulkStoreOpts{}) + err = stStore.BulkDelete(t.Context(), requests, state.BulkStoreOpts{}) require.Error(t, err) _, ok := err.(*state.BulkDeleteRowMismatchError) @@ -636,7 +636,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := stStore.BulkGet(context.Background(), requests, state.BulkGetOpts{}) + resp, err := stStore.BulkGet(t.Context(), requests, state.BulkGetOpts{}) require.Error(t, err) assert.Nil(t, resp) @@ -669,7 +669,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := stStore.BulkGet(context.Background(), requests, state.BulkGetOpts{}) + resp, err := stStore.BulkGet(t.Context(), requests, state.BulkGetOpts{}) require.NoError(t, err) assert.NotNil(t, resp) @@ -685,7 +685,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.Multi(context.Background(), &state.TransactionalStateRequest{ + err = stStore.Multi(t.Context(), &state.TransactionalStateRequest{ Operations: []state.TransactionalStateOperation{}, Metadata: map[string]string{}, }) @@ -715,7 +715,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - err = stStore.Multi(context.Background(), &state.TransactionalStateRequest{ + err = stStore.Multi(t.Context(), &state.TransactionalStateRequest{ Operations: []state.TransactionalStateOperation{ operations[0], operations[1], @@ -734,7 +734,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := stStore.Query(context.Background(), &state.QueryRequest{}) + resp, err := stStore.Query(t.Context(), &state.QueryRequest{}) require.Error(t, err) assert.Nil(t, resp) @@ -774,7 +774,7 @@ func TestComponentCalls(t *testing.T) { require.NoError(t, err) defer cleanup() - resp, err := stStore.Query(context.Background(), request) + resp, err := stStore.Query(t.Context(), request) require.NoError(t, err) assert.NotNil(t, resp) diff --git a/pkg/config/configuration.go b/pkg/config/configuration.go index e9dd53dc5..3adcdb698 100644 --- a/pkg/config/configuration.go +++ b/pkg/config/configuration.go @@ -78,6 +78,8 @@ var defaultFeatures = map[Feature]bool{ } // Configuration is an internal (and duplicate) representation of Dapr's Configuration CRD. +// +//nolint:recvcheck type Configuration struct { metav1.TypeMeta `json:",inline" yaml:",inline"` // See https://github.com/kubernetes/community/blob/master/contributors/devel/sig-architecture/api-conventions.md#metadata diff --git a/pkg/diagnostics/component_monitoring_test.go b/pkg/diagnostics/component_monitoring_test.go index f44ca8cfd..e046e27f9 100644 --- a/pkg/diagnostics/component_monitoring_test.go +++ b/pkg/diagnostics/component_monitoring_test.go @@ -1,7 +1,6 @@ package diagnostics import ( - "context" "testing" "time" @@ -26,8 +25,8 @@ func TestPubSub(t *testing.T) { t.Run("record drop by app or sidecar", func(t *testing.T) { c := componentsMetrics() - c.PubsubIngressEvent(context.Background(), componentName, "drop", "success", "A", 1) - c.PubsubIngressEvent(context.Background(), componentName, "drop", "drop", "A", 1) + c.PubsubIngressEvent(t.Context(), componentName, "drop", "success", "A", 1) + c.PubsubIngressEvent(t.Context(), componentName, "drop", "drop", "A", 1) viewData, _ := view.RetrieveData("component/pubsub_ingress/count") v := view.Find("component/pubsub_ingress/count") @@ -42,7 +41,7 @@ func TestPubSub(t *testing.T) { t.Run("record ingress count", func(t *testing.T) { c := componentsMetrics() - c.PubsubIngressEvent(context.Background(), componentName, "retry", "retry", "A", 0) + c.PubsubIngressEvent(t.Context(), componentName, "retry", "retry", "A", 0) viewData, _ := view.RetrieveData("component/pubsub_ingress/count") v := view.Find("component/pubsub_ingress/count") @@ -53,7 +52,7 @@ func TestPubSub(t *testing.T) { t.Run("record ingress latency", func(t *testing.T) { c := componentsMetrics() - c.PubsubIngressEvent(context.Background(), componentName, "retry", "", "A", 1) + c.PubsubIngressEvent(t.Context(), componentName, "retry", "", "A", 1) viewData, _ := view.RetrieveData("component/pubsub_ingress/latencies") v := view.Find("component/pubsub_ingress/latencies") @@ -66,7 +65,7 @@ func TestPubSub(t *testing.T) { t.Run("record egress latency", func(t *testing.T) { c := componentsMetrics() - c.PubsubEgressEvent(context.Background(), componentName, "A", true, 1) + c.PubsubEgressEvent(t.Context(), componentName, "A", true, 1) viewData, _ := view.RetrieveData("component/pubsub_egress/latencies") v := view.Find("component/pubsub_egress/latencies") @@ -81,7 +80,7 @@ func TestBindings(t *testing.T) { t.Run("record input binding count", func(t *testing.T) { c := componentsMetrics() - c.InputBindingEvent(context.Background(), componentName, false, 0) + c.InputBindingEvent(t.Context(), componentName, false, 0) viewData, _ := view.RetrieveData("component/input_binding/count") v := view.Find("component/input_binding/count") @@ -92,7 +91,7 @@ func TestBindings(t *testing.T) { t.Run("record input binding latency", func(t *testing.T) { c := componentsMetrics() - c.InputBindingEvent(context.Background(), componentName, false, 1) + c.InputBindingEvent(t.Context(), componentName, false, 1) viewData, _ := view.RetrieveData("component/input_binding/latencies") v := view.Find("component/input_binding/count") @@ -105,7 +104,7 @@ func TestBindings(t *testing.T) { t.Run("record output binding count", func(t *testing.T) { c := componentsMetrics() - c.OutputBindingEvent(context.Background(), componentName, "set", false, 0) + c.OutputBindingEvent(t.Context(), componentName, "set", false, 0) viewData, _ := view.RetrieveData("component/output_binding/count") v := view.Find("component/input_binding/count") @@ -116,7 +115,7 @@ func TestBindings(t *testing.T) { t.Run("record output binding latency", func(t *testing.T) { c := componentsMetrics() - c.OutputBindingEvent(context.Background(), componentName, "set", false, 1) + c.OutputBindingEvent(t.Context(), componentName, "set", false, 1) viewData, _ := view.RetrieveData("component/output_binding/latencies") v := view.Find("component/output_binding/latencies") @@ -131,7 +130,7 @@ func TestState(t *testing.T) { t.Run("record state count", func(t *testing.T) { c := componentsMetrics() - c.StateInvoked(context.Background(), componentName, "get", false, 0) + c.StateInvoked(t.Context(), componentName, "get", false, 0) viewData, _ := view.RetrieveData("component/state/count") v := view.Find("component/state/count") @@ -142,7 +141,7 @@ func TestState(t *testing.T) { t.Run("record state latency", func(t *testing.T) { c := componentsMetrics() - c.StateInvoked(context.Background(), componentName, "get", false, 1) + c.StateInvoked(t.Context(), componentName, "get", false, 1) viewData, _ := view.RetrieveData("component/state/latencies") v := view.Find("component/state/latencies") @@ -156,7 +155,7 @@ func TestConfiguration(t *testing.T) { t.Run("record configuration count", func(t *testing.T) { c := componentsMetrics() - c.ConfigurationInvoked(context.Background(), componentName, "get", false, 0) + c.ConfigurationInvoked(t.Context(), componentName, "get", false, 0) viewData, _ := view.RetrieveData("component/configuration/count") v := view.Find("component/configuration/count") @@ -167,7 +166,7 @@ func TestConfiguration(t *testing.T) { t.Run("record configuration latency", func(t *testing.T) { c := componentsMetrics() - c.ConfigurationInvoked(context.Background(), componentName, "get", false, 1) + c.ConfigurationInvoked(t.Context(), componentName, "get", false, 1) viewData, _ := view.RetrieveData("component/configuration/latencies") v := view.Find("component/configuration/latencies") @@ -182,7 +181,7 @@ func TestSecrets(t *testing.T) { t.Run("record secret count", func(t *testing.T) { c := componentsMetrics() - c.SecretInvoked(context.Background(), componentName, "get", false, 0) + c.SecretInvoked(t.Context(), componentName, "get", false, 0) viewData, _ := view.RetrieveData("component/secret/count") v := view.Find("component/secret/count") @@ -193,7 +192,7 @@ func TestSecrets(t *testing.T) { t.Run("record secret latency", func(t *testing.T) { c := componentsMetrics() - c.SecretInvoked(context.Background(), componentName, "get", false, 1) + c.SecretInvoked(t.Context(), componentName, "get", false, 1) viewData, _ := view.RetrieveData("component/secret/latencies") v := view.Find("component/secret/latencies") @@ -208,7 +207,7 @@ func TestConversation(t *testing.T) { t.Run("record conversation count", func(t *testing.T) { c := componentsMetrics() - c.ConversationInvoked(context.Background(), componentName, false, 0) + c.ConversationInvoked(t.Context(), componentName, false, 0) viewData, _ := view.RetrieveData("component/conversation/count") v := view.Find("component/conversation/count") @@ -219,7 +218,7 @@ func TestConversation(t *testing.T) { t.Run("record conversation latency", func(t *testing.T) { c := componentsMetrics() - c.ConversationInvoked(context.Background(), componentName, false, 1) + c.ConversationInvoked(t.Context(), componentName, false, 1) viewData, _ := view.RetrieveData("component/conversation/latencies") v := view.Find("component/conversation/latencies") diff --git a/pkg/diagnostics/grpc_tracing_test.go b/pkg/diagnostics/grpc_tracing_test.go index 4e7d60f57..cf517a055 100644 --- a/pkg/diagnostics/grpc_tracing_test.go +++ b/pkg/diagnostics/grpc_tracing_test.go @@ -19,7 +19,6 @@ import ( "encoding/hex" "errors" "fmt" - "strings" "testing" "github.com/stretchr/testify/assert" @@ -75,7 +74,7 @@ func TestUserDefinedMetadata(t *testing.T) { "no-attr": []string{"value3"}, } - testCtx := grpcMetadata.NewIncomingContext(context.Background(), md) + testCtx := grpcMetadata.NewIncomingContext(t.Context(), md) metadata.SetMetadataInContextUnary(testCtx, nil, nil, func(ctx context.Context, req any) (any, error) { testCtx = ctx return nil, nil @@ -90,7 +89,7 @@ func TestUserDefinedMetadata(t *testing.T) { func TestSpanContextToGRPCMetadata(t *testing.T) { t.Run("empty span context", func(t *testing.T) { - ctx := context.Background() + ctx := t.Context() newCtx := SpanContextToGRPCMetadata(ctx, trace.SpanContext{}) assert.Equal(t, ctx, newCtx) @@ -103,7 +102,7 @@ func TestGRPCTraceUnaryServerInterceptor(t *testing.T) { tp := sdktrace.NewTracerProvider( sdktrace.WithBatcher(exp), ) - defer func() { _ = tp.Shutdown(context.Background()) }() + defer func() { _ = tp.Shutdown(t.Context()) }() otel.SetTracerProvider(tp) interceptor := GRPCTraceUnaryServerInterceptor("fakeAppID", config.TracingSpec{SamplingRate: "1"}) @@ -113,7 +112,7 @@ func TestGRPCTraceUnaryServerInterceptor(t *testing.T) { testTraceBinary := diagUtils.BinaryFromSpanContext(testSpanContext) t.Run("grpc-trace-bin is given", func(t *testing.T) { - ctx := grpcMetadata.NewIncomingContext(context.Background(), grpcMetadata.Pairs("grpc-trace-bin", string(testTraceBinary))) + ctx := grpcMetadata.NewIncomingContext(t.Context(), grpcMetadata.Pairs("grpc-trace-bin", string(testTraceBinary))) fakeInfo := &grpc.UnaryServerInfo{ FullMethod: "/dapr.proto.runtime.v1.Dapr/GetState", } @@ -154,7 +153,7 @@ func TestGRPCTraceUnaryServerInterceptor(t *testing.T) { return nil, errors.New("fake error") } - interceptor(context.Background(), fakeReq, fakeInfo, assertHandler) + interceptor(t.Context(), fakeReq, fakeInfo, assertHandler) sc := span.SpanContext() traceID := sc.TraceID() @@ -178,11 +177,11 @@ func TestGRPCTraceUnaryServerInterceptor(t *testing.T) { return nil, errors.New("fake error") } - interceptor(context.Background(), fakeReq, fakeInfo, assertHandler) + interceptor(t.Context(), fakeReq, fakeInfo, assertHandler) sc := span.SpanContext() spanString := fmt.Sprintf("%v", span) - assert.True(t, strings.Contains(spanString, "CallLocal/targetID/method1")) + assert.Contains(t, spanString, "CallLocal/targetID/method1") traceID := sc.TraceID() spanID := sc.SpanID() assert.NotEmpty(t, hex.EncodeToString(traceID[:])) @@ -201,7 +200,7 @@ func TestGRPCTraceUnaryServerInterceptor(t *testing.T) { ) oldTracerProvider := otel.GetTracerProvider() defer func() { - _ = tp.Shutdown(context.Background()) + _ = tp.Shutdown(t.Context()) // reset tracer provider to older one once the test completes otel.SetTracerProvider(oldTracerProvider) }() @@ -222,11 +221,11 @@ func TestGRPCTraceUnaryServerInterceptor(t *testing.T) { return nil, status.Error(codes.Internal, errors.New("fake status error").Error()) } - interceptor(context.Background(), fakeReq, fakeInfo, assertHandler) + interceptor(t.Context(), fakeReq, fakeInfo, assertHandler) sc := span.SpanContext() spanString := fmt.Sprintf("%v", span) - assert.True(t, strings.Contains(spanString, "CallLocal/targetID/method1")) + assert.Contains(t, spanString, "CallLocal/targetID/method1") traceID := sc.TraceID() spanID := sc.SpanID() assert.NotEmpty(t, hex.EncodeToString(traceID[:])) @@ -240,7 +239,7 @@ func TestGRPCTraceStreamServerInterceptor(t *testing.T) { tp := sdktrace.NewTracerProvider( sdktrace.WithBatcher(exp), ) - defer func() { _ = tp.Shutdown(context.Background()) }() + defer func() { _ = tp.Shutdown(t.Context()) }() otel.SetTracerProvider(tp) interceptor := GRPCTraceStreamServerInterceptor("test", config.TracingSpec{SamplingRate: "1"}) @@ -268,7 +267,7 @@ func TestGRPCTraceStreamServerInterceptor(t *testing.T) { FullMethod: "/dapr.proto.runtime.v1.Dapr/GetState", } - ctx := grpcMetadata.NewIncomingContext(context.Background(), grpcMetadata.Pairs("grpc-trace-bin", string(testTraceBinary))) + ctx := grpcMetadata.NewIncomingContext(t.Context(), grpcMetadata.Pairs("grpc-trace-bin", string(testTraceBinary))) ctx, _ = metadata.SetMetadataInTapHandle(ctx, nil) var span trace.Span @@ -326,7 +325,7 @@ func TestGRPCTraceStreamServerInterceptor(t *testing.T) { FullMethod: "/dapr.proto.internals.v1.ServiceInvocation/CallLocal", } - ctx := grpcMetadata.NewIncomingContext(context.Background(), grpcMetadata.Pairs("grpc-trace-bin", string(testTraceBinary))) + ctx := grpcMetadata.NewIncomingContext(t.Context(), grpcMetadata.Pairs("grpc-trace-bin", string(testTraceBinary))) ctx, _ = metadata.SetMetadataInTapHandle(ctx, nil) var span trace.Span @@ -384,7 +383,7 @@ func TestGRPCTraceStreamServerInterceptor(t *testing.T) { GRPCProxyAppIDKey: "myapp", "grpc-trace-bin": string(testTraceBinary), }) - ctx := grpcMetadata.NewIncomingContext(context.Background(), md) + ctx := grpcMetadata.NewIncomingContext(t.Context(), md) ctx, _ = metadata.SetMetadataInTapHandle(ctx, nil) var span trace.Span @@ -411,7 +410,7 @@ func TestGRPCTraceStreamServerInterceptor(t *testing.T) { md := grpcMetadata.New(map[string]string{ GRPCProxyAppIDKey: "myapp", }) - ctx := grpcMetadata.NewIncomingContext(context.Background(), md) + ctx := grpcMetadata.NewIncomingContext(t.Context(), md) ctx, _ = metadata.SetMetadataInTapHandle(ctx, nil) var span trace.Span diff --git a/pkg/diagnostics/http_monitoring_benchmark_test.go b/pkg/diagnostics/http_monitoring_benchmark_test.go index 285f6edbd..c440db613 100644 --- a/pkg/diagnostics/http_monitoring_benchmark_test.go +++ b/pkg/diagnostics/http_monitoring_benchmark_test.go @@ -39,7 +39,7 @@ func BenchmarkHTTPMiddlewareLowCardinalityNoPathMatching(b *testing.B) { })) // act - for i := 0; i < b.N; i++ { + for i := range b.N { testRequest := fakeOrdersHTTPRequest(requestBody, i) handler.ServeHTTP(httptest.NewRecorder(), testRequest) } @@ -56,7 +56,7 @@ func BenchmarkHTTPMiddlewareHighCardinalityNoPathMatching(b *testing.B) { })) // act - for i := 0; i < b.N; i++ { + for i := range b.N { testRequest := fakeOrdersHTTPRequest(requestBody, i) handler.ServeHTTP(httptest.NewRecorder(), testRequest) } @@ -75,7 +75,7 @@ func BenchmarkHTTPMiddlewareLowCardinalityWithPathMatching(b *testing.B) { })) // act - for i := 0; i < b.N; i++ { + for i := range b.N { testRequest := fakeOrdersHTTPRequest(requestBody, i) handler.ServeHTTP(httptest.NewRecorder(), testRequest) } @@ -92,7 +92,7 @@ func BenchmarkHTTPMiddlewareHighCardinalityWithPathMatching(b *testing.B) { })) // act - for i := 0; i < b.N; i++ { + for i := range b.N { testRequest := fakeOrdersHTTPRequest(requestBody, i) handler.ServeHTTP(httptest.NewRecorder(), testRequest) } diff --git a/pkg/diagnostics/http_tracing_test.go b/pkg/diagnostics/http_tracing_test.go index 338e2661f..dfccaadb2 100644 --- a/pkg/diagnostics/http_tracing_test.go +++ b/pkg/diagnostics/http_tracing_test.go @@ -260,7 +260,7 @@ func TestHTTPTraceMiddleware(t *testing.T) { tp := sdktrace.NewTracerProvider( sdktrace.WithBatcher(exp), ) - defer func() { _ = tp.Shutdown(context.Background()) }() + defer func() { _ = tp.Shutdown(t.Context()) }() otel.SetTracerProvider(tp) t.Run("traceparent is given in request and sampling is enabled", func(t *testing.T) { diff --git a/pkg/diagnostics/metrics_regex_test.go b/pkg/diagnostics/metrics_regex_test.go index bf97f8c1f..eebaa4ab0 100644 --- a/pkg/diagnostics/metrics_regex_test.go +++ b/pkg/diagnostics/metrics_regex_test.go @@ -1,7 +1,6 @@ package diagnostics import ( - "context" "testing" "github.com/stretchr/testify/assert" @@ -51,7 +50,7 @@ func TestRegexRulesSingle(t *testing.T) { view.Unregister(view.Find(statName)) }) - stats.RecordWithTags(context.Background(), + stats.RecordWithTags(t.Context(), diagUtils.WithTags(testStat.Name(), methodKey, "/orders/123"), testStat.M(1)) @@ -74,7 +73,7 @@ func TestRegexRulesSingle(t *testing.T) { s := newGRPCMetrics() s.Init("test", nil) - stats.RecordWithTags(context.Background(), + stats.RecordWithTags(t.Context(), diagUtils.WithTags(testStat.Name(), methodKey, "/siths/123"), testStat.M(1)) @@ -97,10 +96,10 @@ func TestRegexRulesSingle(t *testing.T) { s := newGRPCMetrics() s.Init("test", nil) - stats.RecordWithTags(context.Background(), + stats.RecordWithTags(t.Context(), diagUtils.WithTags(testStat.Name(), methodKey, "/orders/123"), testStat.M(1)) - stats.RecordWithTags(context.Background(), + stats.RecordWithTags(t.Context(), diagUtils.WithTags(testStat.Name(), methodKey, "/lightsabers/123"), testStat.M(1)) diff --git a/pkg/diagnostics/resiliency_monitoring_test.go b/pkg/diagnostics/resiliency_monitoring_test.go index d5b0384e9..8f29e99c4 100644 --- a/pkg/diagnostics/resiliency_monitoring_test.go +++ b/pkg/diagnostics/resiliency_monitoring_test.go @@ -216,7 +216,7 @@ func TestResiliencyCountMonitoringCBStates(t *testing.T) { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") for range 2 { policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, nil }) @@ -232,7 +232,7 @@ func TestResiliencyCountMonitoringCBStates(t *testing.T) { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") for range 3 { policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) @@ -251,7 +251,7 @@ func TestResiliencyCountMonitoringCBStates(t *testing.T) { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") for range 3 { policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) @@ -259,7 +259,7 @@ func TestResiliencyCountMonitoringCBStates(t *testing.T) { // let the circuit breaker to go to half open state (5x cb timeout) time.Sleep(500 * time.Millisecond) policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) @@ -336,7 +336,7 @@ func TestResiliencyActivationsCountMonitoring(t *testing.T) { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") for range 2 { policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, nil }) @@ -349,7 +349,7 @@ func TestResiliencyActivationsCountMonitoring(t *testing.T) { unitFn: func() { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) @@ -370,7 +370,7 @@ func TestResiliencyActivationsCountMonitoring(t *testing.T) { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") for range 2 { - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) @@ -392,12 +392,12 @@ func TestResiliencyActivationsCountMonitoring(t *testing.T) { unitFn: func() { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { time.Sleep(500 * time.Millisecond) return nil, errors.New("fake error") }) - policyRunner = resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner = resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) @@ -421,7 +421,7 @@ func TestResiliencyActivationsCountMonitoring(t *testing.T) { r := createTestResiliency(testResiliencyName, testResiliencyNamespace, "fakeStateStore") for range 2 { policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) @@ -429,7 +429,7 @@ func TestResiliencyActivationsCountMonitoring(t *testing.T) { // let the circuit breaker to go to half open state (5x cb timeout) and then return success to close it time.Sleep(1000 * time.Millisecond) policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, nil }) @@ -437,7 +437,7 @@ func TestResiliencyActivationsCountMonitoring(t *testing.T) { // now open the circuit breaker again for range 2 { policyDef := r.EndpointPolicy("fakeApp", "fakeEndpoint") - policyRunner := resiliency.NewRunner[any](context.Background(), policyDef) + policyRunner := resiliency.NewRunner[any](t.Context(), policyDef) _, _ = policyRunner(func(ctx context.Context) (interface{}, error) { return nil, errors.New("fake error") }) diff --git a/pkg/diagnostics/tracing_test.go b/pkg/diagnostics/tracing_test.go index 2798e0b07..44bf5f322 100644 --- a/pkg/diagnostics/tracing_test.go +++ b/pkg/diagnostics/tracing_test.go @@ -120,7 +120,7 @@ func TestStartInternalCallbackSpan(t *testing.T) { tp := sdktrace.NewTracerProvider( sdktrace.WithBatcher(exp), ) - defer func() { _ = tp.Shutdown(context.Background()) }() + defer func() { _ = tp.Shutdown(t.Context()) }() otel.SetTracerProvider(tp) t.Run("traceparent is provided and sampling is enabled", func(t *testing.T) { @@ -133,7 +133,7 @@ func TestStartInternalCallbackSpan(t *testing.T) { } parent := trace.NewSpanContext(scConfig) - ctx := context.Background() + ctx := t.Context() _, gotSp := StartInternalCallbackSpan(ctx, "testSpanName", parent, traceSpec) sc := gotSp.SpanContext() @@ -153,7 +153,7 @@ func TestStartInternalCallbackSpan(t *testing.T) { } parent := trace.NewSpanContext(scConfig) - ctx := context.Background() + ctx := t.Context() ctx, gotSp := StartInternalCallbackSpan(ctx, "testSpanName", parent, traceSpec) assert.Nil(t, gotSp) @@ -221,9 +221,9 @@ func runTraces(t *testing.T, testName string, numTraces int, samplingRate string sampledCount := 0 for range numTraces { - ctx := context.Background() + ctx := t.Context() if hasParentSpanContext { - traceID, _ := idg.NewIDs(context.Background()) + traceID, _ := idg.NewIDs(t.Context()) scConfig := trace.SpanContextConfig{ TraceID: traceID, SpanID: trace.SpanID{0, 240, 103, 170, 11, 169, 2, 183}, @@ -368,7 +368,7 @@ func defaultIDGenerator() IDGenerator { func TestTraceIDAndStateFromSpan(t *testing.T) { t.Run("non-empty span, id and state are not empty", func(t *testing.T) { idg := defaultIDGenerator() - traceID, _ := idg.NewIDs(context.Background()) + traceID, _ := idg.NewIDs(t.Context()) scConfig := trace.SpanContextConfig{ TraceID: traceID, SpanID: trace.SpanID{0, 240, 103, 170, 11, 169, 2, 183}, @@ -380,7 +380,7 @@ func TestTraceIDAndStateFromSpan(t *testing.T) { scConfig.TraceState = ts parent := trace.NewSpanContext(scConfig) - ctx := context.Background() + ctx := t.Context() ctx = trace.ContextWithRemoteSpanContext(ctx, parent) _, span := tracer.Start(ctx, "testTraceSpan", trace.WithSpanKind(trace.SpanKindClient)) @@ -391,7 +391,7 @@ func TestTraceIDAndStateFromSpan(t *testing.T) { }) t.Run("empty span, id and state are empty", func(t *testing.T) { - span := trace.SpanFromContext(context.Background()) + span := trace.SpanFromContext(t.Context()) id, state := TraceIDAndStateFromSpan(span) assert.Empty(t, id) assert.Empty(t, state) diff --git a/pkg/diagnostics/utils/trace_utils_test.go b/pkg/diagnostics/utils/trace_utils_test.go index 19fa549b1..eb824f141 100644 --- a/pkg/diagnostics/utils/trace_utils_test.go +++ b/pkg/diagnostics/utils/trace_utils_test.go @@ -44,7 +44,7 @@ func TestSpanFromContext(t *testing.T) { }) t.Run("not nil span for context", func(t *testing.T) { - ctx := context.Background() + ctx := t.Context() exp := newOtelFakeExporter() tp := sdktrace.NewTracerProvider(sdktrace.WithBatcher(exp)) tracer := tp.Tracer("dapr-diagnostics-utils-tests") @@ -60,7 +60,7 @@ func TestSpanFromContext(t *testing.T) { }) t.Run("nil span for context", func(t *testing.T) { - ctx := context.Background() + ctx := t.Context() exp := newOtelFakeExporter() _ = sdktrace.NewTracerProvider(sdktrace.WithBatcher(exp)) newCtx := trace.ContextWithSpan(ctx, nil) @@ -72,7 +72,7 @@ func TestSpanFromContext(t *testing.T) { }) t.Run("nil", func(t *testing.T) { - ctx := context.Background() + ctx := t.Context() exp := newOtelFakeExporter() _ = sdktrace.NewTracerProvider(sdktrace.WithBatcher(exp)) newCtx := trace.ContextWithSpan(ctx, nil) diff --git a/pkg/diagnostics/workflow_monitoring_test.go b/pkg/diagnostics/workflow_monitoring_test.go index 7456145b0..245ad4509 100644 --- a/pkg/diagnostics/workflow_monitoring_test.go +++ b/pkg/diagnostics/workflow_monitoring_test.go @@ -1,7 +1,6 @@ package diagnostics import ( - "context" "testing" "github.com/stretchr/testify/assert" @@ -25,7 +24,7 @@ func TestOperations(t *testing.T) { t.Run("Failed Create Operation request count", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), CreateWorkflow, StatusFailed, 0) + w.WorkflowOperationEvent(t.Context(), CreateWorkflow, StatusFailed, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -36,7 +35,7 @@ func TestOperations(t *testing.T) { t.Run("Successful Create Operation request count", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), CreateWorkflow, StatusSuccess, 0) + w.WorkflowOperationEvent(t.Context(), CreateWorkflow, StatusSuccess, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -47,7 +46,7 @@ func TestOperations(t *testing.T) { t.Run("Create Operation request latency", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), CreateWorkflow, StatusSuccess, 1) + w.WorkflowOperationEvent(t.Context(), CreateWorkflow, StatusSuccess, 1) viewData, _ := view.RetrieveData(latencyMetricName) v := view.Find(latencyMetricName) @@ -62,7 +61,7 @@ func TestOperations(t *testing.T) { t.Run("Failed Get Operation Request", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), GetWorkflow, StatusFailed, 0) + w.WorkflowOperationEvent(t.Context(), GetWorkflow, StatusFailed, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -73,7 +72,7 @@ func TestOperations(t *testing.T) { t.Run("Successful Get Operation Request", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), GetWorkflow, StatusSuccess, 0) + w.WorkflowOperationEvent(t.Context(), GetWorkflow, StatusSuccess, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -84,7 +83,7 @@ func TestOperations(t *testing.T) { t.Run("Get Operation request latency", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), GetWorkflow, StatusSuccess, 1) + w.WorkflowOperationEvent(t.Context(), GetWorkflow, StatusSuccess, 1) viewData, _ := view.RetrieveData(latencyMetricName) v := view.Find(latencyMetricName) @@ -99,7 +98,7 @@ func TestOperations(t *testing.T) { t.Run("Failed Add Event request", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), AddEvent, StatusFailed, 0) + w.WorkflowOperationEvent(t.Context(), AddEvent, StatusFailed, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -110,7 +109,7 @@ func TestOperations(t *testing.T) { t.Run("Successful Add Event request", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), AddEvent, StatusSuccess, 0) + w.WorkflowOperationEvent(t.Context(), AddEvent, StatusSuccess, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -121,7 +120,7 @@ func TestOperations(t *testing.T) { t.Run("Add Event Operation latency", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), AddEvent, StatusSuccess, 1) + w.WorkflowOperationEvent(t.Context(), AddEvent, StatusSuccess, 1) viewData, _ := view.RetrieveData(latencyMetricName) v := view.Find(latencyMetricName) @@ -136,7 +135,7 @@ func TestOperations(t *testing.T) { t.Run("Failed Purge workflow request", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), PurgeWorkflow, StatusFailed, 0) + w.WorkflowOperationEvent(t.Context(), PurgeWorkflow, StatusFailed, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -147,7 +146,7 @@ func TestOperations(t *testing.T) { t.Run("Successful Purge workflow request", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), PurgeWorkflow, StatusSuccess, 0) + w.WorkflowOperationEvent(t.Context(), PurgeWorkflow, StatusSuccess, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -158,7 +157,7 @@ func TestOperations(t *testing.T) { t.Run("Purge workflow Operation latency", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowOperationEvent(context.Background(), PurgeWorkflow, StatusSuccess, 1) + w.WorkflowOperationEvent(t.Context(), PurgeWorkflow, StatusSuccess, 1) viewData, _ := view.RetrieveData(latencyMetricName) v := view.Find(latencyMetricName) @@ -179,7 +178,7 @@ func TestExecution(t *testing.T) { t.Run("Failed with retryable error", func(t *testing.T) { w := initWorkflowMetrics() - w.ActivityExecutionEvent(context.Background(), activityName, StatusRecoverable, 0) + w.ActivityExecutionEvent(t.Context(), activityName, StatusRecoverable, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -190,7 +189,7 @@ func TestExecution(t *testing.T) { t.Run("Failed with not-retryable error", func(t *testing.T) { w := initWorkflowMetrics() - w.ActivityExecutionEvent(context.Background(), activityName, StatusFailed, 0) + w.ActivityExecutionEvent(t.Context(), activityName, StatusFailed, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -201,7 +200,7 @@ func TestExecution(t *testing.T) { t.Run("Successful activity execution", func(t *testing.T) { w := initWorkflowMetrics() - w.ActivityExecutionEvent(context.Background(), activityName, StatusSuccess, 0) + w.ActivityExecutionEvent(t.Context(), activityName, StatusSuccess, 0) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -212,7 +211,7 @@ func TestExecution(t *testing.T) { t.Run("activity execution latency", func(t *testing.T) { w := initWorkflowMetrics() - w.ActivityExecutionEvent(context.Background(), activityName, StatusSuccess, 1) + w.ActivityExecutionEvent(t.Context(), activityName, StatusSuccess, 1) viewData, _ := view.RetrieveData(latencyMetricName) v := view.Find(latencyMetricName) @@ -231,7 +230,7 @@ func TestExecution(t *testing.T) { t.Run("Failed with retryable error", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowExecutionEvent(context.Background(), workflowName, StatusRecoverable) + w.WorkflowExecutionEvent(t.Context(), workflowName, StatusRecoverable) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -242,7 +241,7 @@ func TestExecution(t *testing.T) { t.Run("Failed with not-retryable error", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowExecutionEvent(context.Background(), workflowName, StatusFailed) + w.WorkflowExecutionEvent(t.Context(), workflowName, StatusFailed) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -253,7 +252,7 @@ func TestExecution(t *testing.T) { t.Run("Successful workflow execution", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowExecutionEvent(context.Background(), workflowName, StatusSuccess) + w.WorkflowExecutionEvent(t.Context(), workflowName, StatusSuccess) viewData, _ := view.RetrieveData(countMetricName) v := view.Find(countMetricName) @@ -264,7 +263,7 @@ func TestExecution(t *testing.T) { t.Run("workflow execution latency", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowExecutionLatency(context.Background(), workflowName, StatusSuccess, 20) + w.WorkflowExecutionLatency(t.Context(), workflowName, StatusSuccess, 20) viewData, _ := view.RetrieveData(executionLatencyMetricName) v := view.Find(executionLatencyMetricName) @@ -276,7 +275,7 @@ func TestExecution(t *testing.T) { t.Run("workflow scheduling latency", func(t *testing.T) { w := initWorkflowMetrics() - w.WorkflowSchedulingLatency(context.Background(), workflowName, 10) + w.WorkflowSchedulingLatency(t.Context(), workflowName, 10) viewData, _ := view.RetrieveData(schedulingLatencyMetricName) v := view.Find(schedulingLatencyMetricName) diff --git a/pkg/encryption/encryption_test.go b/pkg/encryption/encryption_test.go index 19566db31..e245ca91a 100644 --- a/pkg/encryption/encryption_test.go +++ b/pkg/encryption/encryption_test.go @@ -94,7 +94,7 @@ func TestComponentEncryptionKey(t *testing.T) { secondaryKey := hex.EncodeToString(bytes[:16]) // 128-bit key secretStore := &mockSecretStore{} - secretStore.Init(context.Background(), secretstores.Metadata{Base: metadata.Base{ + secretStore.Init(t.Context(), secretstores.Metadata{Base: metadata.Base{ Properties: map[string]string{ "primaryKey": primaryKey, "secondaryKey": secondaryKey, @@ -158,7 +158,7 @@ func TestComponentEncryptionKey(t *testing.T) { func TestTryGetEncryptionKeyFromMetadataItem(t *testing.T) { t.Run("no secretRef on valid item", func(t *testing.T) { secretStore := &mockSecretStore{} - secretStore.Init(context.Background(), secretstores.Metadata{Base: metadata.Base{ + secretStore.Init(t.Context(), secretstores.Metadata{Base: metadata.Base{ Properties: map[string]string{ "primaryKey": "123", "secondaryKey": "456", diff --git a/pkg/expr/expr_test.go b/pkg/expr/expr_test.go index 4544753d2..0a213f3a8 100644 --- a/pkg/expr/expr_test.go +++ b/pkg/expr/expr_test.go @@ -60,7 +60,7 @@ func BenchmarkEval(b *testing.B) { }, } var r interface{} - for n := 0; n < b.N; n++ { + for range b.N { r, _ = e.Eval(data) } result = r diff --git a/pkg/injector/patcher/sidecar_components_test.go b/pkg/injector/patcher/sidecar_components_test.go index 0764b0509..bcd2eaafa 100644 --- a/pkg/injector/patcher/sidecar_components_test.go +++ b/pkg/injector/patcher/sidecar_components_test.go @@ -228,7 +228,7 @@ func TestComponentsPatch(t *testing.T) { patch, volumeMount := c.componentsPatchOps(componentContainers, Injectable(test.appID, test.componentsList)) patchJSON, _ := json.Marshal(patch) expPatchJSON, _ := json.Marshal(test.expPatch) - assert.Equal(t, string(expPatchJSON), string(patchJSON)) + assert.JSONEq(t, string(expPatchJSON), string(patchJSON)) assert.Equal(t, test.expMount, volumeMount) }) } diff --git a/pkg/injector/service/config.go b/pkg/injector/service/config.go index c92702968..e0718d3f8 100644 --- a/pkg/injector/service/config.go +++ b/pkg/injector/service/config.go @@ -27,6 +27,8 @@ import ( ) // Config represents configuration options for the Dapr Sidecar Injector webhook server. +// +//nolint:recvcheck type Config struct { SidecarImage string `envconfig:"SIDECAR_IMAGE" required:"true"` SidecarImagePullPolicy string `envconfig:"SIDECAR_IMAGE_PULL_POLICY"` diff --git a/pkg/injector/service/injector_test.go b/pkg/injector/service/injector_test.go index 9916d425b..f5d2df7c8 100644 --- a/pkg/injector/service/injector_test.go +++ b/pkg/injector/service/injector_test.go @@ -14,7 +14,6 @@ limitations under the License. package service import ( - "context" "encoding/json" "testing" @@ -131,30 +130,30 @@ func TestAllowedControllersServiceAccountUID(t *testing.T) { Namespace: testCase.namespace, }, } - _, err := client.CoreV1().ServiceAccounts(testCase.namespace).Create(context.TODO(), sa, metav1.CreateOptions{}) + _, err := client.CoreV1().ServiceAccounts(testCase.namespace).Create(t.Context(), sa, metav1.CreateOptions{}) require.NoError(t, err) } t.Run("injector config has no allowed service account", func(t *testing.T) { - uids, err := AllowedControllersServiceAccountUID(context.TODO(), Config{}, client) + uids, err := AllowedControllersServiceAccountUID(t.Context(), Config{}, client) require.NoError(t, err) assert.Len(t, uids, 2) }) t.Run("injector config has a valid allowed service account", func(t *testing.T) { - uids, err := AllowedControllersServiceAccountUID(context.TODO(), Config{AllowedServiceAccounts: "test:test"}, client) + uids, err := AllowedControllersServiceAccountUID(t.Context(), Config{AllowedServiceAccounts: "test:test"}, client) require.NoError(t, err) assert.Len(t, uids, 3) }) t.Run("injector config has a invalid allowed service account", func(t *testing.T) { - uids, err := AllowedControllersServiceAccountUID(context.TODO(), Config{AllowedServiceAccounts: "abc:abc"}, client) + uids, err := AllowedControllersServiceAccountUID(t.Context(), Config{AllowedServiceAccounts: "abc:abc"}, client) require.NoError(t, err) assert.Len(t, uids, 2) }) t.Run("injector config has multiple allowed service accounts", func(t *testing.T) { - uids, err := AllowedControllersServiceAccountUID(context.TODO(), Config{AllowedServiceAccounts: "test:test,abc:abc"}, client) + uids, err := AllowedControllersServiceAccountUID(t.Context(), Config{AllowedServiceAccounts: "test:test,abc:abc"}, client) require.NoError(t, err) assert.Len(t, uids, 3) }) diff --git a/pkg/internal/loader/disk/disk_test.go b/pkg/internal/loader/disk/disk_test.go index 26d9cebda..a19183749 100644 --- a/pkg/internal/loader/disk/disk_test.go +++ b/pkg/internal/loader/disk/disk_test.go @@ -14,7 +14,6 @@ limitations under the License. package disk import ( - "context" "fmt" "io/fs" "os" @@ -52,7 +51,7 @@ spec: value: value2 ` require.NoError(t, os.WriteFile(filepath.Join(tmp, filename), []byte(yaml), fs.FileMode(0o600))) - components, err := request.Load(context.Background()) + components, err := request.Load(t.Context()) require.NoError(t, err) assert.Len(t, components, 1) }) @@ -71,7 +70,7 @@ kind: Component metadata: name: statestore` require.NoError(t, os.WriteFile(filepath.Join(tmp, filename), []byte(yaml), fs.FileMode(0o600))) - components, err := request.Load(context.Background()) + components, err := request.Load(t.Context()) require.NoError(t, err) assert.Empty(t, components) }) @@ -81,7 +80,7 @@ name: statestore` Paths: []string{"test-path-no-exists"}, }) - components, err := request.Load(context.Background()) + components, err := request.Load(t.Context()) require.Error(t, err) assert.Empty(t, components) }) @@ -230,7 +229,7 @@ metadata: Paths: []string{tmp}, AppID: "myappid", }) - components, err := loader.Load(context.Background()) + components, err := loader.Load(t.Context()) assert.Equal(t, test.expErr, err != nil, "%v", err) assert.Equal(t, test.expComps, components) }) diff --git a/pkg/internal/loader/kubernetes/components_test.go b/pkg/internal/loader/kubernetes/components_test.go index 01d64ed0f..2271b33ea 100644 --- a/pkg/internal/loader/kubernetes/components_test.go +++ b/pkg/internal/loader/kubernetes/components_test.go @@ -106,7 +106,7 @@ func TestLoadComponents(t *testing.T) { podName: "testPodName", } - response, err := request.Load(context.Background()) + response, err := request.Load(t.Context()) require.NoError(t, err) assert.NotNil(t, response) assert.Equal(t, "test", response[0].Name) diff --git a/pkg/messaging/direct_messaging_test.go b/pkg/messaging/direct_messaging_test.go index 102a1ab95..3bfc5023f 100644 --- a/pkg/messaging/direct_messaging_test.go +++ b/pkg/messaging/direct_messaging_test.go @@ -194,7 +194,7 @@ func TestInvokeRemote(t *testing.T) { WithMetadata(map[string][]string{invokev1.DestinationIDHeader: {"app1"}}) defer request.Close() - res, _, err := messaging.invokeRemote(context.Background(), "app1", "namespace1", "addr1", request) + res, _, err := messaging.invokeRemote(t.Context(), "app1", "namespace1", "addr1", request) require.NoError(t, err) pd, err := res.ProtoWithData() @@ -214,7 +214,7 @@ func TestInvokeRemote(t *testing.T) { WithMetadata(map[string][]string{invokev1.DestinationIDHeader: {"app1"}}) defer request.Close() - res, _, err := messaging.invokeRemote(context.Background(), "app1", "namespace1", "addr1", request) + res, _, err := messaging.invokeRemote(t.Context(), "app1", "namespace1", "addr1", request) require.NoError(t, err) pd, err := res.ProtoWithData() @@ -239,7 +239,7 @@ func TestInvokeRemote(t *testing.T) { WithMetadata(map[string][]string{invokev1.DestinationIDHeader: {"app1"}}) defer request.Close() - res, _, err := messaging.invokeRemote(context.Background(), "app1", "namespace1", "addr1", request) + res, _, err := messaging.invokeRemote(t.Context(), "app1", "namespace1", "addr1", request) require.NoError(t, err) pd, err := res.ProtoWithData() @@ -257,7 +257,7 @@ func TestInvokeRemote(t *testing.T) { WithMetadata(map[string][]string{invokev1.DestinationIDHeader: {"app1"}}) defer request.Close() - res, _, err := messaging.invokeRemote(context.Background(), "app1", "namespace1", "addr1", request) + res, _, err := messaging.invokeRemote(t.Context(), "app1", "namespace1", "addr1", request) require.Error(t, err) assert.Equal(t, fmt.Sprintf(streamingUnsupportedErr, "app1"), err.Error()) assert.Nil(t, res) @@ -273,7 +273,7 @@ func TestInvokeRemote(t *testing.T) { WithReplay(true) defer request.Close() - res, _, err := messaging.invokeRemote(context.Background(), "app1", "namespace1", "addr1", request) + res, _, err := messaging.invokeRemote(t.Context(), "app1", "namespace1", "addr1", request) require.NoError(t, err) pd, err := res.ProtoWithData() @@ -297,7 +297,7 @@ func TestInvokeRemote(t *testing.T) { WithMetadata(map[string][]string{invokev1.DestinationIDHeader: {"app1"}}) defer request.Close() - res, _, err := messaging.invokeRemote(context.Background(), "app1", "namespace1", "addr1", request) + res, _, err := messaging.invokeRemote(t.Context(), "app1", "namespace1", "addr1", request) require.NoError(t, err) pd, err := res.ProtoWithData() @@ -475,7 +475,7 @@ func TestInvokeRemoteUnaryForHTTPEndpoint(t *testing.T) { channels: (new(channels.Channels)).WithEndpointChannels(map[string]channel.HTTPEndpointAppChannel{"abc": &mockChannel{}}), } - _, err := d.invokeRemoteUnaryForHTTPEndpoint(context.Background(), nil, "abc") + _, err := d.invokeRemoteUnaryForHTTPEndpoint(t.Context(), nil, "abc") require.NoError(t, err) }) @@ -484,7 +484,7 @@ func TestInvokeRemoteUnaryForHTTPEndpoint(t *testing.T) { channels: new(channels.Channels), } - _, err := d.invokeRemoteUnaryForHTTPEndpoint(context.Background(), nil, "abc") + _, err := d.invokeRemoteUnaryForHTTPEndpoint(t.Context(), nil, "abc") require.Error(t, err) }) } diff --git a/pkg/messaging/grpc_proxy_test.go b/pkg/messaging/grpc_proxy_test.go index 483d312c9..ea3f4a76f 100644 --- a/pkg/messaging/grpc_proxy_test.go +++ b/pkg/messaging/grpc_proxy_test.go @@ -101,7 +101,7 @@ func TestSetTelemetryFn(t *testing.T) { }) proxy := p.(*proxy) - ctx := metadata.NewOutgoingContext(context.TODO(), metadata.MD{"a": []string{"b"}}) + ctx := metadata.NewOutgoingContext(t.Context(), metadata.MD{"a": []string{"b"}}) ctx = proxy.telemetryFn(ctx) md, _ := metadata.FromOutgoingContext(ctx) @@ -138,7 +138,7 @@ func TestIntercept(t *testing.T) { }, nil }) - ctx := metadata.NewOutgoingContext(context.TODO(), metadata.MD{"a": []string{"b"}}) + ctx := metadata.NewOutgoingContext(t.Context(), metadata.MD{"a": []string{"b"}}) proxy := p.(*proxy) _, conn, _, teardown, err := proxy.intercept(ctx, "/test") defer teardown(true) @@ -164,7 +164,7 @@ func TestIntercept(t *testing.T) { }, nil }) - ctx := metadata.NewIncomingContext(context.TODO(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"b"}}) + ctx := metadata.NewIncomingContext(t.Context(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"b"}}) proxy := p.(*proxy) _, _, _, _, err := proxy.intercept(ctx, "/test") @@ -190,7 +190,7 @@ func TestIntercept(t *testing.T) { t.Setenv(securityConsts.AppAPITokenEnvVar, "token1") - ctx := metadata.NewIncomingContext(context.TODO(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"a"}, securityConsts.APITokenHeader: []string{"oldtoken"}}) + ctx := metadata.NewIncomingContext(t.Context(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"a"}, securityConsts.APITokenHeader: []string{"oldtoken"}}) proxy := p.(*proxy) ctx, conn, _, teardown, err := proxy.intercept(ctx, "/test") defer teardown(true) @@ -223,7 +223,7 @@ func TestIntercept(t *testing.T) { t.Setenv(securityConsts.AppAPITokenEnvVar, "token1") - ctx := metadata.NewIncomingContext(context.TODO(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"b"}}) + ctx := metadata.NewIncomingContext(t.Context(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"b"}}) proxy := p.(*proxy) ctx, conn, _, teardown, err := proxy.intercept(ctx, "/test") defer teardown(true) @@ -263,7 +263,7 @@ func TestIntercept(t *testing.T) { return ctx }) - ctx := metadata.NewIncomingContext(context.TODO(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"a"}}) + ctx := metadata.NewIncomingContext(t.Context(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"a"}}) proxy := p.(*proxy) _, conn, _, teardown, err := proxy.intercept(ctx, "/test") @@ -284,7 +284,7 @@ func TestIntercept(t *testing.T) { return ctx }) - ctx := metadata.NewIncomingContext(context.TODO(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"a"}}) + ctx := metadata.NewIncomingContext(t.Context(), metadata.MD{diagnostics.GRPCProxyAppIDKey: []string{"a"}}) proxy := p.(*proxy) _, conn, _, teardown, err := proxy.intercept(ctx, "/test") defer teardown(true) diff --git a/pkg/messaging/v1/invoke_method_request_test.go b/pkg/messaging/v1/invoke_method_request_test.go index 2df0518d7..a423e8f14 100644 --- a/pkg/messaging/v1/invoke_method_request_test.go +++ b/pkg/messaging/v1/invoke_method_request_test.go @@ -488,11 +488,11 @@ func TestWithDataObject(t *testing.T) { gotEnc, err := json.Marshal(got) require.NoError(t, err) - assert.Equal(t, []byte(expectJSON), compactJSON(t, gotEnc)) + assert.JSONEq(t, expectJSON, string(compactJSON(t, gotEnc))) data, err := req.RawDataFull() require.NoError(t, err) - assert.Equal(t, []byte(expectJSON), compactJSON(t, data)) + assert.JSONEq(t, expectJSON, string(compactJSON(t, data))) } func TestRequestReplayable(t *testing.T) { diff --git a/pkg/messaging/v1/util_test.go b/pkg/messaging/v1/util_test.go index ab171a94f..367c804f7 100644 --- a/pkg/messaging/v1/util_test.go +++ b/pkg/messaging/v1/util_test.go @@ -14,7 +14,6 @@ limitations under the License. package v1 import ( - "context" "encoding/base64" "fmt" "sort" @@ -49,7 +48,7 @@ func TestInternalMetadataToHTTPHeader(t *testing.T) { expectedKeyNames := []string{"custom-header", "dapr-method", "dapr-scheme", "dapr-path", "dapr-authority", "dapr-grpc-timeout"} savedHeaderKeyNames := []string{} - ctx := context.Background() + ctx := t.Context() InternalMetadataToHTTPHeader(ctx, fakeMetadata, func(k, v string) { savedHeaderKeyNames = append(savedHeaderKeyNames, k) }) @@ -111,7 +110,7 @@ func TestInternalMetadataToGrpcMetadata(t *testing.T) { }, } - ctx := context.Background() + ctx := t.Context() t.Run("without http header conversion for http headers", func(t *testing.T) { convertedMD := InternalMetadataToGrpcMetadata(ctx, httpHeaders, false) @@ -338,7 +337,7 @@ func TestWithCustomGrpcMetadata(t *testing.T) { md[customMetadataKey(i)] = customMetadataValue(i) } - ctx := context.Background() + ctx := t.Context() ctx = WithCustomGRPCMetadata(ctx, md) ctxMd, ok := metadata.FromOutgoingContext(ctx) diff --git a/pkg/metrics/exporter_test.go b/pkg/metrics/exporter_test.go index f7e5aa647..6fb29474e 100644 --- a/pkg/metrics/exporter_test.go +++ b/pkg/metrics/exporter_test.go @@ -47,7 +47,7 @@ func TestMetricsExporter(t *testing.T) { Healthz: healthz.New(), }) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) go func() { errCh <- e.Start(ctx) diff --git a/pkg/operator/api/api_test.go b/pkg/operator/api/api_test.go index bff728435..3d9ad2607 100644 --- a/pkg/operator/api/api_test.go +++ b/pkg/operator/api/api_test.go @@ -94,7 +94,7 @@ func TestProcessComponentSecrets(t *testing.T) { }, } - err := processComponentSecrets(context.Background(), &c, "default", nil) + err := processComponentSecrets(t.Context(), &c, "default", nil) require.NoError(t, err) }) @@ -136,13 +136,13 @@ func TestProcessComponentSecrets(t *testing.T) { }). Build() - err = processComponentSecrets(context.Background(), &c, "default", client) + err = processComponentSecrets(t.Context(), &c, "default", client) require.NoError(t, err) enc := base64.StdEncoding.EncodeToString([]byte("value1")) jsonEnc, _ := json.Marshal(enc) - assert.Equal(t, jsonEnc, c.Spec.Metadata[0].Value.Raw) + assert.JSONEq(t, string(jsonEnc), string(c.Spec.Metadata[0].Value.Raw)) }) t.Run("secret ref exists, default kubernetes secret store, secret extracted", func(t *testing.T) { @@ -183,13 +183,13 @@ func TestProcessComponentSecrets(t *testing.T) { }). Build() - err = processComponentSecrets(context.Background(), &c, "default", client) + err = processComponentSecrets(t.Context(), &c, "default", client) require.NoError(t, err) enc := base64.StdEncoding.EncodeToString([]byte("value1")) jsonEnc, _ := json.Marshal(enc) - assert.Equal(t, jsonEnc, c.Spec.Metadata[0].Value.Raw) + assert.JSONEq(t, string(jsonEnc), string(c.Spec.Metadata[0].Value.Raw)) }) } @@ -587,7 +587,7 @@ func TestListsNamespaced(t *testing.T) { assert.Equal(t, "sub1", sub.Name) assert.Equal(t, "namespace-a", sub.Namespace) - res, err = api.ListSubscriptionsV2(context.TODO(), &operatorv1pb.ListSubscriptionsRequest{ + res, err = api.ListSubscriptionsV2(t.Context(), &operatorv1pb.ListSubscriptionsRequest{ PodName: "baz", Namespace: "namespace-c", }) @@ -692,7 +692,7 @@ func TestListsNamespaced(t *testing.T) { assert.Equal(t, "obj1", endpoint.Name) assert.Equal(t, "namespace-a", endpoint.Namespace) - res, err = api.ListHTTPEndpoints(context.TODO(), &operatorv1pb.ListHTTPEndpointsRequest{ + res, err = api.ListHTTPEndpoints(t.Context(), &operatorv1pb.ListHTTPEndpointsRequest{ Namespace: "namespace-c", }) require.Error(t, err) @@ -719,7 +719,7 @@ func TestProcessHTTPEndpointSecrets(t *testing.T) { }, } t.Run("secret ref exists, not kubernetes secret store, no error", func(t *testing.T) { - err := processHTTPEndpointSecrets(context.Background(), &e, "default", nil) + err := processHTTPEndpointSecrets(t.Context(), &e, "default", nil) require.NoError(t, err) }) @@ -744,11 +744,11 @@ func TestProcessHTTPEndpointSecrets(t *testing.T) { }, }). Build() - require.NoError(t, processHTTPEndpointSecrets(context.Background(), &e, "default", client)) + require.NoError(t, processHTTPEndpointSecrets(t.Context(), &e, "default", client)) enc := base64.StdEncoding.EncodeToString([]byte("value1")) jsonEnc, err := json.Marshal(enc) require.NoError(t, err) - assert.Equal(t, jsonEnc, e.Spec.Headers[0].Value.Raw) + assert.JSONEq(t, string(jsonEnc), string(e.Spec.Headers[0].Value.Raw)) }) t.Run("secret ref exists, default kubernetes secret store, secret extracted", func(t *testing.T) { @@ -773,12 +773,12 @@ func TestProcessHTTPEndpointSecrets(t *testing.T) { }). Build() - require.NoError(t, processHTTPEndpointSecrets(context.Background(), &e, "default", client)) + require.NoError(t, processHTTPEndpointSecrets(t.Context(), &e, "default", client)) enc := base64.StdEncoding.EncodeToString([]byte("value1")) jsonEnc, err := json.Marshal(enc) require.NoError(t, err) - assert.Equal(t, jsonEnc, e.Spec.Headers[0].Value.Raw) + assert.JSONEq(t, string(jsonEnc), string(e.Spec.Headers[0].Value.Raw)) }) } @@ -794,7 +794,7 @@ func Test_Ready(t *testing.T) { close(ch) return ch }, - ctx: context.Background, + ctx: t.Context, expErr: false, }, "if context is cancelled, then expect error": { @@ -803,7 +803,7 @@ func Test_Ready(t *testing.T) { return ch }, ctx: func() context.Context { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) cancel() return ctx }, diff --git a/pkg/operator/api/authz/authz_test.go b/pkg/operator/api/authz/authz_test.go index a68de9277..78d7d3667 100644 --- a/pkg/operator/api/authz/authz_test.go +++ b/pkg/operator/api/authz/authz_test.go @@ -14,7 +14,6 @@ limitations under the License. package authz import ( - "context" "testing" "github.com/spiffe/go-spiffe/v2/spiffeid" @@ -33,7 +32,7 @@ func Test_Request(t *testing.T) { pki := test.GenPKI(t, test.PKIOptions{LeafID: serverID, ClientID: appID}) t.Run("no auth context should error", func(t *testing.T) { - id, err := Request(context.Background(), "ns1") + id, err := Request(t.Context(), "ns1") require.Error(t, err) assert.Equal(t, codes.PermissionDenied, status.Code(err)) assert.Nil(t, id) diff --git a/pkg/operator/api/informer/handler_test.go b/pkg/operator/api/informer/handler_test.go index a6609b4c1..eaa17340a 100644 --- a/pkg/operator/api/informer/handler_test.go +++ b/pkg/operator/api/informer/handler_test.go @@ -38,7 +38,7 @@ func Test_loop(t *testing.T) { appCh: make(chan *Event[compapi.Component]), } - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { h.loop(ctx) close(done) @@ -62,7 +62,7 @@ func Test_loop(t *testing.T) { } go func() { - h.loop(context.Background()) + h.loop(t.Context()) close(done) }() @@ -93,7 +93,7 @@ func Test_loop(t *testing.T) { } go func() { - h.loop(context.Background()) + h.loop(t.Context()) close(done) }() diff --git a/pkg/operator/api/informer/informer_test.go b/pkg/operator/api/informer/informer_test.go index 63c5dc5a3..b366dc8a2 100644 --- a/pkg/operator/api/informer/informer_test.go +++ b/pkg/operator/api/informer/informer_test.go @@ -14,7 +14,6 @@ limitations under the License. package informer import ( - "context" "testing" "time" @@ -46,7 +45,7 @@ func Test_WatchUpdates(t *testing.T) { assert.Equal(t, codes.PermissionDenied, status.Code(err)) assert.Nil(t, appCh) - appCh, err = i.WatchUpdates(context.Background(), "ns2") + appCh, err = i.WatchUpdates(t.Context(), "ns2") require.Error(t, err) assert.Equal(t, codes.PermissionDenied, status.Code(err)) assert.Nil(t, appCh) @@ -65,7 +64,7 @@ func Test_WatchUpdates(t *testing.T) { appCh2, err := i.WatchUpdates(pki.ClientGRPCCtx(t), "ns1") require.NoError(t, err) - i.handleEvent(context.Background(), + i.handleEvent(t.Context(), &compapi.Component{ ObjectMeta: metav1.ObjectMeta{Name: "comp1", Namespace: "ns1"}, }, @@ -80,7 +79,7 @@ func Test_WatchUpdates(t *testing.T) { assert.Equal(c, 1, int(i.batchID.Load())) }, 5*time.Second, 100*time.Millisecond) - i.handleEvent(context.Background(), + i.handleEvent(t.Context(), &compapi.Component{ ObjectMeta: metav1.ObjectMeta{Name: "comp1", Namespace: "ns1"}, Spec: compapi.ComponentSpec{Type: "bindings.redis"}, @@ -96,7 +95,7 @@ func Test_WatchUpdates(t *testing.T) { assert.Equal(c, 2, int(i.batchID.Load())) }, 5*time.Second, 100*time.Millisecond) - i.handleEvent(context.Background(), + i.handleEvent(t.Context(), nil, &compapi.Component{ ObjectMeta: metav1.ObjectMeta{Name: "comp2", Namespace: "ns1"}, diff --git a/pkg/operator/handlers/dapr_handler_test.go b/pkg/operator/handlers/dapr_handler_test.go index e380e317f..4c2c8c456 100644 --- a/pkg/operator/handlers/dapr_handler_test.go +++ b/pkg/operator/handlers/dapr_handler_test.go @@ -1,7 +1,6 @@ package handlers import ( - "context" "reflect" "testing" @@ -96,20 +95,20 @@ func TestIsAnnotatedForDapr(t *testing.T) { func TestDaprService(t *testing.T) { t.Run("invalid empty app id", func(t *testing.T) { d := getDeployment("", "true") - err := getTestDaprHandler().ensureDaprServicePresent(context.TODO(), "default", d) + err := getTestDaprHandler().ensureDaprServicePresent(t.Context(), "default", d) require.Error(t, err) }) t.Run("invalid char app id", func(t *testing.T) { d := getDeployment("myapp@", "true") - err := getTestDaprHandler().ensureDaprServicePresent(context.TODO(), "default", d) + err := getTestDaprHandler().ensureDaprServicePresent(t.Context(), "default", d) require.Error(t, err) }) } func TestCreateDaprServiceAppIDAndMetricsSettings(t *testing.T) { testDaprHandler := getTestDaprHandler() - ctx := context.Background() + ctx := t.Context() myDaprService := types.NamespacedName{ Namespace: "test", Name: "test", @@ -136,7 +135,7 @@ func TestCreateDaprServiceAppIDAndMetricsSettings(t *testing.T) { func TestCreateDaprServiceAppIDAndPortsOverride(t *testing.T) { testDaprHandler := getTestDaprHandler() - ctx := context.Background() + ctx := t.Context() myDaprService := types.NamespacedName{ Namespace: "test", Name: "test", @@ -165,7 +164,7 @@ func TestPatchDaprService(t *testing.T) { cli := fake.NewClientBuilder().WithScheme(s).Build() testDaprHandler.Client = cli - ctx := context.Background() + ctx := t.Context() myDaprService := types.NamespacedName{ Namespace: "test", Name: "test", @@ -342,7 +341,7 @@ func TestInit(t *testing.T) { t.Run("test init dapr handler", func(t *testing.T) { assert.NotNil(t, handler) - err := handler.Init(context.Background()) + err := handler.Init(t.Context()) require.NoError(t, err) diff --git a/pkg/operator/watchdog_test.go b/pkg/operator/watchdog_test.go index 1900e2b6a..770b15597 100644 --- a/pkg/operator/watchdog_test.go +++ b/pkg/operator/watchdog_test.go @@ -82,7 +82,7 @@ func createMockPods(n, daprized, injected, daprdPresent int) (pods []*corev1.Pod } func TestDaprWatchdog_listPods(t *testing.T) { - ctx := context.Background() + ctx := t.Context() rl := ratelimit.NewUnlimited() t.Run("injectorNotPresent", func(t *testing.T) { @@ -235,7 +235,7 @@ func Test_patchPodLabel(t *testing.T) { for _, tc := range tests { ctlClient := fake.NewClientBuilder().WithObjects(tc.pod).Build() t.Run(tc.name, func(t *testing.T) { - if err := patchPodLabel(context.TODO(), ctlClient, tc.pod); (err != nil) != tc.wantErr { + if err := patchPodLabel(t.Context(), ctlClient, tc.pod); (err != nil) != tc.wantErr { t.Fatalf("patchPodLabel() error = %v, wantErr %v", err, tc.wantErr) } if !tc.wantErr { @@ -247,7 +247,7 @@ func Test_patchPodLabel(t *testing.T) { func TestDaprWatchdog_Start(t *testing.T) { // simple test of start - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) cancelled := false defer func() { if !cancelled { diff --git a/pkg/placement/leadership_test.go b/pkg/placement/leadership_test.go index bede82e82..df9e4b332 100644 --- a/pkg/placement/leadership_test.go +++ b/pkg/placement/leadership_test.go @@ -60,7 +60,7 @@ func TestCleanupHeartBeats(t *testing.T) { } func TestMonitorLeadership(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) raftClusterOpts, err := tests.RaftClusterOpts(t) require.NoError(t, err) diff --git a/pkg/placement/membership_test.go b/pkg/placement/membership_test.go index fcede4c6c..3adaf232e 100644 --- a/pkg/placement/membership_test.go +++ b/pkg/placement/membership_test.go @@ -55,7 +55,7 @@ func TestMembershipChangeWorker(t *testing.T) { require.NoError(t, err) setupEach := func(t *testing.T) context.CancelFunc { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) var cancelServer context.CancelFunc serverAddress, testServer, clock, cancelServer = newTestPlacementServer(t, *raftOpts) @@ -513,7 +513,7 @@ func PerformTableUpdateCostTime(t *testing.T) (wastedTime int64) { mockMessage := &v1pb.PlacementTables{Version: "demo"} - ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second) + ctx, cancel := context.WithTimeout(t.Context(), 10*time.Second) defer cancel() var start time.Time diff --git a/pkg/placement/placement_test.go b/pkg/placement/placement_test.go index b0c8d661f..b288dc361 100644 --- a/pkg/placement/placement_test.go +++ b/pkg/placement/placement_test.go @@ -59,7 +59,7 @@ func newTestPlacementServer(t *testing.T, raftOptions raft.Options) (string, *Se testServer.clock = clock serverStopped := make(chan struct{}) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { defer close(serverStopped) err := testServer.Run(ctx) @@ -106,14 +106,14 @@ func newTestClient(t *testing.T, serverAddress string) (*grpc.ClientConn, *net.T client := v1pb.NewPlacementClient(conn) - stream, err := client.ReportDaprStatus(context.Background()) + stream, err := client.ReportDaprStatus(t.Context()) require.NoError(t, err) return conn, tcpConn.(*net.TCPConn), stream } func TestMemberRegistration_NoLeadership(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) defer cancel() raftClusterOpts, err := tests.RaftClusterOpts(t) diff --git a/pkg/placement/raft/ha_test.go b/pkg/placement/raft/ha_test.go index 7889a6875..8e19f1995 100644 --- a/pkg/placement/raft/ha_test.go +++ b/pkg/placement/raft/ha_test.go @@ -277,6 +277,7 @@ func createRaftServer(t *testing.T, nodeID int, peers []PeerInfo) (*Server, <-ch }) require.NoError(t, err) + //nolint:usetesting ctx, cancel := context.WithCancel(context.Background()) stopped := make(chan struct{}) diff --git a/pkg/proto/common/v1/common.pb.go b/pkg/proto/common/v1/common.pb.go index b5f92c4de..1311dfffd 100644 --- a/pkg/proto/common/v1/common.pb.go +++ b/pkg/proto/common/v1/common.pb.go @@ -19,11 +19,12 @@ package common import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" anypb "google.golang.org/protobuf/types/known/anypb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/components/v1/bindings.pb.go b/pkg/proto/components/v1/bindings.pb.go index de86ee3fe..50436eb30 100644 --- a/pkg/proto/components/v1/bindings.pb.go +++ b/pkg/proto/components/v1/bindings.pb.go @@ -19,10 +19,11 @@ package components import ( - protoreflect "google.golang.org/protobuf/reflect/protoreflect" - protoimpl "google.golang.org/protobuf/runtime/protoimpl" reflect "reflect" sync "sync" + + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" ) const ( diff --git a/pkg/proto/components/v1/bindings_grpc.pb.go b/pkg/proto/components/v1/bindings_grpc.pb.go index f16ae441a..745afb9a5 100644 --- a/pkg/proto/components/v1/bindings_grpc.pb.go +++ b/pkg/proto/components/v1/bindings_grpc.pb.go @@ -20,6 +20,7 @@ package components import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/components/v1/common.pb.go b/pkg/proto/components/v1/common.pb.go index 6362d625d..7b88ce842 100644 --- a/pkg/proto/components/v1/common.pb.go +++ b/pkg/proto/components/v1/common.pb.go @@ -19,10 +19,11 @@ package components import ( - protoreflect "google.golang.org/protobuf/reflect/protoreflect" - protoimpl "google.golang.org/protobuf/runtime/protoimpl" reflect "reflect" sync "sync" + + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" ) const ( diff --git a/pkg/proto/components/v1/pubsub.pb.go b/pkg/proto/components/v1/pubsub.pb.go index 94ebe902a..c38825b69 100644 --- a/pkg/proto/components/v1/pubsub.pb.go +++ b/pkg/proto/components/v1/pubsub.pb.go @@ -19,10 +19,11 @@ package components import ( - protoreflect "google.golang.org/protobuf/reflect/protoreflect" - protoimpl "google.golang.org/protobuf/runtime/protoimpl" reflect "reflect" sync "sync" + + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" ) const ( diff --git a/pkg/proto/components/v1/pubsub_grpc.pb.go b/pkg/proto/components/v1/pubsub_grpc.pb.go index 4574eb768..6c5a4190f 100644 --- a/pkg/proto/components/v1/pubsub_grpc.pb.go +++ b/pkg/proto/components/v1/pubsub_grpc.pb.go @@ -20,6 +20,7 @@ package components import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/components/v1/secretstore.pb.go b/pkg/proto/components/v1/secretstore.pb.go index 93ceb8c6f..2fb0e8314 100644 --- a/pkg/proto/components/v1/secretstore.pb.go +++ b/pkg/proto/components/v1/secretstore.pb.go @@ -19,10 +19,11 @@ package components import ( - protoreflect "google.golang.org/protobuf/reflect/protoreflect" - protoimpl "google.golang.org/protobuf/runtime/protoimpl" reflect "reflect" sync "sync" + + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" ) const ( diff --git a/pkg/proto/components/v1/secretstore_grpc.pb.go b/pkg/proto/components/v1/secretstore_grpc.pb.go index 06c432d7f..4a84607ae 100644 --- a/pkg/proto/components/v1/secretstore_grpc.pb.go +++ b/pkg/proto/components/v1/secretstore_grpc.pb.go @@ -20,6 +20,7 @@ package components import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/components/v1/state.pb.go b/pkg/proto/components/v1/state.pb.go index 8a8e00075..71115f2ac 100644 --- a/pkg/proto/components/v1/state.pb.go +++ b/pkg/proto/components/v1/state.pb.go @@ -19,11 +19,12 @@ package components import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" anypb "google.golang.org/protobuf/types/known/anypb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/components/v1/state_grpc.pb.go b/pkg/proto/components/v1/state_grpc.pb.go index 57834b439..9d753929a 100644 --- a/pkg/proto/components/v1/state_grpc.pb.go +++ b/pkg/proto/components/v1/state_grpc.pb.go @@ -20,6 +20,7 @@ package components import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/internals/v1/apiversion.pb.go b/pkg/proto/internals/v1/apiversion.pb.go index cba0c8642..d439a0498 100644 --- a/pkg/proto/internals/v1/apiversion.pb.go +++ b/pkg/proto/internals/v1/apiversion.pb.go @@ -19,10 +19,11 @@ package internals import ( - protoreflect "google.golang.org/protobuf/reflect/protoreflect" - protoimpl "google.golang.org/protobuf/runtime/protoimpl" reflect "reflect" sync "sync" + + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" ) const ( diff --git a/pkg/proto/internals/v1/jobs.pb.go b/pkg/proto/internals/v1/jobs.pb.go index a494112a6..6ea83af9e 100644 --- a/pkg/proto/internals/v1/jobs.pb.go +++ b/pkg/proto/internals/v1/jobs.pb.go @@ -19,11 +19,12 @@ package internals import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" structpb "google.golang.org/protobuf/types/known/structpb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/internals/v1/reminders.pb.go b/pkg/proto/internals/v1/reminders.pb.go index 4d433bf09..6611fc20e 100644 --- a/pkg/proto/internals/v1/reminders.pb.go +++ b/pkg/proto/internals/v1/reminders.pb.go @@ -19,12 +19,13 @@ package internals import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" anypb "google.golang.org/protobuf/types/known/anypb" timestamppb "google.golang.org/protobuf/types/known/timestamppb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/internals/v1/service_invocation.pb.go b/pkg/proto/internals/v1/service_invocation.pb.go index f64910183..bc81a6cb5 100644 --- a/pkg/proto/internals/v1/service_invocation.pb.go +++ b/pkg/proto/internals/v1/service_invocation.pb.go @@ -19,12 +19,13 @@ package internals import ( + reflect "reflect" + sync "sync" + v1 "github.com/dapr/dapr/pkg/proto/common/v1" protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" emptypb "google.golang.org/protobuf/types/known/emptypb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/internals/v1/service_invocation_grpc.pb.go b/pkg/proto/internals/v1/service_invocation_grpc.pb.go index 731d2a763..5860d583b 100644 --- a/pkg/proto/internals/v1/service_invocation_grpc.pb.go +++ b/pkg/proto/internals/v1/service_invocation_grpc.pb.go @@ -20,6 +20,7 @@ package internals import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/internals/v1/status.pb.go b/pkg/proto/internals/v1/status.pb.go index 96395b86c..2bdf0829f 100644 --- a/pkg/proto/internals/v1/status.pb.go +++ b/pkg/proto/internals/v1/status.pb.go @@ -19,11 +19,12 @@ package internals import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" anypb "google.golang.org/protobuf/types/known/anypb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/operator/v1/operator.pb.go b/pkg/proto/operator/v1/operator.pb.go index f7ba1aa61..e09fba2be 100644 --- a/pkg/proto/operator/v1/operator.pb.go +++ b/pkg/proto/operator/v1/operator.pb.go @@ -19,11 +19,12 @@ package operator import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" emptypb "google.golang.org/protobuf/types/known/emptypb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/operator/v1/operator_grpc.pb.go b/pkg/proto/operator/v1/operator_grpc.pb.go index 36eb67717..4714e03bd 100644 --- a/pkg/proto/operator/v1/operator_grpc.pb.go +++ b/pkg/proto/operator/v1/operator_grpc.pb.go @@ -20,6 +20,7 @@ package operator import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/operator/v1/resource.pb.go b/pkg/proto/operator/v1/resource.pb.go index ac3b75360..53e25eed6 100644 --- a/pkg/proto/operator/v1/resource.pb.go +++ b/pkg/proto/operator/v1/resource.pb.go @@ -19,11 +19,12 @@ package operator import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" timestamppb "google.golang.org/protobuf/types/known/timestamppb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/placement/v1/placement.pb.go b/pkg/proto/placement/v1/placement.pb.go index 1ea434d7b..44f8a8d3e 100644 --- a/pkg/proto/placement/v1/placement.pb.go +++ b/pkg/proto/placement/v1/placement.pb.go @@ -19,10 +19,11 @@ package placement import ( - protoreflect "google.golang.org/protobuf/reflect/protoreflect" - protoimpl "google.golang.org/protobuf/runtime/protoimpl" reflect "reflect" sync "sync" + + protoreflect "google.golang.org/protobuf/reflect/protoreflect" + protoimpl "google.golang.org/protobuf/runtime/protoimpl" ) const ( diff --git a/pkg/proto/placement/v1/placement_grpc.pb.go b/pkg/proto/placement/v1/placement_grpc.pb.go index df4372d5c..2fe7776d6 100644 --- a/pkg/proto/placement/v1/placement_grpc.pb.go +++ b/pkg/proto/placement/v1/placement_grpc.pb.go @@ -20,6 +20,7 @@ package placement import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/runtime/v1/appcallback.pb.go b/pkg/proto/runtime/v1/appcallback.pb.go index 3bd818a65..1e0ea5e4c 100644 --- a/pkg/proto/runtime/v1/appcallback.pb.go +++ b/pkg/proto/runtime/v1/appcallback.pb.go @@ -19,14 +19,15 @@ package runtime import ( + reflect "reflect" + sync "sync" + v1 "github.com/dapr/dapr/pkg/proto/common/v1" protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" anypb "google.golang.org/protobuf/types/known/anypb" emptypb "google.golang.org/protobuf/types/known/emptypb" structpb "google.golang.org/protobuf/types/known/structpb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/runtime/v1/appcallback_grpc.pb.go b/pkg/proto/runtime/v1/appcallback_grpc.pb.go index 205212bdb..2a7fa5edb 100644 --- a/pkg/proto/runtime/v1/appcallback_grpc.pb.go +++ b/pkg/proto/runtime/v1/appcallback_grpc.pb.go @@ -20,6 +20,7 @@ package runtime import ( context "context" + v1 "github.com/dapr/dapr/pkg/proto/common/v1" grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" diff --git a/pkg/proto/runtime/v1/dapr.pb.go b/pkg/proto/runtime/v1/dapr.pb.go index f4fa821ad..e29440a54 100644 --- a/pkg/proto/runtime/v1/dapr.pb.go +++ b/pkg/proto/runtime/v1/dapr.pb.go @@ -19,14 +19,15 @@ package runtime import ( + reflect "reflect" + sync "sync" + v1 "github.com/dapr/dapr/pkg/proto/common/v1" protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" anypb "google.golang.org/protobuf/types/known/anypb" emptypb "google.golang.org/protobuf/types/known/emptypb" timestamppb "google.golang.org/protobuf/types/known/timestamppb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/runtime/v1/dapr_grpc.pb.go b/pkg/proto/runtime/v1/dapr_grpc.pb.go index aa65e57a1..b2d5ff34b 100644 --- a/pkg/proto/runtime/v1/dapr_grpc.pb.go +++ b/pkg/proto/runtime/v1/dapr_grpc.pb.go @@ -20,6 +20,7 @@ package runtime import ( context "context" + v1 "github.com/dapr/dapr/pkg/proto/common/v1" grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" diff --git a/pkg/proto/scheduler/v1/failurepolicy.pb.go b/pkg/proto/scheduler/v1/failurepolicy.pb.go index 2d25b7883..e134ee125 100644 --- a/pkg/proto/scheduler/v1/failurepolicy.pb.go +++ b/pkg/proto/scheduler/v1/failurepolicy.pb.go @@ -7,11 +7,12 @@ package scheduler import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" durationpb "google.golang.org/protobuf/types/known/durationpb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/scheduler/v1/scheduler.pb.go b/pkg/proto/scheduler/v1/scheduler.pb.go index 1333468cd..a1934a182 100644 --- a/pkg/proto/scheduler/v1/scheduler.pb.go +++ b/pkg/proto/scheduler/v1/scheduler.pb.go @@ -7,11 +7,12 @@ package scheduler import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" anypb "google.golang.org/protobuf/types/known/anypb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/scheduler/v1/scheduler_grpc.pb.go b/pkg/proto/scheduler/v1/scheduler_grpc.pb.go index 874e1d89f..e9d62a186 100644 --- a/pkg/proto/scheduler/v1/scheduler_grpc.pb.go +++ b/pkg/proto/scheduler/v1/scheduler_grpc.pb.go @@ -8,6 +8,7 @@ package scheduler import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/proto/sentry/v1/sentry.pb.go b/pkg/proto/sentry/v1/sentry.pb.go index 5ae82449d..a91dd7577 100644 --- a/pkg/proto/sentry/v1/sentry.pb.go +++ b/pkg/proto/sentry/v1/sentry.pb.go @@ -19,11 +19,12 @@ package sentry import ( + reflect "reflect" + sync "sync" + protoreflect "google.golang.org/protobuf/reflect/protoreflect" protoimpl "google.golang.org/protobuf/runtime/protoimpl" timestamppb "google.golang.org/protobuf/types/known/timestamppb" - reflect "reflect" - sync "sync" ) const ( diff --git a/pkg/proto/sentry/v1/sentry_grpc.pb.go b/pkg/proto/sentry/v1/sentry_grpc.pb.go index d896a08d3..de7301d81 100644 --- a/pkg/proto/sentry/v1/sentry_grpc.pb.go +++ b/pkg/proto/sentry/v1/sentry_grpc.pb.go @@ -20,6 +20,7 @@ package sentry import ( context "context" + grpc "google.golang.org/grpc" codes "google.golang.org/grpc/codes" status "google.golang.org/grpc/status" diff --git a/pkg/resiliency/noop_test.go b/pkg/resiliency/noop_test.go index 148d6d876..7c0ac1334 100644 --- a/pkg/resiliency/noop_test.go +++ b/pkg/resiliency/noop_test.go @@ -23,7 +23,7 @@ import ( ) func TestNoOp(t *testing.T) { - ctx := context.Background() + ctx := t.Context() policy := NoOp{} tests := []struct { diff --git a/pkg/resiliency/policy_test.go b/pkg/resiliency/policy_test.go index 4ee50a1fa..b28e13a0f 100644 --- a/pkg/resiliency/policy_test.go +++ b/pkg/resiliency/policy_test.go @@ -148,7 +148,7 @@ func TestPolicy(t *testing.T) { "nil policy": nil, } - ctx := context.Background() + ctx := t.Context() for name, tt := range tests { t.Run(name, func(t *testing.T) { called := atomic.Bool{} @@ -204,7 +204,7 @@ func TestPolicyTimeout(t *testing.T) { return nil, nil } - policy := NewRunner[any](context.Background(), &PolicyDefinition{ + policy := NewRunner[any](t.Context(), &PolicyDefinition{ log: testLog, name: "timeout", t: test.timeout, @@ -253,7 +253,7 @@ func TestPolicyRetry(t *testing.T) { return struct{}{}, nil } - policy := NewRunner[struct{}](context.Background(), &PolicyDefinition{ + policy := NewRunner[struct{}](t.Context(), &PolicyDefinition{ log: testLog, name: "retry", t: 10 * time.Millisecond, @@ -319,7 +319,7 @@ func TestPolicyRetryWithMatch(t *testing.T) { match, err := ParseRetryConditionMatch(test.matching) require.NoError(t, err) - policy := NewRunner[struct{}](context.Background(), &PolicyDefinition{ + policy := NewRunner[struct{}](t.Context(), &PolicyDefinition{ log: testLog, name: "retry", t: 10 * time.Millisecond, @@ -363,7 +363,7 @@ func TestPolicyAccumulator(t *testing.T) { r: NewRetry(retry.Config{MaxRetries: 6}, NewRetryConditionMatch()), } var accumulatorCalled int - policy := NewRunnerWithOptions(context.Background(), policyDef, RunnerOpts[int32]{ + policy := NewRunnerWithOptions(t.Context(), policyDef, RunnerOpts[int32]{ Accumulator: func(i int32) { // Only reason for incrementing "val" here is to have something to check for race conditions with "go test -race" val.Add(1) @@ -406,7 +406,7 @@ func TestPolicyDisposer(t *testing.T) { t: 10 * time.Millisecond, r: NewRetry(retry.Config{MaxRetries: 5}, NewRetryConditionMatch()), } - policy := NewRunnerWithOptions(context.Background(), policyDef, RunnerOpts[int32]{ + policy := NewRunnerWithOptions(t.Context(), policyDef, RunnerOpts[int32]{ Disposer: func(i int32) { disposerCalled <- i }, diff --git a/pkg/resiliency/resiliency_test.go b/pkg/resiliency/resiliency_test.go index 111e915ab..6f2bf8e12 100644 --- a/pkg/resiliency/resiliency_test.go +++ b/pkg/resiliency/resiliency_test.go @@ -173,7 +173,7 @@ func getOperatorClient(address string) operatorv1pb.OperatorClient { } func TestPoliciesForTargets(t *testing.T) { - ctx := context.Background() + ctx := t.Context() configs := LoadLocalResiliency(log, "default", "./testdata") assert.Len(t, configs, 1) r := FromConfigurations(log, configs...) @@ -670,7 +670,7 @@ func TestDefaultPoliciesAreUsedIfNoTargetPolicyExists(t *testing.T) { r := FromConfigurations(log, config) // Targeted App - policy := NewRunner[any](context.Background(), + policy := NewRunner[any](t.Context(), r.EndpointPolicy("testApp", "localhost"), ) count := atomic.Int64{} @@ -691,7 +691,7 @@ func TestDefaultPoliciesAreUsedIfNoTargetPolicyExists(t *testing.T) { }, 2) // actorType is not a known target, so we get 1 retry + original call as default circuit breaker trips (consecutiveFailures > 1) // One last one for ActorPostLock which just includes timeouts. - policy = NewRunner[any](context.Background(), + policy = NewRunner[any](t.Context(), r.ActorPostLockPolicy("actorType", "actorID"), ) count.Store(0) @@ -714,7 +714,7 @@ func concurrentPolicyExec(t *testing.T, policyDefFn func(idx int) *PolicyDefinit go func(i int) { defer wg.Done() // Not defined - policy := NewRunner[any](context.Background(), policyDefFn(i)) + policy := NewRunner[any](t.Context(), policyDefFn(i)) count := atomic.Int64{} count.Store(0) policy(func(ctx context.Context) (any, error) { diff --git a/pkg/runtime/hotreload/loader/disk/resource_test.go b/pkg/runtime/hotreload/loader/disk/resource_test.go index 4a66c977c..a8c3c726d 100644 --- a/pkg/runtime/hotreload/loader/disk/resource_test.go +++ b/pkg/runtime/hotreload/loader/disk/resource_test.go @@ -77,7 +77,7 @@ func Test_Disk(t *testing.T) { require.NoError(t, err) errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { errCh <- d.Run(ctx) }() @@ -88,7 +88,7 @@ func Test_Disk(t *testing.T) { assert.Empty(t, store.ListComponents()) - conn, err := d.Components().Stream(context.Background()) + conn, err := d.Components().Stream(t.Context()) require.NoError(t, err) err = os.WriteFile(filepath.Join(dir, "f.yaml"), []byte(strings.Join([]string{comp1, comp2, comp3}, "\n---\n")), 0o600) @@ -154,7 +154,7 @@ func Test_Stream(t *testing.T) { }) errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(func() { cancel() require.NoError(t, <-errCh) @@ -171,7 +171,7 @@ func Test_Stream(t *testing.T) { batcher.Batch(0, struct{}{}) - conn, err := r.Stream(context.Background()) + conn, err := r.Stream(t.Context()) require.NoError(t, err) var events []*loader.Event[componentsapi.Component] @@ -237,7 +237,7 @@ func Test_Stream(t *testing.T) { }) errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(func() { cancel() require.NoError(t, <-errCh) @@ -254,7 +254,7 @@ func Test_Stream(t *testing.T) { batcher.Batch(0, struct{}{}) - conn, err := r.Stream(context.Background()) + conn, err := r.Stream(t.Context()) require.NoError(t, err) var events []*loader.Event[componentsapi.Component] @@ -321,7 +321,7 @@ func Test_Stream(t *testing.T) { }) errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(func() { cancel() require.NoError(t, <-errCh) @@ -338,7 +338,7 @@ func Test_Stream(t *testing.T) { batcher.Batch(0, struct{}{}) - conn, err := r.Stream(context.Background()) + conn, err := r.Stream(t.Context()) require.NoError(t, err) var events []*loader.Event[componentsapi.Component] diff --git a/pkg/runtime/hotreload/loader/operator/resource_test.go b/pkg/runtime/hotreload/loader/operator/resource_test.go index 032feb432..f3b82a146 100644 --- a/pkg/runtime/hotreload/loader/operator/resource_test.go +++ b/pkg/runtime/hotreload/loader/operator/resource_test.go @@ -40,7 +40,7 @@ func Test_generic(t *testing.T) { ) require.NoError(t, r.close()) - ch, err := r.Stream(context.Background()) + ch, err := r.Stream(t.Context()) assert.Nil(t, ch) require.ErrorContains(t, err, "stream is closed") }) @@ -53,7 +53,7 @@ func Test_generic(t *testing.T) { streamer, ) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) cancel() streamer.establishFn = func(context.Context, operatorpb.OperatorClient, string, string) error { @@ -78,7 +78,7 @@ func Test_generic(t *testing.T) { return <-recCh, nil } - ch, err := r.Stream(context.Background()) + ch, err := r.Stream(t.Context()) assert.NotNil(t, ch) require.NoError(t, err) @@ -131,7 +131,7 @@ func Test_generic(t *testing.T) { return nil, errors.New("recv error") } - conn, err := r.Stream(context.Background()) + conn, err := r.Stream(t.Context()) require.NoError(t, err) select { diff --git a/pkg/runtime/hotreload/reconciler/reconciler_test.go b/pkg/runtime/hotreload/reconciler/reconciler_test.go index 249e66705..1a2b313b7 100644 --- a/pkg/runtime/hotreload/reconciler/reconciler_test.go +++ b/pkg/runtime/hotreload/reconciler/reconciler_test.go @@ -54,7 +54,7 @@ func Test_Run(t *testing.T) { r.clock = fakeClock errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { errCh <- r.Run(ctx) }() @@ -111,7 +111,7 @@ func Test_Run(t *testing.T) { r.manager = mngr errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { errCh <- r.Run(ctx) }() @@ -208,7 +208,7 @@ func Test_Run(t *testing.T) { r.manager = mngr errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { errCh <- r.Run(ctx) }() @@ -317,7 +317,7 @@ func Test_reconcile(t *testing.T) { recDone := make(chan struct{}) go func() { defer close(recDone) - r.reconcile(context.Background(), &differ.Result[componentsapi.Component]{ + r.reconcile(t.Context(), &differ.Result[componentsapi.Component]{ Deleted: deleted, Updated: updated, Created: created, @@ -385,21 +385,21 @@ func Test_handleEvent(t *testing.T) { assert.Equal(t, 0, updateCalled) assert.Equal(t, 0, deleteCalled) - r.handleEvent(context.Background(), &loader.Event[componentsapi.Component]{ + r.handleEvent(t.Context(), &loader.Event[componentsapi.Component]{ Type: operator.ResourceEventType_CREATED, Resource: comp1, }) assert.Equal(t, 1, updateCalled) assert.Equal(t, 0, deleteCalled) - r.handleEvent(context.Background(), &loader.Event[componentsapi.Component]{ + r.handleEvent(t.Context(), &loader.Event[componentsapi.Component]{ Type: operator.ResourceEventType_UPDATED, Resource: comp1, }) assert.Equal(t, 2, updateCalled) assert.Equal(t, 0, deleteCalled) - r.handleEvent(context.Background(), &loader.Event[componentsapi.Component]{ + r.handleEvent(t.Context(), &loader.Event[componentsapi.Component]{ Type: operator.ResourceEventType_DELETED, Resource: comp1, }) diff --git a/pkg/runtime/processor/binding/init_test.go b/pkg/runtime/processor/binding/init_test.go index 62e143b1f..5054eef2d 100644 --- a/pkg/runtime/processor/binding/init_test.go +++ b/pkg/runtime/processor/binding/init_test.go @@ -14,7 +14,6 @@ limitations under the License. package binding_test import ( - "context" "testing" "github.com/stretchr/testify/require" @@ -54,7 +53,7 @@ func TestInitBindings(t *testing.T) { c := compapi.Component{} c.ObjectMeta.Name = "testInputBinding" c.Spec.Type = "bindings.testInputBinding" - err := proc.Init(context.TODO(), c) + err := proc.Init(t.Context(), c) require.NoError(t, err) }) @@ -77,7 +76,7 @@ func TestInitBindings(t *testing.T) { c := compapi.Component{} c.ObjectMeta.Name = "testOutputBinding" c.Spec.Type = "bindings.testOutputBinding" - err := proc.Init(context.TODO(), c) + err := proc.Init(t.Context(), c) require.NoError(t, err) }) @@ -108,13 +107,13 @@ func TestInitBindings(t *testing.T) { input := compapi.Component{} input.ObjectMeta.Name = "testinput" input.Spec.Type = "bindings.testinput" - err := proc.Init(context.TODO(), input) + err := proc.Init(t.Context(), input) require.NoError(t, err) output := compapi.Component{} output.ObjectMeta.Name = "testoutput" output.Spec.Type = "bindings.testoutput" - err = proc.Init(context.TODO(), output) + err = proc.Init(t.Context(), output) require.NoError(t, err) }) @@ -134,7 +133,7 @@ func TestInitBindings(t *testing.T) { c := compapi.Component{} c.ObjectMeta.Name = "testNotExistBinding" c.Spec.Type = "bindings.testNotExistBinding" - err := proc.Init(context.TODO(), c) + err := proc.Init(t.Context(), c) require.Error(t, err) }) } diff --git a/pkg/runtime/processor/binding/send_test.go b/pkg/runtime/processor/binding/send_test.go index d5e12ab53..a5528cfae 100644 --- a/pkg/runtime/processor/binding/send_test.go +++ b/pkg/runtime/processor/binding/send_test.go @@ -129,7 +129,7 @@ func TestStartReadingFromBindings(t *testing.T) { m := &rtmock.Binding{} b.compStore.AddInputBinding("test", m) - err := b.StartReadingFromBindings(context.Background()) + err := b.StartReadingFromBindings(t.Context()) require.NoError(t, err) assert.True(t, mockAppChannel.AssertCalled(t, "InvokeMethod", mock.Anything, mock.Anything)) @@ -171,14 +171,14 @@ func TestStartReadingFromBindings(t *testing.T) { }, })) require.NoError(t, b.compStore.CommitPendingComponent()) - err := b.StartReadingFromBindings(context.Background()) + err := b.StartReadingFromBindings(t.Context()) require.NoError(t, err) assert.True(t, mockAppChannel.AssertCalled(t, "InvokeMethod", mock.Anything, mock.Anything)) }) } func TestGetSubscribedBindingsGRPC(t *testing.T) { - secP, err := security.New(context.Background(), security.Options{ + secP, err := security.New(t.Context(), security.Options{ TrustAnchors: []byte("test"), AppID: "test", ControlPlaneTrustDomain: "test.example.com", @@ -190,8 +190,8 @@ func TestGetSubscribedBindingsGRPC(t *testing.T) { Healthz: healthz.New(), }) require.NoError(t, err) - go secP.Run(context.Background()) - sec, err := secP.Handler(context.Background()) + go secP.Run(t.Context()) + sec, err := secP.Handler(t.Context()) require.NoError(t, err) testCases := []struct { @@ -228,7 +228,7 @@ func TestGetSubscribedBindingsGRPC(t *testing.T) { }) defer grpcServer.Stop() // act - resp, _ := b.getSubscribedBindingsGRPC(context.Background()) + resp, _ := b.getSubscribedBindingsGRPC(t.Context()) // assert assert.Equal(t, tc.expectedResponse, resp, "expected response to match") @@ -272,7 +272,7 @@ func TestReadInputBindings(t *testing.T) { b.compStore.AddInputBindingRoute(testInputBindingName, testInputBindingName) mockBinding := rtmock.Binding{} - ctx, cancel := context.WithTimeout(context.Background(), 500*time.Millisecond) + ctx, cancel := context.WithTimeout(t.Context(), 500*time.Millisecond) ch := make(chan bool, 1) mockBinding.ReadErrorCh = ch b.readFromBinding(ctx, testInputBindingName, &mockBinding) @@ -318,7 +318,7 @@ func TestReadInputBindings(t *testing.T) { b.compStore.AddInputBindingRoute(testInputBindingName, testInputBindingName) mockBinding := rtmock.Binding{} - ctx, cancel := context.WithTimeout(context.Background(), 500*time.Millisecond) + ctx, cancel := context.WithTimeout(t.Context(), 500*time.Millisecond) ch := make(chan bool, 1) mockBinding.ReadErrorCh = ch b.readFromBinding(ctx, testInputBindingName, &mockBinding) @@ -364,7 +364,7 @@ func TestReadInputBindings(t *testing.T) { b.compStore.AddInputBindingRoute(testInputBindingName, testInputBindingName) mockBinding := rtmock.Binding{Metadata: map[string]string{"bindings": "input"}} - ctx, cancel := context.WithTimeout(context.Background(), 500*time.Millisecond) + ctx, cancel := context.WithTimeout(t.Context(), 500*time.Millisecond) ch := make(chan bool, 1) mockBinding.ReadErrorCh = ch b.readFromBinding(ctx, testInputBindingName, &mockBinding) @@ -390,7 +390,7 @@ func TestReadInputBindings(t *testing.T) { mockBinding.SetOnReadCloseCh(closeCh) mockBinding.On("Read", mock.MatchedBy(daprt.MatchContextInterface), mock.Anything).Return(nil).Once() - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) b.readFromBinding(ctx, testInputBindingName, mockBinding) time.Sleep(80 * time.Millisecond) cancel() @@ -416,7 +416,7 @@ func TestInvokeOutputBindings(t *testing.T) { }) b.channels = new(channels.Channels).WithAppChannel(mockAppChannel) - _, err := b.SendToOutputBinding(context.Background(), "mockBinding", &bindings.InvokeRequest{ + _, err := b.SendToOutputBinding(t.Context(), "mockBinding", &bindings.InvokeRequest{ Data: []byte(""), }) require.Error(t, err) @@ -435,7 +435,7 @@ func TestInvokeOutputBindings(t *testing.T) { b.compStore.AddOutputBinding("mockBinding", &rtmock.Binding{}) - _, err := b.SendToOutputBinding(context.Background(), "mockBinding", &bindings.InvokeRequest{ + _, err := b.SendToOutputBinding(t.Context(), "mockBinding", &bindings.InvokeRequest{ Data: []byte(""), Operation: bindings.CreateOperation, }) @@ -454,7 +454,7 @@ func TestInvokeOutputBindings(t *testing.T) { b.compStore.AddOutputBinding("mockBinding", &rtmock.Binding{}) - _, err := b.SendToOutputBinding(context.Background(), "mockBinding", &bindings.InvokeRequest{ + _, err := b.SendToOutputBinding(t.Context(), "mockBinding", &bindings.InvokeRequest{ Data: []byte(""), Operation: bindings.GetOperation, }) @@ -476,7 +476,7 @@ func TestBindingTracingHttp(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, mock.Anything).Return(invokev1.NewInvokeMethodResponse(200, "OK", nil), nil) b.channels = new(channels.Channels).WithAppChannel(mockAppChannel) - _, err := b.sendBindingEventToApp(context.Background(), "mockBinding", []byte(""), map[string]string{"traceparent": "00-d97eeaf10b4d00dc6ba794f3a41c5268-09462d216dd14deb-01"}) + _, err := b.sendBindingEventToApp(t.Context(), "mockBinding", []byte(""), map[string]string{"traceparent": "00-d97eeaf10b4d00dc6ba794f3a41c5268-09462d216dd14deb-01"}) require.NoError(t, err) mockAppChannel.AssertCalled(t, "InvokeMethod", mock.Anything, mock.Anything) assert.Len(t, mockAppChannel.Calls, 1) @@ -490,7 +490,7 @@ func TestBindingTracingHttp(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, mock.Anything).Return(invokev1.NewInvokeMethodResponse(204, "OK", nil), nil) b.channels = new(channels.Channels).WithAppChannel(mockAppChannel) - _, err := b.sendBindingEventToApp(context.Background(), "mockBinding", []byte(""), map[string]string{"traceparent": "00-d97eeaf10b4d00dc6ba794f3a41c5268-09462d216dd14deb-01"}) + _, err := b.sendBindingEventToApp(t.Context(), "mockBinding", []byte(""), map[string]string{"traceparent": "00-d97eeaf10b4d00dc6ba794f3a41c5268-09462d216dd14deb-01"}) require.NoError(t, err) mockAppChannel.AssertCalled(t, "InvokeMethod", mock.Anything, mock.Anything) assert.Len(t, mockAppChannel.Calls, 1) @@ -504,7 +504,7 @@ func TestBindingTracingHttp(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, mock.Anything).Return(invokev1.NewInvokeMethodResponse(200, "OK", nil), nil) b.channels = new(channels.Channels).WithAppChannel(mockAppChannel) - _, err := b.sendBindingEventToApp(context.Background(), "mockBinding", []byte(""), map[string]string{"traceparent": "I am not a traceparent"}) + _, err := b.sendBindingEventToApp(t.Context(), "mockBinding", []byte(""), map[string]string{"traceparent": "I am not a traceparent"}) require.NoError(t, err) mockAppChannel.AssertCalled(t, "InvokeMethod", mock.Anything, mock.Anything) assert.Len(t, mockAppChannel.Calls, 1) @@ -560,7 +560,7 @@ func TestBindingResiliency(t *testing.T) { output := componentsV1alpha1.Component{} output.ObjectMeta.Name = "failOutput" output.Spec.Type = "bindings.failingoutput" - err := b.Init(context.TODO(), output) + err := b.Init(t.Context(), output) require.NoError(t, err) t.Run("output binding retries on failure with resiliency", func(t *testing.T) { @@ -568,7 +568,7 @@ func TestBindingResiliency(t *testing.T) { Data: []byte("outputFailingKey"), Operation: "create", } - _, err := b.SendToOutputBinding(context.Background(), "failOutput", req) + _, err := b.SendToOutputBinding(t.Context(), "failOutput", req) require.NoError(t, err) assert.Equal(t, 2, failingBinding.Failure.CallCount("outputFailingKey")) @@ -580,7 +580,7 @@ func TestBindingResiliency(t *testing.T) { Operation: "create", } start := time.Now() - _, err := b.SendToOutputBinding(context.Background(), "failOutput", req) + _, err := b.SendToOutputBinding(t.Context(), "failOutput", req) end := time.Now() require.Error(t, err) @@ -589,7 +589,7 @@ func TestBindingResiliency(t *testing.T) { }) t.Run("input binding retries on failure with resiliency", func(t *testing.T) { - _, err := b.sendBindingEventToApp(context.Background(), "failingInputBinding", []byte("inputFailingKey"), map[string]string{}) + _, err := b.sendBindingEventToApp(t.Context(), "failingInputBinding", []byte("inputFailingKey"), map[string]string{}) require.NoError(t, err) assert.Equal(t, 2, failingChannel.Failure.CallCount("inputFailingKey")) @@ -597,7 +597,7 @@ func TestBindingResiliency(t *testing.T) { t.Run("input binding times out with resiliency", func(t *testing.T) { start := time.Now() - _, err := b.sendBindingEventToApp(context.Background(), "failingInputBinding", []byte("inputTimeoutKey"), map[string]string{}) + _, err := b.sendBindingEventToApp(t.Context(), "failingInputBinding", []byte("inputTimeoutKey"), map[string]string{}) end := time.Now() require.Error(t, err) diff --git a/pkg/runtime/processor/middleware/middleware_test.go b/pkg/runtime/processor/middleware/middleware_test.go index 8203d635c..df9c24697 100644 --- a/pkg/runtime/processor/middleware/middleware_test.go +++ b/pkg/runtime/processor/middleware/middleware_test.go @@ -14,7 +14,6 @@ limitations under the License. package middleware import ( - "context" nethttp "net/http" "testing" @@ -45,7 +44,7 @@ func TestInit(t *testing.T) { HTTP: http.New(), }) - err := m.Init(context.Background(), compapi.Component{ + err := m.Init(t.Context(), compapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "test", }, @@ -76,7 +75,7 @@ func TestInit(t *testing.T) { } }, "mock") - err := m.Init(context.Background(), compapi.Component{ + err := m.Init(t.Context(), compapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "test", }, @@ -107,7 +106,7 @@ func TestInit(t *testing.T) { } }, "mock") - err := m.Init(context.Background(), compapi.Component{ + err := m.Init(t.Context(), compapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "test", }, @@ -135,7 +134,7 @@ func TestInit(t *testing.T) { } }, "notmock") - err := m.Init(context.Background(), compapi.Component{ + err := m.Init(t.Context(), compapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "test", }, @@ -189,7 +188,7 @@ func TestInit(t *testing.T) { assert.Equal(t, 1, rootCalled) assert.Equal(t, 0, middlewareCalled) - err := m.Init(context.Background(), compapi.Component{ + err := m.Init(t.Context(), compapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "test", }, @@ -241,11 +240,11 @@ func TestInit(t *testing.T) { } }, "mock") - require.NoError(t, m.Init(context.Background(), compapi.Component{ + require.NoError(t, m.Init(t.Context(), compapi.Component{ ObjectMeta: metav1.ObjectMeta{Name: "test1"}, Spec: compapi.ComponentSpec{Type: "middleware.http.mock", Version: "v1"}, })) - require.NoError(t, m.Init(context.Background(), compapi.Component{ + require.NoError(t, m.Init(t.Context(), compapi.Component{ ObjectMeta: metav1.ObjectMeta{Name: "test2"}, Spec: compapi.ComponentSpec{Type: "middleware.http.mock", Version: "v1"}, })) diff --git a/pkg/runtime/processor/processor_test.go b/pkg/runtime/processor/processor_test.go index 6d52516ea..6aefa5987 100644 --- a/pkg/runtime/processor/processor_test.go +++ b/pkg/runtime/processor/processor_test.go @@ -112,7 +112,7 @@ func TestProcessComponentsAndDependents(t *testing.T) { } t.Run("test incorrect type", func(t *testing.T) { - err := proc.processComponentAndDependents(context.Background(), incorrectComponentType) + err := proc.processComponentAndDependents(t.Context(), incorrectComponentType) require.Error(t, err, "expected an error") assert.Equal(t, "incorrect type pubsubs.mockPubSub", err.Error(), "expected error strings to match") }) @@ -129,7 +129,7 @@ func TestInitSecretStores(t *testing.T) { "kubernetesMock", ) - err := proc.processComponentAndDependents(context.Background(), componentsapi.Component{ + err := proc.processComponentAndDependents(t.Context(), componentsapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMock", }, @@ -151,7 +151,7 @@ func TestInitSecretStores(t *testing.T) { "kubernetesMock", ) - err := proc.processComponentAndDependents(context.Background(), componentsapi.Component{ + err := proc.processComponentAndDependents(t.Context(), componentsapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMock", }, @@ -176,7 +176,7 @@ func TestInitSecretStores(t *testing.T) { "kubernetesMock", ) - proc.processComponentAndDependents(context.Background(), componentsapi.Component{ + proc.processComponentAndDependents(t.Context(), componentsapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMock", }, @@ -285,7 +285,7 @@ func TestMetadataUUID(t *testing.T) { assert.NotEqual(t, uuid1, uuid2) }) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) } @@ -330,7 +330,7 @@ func TestMetadataPodName(t *testing.T) { assert.Equal(t, "testPodName", consumerID) }) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) } @@ -376,7 +376,7 @@ func TestMetadataNamespace(t *testing.T) { assert.Equal(t, "test.app1", consumerID) }) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) } @@ -424,7 +424,7 @@ func TestMetadataClientID(t *testing.T) { clientIDChan <- k8sClientID }) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) select { @@ -466,7 +466,7 @@ func TestMetadataClientID(t *testing.T) { clientIDChan <- standAloneClientID }) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) appIds := strings.Split(standAloneClientID, " ") assert.Len(t, appIds, 2) @@ -522,7 +522,7 @@ func TestReporter(t *testing.T) { mockPubSub.On("Init", mock.Anything).Return(nil) mockPubSub.On("Close").Return(nil) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) select { @@ -558,7 +558,7 @@ func TestReporter(t *testing.T) { mockPubSub.On("Init", mock.Anything, mock.Anything).Return(errors.New("error")) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.Error(t, err) select { @@ -595,7 +595,7 @@ func TestReporter(t *testing.T) { mockPubSub.On("Init", mock.Anything, mock.Anything).Return(nil) mockPubSub.On("Close").Return(nil) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) // consume the init message @@ -635,7 +635,7 @@ func TestReporter(t *testing.T) { mockPubSub.On("Init", mock.Anything, mock.Anything).Return(nil) mockPubSub.On("Close").Return(errors.New("error")) - err := proc.processComponentAndDependents(context.Background(), pubsubComponent) + err := proc.processComponentAndDependents(t.Context(), pubsubComponent) require.NoError(t, err) // consume the init message @@ -657,7 +657,7 @@ func TestReporter(t *testing.T) { } func TestProcessorWaitGroupError(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) t.Cleanup(func() { cancel() diff --git a/pkg/runtime/processor/pubsub/pubsub_test.go b/pkg/runtime/processor/pubsub/pubsub_test.go index 2949e6f0d..b168a10ec 100644 --- a/pkg/runtime/processor/pubsub/pubsub_test.go +++ b/pkg/runtime/processor/pubsub/pubsub_test.go @@ -143,7 +143,7 @@ func TestInitPubSub(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), matchDaprRequestMethod("dapr/subscribe")).Return(fakeResp, nil) for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -175,7 +175,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -208,7 +208,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -240,7 +240,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -270,7 +270,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - require.NoError(t, ps.Init(context.Background(), comp)) + require.NoError(t, ps.Init(t.Context(), comp)) } // assert @@ -302,7 +302,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -334,7 +334,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -361,7 +361,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -384,7 +384,7 @@ func TestInitPubSub(t *testing.T) { // act for _, comp := range pubsubComponents { - err := ps.Init(context.Background(), comp) + err := ps.Init(t.Context(), comp) require.NoError(t, err) } @@ -458,7 +458,7 @@ func TestConsumerID(t *testing.T) { ComponentStore: compStore, }) - err := ps.Init(context.Background(), pubsubComponent) + err := ps.Init(t.Context(), pubsubComponent) require.NoError(t, err) } diff --git a/pkg/runtime/processor/secret/secret_test.go b/pkg/runtime/processor/secret/secret_test.go index b7bcfe3b1..f283fba0e 100644 --- a/pkg/runtime/processor/secret/secret_test.go +++ b/pkg/runtime/processor/secret/secret_test.go @@ -14,7 +14,6 @@ limitations under the License. package secret import ( - "context" "testing" "github.com/stretchr/testify/assert" @@ -77,7 +76,7 @@ func TestProcessResourceSecrets(t *testing.T) { ) // add Kubernetes component manually - require.NoError(t, sec.Init(context.Background(), componentsapi.Component{ + require.NoError(t, sec.Init(t.Context(), componentsapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: compsecret.BuiltinKubernetesSecretStore, }, @@ -87,7 +86,7 @@ func TestProcessResourceSecrets(t *testing.T) { }, })) - updated, unready := sec.ProcessResource(context.Background(), mockBinding) + updated, unready := sec.ProcessResource(t.Context(), mockBinding) assert.True(t, updated) assert.Equal(t, "value1", mockBinding.Spec.Metadata[0].Value.String()) assert.Empty(t, unready) @@ -120,7 +119,7 @@ func TestProcessResourceSecrets(t *testing.T) { ) // initSecretStore appends Kubernetes component even if kubernetes component is not added - err := sec.Init(context.Background(), componentsapi.Component{ + err := sec.Init(t.Context(), componentsapi.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "mock", }, @@ -131,7 +130,7 @@ func TestProcessResourceSecrets(t *testing.T) { }) require.NoError(t, err) - updated, unready := sec.ProcessResource(context.Background(), mockBinding) + updated, unready := sec.ProcessResource(t.Context(), mockBinding) assert.True(t, updated) assert.Equal(t, "value1", mockBinding.Spec.Metadata[0].Value.String()) assert.Empty(t, unready) @@ -155,7 +154,7 @@ func TestProcessResourceSecrets(t *testing.T) { }), }) - updated, unready := sec.ProcessResource(context.Background(), mockBinding) + updated, unready := sec.ProcessResource(t.Context(), mockBinding) assert.True(t, updated) assert.Equal(t, "ciao mondo", mockBinding.Spec.Metadata[0].Value.String()) assert.Empty(t, unready) @@ -186,7 +185,7 @@ func TestProcessResourceSecrets(t *testing.T) { }), }) - updated, unready := sec.ProcessResource(context.Background(), mockBinding) + updated, unready := sec.ProcessResource(t.Context(), mockBinding) assert.True(t, updated) assert.Equal(t, "", mockBinding.Spec.Metadata[0].Value.String()) assert.Equal(t, "", mockBinding.Spec.Metadata[1].Value.String()) diff --git a/pkg/runtime/processor/state/state_test.go b/pkg/runtime/processor/state/state_test.go index a20e704b8..542d50857 100644 --- a/pkg/runtime/processor/state/state_test.go +++ b/pkg/runtime/processor/state/state_test.go @@ -14,7 +14,6 @@ limitations under the License. package state_test import ( - "context" "crypto/rand" "encoding/hex" "testing" @@ -95,7 +94,7 @@ func TestInitState(t *testing.T) { initMockStateStoreForRegistry(reg, "noerror", primaryKey, nil) // act - err := proc.Init(context.TODO(), mockStateComponent("noerror")) + err := proc.Init(t.Context(), mockStateComponent("noerror")) // assert require.NoError(t, err, "expected no error") @@ -106,7 +105,7 @@ func TestInitState(t *testing.T) { initMockStateStoreForRegistry(reg, "error", primaryKey, assert.AnError) // act - err := proc.Init(context.TODO(), mockStateComponent("error")) + err := proc.Init(t.Context(), mockStateComponent("error")) // assert require.Error(t, err, "expected error") @@ -118,7 +117,7 @@ func TestInitState(t *testing.T) { initMockStateStoreForRegistry(reg, "noencryption", primaryKey, nil) // act - err := proc.Init(context.TODO(), mockStateComponent("noencryption")) + err := proc.Init(t.Context(), mockStateComponent("noencryption")) ok := encryption.EncryptedStateStore("noencryption") // assert @@ -132,7 +131,7 @@ func TestInitState(t *testing.T) { compStore.AddSecretStore("mockSecretStore", &mock.SecretStore{}) - err := proc.Init(context.TODO(), mockStateComponent("encryption")) + err := proc.Init(t.Context(), mockStateComponent("encryption")) ok := encryption.EncryptedStateStore("encryption") // assert diff --git a/pkg/runtime/processor/subscriber/subscriber_test.go b/pkg/runtime/processor/subscriber/subscriber_test.go index 58fae4ca4..ec7396541 100644 --- a/pkg/runtime/processor/subscriber/subscriber_test.go +++ b/pkg/runtime/processor/subscriber/subscriber_test.go @@ -14,7 +14,6 @@ limitations under the License. package subscriber import ( - "context" "encoding/json" "slices" "sync/atomic" @@ -57,9 +56,9 @@ func TestSubscriptionLifecycle(t *testing.T) { mockPubSub1.On("unsubscribed", "topic1").Return(nil) mockPubSub2.On("unsubscribed", "topic2").Return(nil) mockPubSub3.On("unsubscribed", "topic3").Return(nil) - require.NoError(t, mockPubSub1.Init(context.Background(), contribpubsub.Metadata{})) - require.NoError(t, mockPubSub2.Init(context.Background(), contribpubsub.Metadata{})) - require.NoError(t, mockPubSub3.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, mockPubSub1.Init(t.Context(), contribpubsub.Metadata{})) + require.NoError(t, mockPubSub2.Init(t.Context(), contribpubsub.Metadata{})) + require.NoError(t, mockPubSub3.Init(t.Context(), contribpubsub.Metadata{})) compStore := compstore.New() compStore.AddPubSub("mockPubSub1", &rtpubsub.PubsubItem{ @@ -204,7 +203,7 @@ func Test_initProgramaticSubscriptions(t *testing.T) { AppID: TestRuntimeConfigID, Channels: new(channels.Channels), }) - require.NoError(t, subs.initProgramaticSubscriptions(context.Background())) + require.NoError(t, subs.initProgramaticSubscriptions(t.Context())) assert.Empty(t, compStore.ListProgramaticSubscriptions()) }) @@ -219,7 +218,7 @@ func Test_initProgramaticSubscriptions(t *testing.T) { AppID: TestRuntimeConfigID, Channels: new(channels.Channels), }) - require.NoError(t, subs.initProgramaticSubscriptions(context.Background())) + require.NoError(t, subs.initProgramaticSubscriptions(t.Context())) assert.Empty(t, compStore.ListProgramaticSubscriptions()) }) @@ -252,11 +251,11 @@ func Test_initProgramaticSubscriptions(t *testing.T) { WithContentType("application/json") defer fakeResp.Close() - mockAppChannel.On("InvokeMethod", mock.AnythingOfType("context.backgroundCtx"), mock.AnythingOfType("*v1.InvokeMethodRequest")).Return(fakeResp, nil) + mockAppChannel.On("InvokeMethod", mock.AnythingOfType("*context.cancelCtx"), mock.AnythingOfType("*v1.InvokeMethodRequest")).Return(fakeResp, nil) - require.NoError(t, subs.initProgramaticSubscriptions(context.Background())) - require.NoError(t, subs.initProgramaticSubscriptions(context.Background())) - require.NoError(t, subs.initProgramaticSubscriptions(context.Background())) + require.NoError(t, subs.initProgramaticSubscriptions(t.Context())) + require.NoError(t, subs.initProgramaticSubscriptions(t.Context())) + require.NoError(t, subs.initProgramaticSubscriptions(t.Context())) assert.Len(t, compStore.ListProgramaticSubscriptions(), 1) }) } @@ -278,9 +277,9 @@ func TestReloadPubSub(t *testing.T) { mockPubSub1.On("unsubscribed", "topic4").Return(nil) mockPubSub2.On("unsubscribed", "topic5").Return(nil) mockPubSub3.On("unsubscribed", "topic6").Return(nil) - require.NoError(t, mockPubSub1.Init(context.Background(), contribpubsub.Metadata{})) - require.NoError(t, mockPubSub2.Init(context.Background(), contribpubsub.Metadata{})) - require.NoError(t, mockPubSub3.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, mockPubSub1.Init(t.Context(), contribpubsub.Metadata{})) + require.NoError(t, mockPubSub2.Init(t.Context(), contribpubsub.Metadata{})) + require.NoError(t, mockPubSub3.Init(t.Context(), contribpubsub.Metadata{})) compStore := compstore.New() compStore.AddPubSub("mockPubSub1", &rtpubsub.PubsubItem{ diff --git a/pkg/runtime/pubsub/bulkpublish_resiliency_test.go b/pkg/runtime/pubsub/bulkpublish_resiliency_test.go index ef1a33cec..e417734cc 100644 --- a/pkg/runtime/pubsub/bulkpublish_resiliency_test.go +++ b/pkg/runtime/pubsub/bulkpublish_resiliency_test.go @@ -110,7 +110,7 @@ func (m *mockBulkPublisher) BulkPublish(ctx context.Context, req *contribPubsub. } func TestApplyBulkPublishResiliency(t *testing.T) { - ctx := context.Background() + ctx := t.Context() pubsubName := "test-pubsub" bulkMessageEntries := []contribPubsub.BulkMessageEntry{ { diff --git a/pkg/runtime/pubsub/default_bulkpub_test.go b/pkg/runtime/pubsub/default_bulkpub_test.go index ca61e9dd7..2cb3cf2c6 100644 --- a/pkg/runtime/pubsub/default_bulkpub_test.go +++ b/pkg/runtime/pubsub/default_bulkpub_test.go @@ -14,7 +14,6 @@ limitations under the License. package pubsub import ( - "context" "errors" "testing" @@ -97,7 +96,7 @@ func TestBulkPublish_DefaultBulkPublisher(t *testing.T) { } bulkPublisher := NewDefaultBulkPublisher(mockPubSub) - res, err := bulkPublisher.BulkPublish(context.Background(), req) + res, err := bulkPublisher.BulkPublish(t.Context(), req) // Check if the bulk publish method returns an error. if tc.nErrors > 0 { diff --git a/pkg/runtime/pubsub/default_bulksub_test.go b/pkg/runtime/pubsub/default_bulksub_test.go index 9575e579c..bde2c6784 100644 --- a/pkg/runtime/pubsub/default_bulksub_test.go +++ b/pkg/runtime/pubsub/default_bulksub_test.go @@ -72,7 +72,7 @@ func TestFlushMessages(t *testing.T) { return nil, nil } - flushMessages(context.Background(), "topic", tc.messages, tc.msgCbMap, handler) + flushMessages(t.Context(), "topic", tc.messages, tc.msgCbMap, handler) assert.Equal(t, tc.expectedHandlerInvoked, handlerInvoked) }) } @@ -141,7 +141,7 @@ func TestFlushMessages(t *testing.T) { "3": func(err error) { invokedCallbacks["3"] = err }, } - flushMessages(context.Background(), "topic", messages, msgCbMap, handler) + flushMessages(t.Context(), "topic", messages, msgCbMap, handler) for id, err := range invokedCallbacks { if _, ok := tc.entryIdErrMap[id]; ok { diff --git a/pkg/runtime/pubsub/outbox_test.go b/pkg/runtime/pubsub/outbox_test.go index 911a57523..e68526f5d 100644 --- a/pkg/runtime/pubsub/outbox_test.go +++ b/pkg/runtime/pubsub/outbox_test.go @@ -255,7 +255,7 @@ func TestPublishInternal(t *testing.T) { }, }) - _, err := o.PublishInternal(context.Background(), "test", []state.TransactionalStateOperation{ + _, err := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "key", Value: "test", @@ -307,7 +307,7 @@ func TestPublishInternal(t *testing.T) { }, }) - _, err := o.PublishInternal(context.Background(), "test", []state.TransactionalStateOperation{ + _, err := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "key", Value: "test", @@ -361,7 +361,7 @@ func TestPublishInternal(t *testing.T) { }) contentType := "" - _, err := o.PublishInternal(context.TODO(), "test", []state.TransactionalStateOperation{ + _, err := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "key", Value: "test", @@ -431,7 +431,7 @@ func TestPublishInternal(t *testing.T) { require.NoError(t, err) contentType := "application/json" - _, err = o.PublishInternal(context.TODO(), "test", []state.TransactionalStateOperation{ + _, err = o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "key", Value: string(b), @@ -496,7 +496,7 @@ func TestPublishInternal(t *testing.T) { b, err := json.Marshal(&j) require.NoError(t, err) - _, err = o.PublishInternal(context.TODO(), "test", []state.TransactionalStateOperation{ + _, err = o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "key", Value: string(b), @@ -569,7 +569,7 @@ func TestPublishInternal(t *testing.T) { jp, err := json.Marshal(&projection) require.NoError(t, err) - _, err = o.PublishInternal(context.TODO(), "test", []state.TransactionalStateOperation{ + _, err = o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "key", Value: string(b), @@ -587,7 +587,7 @@ func TestPublishInternal(t *testing.T) { t.Run("missing state store", func(t *testing.T) { o := newTestOutbox(nil).(*outboxImpl) - _, err := o.PublishInternal(context.TODO(), "test", []state.TransactionalStateOperation{ + _, err := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "key", Value: "test", @@ -628,7 +628,7 @@ func TestPublishInternal(t *testing.T) { }, }) - _, err := o.PublishInternal(context.TODO(), "test", []state.TransactionalStateOperation{}, "testapp", "", "") + _, err := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{}, "testapp", "", "") require.NoError(t, err) }) @@ -664,7 +664,7 @@ func TestPublishInternal(t *testing.T) { }, }) - _, err := o.PublishInternal(context.TODO(), "test", []state.TransactionalStateOperation{ + _, err := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "1", Value: "hello", @@ -756,12 +756,12 @@ func TestSubscribeToInternalTopics(t *testing.T) { }) const appID = "test" - err := o.SubscribeToInternalTopics(context.Background(), appID) + err := o.SubscribeToInternalTopics(t.Context(), appID) require.NoError(t, err) errCh := make(chan error, 1) go func() { - trs, pErr := o.PublishInternal(context.Background(), "test", []state.TransactionalStateOperation{ + trs, pErr := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "1", Value: "hello", @@ -839,10 +839,10 @@ func TestSubscribeToInternalTopics(t *testing.T) { } const appID = "test" - err := o.SubscribeToInternalTopics(context.Background(), appID) + err := o.SubscribeToInternalTopics(t.Context(), appID) require.NoError(t, err) - trs, pErr := o.PublishInternal(context.Background(), "test", []state.TransactionalStateOperation{ + trs, pErr := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "1", Value: "hello", @@ -910,12 +910,12 @@ func TestSubscribeToInternalTopics(t *testing.T) { }) const appID = "test" - err := o.SubscribeToInternalTopics(context.Background(), appID) + err := o.SubscribeToInternalTopics(t.Context(), appID) require.NoError(t, err) errCh := make(chan error, 1) go func() { - trs, pErr := o.PublishInternal(context.Background(), "test", []state.TransactionalStateOperation{ + trs, pErr := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "1", Value: "hello", @@ -1036,12 +1036,12 @@ func TestSubscribeToInternalTopics(t *testing.T) { }) const appID = "test" - err := o.SubscribeToInternalTopics(context.Background(), appID) + err := o.SubscribeToInternalTopics(t.Context(), appID) require.NoError(t, err) errCh := make(chan error, 1) go func() { - trs, pErr := o.PublishInternal(context.Background(), "test", []state.TransactionalStateOperation{ + trs, pErr := o.PublishInternal(t.Context(), "test", []state.TransactionalStateOperation{ state.SetRequest{ Key: "1", Value: "hello", diff --git a/pkg/runtime/pubsub/publisher/publisher_test.go b/pkg/runtime/pubsub/publisher/publisher_test.go index c66327918..704d839fb 100644 --- a/pkg/runtime/pubsub/publisher/publisher_test.go +++ b/pkg/runtime/pubsub/publisher/publisher_test.go @@ -46,7 +46,7 @@ func TestPublish(t *testing.T) { md := make(map[string]string, 2) md["key"] = "v3" - res, err := ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err := ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestPubsubName, Topic: "topic0", Metadata: md, @@ -64,7 +64,7 @@ func TestPublish(t *testing.T) { assert.Empty(t, res.FailedEntries) compStore.AddPubSub(TestSecondPubsubName, &rtpubsub.PubsubItem{Component: &mockPublishPubSub{}}) - res, err = ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err = ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic1", Entries: []contribpubsub.BulkMessageEntry{ @@ -100,7 +100,7 @@ func TestPublish(t *testing.T) { md := make(map[string]string, 2) md["key"] = "v3" - res, err := ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err := ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestPubsubName, Topic: "topic0", Metadata: md, @@ -122,7 +122,7 @@ func TestPublish(t *testing.T) { ProtectedTopics: []string{"topic1"}, ScopedPublishings: []string{"topic1"}, }) - res, err = ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err = ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic1", Entries: []contribpubsub.BulkMessageEntry{ @@ -157,7 +157,7 @@ func TestPublish(t *testing.T) { md := make(map[string]string, 2) md["key"] = "v3" - res, err := ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err := ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestPubsubName, Topic: "topic5", Metadata: md, @@ -177,7 +177,7 @@ func TestPublish(t *testing.T) { Component: &mockPublishPubSub{}, AllowedTopics: []string{"topic1"}, }) - res, err = ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err = ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic5", Metadata: md, @@ -208,7 +208,7 @@ func TestPublish(t *testing.T) { md := make(map[string]string, 2) md["key"] = "v3" - res, err := ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err := ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestPubsubName, Topic: "topic1", Metadata: md, @@ -228,7 +228,7 @@ func TestPublish(t *testing.T) { Component: &mockPublishPubSub{}, ProtectedTopics: []string{"topic1"}, }) - res, err = ps.BulkPublish(context.Background(), &contribpubsub.BulkPublishRequest{ + res, err = ps.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic1", Metadata: md, @@ -259,7 +259,7 @@ func TestPublish(t *testing.T) { GetPubSubFn: compStore.GetPubSub, }) - err := ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err := ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestPubsubName, Topic: "topic0", Metadata: md, @@ -270,7 +270,7 @@ func TestPublish(t *testing.T) { compStore.AddPubSub(TestSecondPubsubName, &rtpubsub.PubsubItem{ Component: &mockPublishPubSub{}, }) - err = ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err = ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic1", }) @@ -291,7 +291,7 @@ func TestPublish(t *testing.T) { md := make(map[string]string, 2) md["key"] = "v3" - err := ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err := ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestPubsubName, Topic: "topic0", Metadata: md, @@ -303,7 +303,7 @@ func TestPublish(t *testing.T) { ProtectedTopics: []string{"topic1"}, ScopedPublishings: []string{"topic1"}, }) - err = ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err = ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic1", }) @@ -325,7 +325,7 @@ func TestPublish(t *testing.T) { Component: &mockPublishPubSub{}, AllowedTopics: []string{"topic1"}, }) - err := ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err := ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestPubsubName, Topic: "topic5", }) @@ -335,7 +335,7 @@ func TestPublish(t *testing.T) { Component: &mockPublishPubSub{}, AllowedTopics: []string{"topic1"}, }) - err = ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err = ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic5", }) @@ -357,7 +357,7 @@ func TestPublish(t *testing.T) { Component: &mockPublishPubSub{}, ProtectedTopics: []string{"topic1"}, }) - err := ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err := ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestPubsubName, Topic: "topic1", }) @@ -367,7 +367,7 @@ func TestPublish(t *testing.T) { Component: &mockPublishPubSub{}, ProtectedTopics: []string{"topic1"}, }) - err = ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err = ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestSecondPubsubName, Topic: "topic1", }) @@ -388,7 +388,7 @@ func TestNamespacedPublisher(t *testing.T) { Namespace: "ns1", }) - err := ps.Publish(context.Background(), &contribpubsub.PublishRequest{ + err := ps.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: TestPubsubName, Topic: "topic0", }) @@ -462,7 +462,7 @@ func TestPubsubWithResiliency(t *testing.T) { PubsubName: "failPubsub", Topic: "failingTopic", } - err := ps.Publish(context.Background(), req) + err := ps.Publish(t.Context(), req) require.NoError(t, err) assert.Equal(t, 2, failingPubsub.Failure.CallCount("failingTopic")) @@ -495,7 +495,7 @@ func TestPubsubWithResiliency(t *testing.T) { } start := time.Now() - err := ps.Publish(context.Background(), req) + err := ps.Publish(t.Context(), req) end := time.Now() require.Error(t, err) diff --git a/pkg/runtime/pubsub/subscriptions_test.go b/pkg/runtime/pubsub/subscriptions_test.go index 9a41c7127..7056c3320 100644 --- a/pkg/runtime/pubsub/subscriptions_test.go +++ b/pkg/runtime/pubsub/subscriptions_test.go @@ -144,7 +144,7 @@ func (m *mockHTTPSubscriptions) InvokeMethod(ctx context.Context, req *invokev1. func TestHTTPSubscriptions(t *testing.T) { t.Run("topics received, no errors", func(t *testing.T) { m := mockHTTPSubscriptions{} - subs, err := GetSubscriptionsHTTP(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsHTTP(t.Context(), &m, log, resiliency.FromConfigurations(log)) require.NoError(t, err) if assert.Len(t, subs, 1) { assert.Equal(t, "topic1", subs[0].Topic) @@ -163,7 +163,7 @@ func TestHTTPSubscriptions(t *testing.T) { successThreshold: 3, } - subs, err := GetSubscriptionsHTTP(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsHTTP(t.Context(), &m, log, resiliency.FromConfigurations(log)) assert.Equal(t, m.successThreshold, m.callCount) require.NoError(t, err) if assert.Len(t, subs, 1) { @@ -183,7 +183,7 @@ func TestHTTPSubscriptions(t *testing.T) { alwaysError: true, } - _, err := GetSubscriptionsHTTP(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + _, err := GetSubscriptionsHTTP(t.Context(), &m, log, resiliency.FromConfigurations(log)) require.Error(t, err) }) @@ -192,7 +192,7 @@ func TestHTTPSubscriptions(t *testing.T) { successThreshold: 3, } - subs, err := GetSubscriptionsHTTP(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsHTTP(t.Context(), &m, log, resiliency.FromConfigurations(log)) assert.Equal(t, m.successThreshold, m.callCount) require.NoError(t, err) if assert.Len(t, subs, 1) { @@ -212,7 +212,7 @@ func TestHTTPSubscriptions(t *testing.T) { alwaysError: true, } - _, err := GetSubscriptionsHTTP(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + _, err := GetSubscriptionsHTTP(t.Context(), &m, log, resiliency.FromConfigurations(log)) require.Error(t, err) }) } @@ -295,7 +295,7 @@ func (m *mockGRPCSubscriptions) ListTopicSubscriptions(ctx context.Context, in * func TestGRPCSubscriptions(t *testing.T) { t.Run("topics received, no errors", func(t *testing.T) { m := mockGRPCSubscriptions{} - subs, err := GetSubscriptionsGRPC(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsGRPC(t.Context(), &m, log, resiliency.FromConfigurations(log)) require.NoError(t, err) if assert.Len(t, subs, 1) { assert.Equal(t, "topic1", subs[0].Topic) @@ -314,7 +314,7 @@ func TestGRPCSubscriptions(t *testing.T) { successThreshold: 3, } - subs, err := GetSubscriptionsGRPC(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsGRPC(t.Context(), &m, log, resiliency.FromConfigurations(log)) assert.Equal(t, m.successThreshold, m.callCount) require.NoError(t, err) if assert.Len(t, subs, 1) { @@ -335,7 +335,7 @@ func TestGRPCSubscriptions(t *testing.T) { unimplemented: true, } - subs, err := GetSubscriptionsGRPC(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsGRPC(t.Context(), &m, log, resiliency.FromConfigurations(log)) // not implemented error is not retried and is returned as "zero" subscriptions require.NoError(t, err) assert.Equal(t, 1, m.callCount) @@ -347,7 +347,7 @@ func TestGRPCSubscriptions(t *testing.T) { successThreshold: 3, } - subs, err := GetSubscriptionsGRPC(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsGRPC(t.Context(), &m, log, resiliency.FromConfigurations(log)) assert.Equal(t, m.successThreshold, m.callCount) require.NoError(t, err) if assert.Len(t, subs, 1) { @@ -368,7 +368,7 @@ func TestGRPCSubscriptions(t *testing.T) { unimplemented: true, } - subs, err := GetSubscriptionsGRPC(context.TODO(), &m, log, resiliency.FromConfigurations(log)) + subs, err := GetSubscriptionsGRPC(t.Context(), &m, log, resiliency.FromConfigurations(log)) // not implemented error is not retried and is returned as "zero" subscriptions require.NoError(t, err) assert.Equal(t, 1, m.callCount) diff --git a/pkg/runtime/runtime_test.go b/pkg/runtime/runtime_test.go index f6cbff6a8..bb684eda6 100644 --- a/pkg/runtime/runtime_test.go +++ b/pkg/runtime/runtime_test.go @@ -106,7 +106,7 @@ const ( func TestNewRuntime(t *testing.T) { // act - r, err := newDaprRuntime(context.Background(), nil, &internalConfig{ + r, err := newDaprRuntime(t.Context(), nil, &internalConfig{ mode: modes.StandaloneMode, metricsExporter: metrics.New(metrics.Options{ Log: log, @@ -152,7 +152,7 @@ func TestDoProcessComponent(t *testing.T) { // setup ctrl := gomock.NewController(t) mockLockStore := daprt.NewMockStore(ctrl) - mockLockStore.EXPECT().InitLockStore(context.Background(), gomock.Any()).Return(assert.AnError) + mockLockStore.EXPECT().InitLockStore(t.Context(), gomock.Any()).Return(assert.AnError) rt.runtimeConfig.registry.Locks().RegisterComponent( func(_ logger.Logger) lock.Store { @@ -162,7 +162,7 @@ func TestDoProcessComponent(t *testing.T) { ) // act - err := rt.processor.Init(context.Background(), lockComponent) + err := rt.processor.Init(t.Context(), lockComponent) // assert require.Error(t, err, "expected an error") @@ -185,7 +185,7 @@ func TestDoProcessComponent(t *testing.T) { lockComponentV3.Spec.Version = "v3" // act - err := rt.processor.Init(context.Background(), lockComponentV3) + err := rt.processor.Init(t.Context(), lockComponentV3) // assert require.Error(t, err, "expected an error") @@ -196,7 +196,7 @@ func TestDoProcessComponent(t *testing.T) { // setup ctrl := gomock.NewController(t) mockLockStore := daprt.NewMockStore(ctrl) - mockLockStore.EXPECT().InitLockStore(context.Background(), gomock.Any()).Return(nil) + mockLockStore.EXPECT().InitLockStore(t.Context(), gomock.Any()).Return(nil) rt.runtimeConfig.registry.Locks().RegisterComponent( func(_ logger.Logger) lock.Store { @@ -215,7 +215,7 @@ func TestDoProcessComponent(t *testing.T) { }, } // act - err := rt.processor.Init(context.Background(), lockComponentWithWrongStrategy) + err := rt.processor.Init(t.Context(), lockComponentWithWrongStrategy) // assert require.Error(t, err) }) @@ -224,7 +224,7 @@ func TestDoProcessComponent(t *testing.T) { // setup ctrl := gomock.NewController(t) mockLockStore := daprt.NewMockStore(ctrl) - mockLockStore.EXPECT().InitLockStore(context.Background(), gomock.Any()).Return(nil) + mockLockStore.EXPECT().InitLockStore(t.Context(), gomock.Any()).Return(nil) rt.runtimeConfig.registry.Locks().RegisterComponent( func(_ logger.Logger) lock.Store { @@ -234,7 +234,7 @@ func TestDoProcessComponent(t *testing.T) { ) // act - err := rt.processor.Init(context.Background(), lockComponent) + err := rt.processor.Init(t.Context(), lockComponent) // assert require.NoError(t, err, "unexpected error") // get modified key @@ -263,7 +263,7 @@ func TestDoProcessComponent(t *testing.T) { mockPubSub.On("Init", expectedMetadata).Return(assert.AnError) // act - err := rt.processor.Init(context.Background(), pubsubComponent) + err := rt.processor.Init(t.Context(), pubsubComponent) // assert require.Error(t, err, "expected an error") @@ -272,7 +272,7 @@ func TestDoProcessComponent(t *testing.T) { t.Run("test invalid category component", func(t *testing.T) { // act - err := rt.processor.Init(context.Background(), componentsV1alpha1.Component{ + err := rt.processor.Init(t.Context(), componentsV1alpha1.Component{ Spec: componentsV1alpha1.ComponentSpec{ Type: "invalid", }, @@ -301,8 +301,8 @@ func TestFlushOutstandingComponent(t *testing.T) { "kubernetesMock", ) - go rt.processor.Process(context.Background()) - rt.processor.AddPendingComponent(context.Background(), componentsV1alpha1.Component{ + go rt.processor.Process(t.Context()) + rt.processor.AddPendingComponent(t.Context(), componentsV1alpha1.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMock", }, @@ -311,7 +311,7 @@ func TestFlushOutstandingComponent(t *testing.T) { Version: "v1", }, }) - rt.flushOutstandingComponents(context.Background()) + rt.flushOutstandingComponents(t.Context()) assert.True(t, wasCalled) // Make sure that the goroutine was restarted and can flush a second time @@ -323,7 +323,7 @@ func TestFlushOutstandingComponent(t *testing.T) { "kubernetesMock2", ) - rt.processor.AddPendingComponent(context.Background(), componentsV1alpha1.Component{ + rt.processor.AddPendingComponent(t.Context(), componentsV1alpha1.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMock2", }, @@ -332,7 +332,7 @@ func TestFlushOutstandingComponent(t *testing.T) { Version: "v1", }, }) - rt.flushOutstandingComponents(context.Background()) + rt.flushOutstandingComponents(t.Context()) assert.True(t, wasCalled) }) t.Run("flushOutstandingComponents blocks for components with outstanding dependanices", func(t *testing.T) { @@ -376,8 +376,8 @@ func TestFlushOutstandingComponent(t *testing.T) { "kubernetesMockGrandChild", ) - go rt.processor.Process(context.Background()) - rt.processor.AddPendingComponent(context.Background(), componentsV1alpha1.Component{ + go rt.processor.Process(t.Context()) + rt.processor.AddPendingComponent(t.Context(), componentsV1alpha1.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMockGrandChild", }, @@ -398,7 +398,7 @@ func TestFlushOutstandingComponent(t *testing.T) { SecretStore: "kubernetesMockChild", }, }) - rt.processor.AddPendingComponent(context.Background(), componentsV1alpha1.Component{ + rt.processor.AddPendingComponent(t.Context(), componentsV1alpha1.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMockChild", }, @@ -419,7 +419,7 @@ func TestFlushOutstandingComponent(t *testing.T) { SecretStore: "kubernetesMock", }, }) - rt.processor.AddPendingComponent(context.Background(), componentsV1alpha1.Component{ + rt.processor.AddPendingComponent(t.Context(), componentsV1alpha1.Component{ ObjectMeta: metav1.ObjectMeta{ Name: "kubernetesMock", }, @@ -428,7 +428,7 @@ func TestFlushOutstandingComponent(t *testing.T) { Version: "v1", }, }) - rt.flushOutstandingComponents(context.Background()) + rt.flushOutstandingComponents(t.Context()) assert.True(t, wasCalled) assert.True(t, wasCalledChild) assert.True(t, wasCalledGrandChild) @@ -479,7 +479,7 @@ func TestInitNameResolution(t *testing.T) { initMockResolverForRuntime(rt, "anotherResolver", nil) // act - err = rt.initNameResolution(context.Background()) + err = rt.initNameResolution(t.Context()) // assert require.Error(t, err) @@ -499,7 +499,7 @@ func TestInitNameResolution(t *testing.T) { initMockResolverForRuntime(rt, "someResolver", nil) // act - err = rt.initNameResolution(context.Background()) + err = rt.initNameResolution(t.Context()) // assert require.NoError(t, err, "expected no error") @@ -517,7 +517,7 @@ func TestInitNameResolution(t *testing.T) { initMockResolverForRuntime(rt, "mdns", nil) // act - err = rt.initNameResolution(context.Background()) + err = rt.initNameResolution(t.Context()) // assert require.NoError(t, err, "expected no error") @@ -535,7 +535,7 @@ func TestInitNameResolution(t *testing.T) { initMockResolverForRuntime(rt, "mdns", nil) // act - err = rt.initNameResolution(context.Background()) + err = rt.initNameResolution(t.Context()) // assert require.NoError(t, err, "expected no error") @@ -553,7 +553,7 @@ func TestInitNameResolution(t *testing.T) { initMockResolverForRuntime(rt, "kubernetes", nil) // act - err = rt.initNameResolution(context.Background()) + err = rt.initNameResolution(t.Context()) // assert require.NoError(t, err, "expected no error") @@ -571,7 +571,7 @@ func TestInitNameResolution(t *testing.T) { initMockResolverForRuntime(rt, "kubernetes", nil) // act - err = rt.initNameResolution(context.Background()) + err = rt.initNameResolution(t.Context()) // assert require.NoError(t, err, "expected no error") @@ -676,7 +676,7 @@ func TestSetupTracing(t *testing.T) { // Setup tracing with the fake tracer provider store to confirm // the right exporter was registered. tpStore := newFakeTracerProviderStore() - if err := rt.setupTracing(context.Background(), rt.hostAddress, tpStore); tc.expectedErr != "" { + if err := rt.setupTracing(t.Context(), rt.hostAddress, tpStore); tc.expectedErr != "" { assert.Contains(t, err.Error(), tc.expectedErr) } else { require.NoError(t, err) @@ -693,7 +693,7 @@ func TestSetupTracing(t *testing.T) { // We have no way to validate the result, but we can at least // confirm that nothing blows up. if tc.expectedErr == "" { - rt.setupTracing(context.Background(), rt.hostAddress, newOpentelemetryTracerProviderStore()) + rt.setupTracing(t.Context(), rt.hostAddress, newOpentelemetryTracerProviderStore()) } }) } @@ -751,7 +751,7 @@ func TestInitSecretStoresInKubernetesMode(t *testing.T) { defer close(testOk) go func() { // If the test fails, this call blocks forever, eventually causing a timeout - rt.appendBuiltinSecretStore(context.Background()) + rt.appendBuiltinSecretStore(t.Context()) testOk <- struct{}{} }() select { @@ -783,8 +783,8 @@ func TestInitSecretStoresInKubernetesMode(t *testing.T) { } func assertBuiltInSecretStore(t *testing.T, rt *DaprRuntime) { - go rt.processor.Process(context.Background()) - rt.appendBuiltinSecretStore(context.Background()) + go rt.processor.Process(t.Context()) + rt.appendBuiltinSecretStore(t.Context()) assert.Eventually(t, func() bool { _, ok := rt.compStore.GetComponent(secretstoresLoader.BuiltinKubernetesSecretStore) return ok @@ -800,7 +800,7 @@ func NewTestDaprRuntime(t *testing.T, mode modes.DaprMode) (*DaprRuntime, error) func NewTestDaprRuntimeWithID(t *testing.T, mode modes.DaprMode, id string) (*DaprRuntime, error) { testRuntimeConfig := NewTestDaprRuntimeConfig(t, modes.StandaloneMode, string(protocol.HTTPProtocol), 1024) testRuntimeConfig.id = id - rt, err := newDaprRuntime(context.Background(), testSecurity(t), testRuntimeConfig, &config.Configuration{}, &config.AccessControlList{}, resiliency.New(logger.NewLogger("test"))) + rt, err := newDaprRuntime(t.Context(), testSecurity(t), testRuntimeConfig, &config.Configuration{}, &config.AccessControlList{}, resiliency.New(logger.NewLogger("test"))) if err != nil { return nil, err } @@ -812,7 +812,7 @@ func NewTestDaprRuntimeWithID(t *testing.T, mode modes.DaprMode, id string) (*Da func NewTestDaprRuntimeWithProtocol(t *testing.T, mode modes.DaprMode, protocol string, appPort int) (*DaprRuntime, error) { testRuntimeConfig := NewTestDaprRuntimeConfig(t, modes.StandaloneMode, protocol, appPort) - rt, err := newDaprRuntime(context.Background(), testSecurity(t), testRuntimeConfig, &config.Configuration{}, &config.AccessControlList{}, resiliency.New(logger.NewLogger("test"))) + rt, err := newDaprRuntime(t.Context(), testSecurity(t), testRuntimeConfig, &config.Configuration{}, &config.AccessControlList{}, resiliency.New(logger.NewLogger("test"))) if err != nil { return nil, err } @@ -896,7 +896,7 @@ func TestInitActors(t *testing.T) { r.namespace = "" r.runtimeConfig.mTLSEnabled = true - err = r.initActors(context.TODO()) + err = r.initActors(t.Context()) require.Error(t, err) }) @@ -922,7 +922,7 @@ func TestInitActors(t *testing.T) { }) t.Run("placement enable = false", func(t *testing.T) { - r, err := newDaprRuntime(context.Background(), testSecurity(t), &internalConfig{ + r, err := newDaprRuntime(t.Context(), testSecurity(t), &internalConfig{ metricsExporter: metrics.New(metrics.Options{ Log: log, Namespace: metrics.DefaultMetricNamespace, @@ -936,12 +936,12 @@ func TestInitActors(t *testing.T) { defer stopRuntime(t, r) r.channels.Refresh() - err = r.initActors(context.TODO()) + err = r.initActors(t.Context()) require.NoError(t, err) }) t.Run("the state stores can still be initialized normally", func(t *testing.T) { - r, err := newDaprRuntime(context.Background(), testSecurity(t), &internalConfig{ + r, err := newDaprRuntime(t.Context(), testSecurity(t), &internalConfig{ metricsExporter: metrics.New(metrics.Options{ Log: log, Namespace: metrics.DefaultMetricNamespace, @@ -960,7 +960,7 @@ func TestInitActors(t *testing.T) { }) t.Run("the actor store can not be initialized normally", func(t *testing.T) { - r, err := newDaprRuntime(context.Background(), testSecurity(t), &internalConfig{ + r, err := newDaprRuntime(t.Context(), testSecurity(t), &internalConfig{ metricsExporter: metrics.New(metrics.Options{ Log: log, Namespace: metrics.DefaultMetricNamespace, @@ -977,7 +977,7 @@ func TestInitActors(t *testing.T) { name, ok := r.processor.State().ActorStateStoreName() assert.False(t, ok) assert.Equal(t, "", name) - err = r.initActors(context.TODO()) + err = r.initActors(t.Context()) require.NoError(t, err) }) } @@ -1052,7 +1052,7 @@ func TestActorReentrancyConfig(t *testing.T) { mockAppChannel.On("GetAppConfig").Return(&configResp, nil) - r.loadAppConfiguration(context.Background()) + r.loadAppConfiguration(t.Context()) assert.NotNil(t, r.appConfig) @@ -1190,13 +1190,13 @@ func TestCloseWithErrors(t *testing.T) { errCh := make(chan error) go func() { - errCh <- rt.Run(context.Background()) + errCh <- rt.Run(t.Context()) }() - rt.processor.AddPendingComponent(context.Background(), mockOutputBindingComponent) - rt.processor.AddPendingComponent(context.Background(), mockPubSubComponent) - rt.processor.AddPendingComponent(context.Background(), mockStateComponent) - rt.processor.AddPendingComponent(context.Background(), mockSecretsComponent) + rt.processor.AddPendingComponent(t.Context(), mockOutputBindingComponent) + rt.processor.AddPendingComponent(t.Context(), mockPubSubComponent) + rt.processor.AddPendingComponent(t.Context(), mockStateComponent) + rt.processor.AddPendingComponent(t.Context(), mockSecretsComponent) err = rt.runnerCloser.Close() require.Error(t, err) @@ -1227,7 +1227,7 @@ func TestComponentsCallback(t *testing.T) { var callbackInvoked atomic.Bool cfg := NewTestDaprRuntimeConfig(t, modes.StandaloneMode, "http", port) - rt, err := newDaprRuntime(context.Background(), testSecurity(t), cfg, &config.Configuration{}, &config.AccessControlList{}, resiliency.New(logger.NewLogger("test"))) + rt, err := newDaprRuntime(t.Context(), testSecurity(t), cfg, &config.Configuration{}, &config.AccessControlList{}, resiliency.New(logger.NewLogger("test"))) require.NoError(t, err) rt.runtimeConfig.registry = registry.New(registry.NewOptions().WithComponentsCallback(func(components registry.ComponentRegistry) error { callbackInvoked.Store(true) @@ -1236,7 +1236,7 @@ func TestComponentsCallback(t *testing.T) { })) errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { errCh <- rt.Run(ctx) }() @@ -1282,7 +1282,7 @@ func TestGRPCProxy(t *testing.T) { "mdns", // for standalone mode ) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1301,7 +1301,7 @@ func TestGRPCProxy(t *testing.T) { req := &pb.PingRequest{Value: "foo"} t.Run("proxy single streaming request", func(t *testing.T) { - ctx, cancel := context.WithTimeout(context.Background(), time.Second*5) + ctx, cancel := context.WithTimeout(t.Context(), time.Second*5) defer cancel() stream, err := pingStreamClient(ctx, internalPort) require.NoError(t, err) @@ -1315,12 +1315,12 @@ func TestGRPCProxy(t *testing.T) { }) t.Run("proxy concurrent streaming requests", func(t *testing.T) { - ctx1, cancel := context.WithTimeout(context.Background(), time.Second*5) + ctx1, cancel := context.WithTimeout(t.Context(), time.Second*5) defer cancel() stream1, err := pingStreamClient(ctx1, internalPort) require.NoError(t, err) - ctx2, cancel := context.WithTimeout(context.Background(), time.Second) + ctx2, cancel := context.WithTimeout(t.Context(), time.Second) defer cancel() stream2, err := pingStreamClient(ctx2, internalPort) require.NoError(t, err) @@ -1373,7 +1373,7 @@ spec: `), 0o600)) // Use a background context since this is not closed by the test. - ctx := context.Background() + ctx := t.Context() errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1454,7 +1454,7 @@ spec: `), 0o600)) // Use a background context since this is not closed by the test. - ctx := context.Background() + ctx := t.Context() errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1537,7 +1537,7 @@ spec: `), 0o600)) // Use a background context since this is not closed by the test. - ctx := context.Background() + ctx := t.Context() errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1607,7 +1607,7 @@ spec: `), 0o600)) // Use a background context since this is not closed by the test. - ctx := context.Background() + ctx := t.Context() errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1676,7 +1676,7 @@ spec: `), 0o600)) // Use a background context since this is not closed by the test. - ctx := context.Background() + ctx := t.Context() errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1758,11 +1758,11 @@ func TestGetComponentsCapabilitiesMap(t *testing.T) { cSecretStore.ObjectMeta.Name = mockSecretStoreName cSecretStore.Spec.Type = "secretstores.mockSecretStore" - require.NoError(t, rt.processor.Init(context.Background(), cin)) - require.NoError(t, rt.processor.Init(context.Background(), cout)) - require.NoError(t, rt.processor.Init(context.Background(), cPubSub)) - require.NoError(t, rt.processor.Init(context.Background(), cStateStore)) - require.NoError(t, rt.processor.Init(context.Background(), cSecretStore)) + require.NoError(t, rt.processor.Init(t.Context(), cin)) + require.NoError(t, rt.processor.Init(t.Context(), cout)) + require.NoError(t, rt.processor.Init(t.Context(), cPubSub)) + require.NoError(t, rt.processor.Init(t.Context(), cStateStore)) + require.NoError(t, rt.processor.Init(t.Context(), cSecretStore)) capabilities := rt.getComponentsCapabilitesMap() assert.Len(t, capabilities, 5, @@ -1848,7 +1848,7 @@ func TestGracefulShutdownBindings(t *testing.T) { rt, err := NewTestDaprRuntime(t, modes.StandaloneMode) require.NoError(t, err) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1874,8 +1874,8 @@ func TestGracefulShutdownBindings(t *testing.T) { cout := componentsV1alpha1.Component{} cout.ObjectMeta.Name = "testOutputBinding" cout.Spec.Type = "bindings.testOutputBinding" - require.NoError(t, rt.processor.Init(context.Background(), cin)) - require.NoError(t, rt.processor.Init(context.Background(), cout)) + require.NoError(t, rt.processor.Init(t.Context(), cin)) + require.NoError(t, rt.processor.Init(t.Context(), cout)) assert.Len(t, rt.compStore.ListInputBindings(), 1) assert.Len(t, rt.compStore.ListOutputBindings(), 1) @@ -1895,12 +1895,12 @@ func TestBlockShutdownBindings(t *testing.T) { fakeClock := clocktesting.NewFakeClock(time.Now()) rt.clock = fakeClock - rt.appHealthChanged(context.Background(), apphealth.AppStatusHealthy) + rt.appHealthChanged(t.Context(), apphealth.AppStatusHealthy) rt.runtimeConfig.blockShutdownDuration = ptr.Of(time.Millisecond * 100) rt.runtimeConfig.gracefulShutdownDuration = 3 * time.Second - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1930,12 +1930,12 @@ func TestBlockShutdownBindings(t *testing.T) { fakeClock := clocktesting.NewFakeClock(time.Now()) rt.clock = fakeClock - rt.appHealthChanged(context.Background(), apphealth.AppStatusHealthy) + rt.appHealthChanged(t.Context(), apphealth.AppStatusHealthy) rt.runtimeConfig.blockShutdownDuration = ptr.Of(time.Millisecond * 100) rt.runtimeConfig.gracefulShutdownDuration = 3 * time.Second - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -1949,7 +1949,7 @@ func TestBlockShutdownBindings(t *testing.T) { assert.Fail(t, "expected not to return until block timeout is reached") } - rt.appHealthChanged(context.Background(), apphealth.AppStatusUnhealthy) + rt.appHealthChanged(t.Context(), apphealth.AppStatusUnhealthy) select { case <-time.After(rt.runtimeConfig.gracefulShutdownDuration + 2*time.Second): @@ -2007,15 +2007,15 @@ func TestGracefulShutdownPubSub(t *testing.T) { Security: rt.sec, }) - require.NoError(t, rt.processor.Init(context.Background(), cPubSub)) + require.NoError(t, rt.processor.Init(t.Context(), cPubSub)) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) go func() { errCh <- rt.Run(ctx) }() - rt.appHealthChanged(context.Background(), apphealth.AppStatusHealthy) + rt.appHealthChanged(t.Context(), apphealth.AppStatusHealthy) mockPubSub.AssertCalled(t, "Init", mock.Anything) mockPubSub.AssertCalled(t, "Subscribe", mock.AnythingOfType("pubsub.SubscribeRequest"), mock.AnythingOfType("pubsub.Handler")) @@ -2071,7 +2071,7 @@ func TestGracefulShutdownActors(t *testing.T) { rt.namespace = "test" rt.runtimeConfig.appConnectionConfig.Port = -1 - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) errCh := make(chan error) go func() { errCh <- rt.Run(ctx) @@ -2084,7 +2084,7 @@ func TestGracefulShutdownActors(t *testing.T) { } // act - err = rt.processor.Init(context.Background(), mockStateComponent) + err = rt.processor.Init(t.Context(), mockStateComponent) // assert require.NoError(t, err, "expected no error") @@ -2150,11 +2150,11 @@ func TestTraceShutdown(t *testing.T) { } rt.hostAddress = "localhost:3000" tpStore := newOpentelemetryTracerProviderStore() - require.NoError(t, rt.setupTracing(context.Background(), rt.hostAddress, tpStore)) + require.NoError(t, rt.setupTracing(t.Context(), rt.hostAddress, tpStore)) assert.NotNil(t, rt.tracerProvider) errCh := make(chan error) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) go func() { errCh <- rt.Run(ctx) }() @@ -2172,7 +2172,7 @@ func TestTraceShutdown(t *testing.T) { } func testSecurity(t *testing.T) security.Handler { - secP, err := security.New(context.Background(), security.Options{ + secP, err := security.New(t.Context(), security.Options{ TrustAnchors: []byte("test"), AppID: "test", ControlPlaneTrustDomain: "test.example.com", @@ -2184,10 +2184,10 @@ func testSecurity(t *testing.T) security.Handler { Healthz: healthz.New(), }) require.NoError(t, err) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) go secP.Run(ctx) - sec, err := secP.Handler(context.Background()) + sec, err := secP.Handler(t.Context()) require.NoError(t, err) return sec diff --git a/pkg/runtime/subscription/bulkresiliency_test.go b/pkg/runtime/subscription/bulkresiliency_test.go index bf03cae78..f16cec788 100644 --- a/pkg/runtime/subscription/bulkresiliency_test.go +++ b/pkg/runtime/subscription/bulkresiliency_test.go @@ -217,7 +217,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { t.Run("verify Responses when few entries fail even after retries", func(t *testing.T) { mockAppChannel := new(channelt.MockAppChannel) mockAppChannel.Init() - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -254,7 +254,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(resiliencyV1alpha.CircuitBreaker{}, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) mockAppChannel.AssertNumberOfCalls(t, "InvokeMethod", 3) assert.Len(t, *b, 10) @@ -290,7 +290,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("verify Responses when ALL entries fail even after retries", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -329,7 +329,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(resiliencyV1alpha.CircuitBreaker{}, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) mockAppChannel.AssertNumberOfCalls(t, "InvokeMethod", 3) assert.Len(t, *b, 10) @@ -365,7 +365,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("pass ALL entries in second attempt", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -404,7 +404,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(resiliencyV1alpha.CircuitBreaker{}, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) mockAppChannel.AssertNumberOfCalls(t, "InvokeMethod", 2) assert.Len(t, *b, 10) @@ -440,7 +440,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("pass ALL entries in first attempt", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -479,7 +479,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(resiliencyV1alpha.CircuitBreaker{}, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) mockAppChannel.AssertNumberOfCalls(t, "InvokeMethod", 1) assert.Len(t, *b, 10) @@ -515,7 +515,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("fail ALL entries due to timeout", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -556,7 +556,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(resiliencyV1alpha.CircuitBreaker{}, shortTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) assert.Len(t, *b, 10) @@ -580,7 +580,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("verify Responses when ALL entries fail with Circuitbreaker and exhaust retries", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -626,7 +626,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(cb, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse := BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -661,7 +661,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { assert.Equal(t, breaker.ErrOpenState, e) assert.True(t, verifyBulkSubscribeResponses(expectedResponse, *b)) - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) mockAppChannel.AssertNumberOfCalls(t, "InvokeMethod", 2) assert.Len(t, *b, 10) @@ -672,7 +672,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("verify Responses when Partial entries fail with Circuitbreaker and exhaust retries", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -718,7 +718,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(cb, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse := BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -753,7 +753,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { assert.Equal(t, breaker.ErrOpenState, e) assert.True(t, verifyBulkSubscribeResponses(expectedResponse, *b)) - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) mockAppChannel.AssertNumberOfCalls(t, "InvokeMethod", 2) assert.Len(t, *b, 10) @@ -764,7 +764,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("verify Responses when Partial entries Pass with Circuitbreaker half open timeout", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -810,7 +810,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(cb, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse := BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -847,7 +847,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("Partial success with CB and exhaust retries, then act with short half open timeout", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -893,7 +893,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(cb, longTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse := BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -930,7 +930,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { time.Sleep(5 * time.Second) - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse = BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -967,7 +967,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { }) t.Run("Fail all events with timeout and then Open CB - short retries", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -1014,7 +1014,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { policyProvider := createResPolicyProvider(cb, shortTimeout, shortRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.Background(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse := BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -1035,7 +1035,7 @@ func TestBulkSubscribeResiliency(t *testing.T) { assert.Equal(t, breaker.ErrOpenState, e) assert.True(t, verifyBulkSubscribeResponses(expectedResponse, *b)) - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) assert.Len(t, *b, 10) require.Error(t, e) @@ -1049,7 +1049,7 @@ func TestBulkSubscribeResiliencyStateConversionsFromHalfOpen(t *testing.T) { mockAppChannel := new(channelt.MockAppChannel) mockAppChannel.Init() - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -1093,7 +1093,7 @@ func TestBulkSubscribeResiliencyStateConversionsFromHalfOpen(t *testing.T) { policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse := BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -1131,7 +1131,7 @@ func TestBulkSubscribeResiliencyStateConversionsFromHalfOpen(t *testing.T) { time.Sleep(5 * time.Second) // after this time, circuit breaker should be half-open - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse = BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -1169,7 +1169,7 @@ func TestBulkSubscribeResiliencyStateConversionsFromHalfOpen(t *testing.T) { assert.True(t, verifyBulkSubscribeResponses(expectedResponse, *b)) // circuit breaker is open, so no call should go through - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) mockAppChannel.AssertNumberOfCalls(t, "InvokeMethod", 3) assert.Len(t, *b, 10) @@ -1180,7 +1180,7 @@ func TestBulkSubscribeResiliencyStateConversionsFromHalfOpen(t *testing.T) { time.Sleep(5 * time.Second) // after this time, circuit breaker should be half-open - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse = BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -1220,7 +1220,7 @@ func TestBulkSubscribeResiliencyStateConversionsFromHalfOpen(t *testing.T) { func TestBulkSubscribeResiliencyWithLongRetries(t *testing.T) { t.Run("Fail all events with timeout and then Open CB - long retries", func(t *testing.T) { - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -1267,7 +1267,7 @@ func TestBulkSubscribeResiliencyWithLongRetries(t *testing.T) { policyProvider := createResPolicyProvider(cb, shortTimeout, longRetry) policyDef := policyProvider.ComponentInboundPolicy(pubsubName, resiliency.Pubsub) in := getInput() - b, e := ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e := ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) expectedResponse := BulkResponseExpectation{ Responses: []BulkResponseEntryExpectation{ @@ -1288,7 +1288,7 @@ func TestBulkSubscribeResiliencyWithLongRetries(t *testing.T) { assert.Equal(t, breaker.ErrOpenState, e) assert.True(t, verifyBulkSubscribeResponses(expectedResponse, *b)) - b, e = ps.applyBulkSubscribeResiliency(context.TODO(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) + b, e = ps.applyBulkSubscribeResiliency(t.Context(), &in.bscData, in.pbsm, "dlq", orders1, policyDef, true, in.envelope) assert.Len(t, *b, 10) require.Error(t, e) diff --git a/pkg/runtime/subscription/bulksubscription_test.go b/pkg/runtime/subscription/bulksubscription_test.go index f5215bd23..d8dff0722 100644 --- a/pkg/runtime/subscription/bulksubscription_test.go +++ b/pkg/runtime/subscription/bulksubscription_test.go @@ -130,7 +130,7 @@ func TestBulkSubscribe(t *testing.T) { t.Run("bulk Subscribe Message for raw payload", func(t *testing.T) { comp := &mockSubscribePubSub{} - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) resp := contribpubsub.AppBulkResponse{AppResponses: []contribpubsub.AppBulkResponseEntry{{ EntryId: "0", @@ -166,7 +166,7 @@ func TestBulkSubscribe(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - err = comp.Publish(context.TODO(), &contribpubsub.PublishRequest{ + err = comp.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Data: []byte(`{"orderId":"1"}`), @@ -182,7 +182,7 @@ func TestBulkSubscribe(t *testing.T) { t.Run("bulk Subscribe Message for cloud event", func(t *testing.T) { comp := &mockSubscribePubSub{} - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) resp := contribpubsub.AppBulkResponse{AppResponses: []contribpubsub.AppBulkResponseEntry{{ EntryId: "0", @@ -219,7 +219,7 @@ func TestBulkSubscribe(t *testing.T) { order := `{"data":{"orderId":1},"datacontenttype":"application/json","id":"8b540b03-04b5-4871-96ae-c6bde0d5e16d","pubsubname":"orderpubsub","source":"checkout","specversion":"1.0","topic":"orders","traceid":"00-e61de949bb4de415a7af49fc86675648-ffb64972bb907224-01","traceparent":"00-e61de949bb4de415a7af49fc86675648-ffb64972bb907224-01","tracestate":"","type":"com.dapr.event.sent"}` - err = comp.Publish(context.TODO(), &contribpubsub.PublishRequest{ + err = comp.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Data: []byte(order), @@ -237,7 +237,7 @@ func TestBulkSubscribe(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) resp := contribpubsub.AppBulkResponse{AppResponses: []contribpubsub.AppBulkResponseEntry{ {EntryId: "1111111a", Status: contribpubsub.Success}, @@ -273,7 +273,7 @@ func TestBulkSubscribe(t *testing.T) { msgArr := getBulkMessageEntries(2) - comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -298,7 +298,7 @@ func TestBulkSubscribe(t *testing.T) { msgArr = getBulkMessageEntries(3) - comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -325,7 +325,7 @@ func TestBulkSubscribe(t *testing.T) { msgArr = getBulkMessageEntries(4) - comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -350,7 +350,7 @@ func TestBulkSubscribe(t *testing.T) { mockAppChannel3.On("InvokeMethod", mock.MatchedBy(matchContextInterface), mock.Anything).Return(nil, errors.New("Mock error")) msgArr = getBulkMessageEntries(1) - comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -371,7 +371,7 @@ func TestBulkSubscribe(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) resp := contribpubsub.AppBulkResponse{AppResponses: []contribpubsub.AppBulkResponseEntry{ {EntryId: "1111111a", Status: contribpubsub.Success}, @@ -412,7 +412,7 @@ func TestBulkSubscribe(t *testing.T) { msgArr := getBulkMessageEntries(2) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -433,7 +433,7 @@ func TestBulkSubscribe(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) mockAppChannel := new(channelt.MockAppChannel) mockAppChannel.Init() @@ -494,7 +494,7 @@ func TestBulkSubscribe(t *testing.T) { t.Cleanup(ps.Stop) msgArr := getBulkMessageEntries(10) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -536,7 +536,7 @@ func TestBulkSubscribe(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) mockAppChannel := new(channelt.MockAppChannel) mockAppChannel.Init() @@ -578,7 +578,7 @@ func TestBulkSubscribe(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), matchDaprRequestMethod("orders")).Return(respInvoke1, nil) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -610,7 +610,7 @@ func TestBulkSubscribe(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) mockAppChannel := new(channelt.MockAppChannel) mockAppChannel.Init() @@ -657,7 +657,7 @@ func TestBulkSubscribe(t *testing.T) { matchDaprRequestMethod("orders"), ).Return(respInvoke1, nil) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -688,7 +688,7 @@ func TestBulkSubscribe(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) mockAppChannel := new(channelt.MockAppChannel) mockAppChannel.Init() @@ -733,7 +733,7 @@ func TestBulkSubscribe(t *testing.T) { matchDaprRequestMethod("orders"), ).Return(respInvoke1, nil) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -771,7 +771,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) nbei1 := contribpubsub.BulkMessageEntry{EntryId: "1111111a", Event: []byte(`{"orderId":"1"}`)} nbei2 := contribpubsub.BulkMessageEntry{EntryId: "2222222b", Event: []byte(`{"orderId":"2"}`)} @@ -836,7 +836,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -859,7 +859,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { mockServer.BulkResponsePerPath = nil mockServer.Error = status.Error(codes.Unimplemented, "method not implemented") - comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -869,7 +869,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { require.NoError(t, assertItemExistsOnce(comp.GetBulkResponse().Statuses, "1111111a", "2222222b")) mockServer.Error = status.Error(codes.Unknown, "unknown error") - comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -887,7 +887,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) rule1, err := runtimePubsub.CreateRoutingRule(`event.type == "type1"`, orders1) require.NoError(t, err) @@ -976,7 +976,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -1012,7 +1012,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) msgArr := getBulkMessageEntries(4) msgArr[0].EntryId = "" @@ -1071,7 +1071,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -1098,7 +1098,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) msgArr := getBulkMessageEntries(5) responseEntries := make([]*runtimev1pb.TopicEventBulkResponseEntry, 5) @@ -1168,7 +1168,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -1196,7 +1196,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) msgArr := getBulkMessageEntries(5) responseEntries := make([]*runtimev1pb.TopicEventBulkResponseEntry, 5) @@ -1260,7 +1260,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -1284,7 +1284,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) port, err := freeport.GetFreePort() require.NoError(t, err) @@ -1337,7 +1337,7 @@ func TestBulkSubscribeGRPC(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - _, err = comp.BulkPublish(context.TODO(), &contribpubsub.BulkPublishRequest{ + _, err = comp.BulkPublish(t.Context(), &contribpubsub.BulkPublishRequest{ PubsubName: testBulkSubscribePubsub, Topic: "topic0", Entries: msgArr, @@ -1562,7 +1562,7 @@ func TestPubSubDeadLetter(t *testing.T) { comp := &mockSubscribePubSub{ features: []contribpubsub.Feature{contribpubsub.FeatureBulkPublish}, } - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) // Mock send message to app returns error. mockAppChannel := new(channelt.MockAppChannel) @@ -1598,7 +1598,7 @@ func TestPubSubDeadLetter(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - err = comp.Publish(context.TODO(), &contribpubsub.PublishRequest{ + err = comp.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: testDeadLetterPubsub, Topic: "topic0", Data: []byte(`{"id":"1"}`), @@ -1611,7 +1611,7 @@ func TestPubSubDeadLetter(t *testing.T) { t.Run("use dead letter with resiliency", func(t *testing.T) { comp := &mockSubscribePubSub{} - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) // Mock send message to app returns error. mockAppChannel := new(channelt.MockAppChannel) @@ -1646,7 +1646,7 @@ func TestPubSubDeadLetter(t *testing.T) { require.NoError(t, err) t.Cleanup(ps.Stop) - err = comp.Publish(context.TODO(), &contribpubsub.PublishRequest{ + err = comp.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: testDeadLetterPubsub, Topic: "topic0", Data: []byte(`{"id":"1"}`), diff --git a/pkg/runtime/subscription/publish_test.go b/pkg/runtime/subscription/publish_test.go index eedea1c48..441af2e0e 100644 --- a/pkg/runtime/subscription/publish_test.go +++ b/pkg/runtime/subscription/publish_test.go @@ -71,7 +71,7 @@ func TestErrorPublishedNonCloudEventHTTP(t *testing.T) { WithCustomHTTPMetadata(testPubSubMessage.Metadata) defer fakeReq.Close() - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -99,7 +99,7 @@ func TestErrorPublishedNonCloudEventHTTP(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err) @@ -118,7 +118,7 @@ func TestErrorPublishedNonCloudEventHTTP(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err) @@ -138,7 +138,7 @@ func TestErrorPublishedNonCloudEventHTTP(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.Error(t, err) @@ -158,7 +158,7 @@ func TestErrorPublishedNonCloudEventHTTP(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert assert.Equal(t, runtimePubsub.ErrMessageDropped, err) @@ -178,7 +178,7 @@ func TestErrorPublishedNonCloudEventHTTP(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.Error(t, err) @@ -196,7 +196,7 @@ func TestErrorPublishedNonCloudEventHTTP(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.Anything, fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err) @@ -214,7 +214,7 @@ func TestErrorPublishedNonCloudEventGRPC(t *testing.T) { Path: "topic1", } - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -283,7 +283,7 @@ func TestErrorPublishedNonCloudEventGRPC(t *testing.T) { } ps.grpc.SetAppClientConn(&mockClientConn) - err := ps.publishMessageGRPC(context.Background(), testPubSubMessage) + err := ps.publishMessageGRPC(t.Context(), testPubSubMessage) if tc.ExpectError { require.Error(t, err) } else { @@ -316,7 +316,7 @@ func TestOnNewPublishedMessage(t *testing.T) { WithCustomHTTPMetadata(testPubSubMessage.Metadata) defer fakeReq.Close() - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -344,7 +344,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err) @@ -387,7 +387,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReqNoTraceID).Return(fakeResp, nil) // act - err = ps.publishMessageHTTP(context.Background(), message) + err = ps.publishMessageHTTP(t.Context(), message) // assert require.NoError(t, err) @@ -407,7 +407,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err) @@ -427,7 +427,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err) @@ -447,7 +447,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert var cloudEvent map[string]interface{} @@ -470,7 +470,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert assert.Equal(t, runtimePubsub.ErrMessageDropped, err) @@ -490,7 +490,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.Error(t, err, "expected error on unknown status") @@ -510,7 +510,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err, "expected no error on empty status") @@ -530,7 +530,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err, "expected no error on unknown status") @@ -545,7 +545,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(nil, invokeError) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert expectedError := fmt.Errorf("error returned from app channel while sending pub/sub event to app: %w", rterrors.NewRetriable(invokeError)) @@ -565,7 +565,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert require.NoError(t, err, "expected error to be nil") @@ -584,7 +584,7 @@ func TestOnNewPublishedMessage(t *testing.T) { mockAppChannel.On("InvokeMethod", mock.MatchedBy(matchContextInterface), fakeReq).Return(fakeResp, nil) // act - err := ps.publishMessageHTTP(context.Background(), testPubSubMessage) + err := ps.publishMessageHTTP(t.Context(), testPubSubMessage) // assert var cloudEvent map[string]interface{} @@ -743,7 +743,7 @@ func TestOnNewPublishedMessageGRPC(t *testing.T) { port, err := freeport.GetFreePort() require.NoError(t, err) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) t.Cleanup(cancel) comp := inmemory.New(log) require.NoError(t, comp.Init(ctx, contribpubsub.Metadata{})) @@ -793,7 +793,7 @@ func TestOnNewPublishedMessageGRPC(t *testing.T) { ps.grpc = grpc // act - err = ps.publishMessageGRPC(context.Background(), tc.message) + err = ps.publishMessageGRPC(t.Context(), tc.message) // assert if tc.expectedError != nil { @@ -808,7 +808,7 @@ func TestOnNewPublishedMessageGRPC(t *testing.T) { func TestTracingOnNewPublishedMessage(t *testing.T) { t.Run("succeeded to publish message with TraceParent in metadata", func(t *testing.T) { comp := &mockSubscribePubSub{} - require.NoError(t, comp.Init(context.Background(), contribpubsub.Metadata{})) + require.NoError(t, comp.Init(t.Context(), contribpubsub.Metadata{})) resp := contribpubsub.AppResponse{ Status: contribpubsub.Success, @@ -847,7 +847,7 @@ func TestTracingOnNewPublishedMessage(t *testing.T) { traceparent := "00-0af7651916cd43dd8448eb211c80319c-b9c7c989f97918e1-01" traceid := "00-80e1afed08e019fc1110464cfa66635c-7a085853722dc6d2-01" tracestate := "abc=xyz" - err = comp.Publish(context.TODO(), &contribpubsub.PublishRequest{ + err = comp.Publish(t.Context(), &contribpubsub.PublishRequest{ PubsubName: "testpubsub", Topic: "topic0", Data: []byte(`{"orderId":"1"}`), diff --git a/pkg/runtime/wfengine/client.go b/pkg/runtime/wfengine/client.go index 9cbcde227..f08858a48 100644 --- a/pkg/runtime/wfengine/client.go +++ b/pkg/runtime/wfengine/client.go @@ -184,18 +184,15 @@ func (c *client) Get(ctx context.Context, req *workflows.GetRequest) (*workflows res.Workflow.Properties["dapr.workflow.custom_status"] = metadata.GetCustomStatus().GetValue() } - //nolint:protogetter if metadata.Input != nil { res.Workflow.Properties["dapr.workflow.input"] = metadata.GetInput().GetValue() } - //nolint:protogetter if metadata.Output != nil { res.Workflow.Properties["dapr.workflow.output"] = metadata.GetOutput().GetValue() } // Status-specific fields - //nolint:protogetter if metadata.FailureDetails != nil { res.Workflow.Properties["dapr.workflow.failure.error_type"] = metadata.GetFailureDetails().GetErrorType() res.Workflow.Properties["dapr.workflow.failure.error_message"] = metadata.GetFailureDetails().GetErrorMessage() diff --git a/pkg/scheduler/server/internal/authz/authz_test.go b/pkg/scheduler/server/internal/authz/authz_test.go index 16c6643b7..ac60613e9 100644 --- a/pkg/scheduler/server/internal/authz/authz_test.go +++ b/pkg/scheduler/server/internal/authz/authz_test.go @@ -58,7 +58,7 @@ func Test_Metadata(t *testing.T) { nonMTlSCode: ptr.Of(codes.InvalidArgument), }, "no auth context should error": { - ctx: context.Background(), + ctx: t.Context(), meta: &schedulerv1pb.JobMetadata{ AppId: "app1", Namespace: "ns1", @@ -144,7 +144,7 @@ func Test_Initial(t *testing.T) { nonMTlSCode: ptr.Of(codes.InvalidArgument), }, "no auth context should error": { - ctx: context.Background(), + ctx: t.Context(), initial: &schedulerv1pb.WatchJobsRequestInitial{ AppId: "app1", Namespace: "ns1", diff --git a/pkg/scheduler/server/internal/controller/namespace_test.go b/pkg/scheduler/server/internal/controller/namespace_test.go index 76442f07e..6c02e1046 100644 --- a/pkg/scheduler/server/internal/controller/namespace_test.go +++ b/pkg/scheduler/server/internal/controller/namespace_test.go @@ -88,7 +88,7 @@ func Test_Reconcile(t *testing.T) { nsReader: nsReaderBuilder.Build(), } - result, err := nsctrl.Reconcile(context.Background(), req) + result, err := nsctrl.Reconcile(t.Context(), req) assert.Equal(t, ctrl.Result{}, result) assert.Equal(t, test.expErr, err) assert.ElementsMatch(t, test.expDeletePrefixes, prefixCalled) diff --git a/pkg/security/security_test.go b/pkg/security/security_test.go index bb09ce6cf..d472a2598 100644 --- a/pkg/security/security_test.go +++ b/pkg/security/security_test.go @@ -91,7 +91,7 @@ func Test_Start(t *testing.T) { tdFile := filepath.Join(t.TempDir(), "root.pem") require.NoError(t, os.WriteFile(tdFile, root1, 0o600)) - p, err := New(context.Background(), Options{ + p, err := New(t.Context(), Options{ TrustAnchorsFile: &tdFile, AppID: "test", ControlPlaneTrustDomain: "test.example.com", @@ -104,7 +104,7 @@ func Test_Start(t *testing.T) { }) require.NoError(t, err) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) providerStopped := make(chan struct{}) go func() { @@ -175,7 +175,7 @@ func TestCurrentNamespace(t *testing.T) { os.Unsetenv("NAMESPACE") t.Cleanup(func() { if ok { - os.Setenv("NAMESPACE", osns) + t.Setenv("NAMESPACE", osns) } }) ns, err := CurrentNamespaceOrError() diff --git a/pkg/sentry/server/ca/ca_test.go b/pkg/sentry/server/ca/ca_test.go index 1f9735b76..51c2726b5 100644 --- a/pkg/sentry/server/ca/ca_test.go +++ b/pkg/sentry/server/ca/ca_test.go @@ -14,7 +14,6 @@ limitations under the License. package ca import ( - "context" "crypto/ecdsa" "crypto/elliptic" "crypto/rand" @@ -33,7 +32,7 @@ import ( func TestNew(t *testing.T) { t.Run("if no existing bundle exist, new should generate a new bundle", func(t *testing.T) { - os.Setenv("NAMESPACE", "dapr-test") + t.Setenv("NAMESPACE", "dapr-test") t.Cleanup(func() { os.Unsetenv("NAMESPACE") }) @@ -50,7 +49,7 @@ func TestNew(t *testing.T) { Mode: modes.StandaloneMode, } - _, err := New(context.Background(), config) + _, err := New(t.Context(), config) require.NoError(t, err) require.FileExists(t, rootCertPath) @@ -110,7 +109,7 @@ func TestNew(t *testing.T) { require.NoError(t, os.WriteFile(issuerCertPath, issuerFileContents, 0o600)) require.NoError(t, os.WriteFile(issuerKeyPath, issuerKeyFileContents, 0o600)) - caImp, err := New(context.Background(), config) + caImp, err := New(t.Context(), config) require.NoError(t, err) rootCert, err := os.ReadFile(rootCertPath) @@ -161,13 +160,13 @@ func TestSignIdentity(t *testing.T) { require.NoError(t, os.WriteFile(issuerCertPath, issuerFileContents, 0o600)) require.NoError(t, os.WriteFile(issuerKeyPath, issuerKeyFileContents, 0o600)) - ca, err := New(context.Background(), config) + ca, err := New(t.Context(), config) require.NoError(t, err) clientPK, err := ecdsa.GenerateKey(elliptic.P256(), rand.Reader) require.NoError(t, err) - clientCert, err := ca.SignIdentity(context.Background(), &SignRequest{ + clientCert, err := ca.SignIdentity(t.Context(), &SignRequest{ PublicKey: clientPK.Public(), SignatureAlgorithm: x509.ECDSAWithSHA256, TrustDomain: "example.test.dapr.io", diff --git a/pkg/sentry/server/ca/kube_test.go b/pkg/sentry/server/ca/kube_test.go index 57e1aa44f..e188da0a2 100644 --- a/pkg/sentry/server/ca/kube_test.go +++ b/pkg/sentry/server/ca/kube_test.go @@ -14,7 +14,6 @@ limitations under the License. package ca import ( - "context" "crypto/x509" "testing" @@ -258,7 +257,7 @@ func TestKube_get(t *testing.T) { namespace: "dapr-system-test", } - bundle, ok, err := k.get(context.Background()) + bundle, ok, err := k.get(t.Context()) assert.Equal(t, test.expErr, err != nil, "%v", err) assert.Equal(t, test.expOK, ok, "ok") assert.Equal(t, test.expBundle, bundle) diff --git a/pkg/sentry/server/ca/selfhosted_test.go b/pkg/sentry/server/ca/selfhosted_test.go index dc72b10a3..6f31a6155 100644 --- a/pkg/sentry/server/ca/selfhosted_test.go +++ b/pkg/sentry/server/ca/selfhosted_test.go @@ -14,7 +14,6 @@ limitations under the License. package ca import ( - "context" "crypto/x509" "os" "path/filepath" @@ -50,7 +49,7 @@ func TestSelhosted_store(t *testing.T) { }, } - require.NoError(t, s.store(context.Background(), Bundle{ + require.NoError(t, s.store(t.Context(), Bundle{ TrustAnchors: []byte("root"), IssChainPEM: []byte("issuer"), IssKeyPEM: []byte("key"), @@ -178,7 +177,7 @@ func TestSelfhosted_get(t *testing.T) { require.NoError(t, os.WriteFile(keyFile, *test.key, writePerm)) } - got, ok, err := s.get(context.Background()) + got, ok, err := s.get(t.Context()) assert.Equal(t, test.expBundle, got) assert.Equal(t, test.expOk, ok) assert.Equal(t, test.expErr, err != nil, "%v", err) diff --git a/pkg/sentry/server/server_test.go b/pkg/sentry/server/server_test.go index 990e61276..880fa4b84 100644 --- a/pkg/sentry/server/server_test.go +++ b/pkg/sentry/server/server_test.go @@ -267,7 +267,7 @@ func TestRun(t *testing.T) { t.Run(name, func(t *testing.T) { port, err := freeport.GetFreePort() require.NoError(t, err) - ctx, cancel := context.WithCancel(context.Background()) + ctx, cancel := context.WithCancel(t.Context()) opts := Options{ Port: port, Security: test.sec, diff --git a/pkg/sentry/server/validator/internal/common_test.go b/pkg/sentry/server/validator/internal/common_test.go index 4d329f129..f35d8ca78 100644 --- a/pkg/sentry/server/validator/internal/common_test.go +++ b/pkg/sentry/server/validator/internal/common_test.go @@ -14,7 +14,6 @@ limitations under the License. package internal import ( - "context" "errors" "fmt" "testing" @@ -93,7 +92,7 @@ func TestValidate(t *testing.T) { for name, test := range tests { t.Run(name, func(t *testing.T) { - td, err := Validate(context.Background(), test.req) + td, err := Validate(t.Context(), test.req) assert.Equal(t, test.expTD, td) assert.Equal(t, test.expErr, err) }) diff --git a/pkg/sentry/server/validator/kubernetes/kubernetes_test.go b/pkg/sentry/server/validator/kubernetes/kubernetes_test.go index f303d9cf3..ebe904e48 100644 --- a/pkg/sentry/server/validator/kubernetes/kubernetes_test.go +++ b/pkg/sentry/server/validator/kubernetes/kubernetes_test.go @@ -1233,7 +1233,7 @@ func TestValidate(t *testing.T) { client := clientfake.NewClientBuilder().WithScheme(scheme).WithRuntimeObjects(kobjs...).Build() if test.config != nil { - require.NoError(t, client.Create(context.Background(), test.config)) + require.NoError(t, client.Create(t.Context(), test.config)) } k := &kubernetes{ @@ -1245,7 +1245,7 @@ func TestValidate(t *testing.T) { ready: func(_ context.Context) bool { return true }, } - td, err := k.Validate(context.Background(), test.req) + td, err := k.Validate(t.Context(), test.req) assert.Equal(t, test.expErr, err != nil, "%v", err) assert.Equal(t, test.expTD, td) }) diff --git a/tests/e2e/hotreloading/hotreloading_test.go b/tests/e2e/hotreloading/hotreloading_test.go index 61fa0fdd6..8763b4308 100644 --- a/tests/e2e/hotreloading/hotreloading_test.go +++ b/tests/e2e/hotreloading/hotreloading_test.go @@ -17,7 +17,6 @@ limitations under the License. package hotreloading_tests import ( - "context" "fmt" "net/http" "os" @@ -76,7 +75,7 @@ func TestState(t *testing.T) { cl, err := client.New(platform.KubeClient.GetClientConfig(), client.Options{Scheme: scheme}) require.NoError(t, err) - ctx := context.Background() + ctx := t.Context() externalURL := tr.Platform.AcquireAppExternalURL("hotreloading-state") // Wait for app to be available. diff --git a/tests/e2e/metrics/metrics_test.go b/tests/e2e/metrics/metrics_test.go index 72eefe867..dce7a4e0f 100644 --- a/tests/e2e/metrics/metrics_test.go +++ b/tests/e2e/metrics/metrics_test.go @@ -17,7 +17,6 @@ limitations under the License. package metrics_e2e import ( - "context" "encoding/json" "fmt" "io" @@ -260,7 +259,7 @@ func invokeDaprGRPC(t *testing.T, app string, n, daprPort int) { client := pb.NewDaprClient(conn) for i := 0; i < n; i++ { - _, err = client.SaveState(context.Background(), &pb.SaveStateRequest{ + _, err = client.SaveState(t.Context(), &pb.SaveStateRequest{ StoreName: "statestore", States: []*commonv1pb.StateItem{ { diff --git a/tests/integration/framework/binary/binary.go b/tests/integration/framework/binary/binary.go index 4d6754a86..9d2b75cf3 100644 --- a/tests/integration/framework/binary/binary.go +++ b/tests/integration/framework/binary/binary.go @@ -155,6 +155,8 @@ func build(t *testing.T, name string, opts options) { assert.NoError(t, ioout.Close()) assert.NoError(t, ioerr.Close()) + // TODO: @joshvanl: check if we can use `t.Setenv` + //nolint:usetesting assert.NoError(t, os.Setenv(EnvKey(name), binPath)) } else { t.Logf("%q set, using %q pre-built binary", EnvKey(name), EnvValue(name)) diff --git a/tests/integration/framework/iowriter/iowriter_test.go b/tests/integration/framework/iowriter/iowriter_test.go index ce081cf37..1b31a887c 100644 --- a/tests/integration/framework/iowriter/iowriter_test.go +++ b/tests/integration/framework/iowriter/iowriter_test.go @@ -34,15 +34,15 @@ func (m *mockLogger) Log(args ...any) { m.msgs = append(m.msgs, args[0].(string)) } -func (m mockLogger) Name() string { +func (m *mockLogger) Name() string { return "TestLogger" } -func (m mockLogger) Cleanup(fn func()) { +func (m *mockLogger) Cleanup(fn func()) { m.t.Cleanup(fn) } -func (m mockLogger) Failed() bool { +func (m *mockLogger) Failed() bool { return m.failed } diff --git a/tests/integration/framework/metrics/metrics_test.go b/tests/integration/framework/metrics/metrics_test.go index 660000473..583a5d9ce 100644 --- a/tests/integration/framework/metrics/metrics_test.go +++ b/tests/integration/framework/metrics/metrics_test.go @@ -14,7 +14,6 @@ limitations under the License. package metrics import ( - "context" "io" "net/http" "net/http/httptest" @@ -42,7 +41,7 @@ func TestNew(t *testing.T) { })) defer ts.Close() - metrics := New(t, context.Background(), ts.URL) + metrics := New(t, t.Context(), ts.URL) metricsMap := metrics.All() require.NotNil(t, metricsMap) diff --git a/tests/integration/framework/process/http/http.go b/tests/integration/framework/process/http/http.go index bf65c34c0..f9491623c 100644 --- a/tests/integration/framework/process/http/http.go +++ b/tests/integration/framework/process/http/http.go @@ -94,6 +94,7 @@ func (h *HTTP) Run(t *testing.T, ctx context.Context) { } func (h *HTTP) Cleanup(t *testing.T) { + //nolint:usetesting ctx, cancel := context.WithTimeout(context.Background(), time.Second*10) defer cancel() require.NoError(t, errors.Join(h.server.Shutdown(ctx), <-h.srvErrCh)) diff --git a/tests/integration/framework/process/once/once_test.go b/tests/integration/framework/process/once/once_test.go index ac69cf79c..1d9728fed 100644 --- a/tests/integration/framework/process/once/once_test.go +++ b/tests/integration/framework/process/once/once_test.go @@ -51,7 +51,7 @@ func TestOnce(t *testing.T) { } t.Run("Run", func(t *testing.T) { - ctx := context.Background() + ctx := t.Context() // First call should succeed onceProc.Run(t, ctx) diff --git a/tests/integration/integration.go b/tests/integration/integration.go index ab7d895d4..c9dc73162 100644 --- a/tests/integration/integration.go +++ b/tests/integration/integration.go @@ -85,6 +85,9 @@ func RunIntegrationTests(t *testing.T) { t.Log("setting up test case") + // TODO: @joshvanl: update framework to use `t.Context()` which is + // correctly respected on cleanup. + //nolint:usetesting ctx, cancel := context.WithTimeout(context.Background(), 45*time.Second) t.Cleanup(cancel) diff --git a/tests/integration/suite/actors/lock/call/reentry/pertype/http.go b/tests/integration/suite/actors/lock/call/reentry/pertype/http.go index 16cf13beb..c1af93e25 100644 --- a/tests/integration/suite/actors/lock/call/reentry/pertype/http.go +++ b/tests/integration/suite/actors/lock/call/reentry/pertype/http.go @@ -128,7 +128,7 @@ func (h *http) Run(t *testing.T, ctx context.Context) { assert.Equal(t, nethttp.StatusInternalServerError, resp.StatusCode) b, err := io.ReadAll(resp.Body) require.NoError(t, err) - assert.Equal(t, `{"errorCode":"ERR_ACTOR_STACK_DEPTH","message":"maximum stack depth exceeded"}`, string(b)) + assert.JSONEq(t, `{"errorCode":"ERR_ACTOR_STACK_DEPTH","message":"maximum stack depth exceeded"}`, string(b)) require.NoError(t, resp.Body.Close()) assert.Eventually(t, func() bool { return h.app.Daprd().Metrics(t, ctx).MatchMetricAndSum(t, 1, "dapr_error_code_total", "category:actor", "error_code:ERR_ACTOR_STACK_DEPTH") diff --git a/tests/integration/suite/actors/reminders/get.go b/tests/integration/suite/actors/reminders/get.go index 0e10a1fe3..390d20fde 100644 --- a/tests/integration/suite/actors/reminders/get.go +++ b/tests/integration/suite/actors/reminders/get.go @@ -79,5 +79,5 @@ func (g *get) Run(t *testing.T, ctx context.Context) { b, err := io.ReadAll(resp.Body) require.NoError(t, err) require.NoError(t, resp.Body.Close()) - assert.Equal(t, `{"period":"1s","data":"reminderdata","dueTime":"1s"}`, strings.TrimSpace(string(b))) + assert.JSONEq(t, `{"period":"1s","data":"reminderdata","dueTime":"1s"}`, strings.TrimSpace(string(b))) } diff --git a/tests/integration/suite/actors/reminders/scheduler/get.go b/tests/integration/suite/actors/reminders/scheduler/get.go index 674cad6f7..0d0a53a82 100644 --- a/tests/integration/suite/actors/reminders/scheduler/get.go +++ b/tests/integration/suite/actors/reminders/scheduler/get.go @@ -78,5 +78,5 @@ func (g *get) Run(t *testing.T, ctx context.Context) { b, err := io.ReadAll(resp.Body) require.NoError(t, err) require.NoError(t, resp.Body.Close()) - assert.Equal(t, `{"period":"@every 1s","data":"reminderdata","actorID":"1234","actorType":"foo","dueTime":"1s"}`, strings.TrimSpace(string(b))) + assert.JSONEq(t, `{"period":"@every 1s","data":"reminderdata","actorID":"1234","actorType":"foo","dueTime":"1s"}`, strings.TrimSpace(string(b))) } diff --git a/tests/integration/suite/actors/state/http.go b/tests/integration/suite/actors/state/http.go index 68a34d8c9..7e31656ad 100644 --- a/tests/integration/suite/actors/state/http.go +++ b/tests/integration/suite/actors/state/http.go @@ -77,7 +77,7 @@ func (h *http) Run(t *testing.T, ctx context.Context) { b, err = io.ReadAll(resp.Body) require.NoError(t, err) require.NoError(t, resp.Body.Close()) - assert.Equal(t, `{"errorCode":"ERR_ACTOR_INSTANCE_MISSING","message":"actor instance is missing"}`, string(b)) + assert.JSONEq(t, `{"errorCode":"ERR_ACTOR_INSTANCE_MISSING","message":"actor instance is missing"}`, string(b)) assert.Eventually(t, func() bool { return h.app.Daprd().Metrics(t, ctx).MatchMetricAndSum(t, 1, "dapr_error_code_total", "category:actor", "error_code:ERR_ACTOR_INSTANCE_MISSING") }, 5*time.Second, 100*time.Millisecond) diff --git a/tests/integration/suite/actors/timers/callback/grpc.go b/tests/integration/suite/actors/timers/callback/grpc.go index e3c6090e5..df5e3d520 100644 --- a/tests/integration/suite/actors/timers/callback/grpc.go +++ b/tests/integration/suite/actors/timers/callback/grpc.go @@ -51,7 +51,7 @@ func (g *grpc) Setup(t *testing.T) []framework.Option { assert.Equal(t, "/actors/abc/foo/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"aGVsbG8=","callback":"mycallback","dueTime":"0s","period":"10s"}`, string(b)) + assert.JSONEq(t, `{"data":"aGVsbG8=","callback":"mycallback","dueTime":"0s","period":"10s"}`, string(b)) g.called.Add(1) w.WriteHeader(nethttp.StatusOK) }), diff --git a/tests/integration/suite/actors/timers/callback/http.go b/tests/integration/suite/actors/timers/callback/http.go index 5a1a47a48..db74ffd56 100644 --- a/tests/integration/suite/actors/timers/callback/http.go +++ b/tests/integration/suite/actors/timers/callback/http.go @@ -58,7 +58,7 @@ func (h *http) Setup(t *testing.T) []framework.Option { assert.Equal(t, "/actors/abc/foo/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"hello","callback":"mycallback","dueTime":"0s","period":"10s"}`, string(b)) + assert.JSONEq(t, `{"data":"hello","callback":"mycallback","dueTime":"0s","period":"10s"}`, string(b)) } }), ) diff --git a/tests/integration/suite/actors/timers/grpc.go b/tests/integration/suite/actors/timers/grpc.go index 298a477a4..ba2b4703e 100644 --- a/tests/integration/suite/actors/timers/grpc.go +++ b/tests/integration/suite/actors/timers/grpc.go @@ -52,7 +52,7 @@ func (g *grpc) Setup(t *testing.T) []framework.Option { assert.Equal(t, "/actors/abc/foo/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"aGVsbG8=","callback":"","dueTime":"0s","period":"10s"}`, string(b)) + assert.JSONEq(t, `{"data":"aGVsbG8=","callback":"","dueTime":"0s","period":"10s"}`, string(b)) }), ) diff --git a/tests/integration/suite/actors/timers/http.go b/tests/integration/suite/actors/timers/http.go index 56c27829d..da331ec4d 100644 --- a/tests/integration/suite/actors/timers/http.go +++ b/tests/integration/suite/actors/timers/http.go @@ -54,7 +54,7 @@ func (h *http) Setup(t *testing.T) []framework.Option { assert.Equal(t, "/actors/abc/foo/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"hello","callback":"","dueTime":"0s","period":"10s"}`, string(b)) + assert.JSONEq(t, `{"data":"hello","callback":"","dueTime":"0s","period":"10s"}`, string(b)) }), ) diff --git a/tests/integration/suite/actors/timers/remote/grpc.go b/tests/integration/suite/actors/timers/remote/grpc.go index cd10a7c57..48ac62203 100644 --- a/tests/integration/suite/actors/timers/remote/grpc.go +++ b/tests/integration/suite/actors/timers/remote/grpc.go @@ -60,7 +60,7 @@ func (g *grpc) Setup(t *testing.T) []framework.Option { assert.Regexp(t, "/actors/abc/.+/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"aGVsbG8=","callback":"","dueTime":"0s","period":"1s"}`, string(b)) + assert.JSONEq(t, `{"data":"aGVsbG8=","callback":"","dueTime":"0s","period":"1s"}`, string(b)) g.timer1.Add(1) }), ) @@ -82,7 +82,7 @@ func (g *grpc) Setup(t *testing.T) []framework.Option { assert.Regexp(t, "/actors/abc/.+/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"aGVsbG8=","callback":"","dueTime":"0s","period":"1s"}`, string(b)) + assert.JSONEq(t, `{"data":"aGVsbG8=","callback":"","dueTime":"0s","period":"1s"}`, string(b)) g.timer2.Add(1) }), ) diff --git a/tests/integration/suite/actors/timers/remote/http.go b/tests/integration/suite/actors/timers/remote/http.go index 12b8c0749..794274df8 100644 --- a/tests/integration/suite/actors/timers/remote/http.go +++ b/tests/integration/suite/actors/timers/remote/http.go @@ -60,7 +60,7 @@ func (h *http) Setup(t *testing.T) []framework.Option { assert.Regexp(t, "/actors/abc/.+/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"hello","callback":"","dueTime":"0s","period":"1s"}`, string(b)) + assert.JSONEq(t, `{"data":"hello","callback":"","dueTime":"0s","period":"1s"}`, string(b)) h.timer1.Add(1) }), ) @@ -82,7 +82,7 @@ func (h *http) Setup(t *testing.T) []framework.Option { assert.Regexp(t, "/actors/abc/.+/method/timer/foo", r.URL.Path) b, err := io.ReadAll(r.Body) assert.NoError(t, err) - assert.Equal(t, `{"data":"hello","callback":"","dueTime":"0s","period":"1s"}`, string(b)) + assert.JSONEq(t, `{"data":"hello","callback":"","dueTime":"0s","period":"1s"}`, string(b)) h.timer2.Add(1) }), ) diff --git a/tests/integration/suite/daprd/conversation/http/basic.go b/tests/integration/suite/daprd/conversation/http/basic.go index cf94a49fb..befd4b286 100644 --- a/tests/integration/suite/daprd/conversation/http/basic.go +++ b/tests/integration/suite/daprd/conversation/http/basic.go @@ -74,7 +74,7 @@ func (b *basic) Run(t *testing.T, ctx context.Context) { respBody, err := io.ReadAll(resp.Body) require.NoError(t, err) require.NoError(t, resp.Body.Close()) - require.Equal(t, `{"outputs":[{"result":"well hello there"}]}`, string(respBody)) + require.JSONEq(t, `{"outputs":[{"result":"well hello there"}]}`, string(respBody)) }) t.Run("invalid json", func(t *testing.T) { diff --git a/tests/integration/suite/daprd/resiliency/apps/defaultcircuitbreaker.go b/tests/integration/suite/daprd/resiliency/apps/defaultcircuitbreaker.go index 663da1ada..1160f0844 100644 --- a/tests/integration/suite/daprd/resiliency/apps/defaultcircuitbreaker.go +++ b/tests/integration/suite/daprd/resiliency/apps/defaultcircuitbreaker.go @@ -124,7 +124,7 @@ func (d *defaultcircuitbreaker) Run(t *testing.T, ctx context.Context) { // assert cb execution,activation,and current state counts assert.EventuallyWithT(t, func(c *assert.CollectT) { - mtc := d.daprdClient.Metrics(c, context.Background()).All() + mtc := d.daprdClient.Metrics(c, ctx).All() assert.InDelta(c, float64(4), mtc["dapr_resiliency_count|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:closed|target:app_server"], 0) assert.InDelta(c, float64(0), mtc["dapr_resiliency_activations_total|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:open|target:app_server"], 0) assert.InDelta(c, float64(1), mtc["dapr_resiliency_cb_state|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:closed|target:app_server"], 0) @@ -158,7 +158,7 @@ func (d *defaultcircuitbreaker) Run(t *testing.T, ctx context.Context) { // assert cb execution,activation,and current state counts assert.EventuallyWithT(t, func(c *assert.CollectT) { - mtc := d.daprdClient.Metrics(c, context.Background()).All() + mtc := d.daprdClient.Metrics(c, ctx).All() assert.InDelta(c, float64(7), mtc["dapr_resiliency_count|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:closed|target:app_server"], 0) assert.InDelta(c, float64(1), mtc["dapr_resiliency_activations_total|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:open|target:app_server"], 0) assert.InDelta(c, float64(1), mtc["dapr_resiliency_cb_state|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:open|target:app_server"], 0) @@ -177,7 +177,7 @@ func (d *defaultcircuitbreaker) Run(t *testing.T, ctx context.Context) { // make sure gauge is half open assert.EventuallyWithT(t, func(c *assert.CollectT) { - mtc := d.daprdClient.Metrics(c, context.Background()).All() + mtc := d.daprdClient.Metrics(c, ctx).All() assert.InDelta(c, float64(0), mtc["dapr_resiliency_cb_state|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:open|target:app_server"], 0) assert.InDelta(c, float64(1), mtc["dapr_resiliency_cb_state|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:half-open|target:app_server"], 0) }, time.Second*4, 10*time.Millisecond) @@ -192,7 +192,7 @@ func (d *defaultcircuitbreaker) Run(t *testing.T, ctx context.Context) { // make sure gauge is closed assert.EventuallyWithT(t, func(c *assert.CollectT) { - mtc := d.daprdClient.Metrics(c, context.Background()).All() + mtc := d.daprdClient.Metrics(c, ctx).All() assert.InDelta(c, float64(0), mtc["dapr_resiliency_cb_state|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:half-open|target:app_server"], 0) assert.InDelta(c, float64(1), mtc["dapr_resiliency_cb_state|app_id:client|flow_direction:outbound|name:myresiliency|namespace:|policy:circuitbreaker|status:closed|target:app_server"], 0) }, time.Second*4, 10*time.Millisecond) diff --git a/tests/integration/suite/daprd/state/http/customerrors.go b/tests/integration/suite/daprd/state/http/customerrors.go index b879b0db6..3c232f4a8 100644 --- a/tests/integration/suite/daprd/state/http/customerrors.go +++ b/tests/integration/suite/daprd/state/http/customerrors.go @@ -128,7 +128,7 @@ func (c *customerrors) Run(t *testing.T, ctx context.Context) { require.NoError(t, err) require.NoError(t, resp.Body.Close()) assert.Equal(t, 506, resp.StatusCode) - assert.Equal(t, `{"errorCode":"ERR_STATE_GET","message":"fail to get key1 from state store mystore: api error: code = DataLoss desc = get-error"}`, string(b)) + assert.JSONEq(t, `{"errorCode":"ERR_STATE_GET","message":"fail to get key1 from state store mystore: api error: code = DataLoss desc = get-error"}`, string(b)) queryURL := fmt.Sprintf("http://%s/v1.0-alpha1/state/mystore/query", c.daprd.HTTPAddress()) query := strings.NewReader(`{"filter":{"EQ":{"state":"CA"}},"sort":[{"key":"person.id","order":"DESC"}]}`) @@ -140,7 +140,7 @@ func (c *customerrors) Run(t *testing.T, ctx context.Context) { require.NoError(t, err) require.NoError(t, resp.Body.Close()) assert.Equal(t, 509, resp.StatusCode) - assert.Equal(t, `{"errorCode":"query-tag","message":"query-error","details":[{"@type":"type.googleapis.com/google.rpc.ErrorInfo","domain":"query-domain","metadata":null,"reason":"query-reason"}]}`, string(b)) + assert.JSONEq(t, `{"errorCode":"query-tag","message":"query-error","details":[{"@type":"type.googleapis.com/google.rpc.ErrorInfo","domain":"query-domain","metadata":null,"reason":"query-reason"}]}`, string(b)) body := strings.NewReader(`[{"key":"key1","value":"value1"},{"key":"key2","value":"value2"}]`) req, err = nethttp.NewRequestWithContext(ctx, nethttp.MethodPost, stateURL, body) @@ -151,7 +151,7 @@ func (c *customerrors) Run(t *testing.T, ctx context.Context) { require.NoError(t, err) require.NoError(t, resp.Body.Close()) assert.Equal(t, 510, resp.StatusCode) - assert.Equal(t, `{"errorCode":"ERR_STATE_SAVE","message":"failed saving state in state store mystore: api error: code = Canceled desc = bulkset-error"}`, string(b)) + assert.JSONEq(t, `{"errorCode":"ERR_STATE_SAVE","message":"failed saving state in state store mystore: api error: code = Canceled desc = bulkset-error"}`, string(b)) req, err = nethttp.NewRequestWithContext(ctx, nethttp.MethodDelete, stateURL+"/key1", body) require.NoError(t, err) @@ -161,7 +161,7 @@ func (c *customerrors) Run(t *testing.T, ctx context.Context) { require.NoError(t, err) require.NoError(t, resp.Body.Close()) assert.Equal(t, 511, resp.StatusCode) - assert.Equal(t, `{"errorCode":"ERR_STATE_DELETE","message":"failed deleting state with key key1: api error: code = OutOfRange desc = delete-error"}`, string(b)) + assert.JSONEq(t, `{"errorCode":"ERR_STATE_DELETE","message":"failed deleting state with key key1: api error: code = OutOfRange desc = delete-error"}`, string(b)) body = strings.NewReader(`[{"key":"key1","value":"value1"}]`) req, err = nethttp.NewRequestWithContext(ctx, nethttp.MethodPost, stateURL, body) @@ -172,7 +172,7 @@ func (c *customerrors) Run(t *testing.T, ctx context.Context) { require.NoError(t, err) require.NoError(t, resp.Body.Close()) assert.Equal(t, 512, resp.StatusCode) - assert.Equal(t, `{"errorCode":"ERR_STATE_SAVE","message":"failed saving state in state store mystore: api error: code = ResourceExhausted desc = set-error"}`, string(b)) + assert.JSONEq(t, `{"errorCode":"ERR_STATE_SAVE","message":"failed saving state in state store mystore: api error: code = ResourceExhausted desc = set-error"}`, string(b)) assert.EventuallyWithT(t, func(collect *assert.CollectT) { assert.True(collect, c.daprd.Metrics(collect, ctx).MatchMetricAndSum(t, 2, "dapr_error_code_total", "category:state", "error_code:ERR_STATE_SAVE")) diff --git a/tests/integration/suite/daprd/subscriptions/declarative/operator/v1alpha1/deadletter/http.go b/tests/integration/suite/daprd/subscriptions/declarative/operator/v1alpha1/deadletter/http.go index 5134bcdf3..2a3c56ae2 100644 --- a/tests/integration/suite/daprd/subscriptions/declarative/operator/v1alpha1/deadletter/http.go +++ b/tests/integration/suite/daprd/subscriptions/declarative/operator/v1alpha1/deadletter/http.go @@ -132,5 +132,5 @@ func (h *http) Run(t *testing.T, ctx context.Context) { resp := h.sub.Receive(t, ctx) assert.Equal(t, "/b", resp.Route) assert.Equal(t, "a", resp.Extensions()["topic"]) - assert.Equal(t, `{"status": "completed"}`, string(resp.Data())) + assert.JSONEq(t, `{"status": "completed"}`, string(resp.Data())) } diff --git a/tests/integration/suite/daprd/subscriptions/declarative/operator/v2alpha1/deadletter/http.go b/tests/integration/suite/daprd/subscriptions/declarative/operator/v2alpha1/deadletter/http.go index 7d699e70b..9f81a3ba6 100644 --- a/tests/integration/suite/daprd/subscriptions/declarative/operator/v2alpha1/deadletter/http.go +++ b/tests/integration/suite/daprd/subscriptions/declarative/operator/v2alpha1/deadletter/http.go @@ -136,5 +136,5 @@ func (h *http) Run(t *testing.T, ctx context.Context) { resp := h.sub.Receive(t, ctx) assert.Equal(t, "/b", resp.Route) assert.Equal(t, "a", resp.Extensions()["topic"]) - assert.Equal(t, `{"status": "completed"}`, string(resp.Data())) + assert.JSONEq(t, `{"status": "completed"}`, string(resp.Data())) } diff --git a/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v1alpha1/deadletter/http.go b/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v1alpha1/deadletter/http.go index a66229a3d..9b5de3a84 100644 --- a/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v1alpha1/deadletter/http.go +++ b/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v1alpha1/deadletter/http.go @@ -92,5 +92,5 @@ func (h *http) Run(t *testing.T, ctx context.Context) { resp := h.sub.Receive(t, ctx) assert.Equal(t, "/b", resp.Route) assert.Equal(t, "a", resp.Extensions()["topic"]) - assert.Equal(t, `{"status": "completed"}`, string(resp.Data())) + assert.JSONEq(t, `{"status": "completed"}`, string(resp.Data())) } diff --git a/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v2alpha1/deadletter/http.go b/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v2alpha1/deadletter/http.go index 679733e7e..5b60bd12c 100644 --- a/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v2alpha1/deadletter/http.go +++ b/tests/integration/suite/daprd/subscriptions/declarative/selfhosted/v2alpha1/deadletter/http.go @@ -94,5 +94,5 @@ func (h *http) Run(t *testing.T, ctx context.Context) { resp := h.sub.Receive(t, ctx) assert.Equal(t, "/b", resp.Route) assert.Equal(t, "a", resp.Extensions()["topic"]) - assert.Equal(t, `{"status": "completed"}`, string(resp.Data())) + assert.JSONEq(t, `{"status": "completed"}`, string(resp.Data())) } diff --git a/tests/integration/suite/daprd/subscriptions/programmatic/http/v1alpha1/deadletter.go b/tests/integration/suite/daprd/subscriptions/programmatic/http/v1alpha1/deadletter.go index a837fb5c4..416db933d 100644 --- a/tests/integration/suite/daprd/subscriptions/programmatic/http/v1alpha1/deadletter.go +++ b/tests/integration/suite/daprd/subscriptions/programmatic/http/v1alpha1/deadletter.go @@ -85,5 +85,5 @@ func (d *deadletter) Run(t *testing.T, ctx context.Context) { resp := d.sub.Receive(t, ctx) assert.Equal(t, "/b", resp.Route) assert.Equal(t, "a", resp.Extensions()["topic"]) - assert.Equal(t, `{"status": "completed"}`, string(resp.Data())) + assert.JSONEq(t, `{"status": "completed"}`, string(resp.Data())) } diff --git a/tests/integration/suite/daprd/subscriptions/programmatic/http/v2alpha1/deadletter.go b/tests/integration/suite/daprd/subscriptions/programmatic/http/v2alpha1/deadletter.go index 61d43a7bd..f8eebf540 100644 --- a/tests/integration/suite/daprd/subscriptions/programmatic/http/v2alpha1/deadletter.go +++ b/tests/integration/suite/daprd/subscriptions/programmatic/http/v2alpha1/deadletter.go @@ -88,5 +88,5 @@ func (d *deadletter) Run(t *testing.T, ctx context.Context) { resp := d.sub.Receive(t, ctx) assert.Equal(t, "/b", resp.Route) assert.Equal(t, "a", resp.Extensions()["topic"]) - assert.Equal(t, `{"status": "completed"}`, string(resp.Data())) + assert.JSONEq(t, `{"status": "completed"}`, string(resp.Data())) } diff --git a/tests/integration/suite/daprd/tracing/binding/output.go b/tests/integration/suite/daprd/tracing/binding/output.go index 198f5daac..0fca921fe 100644 --- a/tests/integration/suite/daprd/tracing/binding/output.go +++ b/tests/integration/suite/daprd/tracing/binding/output.go @@ -108,7 +108,7 @@ func (b *output) Run(t *testing.T, ctx context.Context) { t.Run("traceparent header provided", func(t *testing.T) { // invoke binding - ctx := context.Background() + ctx := t.Context() reqURL := fmt.Sprintf("http://localhost:%d/v1.0/bindings/http-binding-traceparent", b.daprd.HTTPPort()) req, err := http.NewRequestWithContext(ctx, http.MethodPost, reqURL, strings.NewReader("{\"operation\":\"get\"}")) require.NoError(t, err) diff --git a/tests/integration/suite/scheduler/quorum/notls.go b/tests/integration/suite/scheduler/quorum/notls.go index 2dc73e625..3492fbb1c 100644 --- a/tests/integration/suite/scheduler/quorum/notls.go +++ b/tests/integration/suite/scheduler/quorum/notls.go @@ -163,7 +163,7 @@ func getEtcdKeys(t *testing.T, port int) []*mvccpb.KeyValue { require.NoError(t, err) defer client.Close() - ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second) + ctx, cancel := context.WithTimeout(t.Context(), 30*time.Second) defer cancel() // Get keys with prefix diff --git a/tests/perf/actor_reminder/actor_reminder_test.go b/tests/perf/actor_reminder/actor_reminder_test.go index c38217f34..97aec5436 100644 --- a/tests/perf/actor_reminder/actor_reminder_test.go +++ b/tests/perf/actor_reminder/actor_reminder_test.go @@ -17,7 +17,6 @@ limitations under the License. package actor_reminder_perf import ( - "context" "encoding/json" "fmt" "os" @@ -206,14 +205,14 @@ func TestActorReminderSchedulerRegistrationPerformance(t *testing.T) { cl, err := client.New(platform.KubeClient.GetClientConfig(), client.Options{Scheme: scheme}) require.NoError(t, err) var pod corev1.Pod - err = cl.Get(context.Background(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) + err = cl.Get(t.Context(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) require.NoError(t, err) - err = cl.Delete(context.Background(), &pod) + err = cl.Delete(t.Context(), &pod) require.NoError(t, err) assert.EventuallyWithT(t, func(c *assert.CollectT) { var sts appsv1.StatefulSet - err = cl.Get(context.Background(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server"}, &sts) + err = cl.Get(t.Context(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server"}, &sts) assert.NoError(c, err) assert.Equal(c, int32(1), sts.Status.ReadyReplicas) }, time.Minute, time.Second) @@ -390,14 +389,14 @@ func TestActorReminderSchedulerTriggerPerformance(t *testing.T) { cl, err := client.New(platform.KubeClient.GetClientConfig(), client.Options{Scheme: scheme}) require.NoError(t, err) var pod corev1.Pod - err = cl.Get(context.Background(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) + err = cl.Get(t.Context(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) require.NoError(t, err) - err = cl.Delete(context.Background(), &pod) + err = cl.Delete(t.Context(), &pod) require.NoError(t, err) assert.EventuallyWithT(t, func(c *assert.CollectT) { var sts appsv1.StatefulSet - err = cl.Get(context.Background(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server"}, &sts) + err = cl.Get(t.Context(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server"}, &sts) assert.NoError(c, err) assert.Equal(c, int32(1), sts.Status.ReadyReplicas) }, time.Minute, time.Second) diff --git a/tests/perf/workflows/workflow_test.go b/tests/perf/workflows/workflow_test.go index 56ab3c80c..a05f3f9f8 100644 --- a/tests/perf/workflows/workflow_test.go +++ b/tests/perf/workflows/workflow_test.go @@ -17,7 +17,6 @@ limitations under the License. package workflows import ( - "context" "encoding/json" "fmt" "log" @@ -159,12 +158,12 @@ func testWorkflow(t *testing.T, workflowName string, testAppName string, inputs cl, err := client.New(platform.KubeClient.GetClientConfig(), client.Options{Scheme: scheme}) require.NoError(t, err) var pod corev1.Pod - err = cl.Get(context.Background(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) + err = cl.Get(t.Context(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) require.NoError(t, err) - err = cl.Delete(context.Background(), &pod) + err = cl.Delete(t.Context(), &pod) require.NoError(t, err) assert.EventuallyWithT(t, func(c *assert.CollectT) { - err = cl.Get(context.Background(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) + err = cl.Get(t.Context(), client.ObjectKey{Namespace: kube.DaprTestNamespace, Name: "dapr-scheduler-server-0"}, &pod) if assert.NoError(c, err) { assert.Equal(c, corev1.PodRunning, pod.Status.Phase) } diff --git a/tests/platforms/kubernetes/app_description_test.go b/tests/platforms/kubernetes/app_description_test.go index 773726047..5aaa4e394 100644 --- a/tests/platforms/kubernetes/app_description_test.go +++ b/tests/platforms/kubernetes/app_description_test.go @@ -56,13 +56,13 @@ func TestAppDescription_MarshalJSON(t *testing.T) { app := AppDescription{ IngressEnabled: true, } - os.Setenv(useServiceInternalIP, "false") + t.Setenv(useServiceInternalIP, "false") if !app.ShouldBeExposed() { t.Error("AppDescription.ShouldBeExposed() should evaluate to true when ingress is enabled and internal ip should not be used") } - os.Setenv(useServiceInternalIP, "true") + t.Setenv(useServiceInternalIP, "true") if app.ShouldBeExposed() { t.Error("AppDescription.ShouldBeExposed() should evaluate to false when internal ip should be used") diff --git a/tests/platforms/kubernetes/appmanager_test.go b/tests/platforms/kubernetes/appmanager_test.go index 4b095f8f9..844479230 100644 --- a/tests/platforms/kubernetes/appmanager_test.go +++ b/tests/platforms/kubernetes/appmanager_test.go @@ -14,7 +14,6 @@ limitations under the License. package kubernetes import ( - "context" "fmt" "testing" @@ -75,7 +74,7 @@ func TestDeployApp(t *testing.T) { // assert deploymentClient := client.Deployments(testNamespace) - deployment, _ := deploymentClient.Get(context.TODO(), testApp.AppName, metav1.GetOptions{}) + deployment, _ := deploymentClient.Get(t.Context(), testApp.AppName, metav1.GetOptions{}) assert.NotNil(t, deployment) assert.Equal(t, testApp.AppName, deployment.ObjectMeta.Name) assert.Equal(t, testNamespace, deployment.ObjectMeta.Namespace) @@ -367,7 +366,7 @@ func TestCreateIngressService(t *testing.T) { require.NoError(t, err) // assert serviceClient := client.Services(testNamespace) - obj, _ := serviceClient.Get(context.TODO(), testApp.AppName, metav1.GetOptions{}) + obj, _ := serviceClient.Get(t.Context(), testApp.AppName, metav1.GetOptions{}) assert.NotNil(t, obj) assert.Equal(t, testApp.AppName, obj.ObjectMeta.Name) assert.Equal(t, testNamespace, obj.ObjectMeta.Namespace) @@ -383,7 +382,7 @@ func TestCreateIngressService(t *testing.T) { require.NoError(t, err) // assert serviceClient := client.Services(testNamespace) - obj, _ := serviceClient.Get(context.TODO(), testApp.AppName, metav1.GetOptions{}) + obj, _ := serviceClient.Get(t.Context(), testApp.AppName, metav1.GetOptions{}) assert.NotNil(t, obj) assert.Equal(t, testApp.AppName, obj.ObjectMeta.Name) assert.Equal(t, testNamespace, obj.ObjectMeta.Namespace) diff --git a/tests/platforms/kubernetes/daprcomponent_test.go b/tests/platforms/kubernetes/daprcomponent_test.go index 108006704..c5ddd1052 100644 --- a/tests/platforms/kubernetes/daprcomponent_test.go +++ b/tests/platforms/kubernetes/daprcomponent_test.go @@ -75,6 +75,6 @@ func TestDaprComponentSpec(t *testing.T) { annotations := daprComponent.toComponentSpec().ObjectMeta.Annotations assert.Len(t, annotations, 1) - assert.Equal(t, testContainer, annotations["dapr.io/component-container"]) + assert.JSONEq(t, testContainer, annotations["dapr.io/component-container"]) }) } diff --git a/tests/platforms/kubernetes/portforward.go b/tests/platforms/kubernetes/portforward.go index 56bc20014..e41fbf1d2 100644 --- a/tests/platforms/kubernetes/portforward.go +++ b/tests/platforms/kubernetes/portforward.go @@ -75,7 +75,7 @@ func (p *PodPortForwarder) Connect(name string, targetPorts ...int) ([]int, erro config := p.client.GetClientConfig() ports := make([]int, 0, len(targetPorts)) - for range len(targetPorts) { + for range targetPorts { p, perr := freeport.GetFreePort() if perr != nil { return nil, perr diff --git a/tests/runner/loadtest/k6_client_test.go b/tests/runner/loadtest/k6_client_test.go index ba540c4d3..3d69c387c 100644 --- a/tests/runner/loadtest/k6_client_test.go +++ b/tests/runner/loadtest/k6_client_test.go @@ -15,7 +15,6 @@ package loadtest import ( "bytes" - "context" "io" "net/http" "testing" @@ -59,7 +58,7 @@ func TestK6Client(t *testing.T) { called++ assert.Equal(t, "DELETE", r.Method) }) - require.NoError(t, k6.Delete(context.Background(), k6Name, metav1.DeleteOptions{})) + require.NoError(t, k6.Delete(t.Context(), k6Name, metav1.DeleteOptions{})) assert.Equal(t, 1, called) }) t.Run("Get should call rest GET", func(t *testing.T) { @@ -68,7 +67,7 @@ func TestK6Client(t *testing.T) { called++ assert.Equal(t, "GET", r.Method) }) - _, err = k6.Get(context.Background(), k6Name) + _, err = k6.Get(t.Context(), k6Name) require.NoError(t, err) assert.Equal(t, 1, called) }) @@ -78,7 +77,7 @@ func TestK6Client(t *testing.T) { called++ assert.Equal(t, "POST", r.Method) }) - _, err = k6.Create(context.Background(), &v1.K6{}) + _, err = k6.Create(t.Context(), &v1.K6{}) require.NoError(t, err) assert.Equal(t, 1, called) }) @@ -88,7 +87,7 @@ func TestK6Client(t *testing.T) { called++ assert.Equal(t, "GET", r.Method) }) - _, err = k6.List(context.Background(), metav1.ListOptions{}) + _, err = k6.List(t.Context(), metav1.ListOptions{}) require.NoError(t, err) assert.Equal(t, 1, called) }) diff --git a/tests/runner/loadtest/k6_test.go b/tests/runner/loadtest/k6_test.go index 4c3730fad..38dcc1076 100644 --- a/tests/runner/loadtest/k6_test.go +++ b/tests/runner/loadtest/k6_test.go @@ -157,7 +157,7 @@ func TestK6(t *testing.T) { ) t.Run("With-ish parameters should configure k6 tester", func(t *testing.T) { const envVarKey, envVarValue, fakeImg, parallelism, fakeName = "key", "value", "img", 3, "name" - fakeCtx := context.TODO() + fakeCtx := t.Context() tester := NewK6( script, WithCtx(fakeCtx), diff --git a/utils/utils_test.go b/utils/utils_test.go index b294cd978..174d28ed6 100644 --- a/utils/utils_test.go +++ b/utils/utils_test.go @@ -108,7 +108,7 @@ func TestEnvOrElse(t *testing.T) { const elseValue, fakeEnVar, fakeEnvVarValue = "fakeValue", "envVarThatExists", "envVarValue" defer os.Unsetenv(fakeEnVar) - require.NoError(t, os.Setenv(fakeEnVar, fakeEnvVarValue)) + t.Setenv(fakeEnVar, fakeEnvVarValue) assert.Equal(t, fakeEnvVarValue, GetEnvOrElse(fakeEnVar, elseValue)) }) } @@ -272,7 +272,7 @@ func BenchmarkFilter(b *testing.B) { return n < 50 } - for n := 0; n < b.N; n++ { + for range b.N { Filter(vals, filterFn) } }