dragonfly/pkg/rpc/manager/client/client.go

184 lines
5.2 KiB
Go

/*
* Copyright 2020 The Dragonfly Authors
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
//go:generate mockgen -destination mocks/client_mock.go -source client.go -package mocks
package client
import (
"context"
"errors"
"time"
grpc_middleware "github.com/grpc-ecosystem/go-grpc-middleware"
grpc_zap "github.com/grpc-ecosystem/go-grpc-middleware/logging/zap"
grpc_prometheus "github.com/grpc-ecosystem/go-grpc-prometheus"
"google.golang.org/grpc"
"google.golang.org/grpc/backoff"
"google.golang.org/grpc/credentials/insecure"
logger "d7y.io/dragonfly/v2/internal/dflog"
"d7y.io/dragonfly/v2/pkg/dfnet"
"d7y.io/dragonfly/v2/pkg/reachable"
"d7y.io/dragonfly/v2/pkg/rpc/manager"
)
const (
contextTimeout = 2 * time.Minute
backoffBaseDelay = 1 * time.Second
backoffMultiplier = 1.6
backoffJitter = 0.2
backoffMaxDelay = 10 * time.Second
)
type Client interface {
// Update Seed peer configuration.
UpdateSeedPeer(*manager.UpdateSeedPeerRequest) (*manager.SeedPeer, error)
// Update CDN configuration.
UpdateCDN(*manager.UpdateCDNRequest) (*manager.CDN, error)
// Get Scheduler and Scheduler cluster configuration.
GetScheduler(*manager.GetSchedulerRequest) (*manager.Scheduler, error)
// Update scheduler configuration.
UpdateScheduler(*manager.UpdateSchedulerRequest) (*manager.Scheduler, error)
// List acitve schedulers configuration.
ListSchedulers(*manager.ListSchedulersRequest) (*manager.ListSchedulersResponse, error)
// KeepAlive with manager.
KeepAlive(time.Duration, *manager.KeepAliveRequest)
// Close client connect.
Close() error
}
type client struct {
manager.ManagerClient
conn *grpc.ClientConn
}
func New(target string) (Client, error) {
conn, err := grpc.Dial(
target,
grpc.WithTransportCredentials(insecure.NewCredentials()),
grpc.WithBlock(),
grpc.WithConnectParams(grpc.ConnectParams{
Backoff: backoff.Config{
BaseDelay: backoffBaseDelay,
Multiplier: backoffMultiplier,
Jitter: backoffJitter,
MaxDelay: backoffMaxDelay,
},
}),
grpc.WithStreamInterceptor(grpc_middleware.ChainStreamClient(
grpc_prometheus.StreamClientInterceptor,
grpc_zap.StreamClientInterceptor(logger.GrpcLogger.Desugar()),
)),
)
if err != nil {
return nil, err
}
return &client{
ManagerClient: manager.NewManagerClient(conn),
conn: conn,
}, nil
}
func NewWithAddrs(netAddrs []dfnet.NetAddr) (Client, error) {
for _, netAddr := range netAddrs {
ipReachable := reachable.New(&reachable.Config{Address: netAddr.Addr})
if err := ipReachable.Check(); err == nil {
logger.Infof("use %s address for manager grpc client", netAddr.Addr)
return New(netAddr.Addr)
}
logger.Warnf("%s manager address can not reachable", netAddr.Addr)
}
return nil, errors.New("can not find available manager addresses")
}
func (c *client) UpdateSeedPeer(req *manager.UpdateSeedPeerRequest) (*manager.SeedPeer, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
return c.ManagerClient.UpdateSeedPeer(ctx, req)
}
func (c *client) UpdateCDN(req *manager.UpdateCDNRequest) (*manager.CDN, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
return c.ManagerClient.UpdateCDN(ctx, req)
}
func (c *client) GetScheduler(req *manager.GetSchedulerRequest) (*manager.Scheduler, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
return c.ManagerClient.GetScheduler(ctx, req)
}
func (c *client) UpdateScheduler(req *manager.UpdateSchedulerRequest) (*manager.Scheduler, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
return c.ManagerClient.UpdateScheduler(ctx, req)
}
func (c *client) ListSchedulers(req *manager.ListSchedulersRequest) (*manager.ListSchedulersResponse, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
return c.ManagerClient.ListSchedulers(ctx, req)
}
func (c *client) KeepAlive(interval time.Duration, keepalive *manager.KeepAliveRequest) {
retry:
ctx, cancel := context.WithCancel(context.Background())
stream, err := c.ManagerClient.KeepAlive(ctx)
if err != nil {
time.Sleep(interval)
cancel()
goto retry
}
tick := time.NewTicker(interval)
for {
select {
case <-tick.C:
if err := stream.Send(&manager.KeepAliveRequest{
HostName: keepalive.HostName,
SourceType: keepalive.SourceType,
ClusterId: keepalive.ClusterId,
}); err != nil {
if _, err := stream.CloseAndRecv(); err != nil {
logger.Errorf("hostname %s cluster id %v close and recv stream failed: %v", keepalive.HostName, keepalive.ClusterId, err)
}
cancel()
goto retry
}
}
}
}
func (c *client) Close() error {
return c.conn.Close()
}