dragonfly/scheduler/networktopology/network_topology.go

498 lines
15 KiB
Go

/*
* Copyright 2023 The Dragonfly Authors
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
//go:generate mockgen -destination mocks/network_topology_mock.go -source network_topology.go -package mocks
package networktopology
import (
"context"
"errors"
"sort"
"strconv"
"time"
"github.com/google/uuid"
"github.com/redis/go-redis/v9"
logger "d7y.io/dragonfly/v2/internal/dflog"
"d7y.io/dragonfly/v2/pkg/cache"
"d7y.io/dragonfly/v2/pkg/container/set"
pkgredis "d7y.io/dragonfly/v2/pkg/redis"
"d7y.io/dragonfly/v2/scheduler/config"
"d7y.io/dragonfly/v2/scheduler/resource"
"d7y.io/dragonfly/v2/scheduler/storage"
)
const (
// contextTimeout is the timeout of redis invoke.
contextTimeout = 2 * time.Minute
// snapshotContextTimeout is the timeout of snapshot network topology.
snapshotContextTimeout = 20 * time.Minute
// findProbedCandidateHostsLimit is the limit of find probed candidate hosts.
findProbedCandidateHostsLimit = 50
// defaultScanCountLimit is the predefined amount of work performed with each 'Scan' operation called when retrieve elements from Redis.
defaultScanCountLimit = 64
)
// NetworkTopology is an interface for network topology.
type NetworkTopology interface {
// Started network topology server.
Serve()
// Stop network topology server.
Stop()
// Has to check if there is a connection between source host and destination host.
Has(string, string) bool
// Store stores source host and destination host.
Store(string, string) error
// FindProbedHosts finds the most candidate destination host to be probed, randomly find a range of hosts,
// and then return the host with a smaller probed count.
FindProbedHosts(string) ([]*resource.Host, error)
// DeleteHost deletes source host and all destination host connected to source host.
DeleteHost(string) error
// Probes loads probes interface by source host id and destination host id.
Probes(string, string) Probes
// ProbedCount is the number of times the host has been probed.
ProbedCount(string) (uint64, error)
// Neighbours gets the specified number neighbors of source host for aggregation, by regexp scaning cache
// (if it is not enough for code to work, access redis to get neighbors), then parsing keys and loading host,
// while updating the cache data.
Neighbours(*resource.Host, int) ([]*resource.Host, error)
// Snapshot writes the current network topology to the storage.
Snapshot() error
}
// networkTopology is an implementation of network topology.
type networkTopology struct {
// config is the network topology config.
config config.NetworkTopologyConfig
// rdb is Redis universal client interface.
rdb redis.UniversalClient
// Cache instance.
cache cache.Cache
// resource is resource interface.
resource resource.Resource
// storage is storage interface.
storage storage.Storage
// done channel will be closed when network topology serve stop.
done chan struct{}
}
// New network topology interface.
func NewNetworkTopology(cfg config.NetworkTopologyConfig, rdb redis.UniversalClient, cache cache.Cache, resource resource.Resource, storage storage.Storage) (NetworkTopology, error) {
return &networkTopology{
config: cfg,
rdb: rdb,
cache: cache,
resource: resource,
storage: storage,
done: make(chan struct{}),
}, nil
}
// Started network topology server.
func (nt *networkTopology) Serve() {
logger.Info("collect network topology records")
tick := time.NewTicker(nt.config.CollectInterval)
for {
select {
case <-tick.C:
if err := nt.Snapshot(); err != nil {
logger.Error(err)
break
}
case <-nt.done:
return
}
}
}
// Stop network topology server.
func (nt *networkTopology) Stop() {
close(nt.done)
}
// Has to check if there is a connection between source host and destination host.
func (nt *networkTopology) Has(srcHostID string, destHostID string) bool {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
networkTopologyKey := pkgredis.MakeNetworkTopologyKeyInScheduler(srcHostID, destHostID)
if _, _, ok := nt.cache.GetWithExpiration(networkTopologyKey); ok {
return true
}
networkTopology, err := nt.rdb.HGetAll(ctx, networkTopologyKey).Result()
if err != nil {
logger.Errorf("get networkTopology failed: %s", err.Error())
return false
}
if len(networkTopology) == 0 {
return false
}
// Add cache data.
nt.cache.Set(networkTopologyKey, networkTopology, nt.config.Cache.TTL)
return true
}
// Store stores source host and destination host.
func (nt *networkTopology) Store(srcHostID string, destHostID string) error {
// If the network topology already exists, skip it.
if nt.Has(srcHostID, destHostID) {
return nil
}
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
if err := nt.rdb.HSet(ctx, pkgredis.MakeNetworkTopologyKeyInScheduler(srcHostID, destHostID), "createdAt", time.Now().Format(time.RFC3339Nano)).Err(); err != nil {
return err
}
return nil
}
// FindProbedHosts finds the most candidate destination host to be probed, randomly find a range of hosts,
// and then return the host with a smaller probed count.
func (nt *networkTopology) FindProbedHosts(hostID string) ([]*resource.Host, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
blocklist := set.NewSafeSet[string]()
blocklist.Add(hostID)
candidateHosts := nt.resource.HostManager().LoadRandomHosts(findProbedCandidateHostsLimit, blocklist)
if len(candidateHosts) == 0 {
return nil, errors.New("probed hosts not found")
}
if len(candidateHosts) <= nt.config.Probe.Count {
return candidateHosts, nil
}
var probedCountKeys []string
probedCounts := make(map[string]uint64)
for _, candidateHost := range candidateHosts {
probedCountKey := pkgredis.MakeProbedCountKeyInScheduler(candidateHost.ID)
cache, _, ok := nt.cache.GetWithExpiration(probedCountKey)
if ok {
if probedCount, ok := cache.(uint64); ok {
probedCounts[probedCountKey] = probedCount
}
continue
}
probedCountKeys = append(probedCountKeys, probedCountKey)
}
rawProbedCounts, err := nt.rdb.MGet(ctx, probedCountKeys...).Result()
if err != nil {
return nil, err
}
// Filter invalid probed count. If probed key not exist, the probed count is nil.
for i, rawProbedCount := range rawProbedCounts {
// Initialize the probedCount value of host in redis when the host is first selected as the candidate probe target.
if rawProbedCount == nil {
if err := nt.rdb.Set(ctx, probedCountKeys[i], 0, 0).Err(); err != nil {
return nil, err
}
continue
}
value, ok := rawProbedCount.(string)
if !ok {
return nil, errors.New("invalid value type")
}
probedCount, err := strconv.ParseUint(value, 10, 64)
if err != nil {
return nil, errors.New("invalid probed count")
}
probedCounts[probedCountKeys[i]] = probedCount
// Add cache data.
nt.cache.Set(probedCountKeys[i], probedCount, nt.config.Cache.TTL)
}
// Sort candidate hosts by probed count.
sort.Slice(candidateHosts, func(i, j int) bool {
return probedCounts[pkgredis.MakeProbedCountKeyInScheduler(candidateHosts[i].ID)] < probedCounts[pkgredis.MakeProbedCountKeyInScheduler(candidateHosts[j].ID)]
})
return candidateHosts[:nt.config.Probe.Count], nil
}
// DeleteHost deletes source host and all destination host connected to source host.
func (nt *networkTopology) DeleteHost(hostID string) error {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
deleteKeys := []string{pkgredis.MakeProbedCountKeyInScheduler(hostID)}
srcNetworkTopologyKeys, _, err := nt.rdb.Scan(ctx, 0, pkgredis.MakeNetworkTopologyKeyInScheduler(hostID, "*"), defaultScanCountLimit).Result()
if err != nil {
return err
}
deleteKeys = append(deleteKeys, srcNetworkTopologyKeys...)
destNetworkTopologyKeys, _, err := nt.rdb.Scan(ctx, 0, pkgredis.MakeNetworkTopologyKeyInScheduler("*", hostID), defaultScanCountLimit).Result()
if err != nil {
return err
}
deleteKeys = append(deleteKeys, destNetworkTopologyKeys...)
srcProbesKeys, _, err := nt.rdb.Scan(ctx, 0, pkgredis.MakeProbesKeyInScheduler(hostID, "*"), defaultScanCountLimit).Result()
if err != nil {
return err
}
deleteKeys = append(deleteKeys, srcProbesKeys...)
destProbesKeys, _, err := nt.rdb.Scan(ctx, 0, pkgredis.MakeProbesKeyInScheduler("*", hostID), defaultScanCountLimit).Result()
if err != nil {
return err
}
deleteKeys = append(deleteKeys, destProbesKeys...)
if err := nt.rdb.Del(ctx, deleteKeys...).Err(); err != nil {
return err
}
for _, deleteKey := range deleteKeys {
nt.cache.Delete(deleteKey)
}
return nil
}
// Probes loads probes interface by source host id and destination host id.
func (nt *networkTopology) Probes(srcHostID, destHostID string) Probes {
return NewProbes(nt.config, nt.rdb, nt.cache, srcHostID, destHostID)
}
// ProbedCount is the number of times the host has been probed.
func (nt *networkTopology) ProbedCount(hostID string) (uint64, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
probedCountKey := pkgredis.MakeProbedCountKeyInScheduler(hostID)
if cache, _, ok := nt.cache.GetWithExpiration(probedCountKey); ok {
probedCount, ok := cache.(uint64)
if ok {
return probedCount, nil
}
return uint64(0), errors.New("get probedCount failed")
}
probedCount, err := nt.rdb.Get(ctx, probedCountKey).Uint64()
if err != nil {
return uint64(0), err
}
// Add cache data.
nt.cache.Set(probedCountKey, probedCount, nt.config.Cache.TTL)
return probedCount, nil
}
// Neighbours gets the specified number neighbors of source host for aggregation by regexp scaning cache
// (if it is not enough for code to work, get neighbors from redis), then parsing keys and loading host,
// while updating the cache data.
func (nt *networkTopology) Neighbours(srcHost *resource.Host, n int) ([]*resource.Host, error) {
ctx, cancel := context.WithTimeout(context.Background(), contextTimeout)
defer cancel()
networkTopologyKeys, err := nt.cache.Scan(pkgredis.MakeNetworkTopologyKeyInScheduler(srcHost.ID, "*"), n)
if err != nil {
return nil, err
}
// If it is not enough for code to work, get neighbors from redis.
if len(networkTopologyKeys) < n {
networkTopologyKeys, _, err = nt.rdb.Scan(ctx, 0, pkgredis.MakeNetworkTopologyKeyInScheduler(srcHost.ID, "*"), defaultScanCountLimit).Result()
if err != nil {
return nil, err
}
var networkTopology map[string]string
for _, networkTopologyKey := range networkTopologyKeys {
if networkTopology, err = nt.rdb.HGetAll(ctx, networkTopologyKey).Result(); err != nil {
logger.Error(err)
continue
}
// Add cache data.
nt.cache.Set(networkTopologyKey, networkTopology, nt.config.Cache.TTL)
}
}
neighbours := make([]*resource.Host, 0, n)
for _, networkTopologyKey := range networkTopologyKeys {
if len(neighbours) >= n {
break
}
_, _, _, neighbourID, err := pkgredis.ParseNetworkTopologyKeyInScheduler(networkTopologyKey)
if err != nil {
logger.Error(err)
continue
}
neighbour, loaded := nt.resource.HostManager().Load(neighbourID)
if !loaded {
logger.Errorf("host %s not found", neighbourID)
continue
}
neighbours = append(neighbours, neighbour)
}
return neighbours, nil
}
// Snapshot writes the current network topology to the storage.
func (nt *networkTopology) Snapshot() error {
ctx, cancel := context.WithTimeout(context.Background(), snapshotContextTimeout)
defer cancel()
now := time.Now()
id := uuid.NewString()
probedCountKeys, _, err := nt.rdb.Scan(ctx, 0, pkgredis.MakeProbedCountKeyInScheduler("*"), defaultScanCountLimit).Result()
if err != nil {
return err
}
for _, probedCountKey := range probedCountKeys {
_, _, srcHostID, err := pkgredis.ParseProbedCountKeyInScheduler(probedCountKey)
if err != nil {
logger.Error(err)
continue
}
// Construct destination hosts for network topology.
networkTopologyKeys, _, err := nt.rdb.Scan(ctx, 0, pkgredis.MakeNetworkTopologyKeyInScheduler(srcHostID, "*"), defaultScanCountLimit).Result()
if err != nil {
logger.Error(err)
continue
}
destHosts := make([]storage.DestHost, 0, len(networkTopologyKeys))
for _, networkTopologyKey := range networkTopologyKeys {
_, _, srcHostID, destHostID, err := pkgredis.ParseNetworkTopologyKeyInScheduler(networkTopologyKey)
if err != nil {
logger.Error(err)
continue
}
host, loaded := nt.resource.HostManager().Load(destHostID)
if !loaded {
logger.Errorf("host %s not found", destHostID)
continue
}
ps := nt.Probes(srcHostID, destHostID)
averageRTT, err := ps.AverageRTT()
if err != nil {
logger.Error(err)
continue
}
createdAt, err := ps.CreatedAt()
if err != nil {
logger.Error(err)
continue
}
updatedAt, err := ps.UpdatedAt()
if err != nil {
logger.Error(err)
continue
}
destHost := storage.DestHost{
ID: host.ID,
Type: host.Type.Name(),
Hostname: host.Hostname,
IP: host.IP,
Port: host.Port,
Network: resource.Network{
TCPConnectionCount: host.Network.TCPConnectionCount,
UploadTCPConnectionCount: host.Network.UploadTCPConnectionCount,
Location: host.Network.Location,
IDC: host.Network.IDC,
},
Probes: storage.Probes{
AverageRTT: averageRTT.Nanoseconds(),
CreatedAt: createdAt.UnixNano(),
UpdatedAt: updatedAt.UnixNano(),
},
}
destHosts = append(destHosts, destHost)
}
// Construct source hosts for network topology.
host, loaded := nt.resource.HostManager().Load(srcHostID)
if !loaded {
logger.Errorf("host %s not found", srcHostID)
continue
}
if err = nt.storage.CreateNetworkTopology(storage.NetworkTopology{
ID: id,
Host: storage.SrcHost{
ID: host.ID,
Type: host.Type.Name(),
Hostname: host.Hostname,
IP: host.IP,
Port: host.Port,
Network: resource.Network{
TCPConnectionCount: host.Network.TCPConnectionCount,
UploadTCPConnectionCount: host.Network.UploadTCPConnectionCount,
Location: host.Network.Location,
IDC: host.Network.IDC,
},
},
DestHosts: destHosts,
CreatedAt: now.UnixNano(),
}); err != nil {
logger.Error(err)
continue
}
}
return nil
}