dragonfly/manager/service/job.go

878 lines
27 KiB
Go

/*
* Copyright 2020 The Dragonfly Authors
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package service
import (
"context"
"crypto/x509"
"errors"
"fmt"
"sync"
"time"
machineryv1tasks "github.com/dragonflyoss/machinery/v1/tasks"
"github.com/google/uuid"
"golang.org/x/sync/errgroup"
"gorm.io/gorm"
logger "d7y.io/dragonfly/v2/internal/dflog"
internaljob "d7y.io/dragonfly/v2/internal/job"
"d7y.io/dragonfly/v2/manager/job"
"d7y.io/dragonfly/v2/manager/metrics"
"d7y.io/dragonfly/v2/manager/models"
"d7y.io/dragonfly/v2/manager/types"
pkggc "d7y.io/dragonfly/v2/pkg/gc"
"d7y.io/dragonfly/v2/pkg/idgen"
"d7y.io/dragonfly/v2/pkg/net/http"
"d7y.io/dragonfly/v2/pkg/retry"
"d7y.io/dragonfly/v2/pkg/slices"
"d7y.io/dragonfly/v2/pkg/structure"
pkgtypes "d7y.io/dragonfly/v2/pkg/types"
)
const (
// DefaultGCJobPollingTimeout is the default timeout for polling GC job.
DefaultGCJobPollingTimeout = 30 * time.Minute
// DefaultGCJobPollingInterval is the default interval for polling GC job.
DefaultGCJobPollingInterval = 5 * time.Second
// DefaultGetImageDistributionJobConcurrentCount is the default concurrent count for getting image distribution job.
DefaultGetImageDistributionJobConcurrentCount = 30
)
func (s *service) CreateGCJob(ctx context.Context, json types.CreateGCJobRequest) (*models.Job, error) {
taskID := uuid.NewString()
ctx = context.WithValue(ctx, pkggc.ContextKeyTaskID, taskID)
ctx = context.WithValue(ctx, pkggc.ContextKeyUserID, json.UserID)
// This is a non-block function to run the gc task, which will run the task asynchronously in the backend.
if err := s.gc.Run(ctx, json.Args.Type); err != nil {
logger.Errorf("run gc job failed: %w", err)
return nil, err
}
return s.pollingGCJob(ctx, json.Type, json.UserID, taskID)
}
func (s *service) pollingGCJob(ctx context.Context, jobType string, userID uint, taskID string) (*models.Job, error) {
ctx, cancel := context.WithTimeout(ctx, DefaultGCJobPollingTimeout)
defer cancel()
ticker := time.NewTicker(DefaultGCJobPollingInterval)
defer ticker.Stop()
job := models.Job{}
for {
select {
case <-ctx.Done():
return nil, fmt.Errorf("context done: %w", ctx.Err())
case <-ticker.C:
if err := s.db.WithContext(ctx).First(&job, models.Job{
Type: jobType,
UserID: userID,
TaskID: taskID,
}).Error; err != nil {
if errors.Is(err, gorm.ErrRecordNotFound) {
continue
}
return nil, err
}
// Return the job if the job is in success or failure state, otherwise continue polling.
if job.State == machineryv1tasks.StateSuccess || job.State == machineryv1tasks.StateFailure {
return &job, nil
}
}
}
}
func (s *service) CreateSyncPeersJob(ctx context.Context, json types.CreateSyncPeersJobRequest) error {
schedulers, err := s.findSchedulerInClusters(ctx, json.SchedulerClusterIDs)
if err != nil {
logger.Errorf("find scheduler in clusters failed: %w", err)
return err
}
return s.job.SyncPeers.CreateSyncPeers(ctx, schedulers)
}
func (s *service) CreatePreheatJob(ctx context.Context, json types.CreatePreheatJobRequest) (*models.Job, error) {
if json.Args.Scope == "" {
json.Args.Scope = types.SingleSeedPeerScope
}
if json.Args.ConcurrentCount == 0 {
json.Args.ConcurrentCount = types.DefaultPreheatConcurrentCount
}
if json.Args.Timeout == 0 {
json.Args.Timeout = types.DefaultJobTimeout
}
if json.Args.FilteredQueryParams == "" {
json.Args.FilteredQueryParams = http.RawDefaultFilteredQueryParams
}
args, err := structure.StructToMap(json.Args)
if err != nil {
logger.Errorf("convert preheat args to map failed: %w", err)
return nil, err
}
candidateSchedulers, err := s.findAllCandidateSchedulersInClusters(ctx, json.SchedulerClusterIDs, []string{types.SchedulerFeaturePreheat})
if err != nil {
logger.Errorf("find candidate schedulers in clusters failed: %w", err)
return nil, err
}
groupJobState, err := s.job.CreatePreheat(ctx, candidateSchedulers, json.Args)
if err != nil {
logger.Errorf("create preheat job failed: %w", err)
return nil, err
}
var candidateSchedulerClusters []models.SchedulerCluster
for _, candidateScheduler := range candidateSchedulers {
candidateSchedulerClusters = append(candidateSchedulerClusters, candidateScheduler.SchedulerCluster)
}
job := models.Job{
TaskID: groupJobState.GroupUUID,
BIO: json.BIO,
Type: json.Type,
State: groupJobState.State,
Args: args,
UserID: json.UserID,
SchedulerClusters: candidateSchedulerClusters,
}
if err := s.db.WithContext(ctx).Create(&job).Error; err != nil {
logger.Errorf("create preheat job failed: %w", err)
return nil, err
}
go s.pollingJob(context.Background(), internaljob.PreheatJob, job.ID, job.TaskID, 30, 300, 16)
return &job, nil
}
func (s *service) CreateGetTaskJob(ctx context.Context, json types.CreateGetTaskJobRequest) (*models.Job, error) {
if json.Args.FilteredQueryParams == "" {
json.Args.FilteredQueryParams = http.RawDefaultFilteredQueryParams
}
args, err := structure.StructToMap(json.Args)
if err != nil {
logger.Errorf("convert get task args to map failed: %w", err)
return nil, err
}
schedulers, err := s.findAllSchedulersInClusters(ctx, json.SchedulerClusterIDs)
if err != nil {
logger.Errorf("find schedulers in clusters failed: %w", err)
return nil, err
}
groupJobState, err := s.job.CreateGetTask(ctx, schedulers, json.Args)
if err != nil {
logger.Errorf("create get task job failed: %w", err)
return nil, err
}
var schedulerClusters []models.SchedulerCluster
for _, scheduler := range schedulers {
schedulerClusters = append(schedulerClusters, scheduler.SchedulerCluster)
}
job := models.Job{
TaskID: groupJobState.GroupUUID,
BIO: json.BIO,
Type: json.Type,
State: groupJobState.State,
Args: args,
UserID: json.UserID,
SchedulerClusters: schedulerClusters,
}
if err := s.db.WithContext(ctx).Create(&job).Error; err != nil {
logger.Errorf("create get task job failed: %w", err)
return nil, err
}
go s.pollingJob(context.Background(), internaljob.GetTaskJob, job.ID, job.TaskID, 30, 300, 16)
logger.Infof("create get task job %s for task %s in scheduler clusters %v", job.ID, job.TaskID, json.SchedulerClusterIDs)
return &job, nil
}
func (s *service) CreateGetImageDistributionJob(ctx context.Context, json types.CreateGetImageDistributionJobRequest) (*types.CreateGetImageDistributionJobResponse, error) {
imageLayers, err := s.getImageLayers(ctx, json)
if err != nil {
err = fmt.Errorf("get image layers failed: %w", err)
logger.Error(err)
return nil, err
}
var layers []internaljob.PreheatRequest
for _, imageLayer := range imageLayers {
for _, url := range imageLayer.URLs {
layers = append(layers, internaljob.PreheatRequest{
URL: url,
PieceLength: imageLayer.PieceLength,
Tag: imageLayer.Tag,
Application: imageLayer.Application,
FilteredQueryParams: imageLayer.FilteredQueryParams,
})
}
}
if len(layers) == 0 {
err = errors.New("no valid image layers found")
logger.Error(err)
return nil, err
}
image := types.Image{Layers: make([]types.Layer, 0, len(layers))}
for _, file := range layers {
image.Layers = append(image.Layers, types.Layer{URL: file.URL})
}
schedulers, err := s.findAllSchedulersInClusters(ctx, json.SchedulerClusterIDs)
if err != nil {
err = fmt.Errorf("find schedulers in clusters failed: %w", err)
logger.Error(err)
return nil, err
}
// Create multiple get task jobs synchronously for each layer and
// extract peers from the jobs.
jobs := s.createGetTaskJobsSync(ctx, layers, json, schedulers)
peers, err := s.extractPeersFromJobs(jobs)
if err != nil {
err = fmt.Errorf("extract peers from jobs failed: %w", err)
logger.Error(err)
return nil, err
}
return &types.CreateGetImageDistributionJobResponse{
Image: image,
Peers: peers,
}, nil
}
func (s *service) getImageLayers(ctx context.Context, json types.CreateGetImageDistributionJobRequest) ([]internaljob.PreheatRequest, error) {
var certPool *x509.CertPool
if len(s.config.Job.Preheat.TLS.CACert) != 0 {
certPool = x509.NewCertPool()
if !certPool.AppendCertsFromPEM([]byte(s.config.Job.Preheat.TLS.CACert)) {
return nil, errors.New("invalid CA Cert")
}
}
layers, err := job.GetImageLayers(ctx, types.PreheatArgs{
Type: job.PreheatImageType.String(),
URL: json.Args.URL,
PieceLength: json.Args.PieceLength,
Tag: json.Args.Tag,
Application: json.Args.Application,
FilteredQueryParams: json.Args.FilteredQueryParams,
Headers: json.Args.Headers,
Username: json.Args.Username,
Password: json.Args.Password,
Platform: json.Args.Platform,
}, s.config.Job.Preheat.RegistryTimeout, certPool, s.config.Job.Preheat.TLS.InsecureSkipVerify)
if err != nil {
return nil, fmt.Errorf("get image layers failed: %w", err)
}
if len(layers) == 0 {
return nil, errors.New("no valid image layers found")
}
return layers, nil
}
func (s *service) createGetTaskJobsSync(ctx context.Context, layers []internaljob.PreheatRequest, json types.CreateGetImageDistributionJobRequest, schedulers []models.Scheduler) []*models.Job {
var mu sync.Mutex
jobs := make([]*models.Job, 0, len(layers))
eg, ctx := errgroup.WithContext(ctx)
eg.SetLimit(DefaultGetImageDistributionJobConcurrentCount)
for _, file := range layers {
eg.Go(func() error {
job, err := s.createGetTaskJobSync(ctx, types.CreateGetTaskJobRequest{
BIO: json.BIO,
Type: internaljob.GetTaskJob,
Args: types.GetTaskArgs{
URL: file.URL,
PieceLength: file.PieceLength,
Tag: file.Tag,
Application: file.Application,
FilteredQueryParams: file.FilteredQueryParams,
},
SchedulerClusterIDs: json.SchedulerClusterIDs,
}, schedulers)
if err != nil {
logger.Warnf("failed to create task job for image layer %s: %w", file.URL, err)
return nil
}
mu.Lock()
jobs = append(jobs, job)
mu.Unlock()
return nil
})
}
// If any of the goroutines return an error, ignore it and continue processing.
if err := eg.Wait(); err != nil {
logger.Errorf("failed to create get task jobs: %w", err)
}
return jobs
}
func (s *service) createGetTaskJobSync(ctx context.Context, json types.CreateGetTaskJobRequest, schedulers []models.Scheduler) (*models.Job, error) {
if json.Args.FilteredQueryParams == "" {
json.Args.FilteredQueryParams = http.RawDefaultFilteredQueryParams
}
args, err := structure.StructToMap(json.Args)
if err != nil {
return nil, err
}
groupJobState, err := s.job.CreateGetTask(ctx, schedulers, json.Args)
if err != nil {
return nil, err
}
var schedulerClusters []models.SchedulerCluster
for _, scheduler := range schedulers {
schedulerClusters = append(schedulerClusters, scheduler.SchedulerCluster)
}
job := models.Job{
TaskID: groupJobState.GroupUUID,
BIO: json.BIO,
Type: json.Type,
State: groupJobState.State,
Args: args,
UserID: json.UserID,
SchedulerClusters: schedulerClusters,
}
if err := s.db.WithContext(ctx).Create(&job).Error; err != nil {
return nil, err
}
s.pollingJob(context.Background(), internaljob.GetTaskJob, job.ID, job.TaskID, 3, 5, 4)
if err := s.db.WithContext(ctx).Preload("SeedPeerClusters").Preload("SchedulerClusters").First(&job, job.ID).Error; err != nil {
return nil, err
}
return &job, nil
}
func (s *service) extractPeersFromJobs(jobs []*models.Job) ([]types.Peer, error) {
m := make(map[string]*types.Peer, len(jobs))
for _, job := range jobs {
if job.State != machineryv1tasks.StateSuccess {
continue
}
jobStates, ok := job.Result["job_states"].([]any)
if !ok {
logger.Warnf("job %s has no job_states in result", job.ID)
continue
}
url, ok := job.Args["url"].(string)
if !ok {
logger.Warnf("job %s has no url in args", job.ID)
continue
}
for _, jobState := range jobStates {
jobState, ok := jobState.(map[string]any)
if !ok {
logger.Warnf("job %s has invalid job_state in result", job.ID)
continue
}
results, ok := jobState["results"].([]any)
if !ok {
logger.Warnf("job %s has no results in job_state", job.ID)
continue
}
for _, result := range results {
result, ok := result.(map[string]any)
if !ok {
logger.Warnf("job %s has invalid result in job_state", job.ID)
continue
}
schedulerClusterID, ok := result["scheduler_cluster_id"].(float64)
if !ok {
logger.Warnf("job %s has no scheduler_cluster_id in result", job.ID)
continue
}
peers, ok := result["peers"].([]any)
if !ok {
logger.Warnf("job %s has no peers in result", job.ID)
continue
}
for _, peer := range peers {
peer, ok := peer.(map[string]any)
if !ok {
logger.Warnf("job %s has invalid peer in result", job.ID)
continue
}
id, ok := peer["id"].(string)
if !ok {
logger.Warnf("job %s has invalid peer id in result", job.ID)
continue
}
hostType, ok := peer["host_type"].(string)
if !ok {
logger.Warnf("job %s has no host_type in result for peer %s", job.ID, id)
continue
}
// Only collect normal peers and skip seed peers.
if hostType != pkgtypes.HostTypeNormalName {
continue
}
ip, ok := peer["ip"].(string)
if !ok {
logger.Warnf("job %s has no ip in result for peer %s", job.ID, id)
continue
}
hostname, ok := peer["hostname"].(string)
if !ok {
logger.Warnf("job %s has no hostname in result for peer %s", job.ID, id)
continue
}
hostID := idgen.HostIDV2(ip, hostname, false)
p, found := m[hostID]
if !found {
m[hostID] = &types.Peer{
IP: ip,
Hostname: hostname,
CachedLayers: []types.Layer{{URL: url}},
SchedulerClusterID: uint(schedulerClusterID),
}
} else {
if slices.Contains(p.CachedLayers, types.Layer{URL: url}) {
continue
}
p.CachedLayers = append(p.CachedLayers, types.Layer{URL: url})
}
}
}
}
}
peers := make([]types.Peer, 0, len(m))
for _, peer := range m {
peers = append(peers, *peer)
}
return peers, nil
}
func (s *service) CreateDeleteTaskJob(ctx context.Context, json types.CreateDeleteTaskJobRequest) (*models.Job, error) {
if json.Args.Timeout == 0 {
json.Args.Timeout = types.DefaultJobTimeout
}
if json.Args.FilteredQueryParams == "" {
json.Args.FilteredQueryParams = http.RawDefaultFilteredQueryParams
}
args, err := structure.StructToMap(json.Args)
if err != nil {
logger.Errorf("convert delete task args to map failed: %w", err)
return nil, err
}
schedulers, err := s.findAllSchedulersInClusters(ctx, json.SchedulerClusterIDs)
if err != nil {
logger.Errorf("find schedulers in clusters failed: %w", err)
return nil, err
}
groupJobState, err := s.job.CreateDeleteTask(ctx, schedulers, json.Args)
if err != nil {
logger.Errorf("create delete task job failed: %w", err)
return nil, err
}
var schedulerClusters []models.SchedulerCluster
for _, scheduler := range schedulers {
schedulerClusters = append(schedulerClusters, scheduler.SchedulerCluster)
}
job := models.Job{
TaskID: groupJobState.GroupUUID,
BIO: json.BIO,
Type: json.Type,
State: groupJobState.State,
Args: args,
UserID: json.UserID,
SchedulerClusters: schedulerClusters,
}
if err := s.db.WithContext(ctx).Create(&job).Error; err != nil {
logger.Errorf("create delete task job failed: %w", err)
return nil, err
}
go s.pollingJob(context.Background(), internaljob.DeleteTaskJob, job.ID, job.TaskID, 30, 300, 16)
return &job, nil
}
func (s *service) findSchedulerInClusters(ctx context.Context, schedulerClusterIDs []uint) ([]models.Scheduler, error) {
var activeSchedulers []models.Scheduler
if len(schedulerClusterIDs) != 0 {
// Find the scheduler clusters by request.
for _, schedulerClusterID := range schedulerClusterIDs {
schedulerCluster := models.SchedulerCluster{}
if err := s.db.WithContext(ctx).First(&schedulerCluster, schedulerClusterID).Error; err != nil {
return nil, fmt.Errorf("scheduler cluster id %d: %w", schedulerClusterID, err)
}
scheduler := models.Scheduler{}
if err := s.db.WithContext(ctx).Preload("SchedulerCluster").First(&scheduler, models.Scheduler{
SchedulerClusterID: schedulerCluster.ID,
State: models.SchedulerStateActive,
}).Error; err != nil {
return nil, fmt.Errorf("scheduler cluster id %d: %w", schedulerClusterID, err)
}
activeSchedulers = append(activeSchedulers, scheduler)
}
} else {
// Find all of the scheduler clusters that has active scheduler.
var schedulerClusters []models.SchedulerCluster
if err := s.db.WithContext(ctx).Find(&schedulerClusters).Error; err != nil {
return nil, err
}
for _, schedulerCluster := range schedulerClusters {
scheduler := models.Scheduler{}
if err := s.db.WithContext(ctx).Preload("SchedulerCluster").First(&scheduler, models.Scheduler{
SchedulerClusterID: schedulerCluster.ID,
State: models.SchedulerStateActive,
}).Error; err != nil {
continue
}
activeSchedulers = append(activeSchedulers, scheduler)
}
}
if len(activeSchedulers) == 0 {
return nil, errors.New("active schedulers not found")
}
return activeSchedulers, nil
}
func (s *service) findAllSchedulersInClusters(ctx context.Context, schedulerClusterIDs []uint) ([]models.Scheduler, error) {
var activeSchedulers []models.Scheduler
if len(schedulerClusterIDs) != 0 {
// Find the scheduler clusters by request.
for _, schedulerClusterID := range schedulerClusterIDs {
schedulerCluster := models.SchedulerCluster{}
if err := s.db.WithContext(ctx).First(&schedulerCluster, schedulerClusterID).Error; err != nil {
return nil, fmt.Errorf("scheduler cluster id %d: %w", schedulerClusterID, err)
}
var schedulers []models.Scheduler
if err := s.db.WithContext(ctx).Preload("SchedulerCluster").Find(&schedulers, models.Scheduler{
SchedulerClusterID: schedulerCluster.ID,
State: models.SchedulerStateActive,
}).Error; err != nil {
return nil, fmt.Errorf("scheduler cluster id %d: %w", schedulerClusterID, err)
}
activeSchedulers = append(activeSchedulers, schedulers...)
}
} else {
// Find all of the scheduler clusters that has active schedulers.
var schedulerClusters []models.SchedulerCluster
if err := s.db.WithContext(ctx).Find(&schedulerClusters).Error; err != nil {
return nil, err
}
for _, schedulerCluster := range schedulerClusters {
var schedulers []models.Scheduler
if err := s.db.WithContext(ctx).Preload("SchedulerCluster").Find(&schedulers, models.Scheduler{
SchedulerClusterID: schedulerCluster.ID,
State: models.SchedulerStateActive,
}).Error; err != nil {
continue
}
activeSchedulers = append(activeSchedulers, schedulers...)
}
}
if len(activeSchedulers) == 0 {
return nil, errors.New("active schedulers not found")
}
return activeSchedulers, nil
}
func (s *service) findAllCandidateSchedulersInClusters(ctx context.Context, schedulerClusterIDs []uint, features []string) ([]models.Scheduler, error) {
var candidateSchedulers []models.Scheduler
if len(schedulerClusterIDs) != 0 {
// Find the scheduler clusters by request.
for _, schedulerClusterID := range schedulerClusterIDs {
schedulerCluster := models.SchedulerCluster{}
if err := s.db.WithContext(ctx).First(&schedulerCluster, schedulerClusterID).Error; err != nil {
return nil, fmt.Errorf("scheduler cluster id %d: %w", schedulerClusterID, err)
}
var schedulers []models.Scheduler
if err := s.db.WithContext(ctx).Preload("SchedulerCluster").Find(&schedulers, models.Scheduler{
SchedulerClusterID: schedulerCluster.ID,
State: models.SchedulerStateActive,
}).Error; err != nil {
return nil, fmt.Errorf("scheduler cluster id %d: %w", schedulerClusterID, err)
}
for _, scheduler := range schedulers {
// If the features is empty, return the first scheduler directly.
if len(features) == 0 {
candidateSchedulers = append(candidateSchedulers, scheduler)
break
}
// Scan the schedulers to find the first scheduler that supports feature.
if slices.Contains(scheduler.Features, features...) {
candidateSchedulers = append(candidateSchedulers, scheduler)
break
}
}
}
} else {
// Find all of the scheduler clusters that has active schedulers.
var candidateSchedulerClusters []models.SchedulerCluster
if err := s.db.WithContext(ctx).Find(&candidateSchedulerClusters).Error; err != nil {
return nil, err
}
for _, candidateSchedulerCluster := range candidateSchedulerClusters {
var schedulers []models.Scheduler
if err := s.db.WithContext(ctx).Preload("SchedulerCluster").Find(&schedulers, models.Scheduler{
SchedulerClusterID: candidateSchedulerCluster.ID,
State: models.SchedulerStateActive,
}).Error; err != nil {
continue
}
for _, scheduler := range schedulers {
// If the features is empty, return the first scheduler directly.
if len(features) == 0 {
candidateSchedulers = append(candidateSchedulers, scheduler)
break
}
// Scan the schedulers to find the first scheduler that supports feature.
if slices.Contains(scheduler.Features, features...) {
candidateSchedulers = append(candidateSchedulers, scheduler)
break
}
}
}
}
if len(candidateSchedulers) == 0 {
return nil, errors.New("candidate schedulers not found")
}
return candidateSchedulers, nil
}
func (s *service) pollingJob(ctx context.Context, name string, id uint, groupUUID string, initBackoff float64, maxBackoff float64, maxAttempts int) {
var (
job models.Job
log = logger.WithGroupAndJobID(groupUUID, fmt.Sprint(id))
)
if _, _, err := retry.Run(ctx, initBackoff, maxBackoff, maxAttempts, func() (any, bool, error) {
groupJob, err := s.job.GetGroupJobState(name, groupUUID)
if err != nil {
err = fmt.Errorf("get group job state failed: %w", err)
log.Error(err)
return nil, false, err
}
result, err := structure.StructToMap(groupJob)
if err != nil {
err = fmt.Errorf("convert group job state to map failed: %w", err)
log.Error(err)
return nil, false, err
}
if err := s.db.WithContext(ctx).First(&job, id).Updates(models.Job{
State: groupJob.State,
Result: result,
}).Error; err != nil {
err = fmt.Errorf("update job state failed: %w", err)
log.Error(err)
return nil, true, err
}
switch job.State {
case machineryv1tasks.StateSuccess:
// Collect CreateJobSuccessCount. metrics.
metrics.CreateJobSuccessCount.WithLabelValues(name).Inc()
log.Info("polling group succeeded")
return nil, true, nil
case machineryv1tasks.StateFailure:
log.Error("polling group failed")
return nil, true, nil
default:
msg := fmt.Sprintf("polling job state is %s", job.State)
log.Info(msg)
return nil, false, errors.New(msg)
}
}); err != nil {
err = fmt.Errorf("polling group job failed: %w", err)
log.Error(err)
}
// Polling timeout and failed.
if job.State != machineryv1tasks.StateSuccess && job.State != machineryv1tasks.StateFailure {
job := models.Job{}
if err := s.db.WithContext(ctx).First(&job, id).Updates(models.Job{
State: machineryv1tasks.StateFailure,
}).Error; err != nil {
err = fmt.Errorf("update job state to failure failed: %w", err)
log.Error(err)
}
log.Error("polling group timeout")
}
}
func (s *service) DestroyJob(ctx context.Context, id uint) error {
job := models.Job{}
if err := s.db.WithContext(ctx).First(&job, id).Error; err != nil {
return err
}
if err := s.db.WithContext(ctx).Unscoped().Delete(&models.Job{}, id).Error; err != nil {
return err
}
return nil
}
func (s *service) UpdateJob(ctx context.Context, id uint, json types.UpdateJobRequest) (*models.Job, error) {
job := models.Job{}
if err := s.db.WithContext(ctx).Preload("SeedPeerClusters").Preload("SchedulerClusters").First(&job, id).Updates(models.Job{
BIO: json.BIO,
UserID: json.UserID,
}).Error; err != nil {
return nil, err
}
return &job, nil
}
func (s *service) GetJob(ctx context.Context, id uint) (*models.Job, error) {
job := models.Job{}
if err := s.db.WithContext(ctx).Preload("SeedPeerClusters").Preload("SchedulerClusters").First(&job, id).Error; err != nil {
return nil, err
}
return &job, nil
}
func (s *service) GetJobs(ctx context.Context, q types.GetJobsQuery) ([]models.Job, int64, error) {
var count int64
var jobs []models.Job
if err := s.db.WithContext(ctx).Scopes(models.Paginate(q.Page, q.PerPage)).Where(&models.Job{
Type: q.Type,
State: q.State,
UserID: q.UserID,
}).Order("created_at DESC").Find(&jobs).Limit(-1).Offset(-1).Count(&count).Error; err != nil {
return nil, 0, err
}
return jobs, count, nil
}
func (s *service) AddJobToSchedulerClusters(ctx context.Context, id, schedulerClusterIDs []uint) error {
job := models.Job{}
if err := s.db.WithContext(ctx).First(&job, id).Error; err != nil {
return err
}
var schedulerClusters []*models.SchedulerCluster
for _, schedulerClusterID := range schedulerClusterIDs {
schedulerCluster := models.SchedulerCluster{}
if err := s.db.WithContext(ctx).First(&schedulerCluster, schedulerClusterID).Error; err != nil {
return err
}
schedulerClusters = append(schedulerClusters, &schedulerCluster)
}
if err := s.db.WithContext(ctx).Model(&job).Association("SchedulerClusters").Append(schedulerClusters); err != nil {
return err
}
return nil
}
func (s *service) AddJobToSeedPeerClusters(ctx context.Context, id, seedPeerClusterIDs []uint) error {
job := models.Job{}
if err := s.db.WithContext(ctx).First(&job, id).Error; err != nil {
return err
}
var seedPeerClusters []*models.SeedPeerCluster
for _, seedPeerClusterID := range seedPeerClusterIDs {
seedPeerCluster := models.SeedPeerCluster{}
if err := s.db.WithContext(ctx).First(&seedPeerCluster, seedPeerClusterID).Error; err != nil {
return err
}
seedPeerClusters = append(seedPeerClusters, &seedPeerCluster)
}
if err := s.db.WithContext(ctx).Model(&job).Association("SeedPeerClusters").Append(seedPeerClusters); err != nil {
return err
}
return nil
}