From 63e466ac0dbff92314856de1790f57001c8a8149 Mon Sep 17 00:00:00 2001 From: lenhattan86 Date: Mon, 7 Jun 2021 15:37:27 -0700 Subject: [PATCH] add restartInstances with sla requirements --- auroraAPI.thrift | 3 + gen-go/apache/aurora/GoUnusedProtection__.go | 2 +- gen-go/apache/aurora/auroraAPI-consts.go | 2 +- gen-go/apache/aurora/auroraAPI.go | 1448 +++++++++++------ .../aurora_admin-remote.go | 703 ++++---- .../aurora_scheduler_manager-remote.go | 501 +++--- .../read_only_scheduler-remote.go | 148 +- realis.go | 14 + 8 files changed, 1689 insertions(+), 1132 deletions(-) diff --git a/auroraAPI.thrift b/auroraAPI.thrift index 3e43f6f..cda37af 100644 --- a/auroraAPI.thrift +++ b/auroraAPI.thrift @@ -1163,6 +1163,9 @@ service AuroraSchedulerManager extends ReadOnlyScheduler { /** Restarts a batch of shards. */ Response restartShards(5: JobKey job, 3: set shardIds) + /** Restarts a batch of shards if sla requirement is satisfied. */ + Response slaRestartShards(5: JobKey job, 3: set shardIds, 2: SlaPolicy slaPolicy) + /** Initiates a kill on tasks. */ Response killTasks(4: JobKey job, 5: set instances, 6: string message) diff --git a/gen-go/apache/aurora/GoUnusedProtection__.go b/gen-go/apache/aurora/GoUnusedProtection__.go index 462b3b4..78d7a37 100644 --- a/gen-go/apache/aurora/GoUnusedProtection__.go +++ b/gen-go/apache/aurora/GoUnusedProtection__.go @@ -1,4 +1,4 @@ -// Code generated by Thrift Compiler (0.14.0). DO NOT EDIT. +// Code generated by Thrift Compiler (0.14.1). DO NOT EDIT. package aurora diff --git a/gen-go/apache/aurora/auroraAPI-consts.go b/gen-go/apache/aurora/auroraAPI-consts.go index d4a319e..30a3f98 100644 --- a/gen-go/apache/aurora/auroraAPI-consts.go +++ b/gen-go/apache/aurora/auroraAPI-consts.go @@ -1,4 +1,4 @@ -// Code generated by Thrift Compiler (0.14.0). DO NOT EDIT. +// Code generated by Thrift Compiler (0.14.1). DO NOT EDIT. package aurora diff --git a/gen-go/apache/aurora/auroraAPI.go b/gen-go/apache/aurora/auroraAPI.go index 914d268..eb8b81c 100644 --- a/gen-go/apache/aurora/auroraAPI.go +++ b/gen-go/apache/aurora/auroraAPI.go @@ -1,4 +1,4 @@ -// Code generated by Thrift Compiler (0.14.0). DO NOT EDIT. +// Code generated by Thrift Compiler (0.14.1). DO NOT EDIT. package aurora @@ -18903,67 +18903,67 @@ func (p *ExplicitReconciliationSettings) String() string { type ReadOnlyScheduler interface { // Returns a summary of the jobs grouped by role. - GetRoleSummary(ctx context.Context) (r *Response, err error) + GetRoleSummary(ctx context.Context) (_r *Response, _err error) // Returns a summary of jobs, optionally only those owned by a specific role. // // Parameters: // - Role - GetJobSummary(ctx context.Context, role string) (r *Response, err error) + GetJobSummary(ctx context.Context, role string) (_r *Response, _err error) // Fetches the status of tasks. // // Parameters: // - Query - GetTasksStatus(ctx context.Context, query *TaskQuery) (r *Response, err error) + GetTasksStatus(ctx context.Context, query *TaskQuery) (_r *Response, _err error) // Same as getTaskStatus but without the TaskConfig.ExecutorConfig data set. // This is an interim solution until we have a better way to query TaskConfigs (AURORA-541). // // Parameters: // - Query - GetTasksWithoutConfigs(ctx context.Context, query *TaskQuery) (r *Response, err error) + GetTasksWithoutConfigs(ctx context.Context, query *TaskQuery) (_r *Response, _err error) // Returns user-friendly reasons (if available) for tasks retained in PENDING state. // // Parameters: // - Query - GetPendingReason(ctx context.Context, query *TaskQuery) (r *Response, err error) + GetPendingReason(ctx context.Context, query *TaskQuery) (_r *Response, _err error) // Fetches the configuration summary of active tasks for the specified job. // // Parameters: // - Job - GetConfigSummary(ctx context.Context, job *JobKey) (r *Response, err error) + GetConfigSummary(ctx context.Context, job *JobKey) (_r *Response, _err error) // Fetches the status of jobs. // ownerRole is optional, in which case all jobs are returned. // // Parameters: // - OwnerRole - GetJobs(ctx context.Context, ownerRole string) (r *Response, err error) + GetJobs(ctx context.Context, ownerRole string) (_r *Response, _err error) // Fetches the quota allocated for a user. // // Parameters: // - OwnerRole - GetQuota(ctx context.Context, ownerRole string) (r *Response, err error) + GetQuota(ctx context.Context, ownerRole string) (_r *Response, _err error) // Populates fields in a job configuration as though it were about to be run. // This can be used to diff a configuration running tasks. // // Parameters: // - Description - PopulateJobConfig(ctx context.Context, description *JobConfiguration) (r *Response, err error) + PopulateJobConfig(ctx context.Context, description *JobConfiguration) (_r *Response, _err error) // Gets job update summaries. // // Parameters: // - JobUpdateQuery - GetJobUpdateSummaries(ctx context.Context, jobUpdateQuery *JobUpdateQuery) (r *Response, err error) + GetJobUpdateSummaries(ctx context.Context, jobUpdateQuery *JobUpdateQuery) (_r *Response, _err error) // Gets job update details. // // Parameters: // - Query - GetJobUpdateDetails(ctx context.Context, query *JobUpdateQuery) (r *Response, err error) + GetJobUpdateDetails(ctx context.Context, query *JobUpdateQuery) (_r *Response, _err error) // Gets the diff between client (desired) and server (current) job states. // // Parameters: // - Request - GetJobUpdateDiff(ctx context.Context, request *JobUpdateRequest) (r *Response, err error) + GetJobUpdateDiff(ctx context.Context, request *JobUpdateRequest) (_r *Response, _err error) // Gets tier configurations. - GetTierConfigs(ctx context.Context) (r *Response, err error) + GetTierConfigs(ctx context.Context) (_r *Response, _err error) } type ReadOnlySchedulerClient struct { @@ -19002,50 +19002,50 @@ func (p *ReadOnlySchedulerClient) SetLastResponseMeta_(meta thrift.ResponseMeta) } // Returns a summary of the jobs grouped by role. -func (p *ReadOnlySchedulerClient) GetRoleSummary(ctx context.Context) (r *Response, err error) { +func (p *ReadOnlySchedulerClient) GetRoleSummary(ctx context.Context) (_r *Response, _err error) { var _args103 ReadOnlySchedulerGetRoleSummaryArgs - var _result104 ReadOnlySchedulerGetRoleSummaryResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getRoleSummary", &_args103, &_result104) - p.SetLastResponseMeta_(meta) - if err != nil { + var _result105 ReadOnlySchedulerGetRoleSummaryResult + var _meta104 thrift.ResponseMeta + _meta104, _err = p.Client_().Call(ctx, "getRoleSummary", &_args103, &_result105) + p.SetLastResponseMeta_(_meta104) + if _err != nil { return } - return _result104.GetSuccess(), nil + return _result105.GetSuccess(), nil } // Returns a summary of jobs, optionally only those owned by a specific role. // // Parameters: // - Role -func (p *ReadOnlySchedulerClient) GetJobSummary(ctx context.Context, role string) (r *Response, err error) { - var _args105 ReadOnlySchedulerGetJobSummaryArgs - _args105.Role = role - var _result106 ReadOnlySchedulerGetJobSummaryResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getJobSummary", &_args105, &_result106) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetJobSummary(ctx context.Context, role string) (_r *Response, _err error) { + var _args106 ReadOnlySchedulerGetJobSummaryArgs + _args106.Role = role + var _result108 ReadOnlySchedulerGetJobSummaryResult + var _meta107 thrift.ResponseMeta + _meta107, _err = p.Client_().Call(ctx, "getJobSummary", &_args106, &_result108) + p.SetLastResponseMeta_(_meta107) + if _err != nil { return } - return _result106.GetSuccess(), nil + return _result108.GetSuccess(), nil } // Fetches the status of tasks. // // Parameters: // - Query -func (p *ReadOnlySchedulerClient) GetTasksStatus(ctx context.Context, query *TaskQuery) (r *Response, err error) { - var _args107 ReadOnlySchedulerGetTasksStatusArgs - _args107.Query = query - var _result108 ReadOnlySchedulerGetTasksStatusResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getTasksStatus", &_args107, &_result108) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetTasksStatus(ctx context.Context, query *TaskQuery) (_r *Response, _err error) { + var _args109 ReadOnlySchedulerGetTasksStatusArgs + _args109.Query = query + var _result111 ReadOnlySchedulerGetTasksStatusResult + var _meta110 thrift.ResponseMeta + _meta110, _err = p.Client_().Call(ctx, "getTasksStatus", &_args109, &_result111) + p.SetLastResponseMeta_(_meta110) + if _err != nil { return } - return _result108.GetSuccess(), nil + return _result111.GetSuccess(), nil } // Same as getTaskStatus but without the TaskConfig.ExecutorConfig data set. @@ -19053,51 +19053,51 @@ func (p *ReadOnlySchedulerClient) GetTasksStatus(ctx context.Context, query *Tas // // Parameters: // - Query -func (p *ReadOnlySchedulerClient) GetTasksWithoutConfigs(ctx context.Context, query *TaskQuery) (r *Response, err error) { - var _args109 ReadOnlySchedulerGetTasksWithoutConfigsArgs - _args109.Query = query - var _result110 ReadOnlySchedulerGetTasksWithoutConfigsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getTasksWithoutConfigs", &_args109, &_result110) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetTasksWithoutConfigs(ctx context.Context, query *TaskQuery) (_r *Response, _err error) { + var _args112 ReadOnlySchedulerGetTasksWithoutConfigsArgs + _args112.Query = query + var _result114 ReadOnlySchedulerGetTasksWithoutConfigsResult + var _meta113 thrift.ResponseMeta + _meta113, _err = p.Client_().Call(ctx, "getTasksWithoutConfigs", &_args112, &_result114) + p.SetLastResponseMeta_(_meta113) + if _err != nil { return } - return _result110.GetSuccess(), nil + return _result114.GetSuccess(), nil } // Returns user-friendly reasons (if available) for tasks retained in PENDING state. // // Parameters: // - Query -func (p *ReadOnlySchedulerClient) GetPendingReason(ctx context.Context, query *TaskQuery) (r *Response, err error) { - var _args111 ReadOnlySchedulerGetPendingReasonArgs - _args111.Query = query - var _result112 ReadOnlySchedulerGetPendingReasonResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getPendingReason", &_args111, &_result112) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetPendingReason(ctx context.Context, query *TaskQuery) (_r *Response, _err error) { + var _args115 ReadOnlySchedulerGetPendingReasonArgs + _args115.Query = query + var _result117 ReadOnlySchedulerGetPendingReasonResult + var _meta116 thrift.ResponseMeta + _meta116, _err = p.Client_().Call(ctx, "getPendingReason", &_args115, &_result117) + p.SetLastResponseMeta_(_meta116) + if _err != nil { return } - return _result112.GetSuccess(), nil + return _result117.GetSuccess(), nil } // Fetches the configuration summary of active tasks for the specified job. // // Parameters: // - Job -func (p *ReadOnlySchedulerClient) GetConfigSummary(ctx context.Context, job *JobKey) (r *Response, err error) { - var _args113 ReadOnlySchedulerGetConfigSummaryArgs - _args113.Job = job - var _result114 ReadOnlySchedulerGetConfigSummaryResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getConfigSummary", &_args113, &_result114) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetConfigSummary(ctx context.Context, job *JobKey) (_r *Response, _err error) { + var _args118 ReadOnlySchedulerGetConfigSummaryArgs + _args118.Job = job + var _result120 ReadOnlySchedulerGetConfigSummaryResult + var _meta119 thrift.ResponseMeta + _meta119, _err = p.Client_().Call(ctx, "getConfigSummary", &_args118, &_result120) + p.SetLastResponseMeta_(_meta119) + if _err != nil { return } - return _result114.GetSuccess(), nil + return _result120.GetSuccess(), nil } // Fetches the status of jobs. @@ -19105,34 +19105,34 @@ func (p *ReadOnlySchedulerClient) GetConfigSummary(ctx context.Context, job *Job // // Parameters: // - OwnerRole -func (p *ReadOnlySchedulerClient) GetJobs(ctx context.Context, ownerRole string) (r *Response, err error) { - var _args115 ReadOnlySchedulerGetJobsArgs - _args115.OwnerRole = ownerRole - var _result116 ReadOnlySchedulerGetJobsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getJobs", &_args115, &_result116) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetJobs(ctx context.Context, ownerRole string) (_r *Response, _err error) { + var _args121 ReadOnlySchedulerGetJobsArgs + _args121.OwnerRole = ownerRole + var _result123 ReadOnlySchedulerGetJobsResult + var _meta122 thrift.ResponseMeta + _meta122, _err = p.Client_().Call(ctx, "getJobs", &_args121, &_result123) + p.SetLastResponseMeta_(_meta122) + if _err != nil { return } - return _result116.GetSuccess(), nil + return _result123.GetSuccess(), nil } // Fetches the quota allocated for a user. // // Parameters: // - OwnerRole -func (p *ReadOnlySchedulerClient) GetQuota(ctx context.Context, ownerRole string) (r *Response, err error) { - var _args117 ReadOnlySchedulerGetQuotaArgs - _args117.OwnerRole = ownerRole - var _result118 ReadOnlySchedulerGetQuotaResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getQuota", &_args117, &_result118) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetQuota(ctx context.Context, ownerRole string) (_r *Response, _err error) { + var _args124 ReadOnlySchedulerGetQuotaArgs + _args124.OwnerRole = ownerRole + var _result126 ReadOnlySchedulerGetQuotaResult + var _meta125 thrift.ResponseMeta + _meta125, _err = p.Client_().Call(ctx, "getQuota", &_args124, &_result126) + p.SetLastResponseMeta_(_meta125) + if _err != nil { return } - return _result118.GetSuccess(), nil + return _result126.GetSuccess(), nil } // Populates fields in a job configuration as though it were about to be run. @@ -19140,81 +19140,81 @@ func (p *ReadOnlySchedulerClient) GetQuota(ctx context.Context, ownerRole string // // Parameters: // - Description -func (p *ReadOnlySchedulerClient) PopulateJobConfig(ctx context.Context, description *JobConfiguration) (r *Response, err error) { - var _args119 ReadOnlySchedulerPopulateJobConfigArgs - _args119.Description = description - var _result120 ReadOnlySchedulerPopulateJobConfigResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "populateJobConfig", &_args119, &_result120) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) PopulateJobConfig(ctx context.Context, description *JobConfiguration) (_r *Response, _err error) { + var _args127 ReadOnlySchedulerPopulateJobConfigArgs + _args127.Description = description + var _result129 ReadOnlySchedulerPopulateJobConfigResult + var _meta128 thrift.ResponseMeta + _meta128, _err = p.Client_().Call(ctx, "populateJobConfig", &_args127, &_result129) + p.SetLastResponseMeta_(_meta128) + if _err != nil { return } - return _result120.GetSuccess(), nil + return _result129.GetSuccess(), nil } // Gets job update summaries. // // Parameters: // - JobUpdateQuery -func (p *ReadOnlySchedulerClient) GetJobUpdateSummaries(ctx context.Context, jobUpdateQuery *JobUpdateQuery) (r *Response, err error) { - var _args121 ReadOnlySchedulerGetJobUpdateSummariesArgs - _args121.JobUpdateQuery = jobUpdateQuery - var _result122 ReadOnlySchedulerGetJobUpdateSummariesResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getJobUpdateSummaries", &_args121, &_result122) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetJobUpdateSummaries(ctx context.Context, jobUpdateQuery *JobUpdateQuery) (_r *Response, _err error) { + var _args130 ReadOnlySchedulerGetJobUpdateSummariesArgs + _args130.JobUpdateQuery = jobUpdateQuery + var _result132 ReadOnlySchedulerGetJobUpdateSummariesResult + var _meta131 thrift.ResponseMeta + _meta131, _err = p.Client_().Call(ctx, "getJobUpdateSummaries", &_args130, &_result132) + p.SetLastResponseMeta_(_meta131) + if _err != nil { return } - return _result122.GetSuccess(), nil + return _result132.GetSuccess(), nil } // Gets job update details. // // Parameters: // - Query -func (p *ReadOnlySchedulerClient) GetJobUpdateDetails(ctx context.Context, query *JobUpdateQuery) (r *Response, err error) { - var _args123 ReadOnlySchedulerGetJobUpdateDetailsArgs - _args123.Query = query - var _result124 ReadOnlySchedulerGetJobUpdateDetailsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getJobUpdateDetails", &_args123, &_result124) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetJobUpdateDetails(ctx context.Context, query *JobUpdateQuery) (_r *Response, _err error) { + var _args133 ReadOnlySchedulerGetJobUpdateDetailsArgs + _args133.Query = query + var _result135 ReadOnlySchedulerGetJobUpdateDetailsResult + var _meta134 thrift.ResponseMeta + _meta134, _err = p.Client_().Call(ctx, "getJobUpdateDetails", &_args133, &_result135) + p.SetLastResponseMeta_(_meta134) + if _err != nil { return } - return _result124.GetSuccess(), nil + return _result135.GetSuccess(), nil } // Gets the diff between client (desired) and server (current) job states. // // Parameters: // - Request -func (p *ReadOnlySchedulerClient) GetJobUpdateDiff(ctx context.Context, request *JobUpdateRequest) (r *Response, err error) { - var _args125 ReadOnlySchedulerGetJobUpdateDiffArgs - _args125.Request = request - var _result126 ReadOnlySchedulerGetJobUpdateDiffResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getJobUpdateDiff", &_args125, &_result126) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetJobUpdateDiff(ctx context.Context, request *JobUpdateRequest) (_r *Response, _err error) { + var _args136 ReadOnlySchedulerGetJobUpdateDiffArgs + _args136.Request = request + var _result138 ReadOnlySchedulerGetJobUpdateDiffResult + var _meta137 thrift.ResponseMeta + _meta137, _err = p.Client_().Call(ctx, "getJobUpdateDiff", &_args136, &_result138) + p.SetLastResponseMeta_(_meta137) + if _err != nil { return } - return _result126.GetSuccess(), nil + return _result138.GetSuccess(), nil } // Gets tier configurations. -func (p *ReadOnlySchedulerClient) GetTierConfigs(ctx context.Context) (r *Response, err error) { - var _args127 ReadOnlySchedulerGetTierConfigsArgs - var _result128 ReadOnlySchedulerGetTierConfigsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "getTierConfigs", &_args127, &_result128) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *ReadOnlySchedulerClient) GetTierConfigs(ctx context.Context) (_r *Response, _err error) { + var _args139 ReadOnlySchedulerGetTierConfigsArgs + var _result141 ReadOnlySchedulerGetTierConfigsResult + var _meta140 thrift.ResponseMeta + _meta140, _err = p.Client_().Call(ctx, "getTierConfigs", &_args139, &_result141) + p.SetLastResponseMeta_(_meta140) + if _err != nil { return } - return _result128.GetSuccess(), nil + return _result141.GetSuccess(), nil } type ReadOnlySchedulerProcessor struct { @@ -19237,21 +19237,21 @@ func (p *ReadOnlySchedulerProcessor) ProcessorMap() map[string]thrift.TProcessor func NewReadOnlySchedulerProcessor(handler ReadOnlyScheduler) *ReadOnlySchedulerProcessor { - self129 := &ReadOnlySchedulerProcessor{handler:handler, processorMap:make(map[string]thrift.TProcessorFunction)} - self129.processorMap["getRoleSummary"] = &readOnlySchedulerProcessorGetRoleSummary{handler:handler} - self129.processorMap["getJobSummary"] = &readOnlySchedulerProcessorGetJobSummary{handler:handler} - self129.processorMap["getTasksStatus"] = &readOnlySchedulerProcessorGetTasksStatus{handler:handler} - self129.processorMap["getTasksWithoutConfigs"] = &readOnlySchedulerProcessorGetTasksWithoutConfigs{handler:handler} - self129.processorMap["getPendingReason"] = &readOnlySchedulerProcessorGetPendingReason{handler:handler} - self129.processorMap["getConfigSummary"] = &readOnlySchedulerProcessorGetConfigSummary{handler:handler} - self129.processorMap["getJobs"] = &readOnlySchedulerProcessorGetJobs{handler:handler} - self129.processorMap["getQuota"] = &readOnlySchedulerProcessorGetQuota{handler:handler} - self129.processorMap["populateJobConfig"] = &readOnlySchedulerProcessorPopulateJobConfig{handler:handler} - self129.processorMap["getJobUpdateSummaries"] = &readOnlySchedulerProcessorGetJobUpdateSummaries{handler:handler} - self129.processorMap["getJobUpdateDetails"] = &readOnlySchedulerProcessorGetJobUpdateDetails{handler:handler} - self129.processorMap["getJobUpdateDiff"] = &readOnlySchedulerProcessorGetJobUpdateDiff{handler:handler} - self129.processorMap["getTierConfigs"] = &readOnlySchedulerProcessorGetTierConfigs{handler:handler} -return self129 + self142 := &ReadOnlySchedulerProcessor{handler:handler, processorMap:make(map[string]thrift.TProcessorFunction)} + self142.processorMap["getRoleSummary"] = &readOnlySchedulerProcessorGetRoleSummary{handler:handler} + self142.processorMap["getJobSummary"] = &readOnlySchedulerProcessorGetJobSummary{handler:handler} + self142.processorMap["getTasksStatus"] = &readOnlySchedulerProcessorGetTasksStatus{handler:handler} + self142.processorMap["getTasksWithoutConfigs"] = &readOnlySchedulerProcessorGetTasksWithoutConfigs{handler:handler} + self142.processorMap["getPendingReason"] = &readOnlySchedulerProcessorGetPendingReason{handler:handler} + self142.processorMap["getConfigSummary"] = &readOnlySchedulerProcessorGetConfigSummary{handler:handler} + self142.processorMap["getJobs"] = &readOnlySchedulerProcessorGetJobs{handler:handler} + self142.processorMap["getQuota"] = &readOnlySchedulerProcessorGetQuota{handler:handler} + self142.processorMap["populateJobConfig"] = &readOnlySchedulerProcessorPopulateJobConfig{handler:handler} + self142.processorMap["getJobUpdateSummaries"] = &readOnlySchedulerProcessorGetJobUpdateSummaries{handler:handler} + self142.processorMap["getJobUpdateDetails"] = &readOnlySchedulerProcessorGetJobUpdateDetails{handler:handler} + self142.processorMap["getJobUpdateDiff"] = &readOnlySchedulerProcessorGetJobUpdateDiff{handler:handler} + self142.processorMap["getTierConfigs"] = &readOnlySchedulerProcessorGetTierConfigs{handler:handler} +return self142 } func (p *ReadOnlySchedulerProcessor) Process(ctx context.Context, iprot, oprot thrift.TProtocol) (success bool, err thrift.TException) { @@ -19262,12 +19262,12 @@ func (p *ReadOnlySchedulerProcessor) Process(ctx context.Context, iprot, oprot t } iprot.Skip(ctx, thrift.STRUCT) iprot.ReadMessageEnd(ctx) - x130 := thrift.NewTApplicationException(thrift.UNKNOWN_METHOD, "Unknown function " + name) + x143 := thrift.NewTApplicationException(thrift.UNKNOWN_METHOD, "Unknown function " + name) oprot.WriteMessageBegin(ctx, name, thrift.EXCEPTION, seqId) - x130.Write(ctx, oprot) + x143.Write(ctx, oprot) oprot.WriteMessageEnd(ctx) oprot.Flush(ctx) - return false, x130 + return false, x143 } @@ -22770,88 +22770,95 @@ ReadOnlyScheduler // // Parameters: // - Description - CreateJob(ctx context.Context, description *JobConfiguration) (r *Response, err error) + CreateJob(ctx context.Context, description *JobConfiguration) (_r *Response, _err error) // Enters a job into the cron schedule, without actually starting the job. // If the job is already present in the schedule, this will update the schedule entry with the new // configuration. // // Parameters: // - Description - ScheduleCronJob(ctx context.Context, description *JobConfiguration) (r *Response, err error) + ScheduleCronJob(ctx context.Context, description *JobConfiguration) (_r *Response, _err error) // Removes a job from the cron schedule. The request will be denied if the job was not previously // scheduled with scheduleCronJob. // // Parameters: // - Job - DescheduleCronJob(ctx context.Context, job *JobKey) (r *Response, err error) + DescheduleCronJob(ctx context.Context, job *JobKey) (_r *Response, _err error) // Starts a cron job immediately. The request will be denied if the specified job does not // exist for the role account, or the job is not a cron job. // // Parameters: // - Job - StartCronJob(ctx context.Context, job *JobKey) (r *Response, err error) + StartCronJob(ctx context.Context, job *JobKey) (_r *Response, _err error) // Restarts a batch of shards. // // Parameters: // - Job // - ShardIds - RestartShards(ctx context.Context, job *JobKey, shardIds []int32) (r *Response, err error) + RestartShards(ctx context.Context, job *JobKey, shardIds []int32) (_r *Response, _err error) + // Restarts a batch of shards if sla requirement is satisfied. + // + // Parameters: + // - Job + // - ShardIds + // - SlaPolicy + SlaRestartShards(ctx context.Context, job *JobKey, shardIds []int32, slaPolicy *SlaPolicy) (_r *Response, _err error) // Initiates a kill on tasks. // // Parameters: // - Job // - Instances // - Message - KillTasks(ctx context.Context, job *JobKey, instances []int32, message string) (r *Response, err error) + KillTasks(ctx context.Context, job *JobKey, instances []int32, message string) (_r *Response, _err error) // Adds new instances with the TaskConfig of the existing instance pointed by the key. // // Parameters: // - Key // - Count - AddInstances(ctx context.Context, key *InstanceKey, count int32) (r *Response, err error) + AddInstances(ctx context.Context, key *InstanceKey, count int32) (_r *Response, _err error) // Replaces the template (configuration) for the existing cron job. // The cron job template (configuration) must exist for the call to succeed. // // Parameters: // - Config - ReplaceCronTemplate(ctx context.Context, config *JobConfiguration) (r *Response, err error) + ReplaceCronTemplate(ctx context.Context, config *JobConfiguration) (_r *Response, _err error) // Starts update of the existing service job. // // Parameters: // - Request: A description of how to change the job. // - Message: A user-specified message to include with the induced job update state change. - StartJobUpdate(ctx context.Context, request *JobUpdateRequest, message string) (r *Response, err error) + StartJobUpdate(ctx context.Context, request *JobUpdateRequest, message string) (_r *Response, _err error) // Pauses the specified job update. Can be resumed by resumeUpdate call. // // Parameters: // - Key: The update to pause. // - Message: A user-specified message to include with the induced job update state change. - PauseJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) + PauseJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) // Resumes progress of a previously paused job update. // // Parameters: // - Key: The update to resume. // - Message: A user-specified message to include with the induced job update state change. - ResumeJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) + ResumeJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) // Permanently aborts the job update. Does not remove the update history. // // Parameters: // - Key: The update to abort. // - Message: A user-specified message to include with the induced job update state change. - AbortJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) + AbortJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) // Rollbacks the specified active job update to the initial state. // // Parameters: // - Key: The update to rollback. // - Message: A user-specified message to include with the induced job update state change. - RollbackJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) + RollbackJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) // Allows progress of the job update in case blockIfNoPulsesAfterMs is specified in // JobUpdateSettings. Unblocks progress if the update was previously blocked. // Responds with ResponseCode.INVALID_REQUEST in case an unknown update key is specified. // // Parameters: // - Key - PulseJobUpdate(ctx context.Context, key *JobUpdateKey) (r *Response, err error) + PulseJobUpdate(ctx context.Context, key *JobUpdateKey) (_r *Response, _err error) } type AuroraSchedulerManagerClient struct { @@ -22876,17 +22883,17 @@ func NewAuroraSchedulerManagerClient(c thrift.TClient) *AuroraSchedulerManagerCl // // Parameters: // - Description -func (p *AuroraSchedulerManagerClient) CreateJob(ctx context.Context, description *JobConfiguration) (r *Response, err error) { - var _args182 AuroraSchedulerManagerCreateJobArgs - _args182.Description = description - var _result183 AuroraSchedulerManagerCreateJobResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "createJob", &_args182, &_result183) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) CreateJob(ctx context.Context, description *JobConfiguration) (_r *Response, _err error) { + var _args195 AuroraSchedulerManagerCreateJobArgs + _args195.Description = description + var _result197 AuroraSchedulerManagerCreateJobResult + var _meta196 thrift.ResponseMeta + _meta196, _err = p.Client_().Call(ctx, "createJob", &_args195, &_result197) + p.SetLastResponseMeta_(_meta196) + if _err != nil { return } - return _result183.GetSuccess(), nil + return _result197.GetSuccess(), nil } // Enters a job into the cron schedule, without actually starting the job. @@ -22895,17 +22902,17 @@ func (p *AuroraSchedulerManagerClient) CreateJob(ctx context.Context, descriptio // // Parameters: // - Description -func (p *AuroraSchedulerManagerClient) ScheduleCronJob(ctx context.Context, description *JobConfiguration) (r *Response, err error) { - var _args184 AuroraSchedulerManagerScheduleCronJobArgs - _args184.Description = description - var _result185 AuroraSchedulerManagerScheduleCronJobResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "scheduleCronJob", &_args184, &_result185) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) ScheduleCronJob(ctx context.Context, description *JobConfiguration) (_r *Response, _err error) { + var _args198 AuroraSchedulerManagerScheduleCronJobArgs + _args198.Description = description + var _result200 AuroraSchedulerManagerScheduleCronJobResult + var _meta199 thrift.ResponseMeta + _meta199, _err = p.Client_().Call(ctx, "scheduleCronJob", &_args198, &_result200) + p.SetLastResponseMeta_(_meta199) + if _err != nil { return } - return _result185.GetSuccess(), nil + return _result200.GetSuccess(), nil } // Removes a job from the cron schedule. The request will be denied if the job was not previously @@ -22913,17 +22920,17 @@ func (p *AuroraSchedulerManagerClient) ScheduleCronJob(ctx context.Context, desc // // Parameters: // - Job -func (p *AuroraSchedulerManagerClient) DescheduleCronJob(ctx context.Context, job *JobKey) (r *Response, err error) { - var _args186 AuroraSchedulerManagerDescheduleCronJobArgs - _args186.Job = job - var _result187 AuroraSchedulerManagerDescheduleCronJobResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "descheduleCronJob", &_args186, &_result187) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) DescheduleCronJob(ctx context.Context, job *JobKey) (_r *Response, _err error) { + var _args201 AuroraSchedulerManagerDescheduleCronJobArgs + _args201.Job = job + var _result203 AuroraSchedulerManagerDescheduleCronJobResult + var _meta202 thrift.ResponseMeta + _meta202, _err = p.Client_().Call(ctx, "descheduleCronJob", &_args201, &_result203) + p.SetLastResponseMeta_(_meta202) + if _err != nil { return } - return _result187.GetSuccess(), nil + return _result203.GetSuccess(), nil } // Starts a cron job immediately. The request will be denied if the specified job does not @@ -22931,17 +22938,17 @@ func (p *AuroraSchedulerManagerClient) DescheduleCronJob(ctx context.Context, jo // // Parameters: // - Job -func (p *AuroraSchedulerManagerClient) StartCronJob(ctx context.Context, job *JobKey) (r *Response, err error) { - var _args188 AuroraSchedulerManagerStartCronJobArgs - _args188.Job = job - var _result189 AuroraSchedulerManagerStartCronJobResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "startCronJob", &_args188, &_result189) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) StartCronJob(ctx context.Context, job *JobKey) (_r *Response, _err error) { + var _args204 AuroraSchedulerManagerStartCronJobArgs + _args204.Job = job + var _result206 AuroraSchedulerManagerStartCronJobResult + var _meta205 thrift.ResponseMeta + _meta205, _err = p.Client_().Call(ctx, "startCronJob", &_args204, &_result206) + p.SetLastResponseMeta_(_meta205) + if _err != nil { return } - return _result189.GetSuccess(), nil + return _result206.GetSuccess(), nil } // Restarts a batch of shards. @@ -22949,18 +22956,39 @@ func (p *AuroraSchedulerManagerClient) StartCronJob(ctx context.Context, job *Jo // Parameters: // - Job // - ShardIds -func (p *AuroraSchedulerManagerClient) RestartShards(ctx context.Context, job *JobKey, shardIds []int32) (r *Response, err error) { - var _args190 AuroraSchedulerManagerRestartShardsArgs - _args190.Job = job - _args190.ShardIds = shardIds - var _result191 AuroraSchedulerManagerRestartShardsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "restartShards", &_args190, &_result191) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) RestartShards(ctx context.Context, job *JobKey, shardIds []int32) (_r *Response, _err error) { + var _args207 AuroraSchedulerManagerRestartShardsArgs + _args207.Job = job + _args207.ShardIds = shardIds + var _result209 AuroraSchedulerManagerRestartShardsResult + var _meta208 thrift.ResponseMeta + _meta208, _err = p.Client_().Call(ctx, "restartShards", &_args207, &_result209) + p.SetLastResponseMeta_(_meta208) + if _err != nil { return } - return _result191.GetSuccess(), nil + return _result209.GetSuccess(), nil +} + +// Restarts a batch of shards if sla requirement is satisfied. +// +// Parameters: +// - Job +// - ShardIds +// - SlaPolicy +func (p *AuroraSchedulerManagerClient) SlaRestartShards(ctx context.Context, job *JobKey, shardIds []int32, slaPolicy *SlaPolicy) (_r *Response, _err error) { + var _args210 AuroraSchedulerManagerSlaRestartShardsArgs + _args210.Job = job + _args210.ShardIds = shardIds + _args210.SlaPolicy = slaPolicy + var _result212 AuroraSchedulerManagerSlaRestartShardsResult + var _meta211 thrift.ResponseMeta + _meta211, _err = p.Client_().Call(ctx, "slaRestartShards", &_args210, &_result212) + p.SetLastResponseMeta_(_meta211) + if _err != nil { + return + } + return _result212.GetSuccess(), nil } // Initiates a kill on tasks. @@ -22969,19 +22997,19 @@ func (p *AuroraSchedulerManagerClient) RestartShards(ctx context.Context, job *J // - Job // - Instances // - Message -func (p *AuroraSchedulerManagerClient) KillTasks(ctx context.Context, job *JobKey, instances []int32, message string) (r *Response, err error) { - var _args192 AuroraSchedulerManagerKillTasksArgs - _args192.Job = job - _args192.Instances = instances - _args192.Message = message - var _result193 AuroraSchedulerManagerKillTasksResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "killTasks", &_args192, &_result193) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) KillTasks(ctx context.Context, job *JobKey, instances []int32, message string) (_r *Response, _err error) { + var _args213 AuroraSchedulerManagerKillTasksArgs + _args213.Job = job + _args213.Instances = instances + _args213.Message = message + var _result215 AuroraSchedulerManagerKillTasksResult + var _meta214 thrift.ResponseMeta + _meta214, _err = p.Client_().Call(ctx, "killTasks", &_args213, &_result215) + p.SetLastResponseMeta_(_meta214) + if _err != nil { return } - return _result193.GetSuccess(), nil + return _result215.GetSuccess(), nil } // Adds new instances with the TaskConfig of the existing instance pointed by the key. @@ -22989,18 +23017,18 @@ func (p *AuroraSchedulerManagerClient) KillTasks(ctx context.Context, job *JobKe // Parameters: // - Key // - Count -func (p *AuroraSchedulerManagerClient) AddInstances(ctx context.Context, key *InstanceKey, count int32) (r *Response, err error) { - var _args194 AuroraSchedulerManagerAddInstancesArgs - _args194.Key = key - _args194.Count = count - var _result195 AuroraSchedulerManagerAddInstancesResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "addInstances", &_args194, &_result195) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) AddInstances(ctx context.Context, key *InstanceKey, count int32) (_r *Response, _err error) { + var _args216 AuroraSchedulerManagerAddInstancesArgs + _args216.Key = key + _args216.Count = count + var _result218 AuroraSchedulerManagerAddInstancesResult + var _meta217 thrift.ResponseMeta + _meta217, _err = p.Client_().Call(ctx, "addInstances", &_args216, &_result218) + p.SetLastResponseMeta_(_meta217) + if _err != nil { return } - return _result195.GetSuccess(), nil + return _result218.GetSuccess(), nil } // Replaces the template (configuration) for the existing cron job. @@ -23008,17 +23036,17 @@ func (p *AuroraSchedulerManagerClient) AddInstances(ctx context.Context, key *In // // Parameters: // - Config -func (p *AuroraSchedulerManagerClient) ReplaceCronTemplate(ctx context.Context, config *JobConfiguration) (r *Response, err error) { - var _args196 AuroraSchedulerManagerReplaceCronTemplateArgs - _args196.Config = config - var _result197 AuroraSchedulerManagerReplaceCronTemplateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "replaceCronTemplate", &_args196, &_result197) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) ReplaceCronTemplate(ctx context.Context, config *JobConfiguration) (_r *Response, _err error) { + var _args219 AuroraSchedulerManagerReplaceCronTemplateArgs + _args219.Config = config + var _result221 AuroraSchedulerManagerReplaceCronTemplateResult + var _meta220 thrift.ResponseMeta + _meta220, _err = p.Client_().Call(ctx, "replaceCronTemplate", &_args219, &_result221) + p.SetLastResponseMeta_(_meta220) + if _err != nil { return } - return _result197.GetSuccess(), nil + return _result221.GetSuccess(), nil } // Starts update of the existing service job. @@ -23026,18 +23054,18 @@ func (p *AuroraSchedulerManagerClient) ReplaceCronTemplate(ctx context.Context, // Parameters: // - Request: A description of how to change the job. // - Message: A user-specified message to include with the induced job update state change. -func (p *AuroraSchedulerManagerClient) StartJobUpdate(ctx context.Context, request *JobUpdateRequest, message string) (r *Response, err error) { - var _args198 AuroraSchedulerManagerStartJobUpdateArgs - _args198.Request = request - _args198.Message = message - var _result199 AuroraSchedulerManagerStartJobUpdateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "startJobUpdate", &_args198, &_result199) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) StartJobUpdate(ctx context.Context, request *JobUpdateRequest, message string) (_r *Response, _err error) { + var _args222 AuroraSchedulerManagerStartJobUpdateArgs + _args222.Request = request + _args222.Message = message + var _result224 AuroraSchedulerManagerStartJobUpdateResult + var _meta223 thrift.ResponseMeta + _meta223, _err = p.Client_().Call(ctx, "startJobUpdate", &_args222, &_result224) + p.SetLastResponseMeta_(_meta223) + if _err != nil { return } - return _result199.GetSuccess(), nil + return _result224.GetSuccess(), nil } // Pauses the specified job update. Can be resumed by resumeUpdate call. @@ -23045,18 +23073,18 @@ func (p *AuroraSchedulerManagerClient) StartJobUpdate(ctx context.Context, reque // Parameters: // - Key: The update to pause. // - Message: A user-specified message to include with the induced job update state change. -func (p *AuroraSchedulerManagerClient) PauseJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) { - var _args200 AuroraSchedulerManagerPauseJobUpdateArgs - _args200.Key = key - _args200.Message = message - var _result201 AuroraSchedulerManagerPauseJobUpdateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "pauseJobUpdate", &_args200, &_result201) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) PauseJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) { + var _args225 AuroraSchedulerManagerPauseJobUpdateArgs + _args225.Key = key + _args225.Message = message + var _result227 AuroraSchedulerManagerPauseJobUpdateResult + var _meta226 thrift.ResponseMeta + _meta226, _err = p.Client_().Call(ctx, "pauseJobUpdate", &_args225, &_result227) + p.SetLastResponseMeta_(_meta226) + if _err != nil { return } - return _result201.GetSuccess(), nil + return _result227.GetSuccess(), nil } // Resumes progress of a previously paused job update. @@ -23064,18 +23092,18 @@ func (p *AuroraSchedulerManagerClient) PauseJobUpdate(ctx context.Context, key * // Parameters: // - Key: The update to resume. // - Message: A user-specified message to include with the induced job update state change. -func (p *AuroraSchedulerManagerClient) ResumeJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) { - var _args202 AuroraSchedulerManagerResumeJobUpdateArgs - _args202.Key = key - _args202.Message = message - var _result203 AuroraSchedulerManagerResumeJobUpdateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "resumeJobUpdate", &_args202, &_result203) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) ResumeJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) { + var _args228 AuroraSchedulerManagerResumeJobUpdateArgs + _args228.Key = key + _args228.Message = message + var _result230 AuroraSchedulerManagerResumeJobUpdateResult + var _meta229 thrift.ResponseMeta + _meta229, _err = p.Client_().Call(ctx, "resumeJobUpdate", &_args228, &_result230) + p.SetLastResponseMeta_(_meta229) + if _err != nil { return } - return _result203.GetSuccess(), nil + return _result230.GetSuccess(), nil } // Permanently aborts the job update. Does not remove the update history. @@ -23083,18 +23111,18 @@ func (p *AuroraSchedulerManagerClient) ResumeJobUpdate(ctx context.Context, key // Parameters: // - Key: The update to abort. // - Message: A user-specified message to include with the induced job update state change. -func (p *AuroraSchedulerManagerClient) AbortJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) { - var _args204 AuroraSchedulerManagerAbortJobUpdateArgs - _args204.Key = key - _args204.Message = message - var _result205 AuroraSchedulerManagerAbortJobUpdateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "abortJobUpdate", &_args204, &_result205) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) AbortJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) { + var _args231 AuroraSchedulerManagerAbortJobUpdateArgs + _args231.Key = key + _args231.Message = message + var _result233 AuroraSchedulerManagerAbortJobUpdateResult + var _meta232 thrift.ResponseMeta + _meta232, _err = p.Client_().Call(ctx, "abortJobUpdate", &_args231, &_result233) + p.SetLastResponseMeta_(_meta232) + if _err != nil { return } - return _result205.GetSuccess(), nil + return _result233.GetSuccess(), nil } // Rollbacks the specified active job update to the initial state. @@ -23102,18 +23130,18 @@ func (p *AuroraSchedulerManagerClient) AbortJobUpdate(ctx context.Context, key * // Parameters: // - Key: The update to rollback. // - Message: A user-specified message to include with the induced job update state change. -func (p *AuroraSchedulerManagerClient) RollbackJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (r *Response, err error) { - var _args206 AuroraSchedulerManagerRollbackJobUpdateArgs - _args206.Key = key - _args206.Message = message - var _result207 AuroraSchedulerManagerRollbackJobUpdateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "rollbackJobUpdate", &_args206, &_result207) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) RollbackJobUpdate(ctx context.Context, key *JobUpdateKey, message string) (_r *Response, _err error) { + var _args234 AuroraSchedulerManagerRollbackJobUpdateArgs + _args234.Key = key + _args234.Message = message + var _result236 AuroraSchedulerManagerRollbackJobUpdateResult + var _meta235 thrift.ResponseMeta + _meta235, _err = p.Client_().Call(ctx, "rollbackJobUpdate", &_args234, &_result236) + p.SetLastResponseMeta_(_meta235) + if _err != nil { return } - return _result207.GetSuccess(), nil + return _result236.GetSuccess(), nil } // Allows progress of the job update in case blockIfNoPulsesAfterMs is specified in @@ -23122,17 +23150,17 @@ func (p *AuroraSchedulerManagerClient) RollbackJobUpdate(ctx context.Context, ke // // Parameters: // - Key -func (p *AuroraSchedulerManagerClient) PulseJobUpdate(ctx context.Context, key *JobUpdateKey) (r *Response, err error) { - var _args208 AuroraSchedulerManagerPulseJobUpdateArgs - _args208.Key = key - var _result209 AuroraSchedulerManagerPulseJobUpdateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "pulseJobUpdate", &_args208, &_result209) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraSchedulerManagerClient) PulseJobUpdate(ctx context.Context, key *JobUpdateKey) (_r *Response, _err error) { + var _args237 AuroraSchedulerManagerPulseJobUpdateArgs + _args237.Key = key + var _result239 AuroraSchedulerManagerPulseJobUpdateResult + var _meta238 thrift.ResponseMeta + _meta238, _err = p.Client_().Call(ctx, "pulseJobUpdate", &_args237, &_result239) + p.SetLastResponseMeta_(_meta238) + if _err != nil { return } - return _result209.GetSuccess(), nil + return _result239.GetSuccess(), nil } type AuroraSchedulerManagerProcessor struct { @@ -23140,22 +23168,23 @@ type AuroraSchedulerManagerProcessor struct { } func NewAuroraSchedulerManagerProcessor(handler AuroraSchedulerManager) *AuroraSchedulerManagerProcessor { - self210 := &AuroraSchedulerManagerProcessor{NewReadOnlySchedulerProcessor(handler)} - self210.AddToProcessorMap("createJob", &auroraSchedulerManagerProcessorCreateJob{handler:handler}) - self210.AddToProcessorMap("scheduleCronJob", &auroraSchedulerManagerProcessorScheduleCronJob{handler:handler}) - self210.AddToProcessorMap("descheduleCronJob", &auroraSchedulerManagerProcessorDescheduleCronJob{handler:handler}) - self210.AddToProcessorMap("startCronJob", &auroraSchedulerManagerProcessorStartCronJob{handler:handler}) - self210.AddToProcessorMap("restartShards", &auroraSchedulerManagerProcessorRestartShards{handler:handler}) - self210.AddToProcessorMap("killTasks", &auroraSchedulerManagerProcessorKillTasks{handler:handler}) - self210.AddToProcessorMap("addInstances", &auroraSchedulerManagerProcessorAddInstances{handler:handler}) - self210.AddToProcessorMap("replaceCronTemplate", &auroraSchedulerManagerProcessorReplaceCronTemplate{handler:handler}) - self210.AddToProcessorMap("startJobUpdate", &auroraSchedulerManagerProcessorStartJobUpdate{handler:handler}) - self210.AddToProcessorMap("pauseJobUpdate", &auroraSchedulerManagerProcessorPauseJobUpdate{handler:handler}) - self210.AddToProcessorMap("resumeJobUpdate", &auroraSchedulerManagerProcessorResumeJobUpdate{handler:handler}) - self210.AddToProcessorMap("abortJobUpdate", &auroraSchedulerManagerProcessorAbortJobUpdate{handler:handler}) - self210.AddToProcessorMap("rollbackJobUpdate", &auroraSchedulerManagerProcessorRollbackJobUpdate{handler:handler}) - self210.AddToProcessorMap("pulseJobUpdate", &auroraSchedulerManagerProcessorPulseJobUpdate{handler:handler}) - return self210 + self240 := &AuroraSchedulerManagerProcessor{NewReadOnlySchedulerProcessor(handler)} + self240.AddToProcessorMap("createJob", &auroraSchedulerManagerProcessorCreateJob{handler:handler}) + self240.AddToProcessorMap("scheduleCronJob", &auroraSchedulerManagerProcessorScheduleCronJob{handler:handler}) + self240.AddToProcessorMap("descheduleCronJob", &auroraSchedulerManagerProcessorDescheduleCronJob{handler:handler}) + self240.AddToProcessorMap("startCronJob", &auroraSchedulerManagerProcessorStartCronJob{handler:handler}) + self240.AddToProcessorMap("restartShards", &auroraSchedulerManagerProcessorRestartShards{handler:handler}) + self240.AddToProcessorMap("slaRestartShards", &auroraSchedulerManagerProcessorSlaRestartShards{handler:handler}) + self240.AddToProcessorMap("killTasks", &auroraSchedulerManagerProcessorKillTasks{handler:handler}) + self240.AddToProcessorMap("addInstances", &auroraSchedulerManagerProcessorAddInstances{handler:handler}) + self240.AddToProcessorMap("replaceCronTemplate", &auroraSchedulerManagerProcessorReplaceCronTemplate{handler:handler}) + self240.AddToProcessorMap("startJobUpdate", &auroraSchedulerManagerProcessorStartJobUpdate{handler:handler}) + self240.AddToProcessorMap("pauseJobUpdate", &auroraSchedulerManagerProcessorPauseJobUpdate{handler:handler}) + self240.AddToProcessorMap("resumeJobUpdate", &auroraSchedulerManagerProcessorResumeJobUpdate{handler:handler}) + self240.AddToProcessorMap("abortJobUpdate", &auroraSchedulerManagerProcessorAbortJobUpdate{handler:handler}) + self240.AddToProcessorMap("rollbackJobUpdate", &auroraSchedulerManagerProcessorRollbackJobUpdate{handler:handler}) + self240.AddToProcessorMap("pulseJobUpdate", &auroraSchedulerManagerProcessorPulseJobUpdate{handler:handler}) + return self240 } type auroraSchedulerManagerProcessorCreateJob struct { @@ -23553,6 +23582,85 @@ func (p *auroraSchedulerManagerProcessorRestartShards) Process(ctx context.Conte return true, err } +type auroraSchedulerManagerProcessorSlaRestartShards struct { + handler AuroraSchedulerManager +} + +func (p *auroraSchedulerManagerProcessorSlaRestartShards) Process(ctx context.Context, seqId int32, iprot, oprot thrift.TProtocol) (success bool, err thrift.TException) { + args := AuroraSchedulerManagerSlaRestartShardsArgs{} + var err2 error + if err2 = args.Read(ctx, iprot); err2 != nil { + iprot.ReadMessageEnd(ctx) + x := thrift.NewTApplicationException(thrift.PROTOCOL_ERROR, err2.Error()) + oprot.WriteMessageBegin(ctx, "slaRestartShards", thrift.EXCEPTION, seqId) + x.Write(ctx, oprot) + oprot.WriteMessageEnd(ctx) + oprot.Flush(ctx) + return false, thrift.WrapTException(err2) + } + iprot.ReadMessageEnd(ctx) + + tickerCancel := func() {} + // Start a goroutine to do server side connectivity check. + if thrift.ServerConnectivityCheckInterval > 0 { + var cancel context.CancelFunc + ctx, cancel = context.WithCancel(ctx) + defer cancel() + var tickerCtx context.Context + tickerCtx, tickerCancel = context.WithCancel(context.Background()) + defer tickerCancel() + go func(ctx context.Context, cancel context.CancelFunc) { + ticker := time.NewTicker(thrift.ServerConnectivityCheckInterval) + defer ticker.Stop() + for { + select { + case <-ctx.Done(): + return + case <-ticker.C: + if !iprot.Transport().IsOpen() { + cancel() + return + } + } + } + }(tickerCtx, cancel) + } + + result := AuroraSchedulerManagerSlaRestartShardsResult{} + var retval *Response + if retval, err2 = p.handler.SlaRestartShards(ctx, args.Job, args.ShardIds, args.SlaPolicy); err2 != nil { + tickerCancel() + if err2 == thrift.ErrAbandonRequest { + return false, thrift.WrapTException(err2) + } + x := thrift.NewTApplicationException(thrift.INTERNAL_ERROR, "Internal error processing slaRestartShards: " + err2.Error()) + oprot.WriteMessageBegin(ctx, "slaRestartShards", thrift.EXCEPTION, seqId) + x.Write(ctx, oprot) + oprot.WriteMessageEnd(ctx) + oprot.Flush(ctx) + return true, thrift.WrapTException(err2) + } else { + result.Success = retval + } + tickerCancel() + if err2 = oprot.WriteMessageBegin(ctx, "slaRestartShards", thrift.REPLY, seqId); err2 != nil { + err = thrift.WrapTException(err2) + } + if err2 = result.Write(ctx, oprot); err == nil && err2 != nil { + err = thrift.WrapTException(err2) + } + if err2 = oprot.WriteMessageEnd(ctx); err == nil && err2 != nil { + err = thrift.WrapTException(err2) + } + if err2 = oprot.Flush(ctx); err == nil && err2 != nil { + err = thrift.WrapTException(err2) + } + if err != nil { + return + } + return true, err +} + type auroraSchedulerManagerProcessorKillTasks struct { handler AuroraSchedulerManager } @@ -25154,13 +25262,13 @@ func (p *AuroraSchedulerManagerRestartShardsArgs) ReadField3(ctx context.Contex tSet := make([]int32, 0, size) p.ShardIds = tSet for i := 0; i < size; i ++ { -var _elem211 int32 +var _elem241 int32 if v, err := iprot.ReadI32(ctx); err != nil { return thrift.PrependError("error reading field 0: ", err) } else { - _elem211 = v + _elem241 = v } - p.ShardIds = append(p.ShardIds, _elem211) + p.ShardIds = append(p.ShardIds, _elem241) } if err := iprot.ReadSetEnd(ctx); err != nil { return thrift.PrependError("error reading set end: ", err) @@ -25328,6 +25436,316 @@ func (p *AuroraSchedulerManagerRestartShardsResult) String() string { return fmt.Sprintf("AuroraSchedulerManagerRestartShardsResult(%+v)", *p) } +// Attributes: +// - Job +// - ShardIds +// - SlaPolicy +type AuroraSchedulerManagerSlaRestartShardsArgs struct { + // unused field # 1 + SlaPolicy *SlaPolicy `thrift:"slaPolicy,2" db:"slaPolicy" json:"slaPolicy"` + ShardIds []int32 `thrift:"shardIds,3" db:"shardIds" json:"shardIds"` + // unused field # 4 + Job *JobKey `thrift:"job,5" db:"job" json:"job"` +} + +func NewAuroraSchedulerManagerSlaRestartShardsArgs() *AuroraSchedulerManagerSlaRestartShardsArgs { + return &AuroraSchedulerManagerSlaRestartShardsArgs{} +} + +var AuroraSchedulerManagerSlaRestartShardsArgs_Job_DEFAULT *JobKey +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) GetJob() *JobKey { + if !p.IsSetJob() { + return AuroraSchedulerManagerSlaRestartShardsArgs_Job_DEFAULT + } +return p.Job +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) GetShardIds() []int32 { + return p.ShardIds +} +var AuroraSchedulerManagerSlaRestartShardsArgs_SlaPolicy_DEFAULT *SlaPolicy +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) GetSlaPolicy() *SlaPolicy { + if !p.IsSetSlaPolicy() { + return AuroraSchedulerManagerSlaRestartShardsArgs_SlaPolicy_DEFAULT + } +return p.SlaPolicy +} +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) IsSetJob() bool { + return p.Job != nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) IsSetSlaPolicy() bool { + return p.SlaPolicy != nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) Read(ctx context.Context, iprot thrift.TProtocol) error { + if _, err := iprot.ReadStructBegin(ctx); err != nil { + return thrift.PrependError(fmt.Sprintf("%T read error: ", p), err) + } + + + for { + _, fieldTypeId, fieldId, err := iprot.ReadFieldBegin(ctx) + if err != nil { + return thrift.PrependError(fmt.Sprintf("%T field %d read error: ", p, fieldId), err) + } + if fieldTypeId == thrift.STOP { break; } + switch fieldId { + case 5: + if fieldTypeId == thrift.STRUCT { + if err := p.ReadField5(ctx, iprot); err != nil { + return err + } + } else { + if err := iprot.Skip(ctx, fieldTypeId); err != nil { + return err + } + } + case 3: + if fieldTypeId == thrift.SET { + if err := p.ReadField3(ctx, iprot); err != nil { + return err + } + } else { + if err := iprot.Skip(ctx, fieldTypeId); err != nil { + return err + } + } + case 2: + if fieldTypeId == thrift.STRUCT { + if err := p.ReadField2(ctx, iprot); err != nil { + return err + } + } else { + if err := iprot.Skip(ctx, fieldTypeId); err != nil { + return err + } + } + default: + if err := iprot.Skip(ctx, fieldTypeId); err != nil { + return err + } + } + if err := iprot.ReadFieldEnd(ctx); err != nil { + return err + } + } + if err := iprot.ReadStructEnd(ctx); err != nil { + return thrift.PrependError(fmt.Sprintf("%T read struct end error: ", p), err) + } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) ReadField5(ctx context.Context, iprot thrift.TProtocol) error { + p.Job = &JobKey{} + if err := p.Job.Read(ctx, iprot); err != nil { + return thrift.PrependError(fmt.Sprintf("%T error reading struct: ", p.Job), err) + } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) ReadField3(ctx context.Context, iprot thrift.TProtocol) error { + _, size, err := iprot.ReadSetBegin(ctx) + if err != nil { + return thrift.PrependError("error reading set begin: ", err) + } + tSet := make([]int32, 0, size) + p.ShardIds = tSet + for i := 0; i < size; i ++ { +var _elem242 int32 + if v, err := iprot.ReadI32(ctx); err != nil { + return thrift.PrependError("error reading field 0: ", err) +} else { + _elem242 = v +} + p.ShardIds = append(p.ShardIds, _elem242) + } + if err := iprot.ReadSetEnd(ctx); err != nil { + return thrift.PrependError("error reading set end: ", err) + } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) ReadField2(ctx context.Context, iprot thrift.TProtocol) error { + p.SlaPolicy = &SlaPolicy{} + if err := p.SlaPolicy.Read(ctx, iprot); err != nil { + return thrift.PrependError(fmt.Sprintf("%T error reading struct: ", p.SlaPolicy), err) + } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) Write(ctx context.Context, oprot thrift.TProtocol) error { + if err := oprot.WriteStructBegin(ctx, "slaRestartShards_args"); err != nil { + return thrift.PrependError(fmt.Sprintf("%T write struct begin error: ", p), err) } + if p != nil { + if err := p.writeField2(ctx, oprot); err != nil { return err } + if err := p.writeField3(ctx, oprot); err != nil { return err } + if err := p.writeField5(ctx, oprot); err != nil { return err } + } + if err := oprot.WriteFieldStop(ctx); err != nil { + return thrift.PrependError("write field stop error: ", err) } + if err := oprot.WriteStructEnd(ctx); err != nil { + return thrift.PrependError("write struct stop error: ", err) } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) writeField2(ctx context.Context, oprot thrift.TProtocol) (err error) { + if err := oprot.WriteFieldBegin(ctx, "slaPolicy", thrift.STRUCT, 2); err != nil { + return thrift.PrependError(fmt.Sprintf("%T write field begin error 2:slaPolicy: ", p), err) } + if err := p.SlaPolicy.Write(ctx, oprot); err != nil { + return thrift.PrependError(fmt.Sprintf("%T error writing struct: ", p.SlaPolicy), err) + } + if err := oprot.WriteFieldEnd(ctx); err != nil { + return thrift.PrependError(fmt.Sprintf("%T write field end error 2:slaPolicy: ", p), err) } + return err +} + +func (p *AuroraSchedulerManagerSlaRestartShardsArgs) writeField3(ctx context.Context, oprot thrift.TProtocol) (err error) { + if err := oprot.WriteFieldBegin(ctx, "shardIds", thrift.SET, 3); err != nil { + return thrift.PrependError(fmt.Sprintf("%T write field begin error 3:shardIds: ", p), err) } + if err := oprot.WriteSetBegin(ctx, thrift.I32, len(p.ShardIds)); err != nil { + return thrift.PrependError("error writing set begin: ", err) + } + for i := 0; i" + } + return fmt.Sprintf("AuroraSchedulerManagerSlaRestartShardsArgs(%+v)", *p) +} + +// Attributes: +// - Success +type AuroraSchedulerManagerSlaRestartShardsResult struct { + Success *Response `thrift:"success,0" db:"success" json:"success,omitempty"` +} + +func NewAuroraSchedulerManagerSlaRestartShardsResult() *AuroraSchedulerManagerSlaRestartShardsResult { + return &AuroraSchedulerManagerSlaRestartShardsResult{} +} + +var AuroraSchedulerManagerSlaRestartShardsResult_Success_DEFAULT *Response +func (p *AuroraSchedulerManagerSlaRestartShardsResult) GetSuccess() *Response { + if !p.IsSetSuccess() { + return AuroraSchedulerManagerSlaRestartShardsResult_Success_DEFAULT + } +return p.Success +} +func (p *AuroraSchedulerManagerSlaRestartShardsResult) IsSetSuccess() bool { + return p.Success != nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsResult) Read(ctx context.Context, iprot thrift.TProtocol) error { + if _, err := iprot.ReadStructBegin(ctx); err != nil { + return thrift.PrependError(fmt.Sprintf("%T read error: ", p), err) + } + + + for { + _, fieldTypeId, fieldId, err := iprot.ReadFieldBegin(ctx) + if err != nil { + return thrift.PrependError(fmt.Sprintf("%T field %d read error: ", p, fieldId), err) + } + if fieldTypeId == thrift.STOP { break; } + switch fieldId { + case 0: + if fieldTypeId == thrift.STRUCT { + if err := p.ReadField0(ctx, iprot); err != nil { + return err + } + } else { + if err := iprot.Skip(ctx, fieldTypeId); err != nil { + return err + } + } + default: + if err := iprot.Skip(ctx, fieldTypeId); err != nil { + return err + } + } + if err := iprot.ReadFieldEnd(ctx); err != nil { + return err + } + } + if err := iprot.ReadStructEnd(ctx); err != nil { + return thrift.PrependError(fmt.Sprintf("%T read struct end error: ", p), err) + } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsResult) ReadField0(ctx context.Context, iprot thrift.TProtocol) error { + p.Success = &Response{} + if err := p.Success.Read(ctx, iprot); err != nil { + return thrift.PrependError(fmt.Sprintf("%T error reading struct: ", p.Success), err) + } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsResult) Write(ctx context.Context, oprot thrift.TProtocol) error { + if err := oprot.WriteStructBegin(ctx, "slaRestartShards_result"); err != nil { + return thrift.PrependError(fmt.Sprintf("%T write struct begin error: ", p), err) } + if p != nil { + if err := p.writeField0(ctx, oprot); err != nil { return err } + } + if err := oprot.WriteFieldStop(ctx); err != nil { + return thrift.PrependError("write field stop error: ", err) } + if err := oprot.WriteStructEnd(ctx); err != nil { + return thrift.PrependError("write struct stop error: ", err) } + return nil +} + +func (p *AuroraSchedulerManagerSlaRestartShardsResult) writeField0(ctx context.Context, oprot thrift.TProtocol) (err error) { + if p.IsSetSuccess() { + if err := oprot.WriteFieldBegin(ctx, "success", thrift.STRUCT, 0); err != nil { + return thrift.PrependError(fmt.Sprintf("%T write field begin error 0:success: ", p), err) } + if err := p.Success.Write(ctx, oprot); err != nil { + return thrift.PrependError(fmt.Sprintf("%T error writing struct: ", p.Success), err) + } + if err := oprot.WriteFieldEnd(ctx); err != nil { + return thrift.PrependError(fmt.Sprintf("%T write field end error 0:success: ", p), err) } + } + return err +} + +func (p *AuroraSchedulerManagerSlaRestartShardsResult) String() string { + if p == nil { + return "" + } + return fmt.Sprintf("AuroraSchedulerManagerSlaRestartShardsResult(%+v)", *p) +} + // Attributes: // - Job // - Instances @@ -25436,13 +25854,13 @@ func (p *AuroraSchedulerManagerKillTasksArgs) ReadField5(ctx context.Context, i tSet := make([]int32, 0, size) p.Instances = tSet for i := 0; i < size; i ++ { -var _elem212 int32 +var _elem243 int32 if v, err := iprot.ReadI32(ctx); err != nil { return thrift.PrependError("error reading field 0: ", err) } else { - _elem212 = v + _elem243 = v } - p.Instances = append(p.Instances, _elem212) + p.Instances = append(p.Instances, _elem243) } if err := iprot.ReadSetEnd(ctx); err != nil { return thrift.PrependError("error reading set end: ", err) @@ -27444,7 +27862,7 @@ AuroraSchedulerManager // Parameters: // - OwnerRole // - Quota - SetQuota(ctx context.Context, ownerRole string, quota *ResourceAggregate) (r *Response, err error) + SetQuota(ctx context.Context, ownerRole string, quota *ResourceAggregate) (_r *Response, _err error) // Forces a task into a specific state. This does not guarantee the task will enter the given // state, as the task must still transition within the bounds of the state machine. However, // it attempts to enter that state via the state machine. @@ -27452,50 +27870,50 @@ AuroraSchedulerManager // Parameters: // - TaskId // - Status - ForceTaskState(ctx context.Context, taskId string, status ScheduleStatus) (r *Response, err error) + ForceTaskState(ctx context.Context, taskId string, status ScheduleStatus) (_r *Response, _err error) // Immediately writes a storage snapshot to disk. - PerformBackup(ctx context.Context) (r *Response, err error) + PerformBackup(ctx context.Context) (_r *Response, _err error) // Lists backups that are available for recovery. - ListBackups(ctx context.Context) (r *Response, err error) + ListBackups(ctx context.Context) (_r *Response, _err error) // Loads a backup to an in-memory storage. This must precede all other recovery operations. // // Parameters: // - BackupId - StageRecovery(ctx context.Context, backupId string) (r *Response, err error) + StageRecovery(ctx context.Context, backupId string) (_r *Response, _err error) // Queries for tasks in a staged recovery. // // Parameters: // - Query - QueryRecovery(ctx context.Context, query *TaskQuery) (r *Response, err error) + QueryRecovery(ctx context.Context, query *TaskQuery) (_r *Response, _err error) // Deletes tasks from a staged recovery. // // Parameters: // - Query - DeleteRecoveryTasks(ctx context.Context, query *TaskQuery) (r *Response, err error) + DeleteRecoveryTasks(ctx context.Context, query *TaskQuery) (_r *Response, _err error) // Commits a staged recovery, completely replacing the previous storage state. - CommitRecovery(ctx context.Context) (r *Response, err error) + CommitRecovery(ctx context.Context) (_r *Response, _err error) // Unloads (aborts) a staged recovery. - UnloadRecovery(ctx context.Context) (r *Response, err error) + UnloadRecovery(ctx context.Context) (_r *Response, _err error) // Put the given hosts into maintenance mode. // // Parameters: // - Hosts - StartMaintenance(ctx context.Context, hosts *Hosts) (r *Response, err error) + StartMaintenance(ctx context.Context, hosts *Hosts) (_r *Response, _err error) // Ask scheduler to begin moving tasks scheduled on given hosts. // // Parameters: // - Hosts - DrainHosts(ctx context.Context, hosts *Hosts) (r *Response, err error) + DrainHosts(ctx context.Context, hosts *Hosts) (_r *Response, _err error) // Retrieve the current maintenance states for a group of hosts. // // Parameters: // - Hosts - MaintenanceStatus(ctx context.Context, hosts *Hosts) (r *Response, err error) + MaintenanceStatus(ctx context.Context, hosts *Hosts) (_r *Response, _err error) // Set the given hosts back into serving mode. // // Parameters: // - Hosts - EndMaintenance(ctx context.Context, hosts *Hosts) (r *Response, err error) + EndMaintenance(ctx context.Context, hosts *Hosts) (_r *Response, _err error) // Ask scheduler to put hosts into DRAINING mode and move scheduled tasks off of the hosts // such that its SLA requirements are satisfied. Use defaultSlaPolicy if it is not set for a task. // @@ -27504,23 +27922,23 @@ AuroraSchedulerManager // - Hosts // - DefaultSlaPolicy // - TimeoutSecs - SlaDrainHosts(ctx context.Context, hosts *Hosts, defaultSlaPolicy *SlaPolicy, timeoutSecs int64) (r *Response, err error) + SlaDrainHosts(ctx context.Context, hosts *Hosts, defaultSlaPolicy *SlaPolicy, timeoutSecs int64) (_r *Response, _err error) // Start a storage snapshot and block until it completes. - Snapshot(ctx context.Context) (r *Response, err error) + Snapshot(ctx context.Context) (_r *Response, _err error) // Tell scheduler to trigger an explicit task reconciliation with the given settings. // // Parameters: // - Settings - TriggerExplicitTaskReconciliation(ctx context.Context, settings *ExplicitReconciliationSettings) (r *Response, err error) + TriggerExplicitTaskReconciliation(ctx context.Context, settings *ExplicitReconciliationSettings) (_r *Response, _err error) // Tell scheduler to trigger an implicit task reconciliation. - TriggerImplicitTaskReconciliation(ctx context.Context) (r *Response, err error) + TriggerImplicitTaskReconciliation(ctx context.Context) (_r *Response, _err error) // Force prune any (terminal) tasks that match the query. If no statuses are supplied with the // query, it will default to all terminal task states. If statuses are supplied, they must be // terminal states. // // Parameters: // - Query - PruneTasks(ctx context.Context, query *TaskQuery) (r *Response, err error) + PruneTasks(ctx context.Context, query *TaskQuery) (_r *Response, _err error) } type AuroraAdminClient struct { @@ -27545,18 +27963,18 @@ func NewAuroraAdminClient(c thrift.TClient) *AuroraAdminClient { // Parameters: // - OwnerRole // - Quota -func (p *AuroraAdminClient) SetQuota(ctx context.Context, ownerRole string, quota *ResourceAggregate) (r *Response, err error) { - var _args367 AuroraAdminSetQuotaArgs - _args367.OwnerRole = ownerRole - _args367.Quota = quota - var _result368 AuroraAdminSetQuotaResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "setQuota", &_args367, &_result368) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) SetQuota(ctx context.Context, ownerRole string, quota *ResourceAggregate) (_r *Response, _err error) { + var _args416 AuroraAdminSetQuotaArgs + _args416.OwnerRole = ownerRole + _args416.Quota = quota + var _result418 AuroraAdminSetQuotaResult + var _meta417 thrift.ResponseMeta + _meta417, _err = p.Client_().Call(ctx, "setQuota", &_args416, &_result418) + p.SetLastResponseMeta_(_meta417) + if _err != nil { return } - return _result368.GetSuccess(), nil + return _result418.GetSuccess(), nil } // Forces a task into a specific state. This does not guarantee the task will enter the given @@ -27566,189 +27984,189 @@ func (p *AuroraAdminClient) SetQuota(ctx context.Context, ownerRole string, quot // Parameters: // - TaskId // - Status -func (p *AuroraAdminClient) ForceTaskState(ctx context.Context, taskId string, status ScheduleStatus) (r *Response, err error) { - var _args369 AuroraAdminForceTaskStateArgs - _args369.TaskId = taskId - _args369.Status = status - var _result370 AuroraAdminForceTaskStateResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "forceTaskState", &_args369, &_result370) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) ForceTaskState(ctx context.Context, taskId string, status ScheduleStatus) (_r *Response, _err error) { + var _args419 AuroraAdminForceTaskStateArgs + _args419.TaskId = taskId + _args419.Status = status + var _result421 AuroraAdminForceTaskStateResult + var _meta420 thrift.ResponseMeta + _meta420, _err = p.Client_().Call(ctx, "forceTaskState", &_args419, &_result421) + p.SetLastResponseMeta_(_meta420) + if _err != nil { return } - return _result370.GetSuccess(), nil + return _result421.GetSuccess(), nil } // Immediately writes a storage snapshot to disk. -func (p *AuroraAdminClient) PerformBackup(ctx context.Context) (r *Response, err error) { - var _args371 AuroraAdminPerformBackupArgs - var _result372 AuroraAdminPerformBackupResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "performBackup", &_args371, &_result372) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) PerformBackup(ctx context.Context) (_r *Response, _err error) { + var _args422 AuroraAdminPerformBackupArgs + var _result424 AuroraAdminPerformBackupResult + var _meta423 thrift.ResponseMeta + _meta423, _err = p.Client_().Call(ctx, "performBackup", &_args422, &_result424) + p.SetLastResponseMeta_(_meta423) + if _err != nil { return } - return _result372.GetSuccess(), nil + return _result424.GetSuccess(), nil } // Lists backups that are available for recovery. -func (p *AuroraAdminClient) ListBackups(ctx context.Context) (r *Response, err error) { - var _args373 AuroraAdminListBackupsArgs - var _result374 AuroraAdminListBackupsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "listBackups", &_args373, &_result374) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) ListBackups(ctx context.Context) (_r *Response, _err error) { + var _args425 AuroraAdminListBackupsArgs + var _result427 AuroraAdminListBackupsResult + var _meta426 thrift.ResponseMeta + _meta426, _err = p.Client_().Call(ctx, "listBackups", &_args425, &_result427) + p.SetLastResponseMeta_(_meta426) + if _err != nil { return } - return _result374.GetSuccess(), nil + return _result427.GetSuccess(), nil } // Loads a backup to an in-memory storage. This must precede all other recovery operations. // // Parameters: // - BackupId -func (p *AuroraAdminClient) StageRecovery(ctx context.Context, backupId string) (r *Response, err error) { - var _args375 AuroraAdminStageRecoveryArgs - _args375.BackupId = backupId - var _result376 AuroraAdminStageRecoveryResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "stageRecovery", &_args375, &_result376) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) StageRecovery(ctx context.Context, backupId string) (_r *Response, _err error) { + var _args428 AuroraAdminStageRecoveryArgs + _args428.BackupId = backupId + var _result430 AuroraAdminStageRecoveryResult + var _meta429 thrift.ResponseMeta + _meta429, _err = p.Client_().Call(ctx, "stageRecovery", &_args428, &_result430) + p.SetLastResponseMeta_(_meta429) + if _err != nil { return } - return _result376.GetSuccess(), nil + return _result430.GetSuccess(), nil } // Queries for tasks in a staged recovery. // // Parameters: // - Query -func (p *AuroraAdminClient) QueryRecovery(ctx context.Context, query *TaskQuery) (r *Response, err error) { - var _args377 AuroraAdminQueryRecoveryArgs - _args377.Query = query - var _result378 AuroraAdminQueryRecoveryResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "queryRecovery", &_args377, &_result378) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) QueryRecovery(ctx context.Context, query *TaskQuery) (_r *Response, _err error) { + var _args431 AuroraAdminQueryRecoveryArgs + _args431.Query = query + var _result433 AuroraAdminQueryRecoveryResult + var _meta432 thrift.ResponseMeta + _meta432, _err = p.Client_().Call(ctx, "queryRecovery", &_args431, &_result433) + p.SetLastResponseMeta_(_meta432) + if _err != nil { return } - return _result378.GetSuccess(), nil + return _result433.GetSuccess(), nil } // Deletes tasks from a staged recovery. // // Parameters: // - Query -func (p *AuroraAdminClient) DeleteRecoveryTasks(ctx context.Context, query *TaskQuery) (r *Response, err error) { - var _args379 AuroraAdminDeleteRecoveryTasksArgs - _args379.Query = query - var _result380 AuroraAdminDeleteRecoveryTasksResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "deleteRecoveryTasks", &_args379, &_result380) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) DeleteRecoveryTasks(ctx context.Context, query *TaskQuery) (_r *Response, _err error) { + var _args434 AuroraAdminDeleteRecoveryTasksArgs + _args434.Query = query + var _result436 AuroraAdminDeleteRecoveryTasksResult + var _meta435 thrift.ResponseMeta + _meta435, _err = p.Client_().Call(ctx, "deleteRecoveryTasks", &_args434, &_result436) + p.SetLastResponseMeta_(_meta435) + if _err != nil { return } - return _result380.GetSuccess(), nil + return _result436.GetSuccess(), nil } // Commits a staged recovery, completely replacing the previous storage state. -func (p *AuroraAdminClient) CommitRecovery(ctx context.Context) (r *Response, err error) { - var _args381 AuroraAdminCommitRecoveryArgs - var _result382 AuroraAdminCommitRecoveryResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "commitRecovery", &_args381, &_result382) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) CommitRecovery(ctx context.Context) (_r *Response, _err error) { + var _args437 AuroraAdminCommitRecoveryArgs + var _result439 AuroraAdminCommitRecoveryResult + var _meta438 thrift.ResponseMeta + _meta438, _err = p.Client_().Call(ctx, "commitRecovery", &_args437, &_result439) + p.SetLastResponseMeta_(_meta438) + if _err != nil { return } - return _result382.GetSuccess(), nil + return _result439.GetSuccess(), nil } // Unloads (aborts) a staged recovery. -func (p *AuroraAdminClient) UnloadRecovery(ctx context.Context) (r *Response, err error) { - var _args383 AuroraAdminUnloadRecoveryArgs - var _result384 AuroraAdminUnloadRecoveryResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "unloadRecovery", &_args383, &_result384) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) UnloadRecovery(ctx context.Context) (_r *Response, _err error) { + var _args440 AuroraAdminUnloadRecoveryArgs + var _result442 AuroraAdminUnloadRecoveryResult + var _meta441 thrift.ResponseMeta + _meta441, _err = p.Client_().Call(ctx, "unloadRecovery", &_args440, &_result442) + p.SetLastResponseMeta_(_meta441) + if _err != nil { return } - return _result384.GetSuccess(), nil + return _result442.GetSuccess(), nil } // Put the given hosts into maintenance mode. // // Parameters: // - Hosts -func (p *AuroraAdminClient) StartMaintenance(ctx context.Context, hosts *Hosts) (r *Response, err error) { - var _args385 AuroraAdminStartMaintenanceArgs - _args385.Hosts = hosts - var _result386 AuroraAdminStartMaintenanceResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "startMaintenance", &_args385, &_result386) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) StartMaintenance(ctx context.Context, hosts *Hosts) (_r *Response, _err error) { + var _args443 AuroraAdminStartMaintenanceArgs + _args443.Hosts = hosts + var _result445 AuroraAdminStartMaintenanceResult + var _meta444 thrift.ResponseMeta + _meta444, _err = p.Client_().Call(ctx, "startMaintenance", &_args443, &_result445) + p.SetLastResponseMeta_(_meta444) + if _err != nil { return } - return _result386.GetSuccess(), nil + return _result445.GetSuccess(), nil } // Ask scheduler to begin moving tasks scheduled on given hosts. // // Parameters: // - Hosts -func (p *AuroraAdminClient) DrainHosts(ctx context.Context, hosts *Hosts) (r *Response, err error) { - var _args387 AuroraAdminDrainHostsArgs - _args387.Hosts = hosts - var _result388 AuroraAdminDrainHostsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "drainHosts", &_args387, &_result388) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) DrainHosts(ctx context.Context, hosts *Hosts) (_r *Response, _err error) { + var _args446 AuroraAdminDrainHostsArgs + _args446.Hosts = hosts + var _result448 AuroraAdminDrainHostsResult + var _meta447 thrift.ResponseMeta + _meta447, _err = p.Client_().Call(ctx, "drainHosts", &_args446, &_result448) + p.SetLastResponseMeta_(_meta447) + if _err != nil { return } - return _result388.GetSuccess(), nil + return _result448.GetSuccess(), nil } // Retrieve the current maintenance states for a group of hosts. // // Parameters: // - Hosts -func (p *AuroraAdminClient) MaintenanceStatus(ctx context.Context, hosts *Hosts) (r *Response, err error) { - var _args389 AuroraAdminMaintenanceStatusArgs - _args389.Hosts = hosts - var _result390 AuroraAdminMaintenanceStatusResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "maintenanceStatus", &_args389, &_result390) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) MaintenanceStatus(ctx context.Context, hosts *Hosts) (_r *Response, _err error) { + var _args449 AuroraAdminMaintenanceStatusArgs + _args449.Hosts = hosts + var _result451 AuroraAdminMaintenanceStatusResult + var _meta450 thrift.ResponseMeta + _meta450, _err = p.Client_().Call(ctx, "maintenanceStatus", &_args449, &_result451) + p.SetLastResponseMeta_(_meta450) + if _err != nil { return } - return _result390.GetSuccess(), nil + return _result451.GetSuccess(), nil } // Set the given hosts back into serving mode. // // Parameters: // - Hosts -func (p *AuroraAdminClient) EndMaintenance(ctx context.Context, hosts *Hosts) (r *Response, err error) { - var _args391 AuroraAdminEndMaintenanceArgs - _args391.Hosts = hosts - var _result392 AuroraAdminEndMaintenanceResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "endMaintenance", &_args391, &_result392) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) EndMaintenance(ctx context.Context, hosts *Hosts) (_r *Response, _err error) { + var _args452 AuroraAdminEndMaintenanceArgs + _args452.Hosts = hosts + var _result454 AuroraAdminEndMaintenanceResult + var _meta453 thrift.ResponseMeta + _meta453, _err = p.Client_().Call(ctx, "endMaintenance", &_args452, &_result454) + p.SetLastResponseMeta_(_meta453) + if _err != nil { return } - return _result392.GetSuccess(), nil + return _result454.GetSuccess(), nil } // Ask scheduler to put hosts into DRAINING mode and move scheduled tasks off of the hosts @@ -27759,62 +28177,62 @@ func (p *AuroraAdminClient) EndMaintenance(ctx context.Context, hosts *Hosts) (r // - Hosts // - DefaultSlaPolicy // - TimeoutSecs -func (p *AuroraAdminClient) SlaDrainHosts(ctx context.Context, hosts *Hosts, defaultSlaPolicy *SlaPolicy, timeoutSecs int64) (r *Response, err error) { - var _args393 AuroraAdminSlaDrainHostsArgs - _args393.Hosts = hosts - _args393.DefaultSlaPolicy = defaultSlaPolicy - _args393.TimeoutSecs = timeoutSecs - var _result394 AuroraAdminSlaDrainHostsResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "slaDrainHosts", &_args393, &_result394) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) SlaDrainHosts(ctx context.Context, hosts *Hosts, defaultSlaPolicy *SlaPolicy, timeoutSecs int64) (_r *Response, _err error) { + var _args455 AuroraAdminSlaDrainHostsArgs + _args455.Hosts = hosts + _args455.DefaultSlaPolicy = defaultSlaPolicy + _args455.TimeoutSecs = timeoutSecs + var _result457 AuroraAdminSlaDrainHostsResult + var _meta456 thrift.ResponseMeta + _meta456, _err = p.Client_().Call(ctx, "slaDrainHosts", &_args455, &_result457) + p.SetLastResponseMeta_(_meta456) + if _err != nil { return } - return _result394.GetSuccess(), nil + return _result457.GetSuccess(), nil } // Start a storage snapshot and block until it completes. -func (p *AuroraAdminClient) Snapshot(ctx context.Context) (r *Response, err error) { - var _args395 AuroraAdminSnapshotArgs - var _result396 AuroraAdminSnapshotResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "snapshot", &_args395, &_result396) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) Snapshot(ctx context.Context) (_r *Response, _err error) { + var _args458 AuroraAdminSnapshotArgs + var _result460 AuroraAdminSnapshotResult + var _meta459 thrift.ResponseMeta + _meta459, _err = p.Client_().Call(ctx, "snapshot", &_args458, &_result460) + p.SetLastResponseMeta_(_meta459) + if _err != nil { return } - return _result396.GetSuccess(), nil + return _result460.GetSuccess(), nil } // Tell scheduler to trigger an explicit task reconciliation with the given settings. // // Parameters: // - Settings -func (p *AuroraAdminClient) TriggerExplicitTaskReconciliation(ctx context.Context, settings *ExplicitReconciliationSettings) (r *Response, err error) { - var _args397 AuroraAdminTriggerExplicitTaskReconciliationArgs - _args397.Settings = settings - var _result398 AuroraAdminTriggerExplicitTaskReconciliationResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "triggerExplicitTaskReconciliation", &_args397, &_result398) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) TriggerExplicitTaskReconciliation(ctx context.Context, settings *ExplicitReconciliationSettings) (_r *Response, _err error) { + var _args461 AuroraAdminTriggerExplicitTaskReconciliationArgs + _args461.Settings = settings + var _result463 AuroraAdminTriggerExplicitTaskReconciliationResult + var _meta462 thrift.ResponseMeta + _meta462, _err = p.Client_().Call(ctx, "triggerExplicitTaskReconciliation", &_args461, &_result463) + p.SetLastResponseMeta_(_meta462) + if _err != nil { return } - return _result398.GetSuccess(), nil + return _result463.GetSuccess(), nil } // Tell scheduler to trigger an implicit task reconciliation. -func (p *AuroraAdminClient) TriggerImplicitTaskReconciliation(ctx context.Context) (r *Response, err error) { - var _args399 AuroraAdminTriggerImplicitTaskReconciliationArgs - var _result400 AuroraAdminTriggerImplicitTaskReconciliationResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "triggerImplicitTaskReconciliation", &_args399, &_result400) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) TriggerImplicitTaskReconciliation(ctx context.Context) (_r *Response, _err error) { + var _args464 AuroraAdminTriggerImplicitTaskReconciliationArgs + var _result466 AuroraAdminTriggerImplicitTaskReconciliationResult + var _meta465 thrift.ResponseMeta + _meta465, _err = p.Client_().Call(ctx, "triggerImplicitTaskReconciliation", &_args464, &_result466) + p.SetLastResponseMeta_(_meta465) + if _err != nil { return } - return _result400.GetSuccess(), nil + return _result466.GetSuccess(), nil } // Force prune any (terminal) tasks that match the query. If no statuses are supplied with the @@ -27823,17 +28241,17 @@ func (p *AuroraAdminClient) TriggerImplicitTaskReconciliation(ctx context.Contex // // Parameters: // - Query -func (p *AuroraAdminClient) PruneTasks(ctx context.Context, query *TaskQuery) (r *Response, err error) { - var _args401 AuroraAdminPruneTasksArgs - _args401.Query = query - var _result402 AuroraAdminPruneTasksResult - var meta thrift.ResponseMeta - meta, err = p.Client_().Call(ctx, "pruneTasks", &_args401, &_result402) - p.SetLastResponseMeta_(meta) - if err != nil { +func (p *AuroraAdminClient) PruneTasks(ctx context.Context, query *TaskQuery) (_r *Response, _err error) { + var _args467 AuroraAdminPruneTasksArgs + _args467.Query = query + var _result469 AuroraAdminPruneTasksResult + var _meta468 thrift.ResponseMeta + _meta468, _err = p.Client_().Call(ctx, "pruneTasks", &_args467, &_result469) + p.SetLastResponseMeta_(_meta468) + if _err != nil { return } - return _result402.GetSuccess(), nil + return _result469.GetSuccess(), nil } type AuroraAdminProcessor struct { @@ -27841,26 +28259,26 @@ type AuroraAdminProcessor struct { } func NewAuroraAdminProcessor(handler AuroraAdmin) *AuroraAdminProcessor { - self403 := &AuroraAdminProcessor{NewAuroraSchedulerManagerProcessor(handler)} - self403.AddToProcessorMap("setQuota", &auroraAdminProcessorSetQuota{handler:handler}) - self403.AddToProcessorMap("forceTaskState", &auroraAdminProcessorForceTaskState{handler:handler}) - self403.AddToProcessorMap("performBackup", &auroraAdminProcessorPerformBackup{handler:handler}) - self403.AddToProcessorMap("listBackups", &auroraAdminProcessorListBackups{handler:handler}) - self403.AddToProcessorMap("stageRecovery", &auroraAdminProcessorStageRecovery{handler:handler}) - self403.AddToProcessorMap("queryRecovery", &auroraAdminProcessorQueryRecovery{handler:handler}) - self403.AddToProcessorMap("deleteRecoveryTasks", &auroraAdminProcessorDeleteRecoveryTasks{handler:handler}) - self403.AddToProcessorMap("commitRecovery", &auroraAdminProcessorCommitRecovery{handler:handler}) - self403.AddToProcessorMap("unloadRecovery", &auroraAdminProcessorUnloadRecovery{handler:handler}) - self403.AddToProcessorMap("startMaintenance", &auroraAdminProcessorStartMaintenance{handler:handler}) - self403.AddToProcessorMap("drainHosts", &auroraAdminProcessorDrainHosts{handler:handler}) - self403.AddToProcessorMap("maintenanceStatus", &auroraAdminProcessorMaintenanceStatus{handler:handler}) - self403.AddToProcessorMap("endMaintenance", &auroraAdminProcessorEndMaintenance{handler:handler}) - self403.AddToProcessorMap("slaDrainHosts", &auroraAdminProcessorSlaDrainHosts{handler:handler}) - self403.AddToProcessorMap("snapshot", &auroraAdminProcessorSnapshot{handler:handler}) - self403.AddToProcessorMap("triggerExplicitTaskReconciliation", &auroraAdminProcessorTriggerExplicitTaskReconciliation{handler:handler}) - self403.AddToProcessorMap("triggerImplicitTaskReconciliation", &auroraAdminProcessorTriggerImplicitTaskReconciliation{handler:handler}) - self403.AddToProcessorMap("pruneTasks", &auroraAdminProcessorPruneTasks{handler:handler}) - return self403 + self470 := &AuroraAdminProcessor{NewAuroraSchedulerManagerProcessor(handler)} + self470.AddToProcessorMap("setQuota", &auroraAdminProcessorSetQuota{handler:handler}) + self470.AddToProcessorMap("forceTaskState", &auroraAdminProcessorForceTaskState{handler:handler}) + self470.AddToProcessorMap("performBackup", &auroraAdminProcessorPerformBackup{handler:handler}) + self470.AddToProcessorMap("listBackups", &auroraAdminProcessorListBackups{handler:handler}) + self470.AddToProcessorMap("stageRecovery", &auroraAdminProcessorStageRecovery{handler:handler}) + self470.AddToProcessorMap("queryRecovery", &auroraAdminProcessorQueryRecovery{handler:handler}) + self470.AddToProcessorMap("deleteRecoveryTasks", &auroraAdminProcessorDeleteRecoveryTasks{handler:handler}) + self470.AddToProcessorMap("commitRecovery", &auroraAdminProcessorCommitRecovery{handler:handler}) + self470.AddToProcessorMap("unloadRecovery", &auroraAdminProcessorUnloadRecovery{handler:handler}) + self470.AddToProcessorMap("startMaintenance", &auroraAdminProcessorStartMaintenance{handler:handler}) + self470.AddToProcessorMap("drainHosts", &auroraAdminProcessorDrainHosts{handler:handler}) + self470.AddToProcessorMap("maintenanceStatus", &auroraAdminProcessorMaintenanceStatus{handler:handler}) + self470.AddToProcessorMap("endMaintenance", &auroraAdminProcessorEndMaintenance{handler:handler}) + self470.AddToProcessorMap("slaDrainHosts", &auroraAdminProcessorSlaDrainHosts{handler:handler}) + self470.AddToProcessorMap("snapshot", &auroraAdminProcessorSnapshot{handler:handler}) + self470.AddToProcessorMap("triggerExplicitTaskReconciliation", &auroraAdminProcessorTriggerExplicitTaskReconciliation{handler:handler}) + self470.AddToProcessorMap("triggerImplicitTaskReconciliation", &auroraAdminProcessorTriggerImplicitTaskReconciliation{handler:handler}) + self470.AddToProcessorMap("pruneTasks", &auroraAdminProcessorPruneTasks{handler:handler}) + return self470 } type auroraAdminProcessorSetQuota struct { diff --git a/gen-go/apache/aurora/aurora_admin-remote/aurora_admin-remote.go b/gen-go/apache/aurora/aurora_admin-remote/aurora_admin-remote.go index d2d4441..e4a6f3d 100755 --- a/gen-go/apache/aurora/aurora_admin-remote/aurora_admin-remote.go +++ b/gen-go/apache/aurora/aurora_admin-remote/aurora_admin-remote.go @@ -1,4 +1,4 @@ -// Code generated by Thrift Compiler (0.14.0). DO NOT EDIT. +// Code generated by Thrift Compiler (0.14.1). DO NOT EDIT. package main @@ -13,7 +13,7 @@ import ( "strconv" "strings" "github.com/apache/thrift/lib/go/thrift" - "apache/aurora" + "aurora" ) var _ = aurora.GoUnusedProtection__ @@ -45,6 +45,7 @@ func Usage() { fmt.Fprintln(os.Stderr, " Response descheduleCronJob(JobKey job)") fmt.Fprintln(os.Stderr, " Response startCronJob(JobKey job)") fmt.Fprintln(os.Stderr, " Response restartShards(JobKey job, shardIds)") + fmt.Fprintln(os.Stderr, " Response slaRestartShards(JobKey job, shardIds, SlaPolicy slaPolicy)") fmt.Fprintln(os.Stderr, " Response killTasks(JobKey job, instances, string message)") fmt.Fprintln(os.Stderr, " Response addInstances(InstanceKey key, i32 count)") fmt.Fprintln(os.Stderr, " Response replaceCronTemplate(JobConfiguration config)") @@ -195,19 +196,19 @@ func main() { } argvalue0 := flag.Arg(1) value0 := argvalue0 - arg405 := flag.Arg(2) - mbTrans406 := thrift.NewTMemoryBufferLen(len(arg405)) - defer mbTrans406.Close() - _, err407 := mbTrans406.WriteString(arg405) - if err407 != nil { + arg472 := flag.Arg(2) + mbTrans473 := thrift.NewTMemoryBufferLen(len(arg472)) + defer mbTrans473.Close() + _, err474 := mbTrans473.WriteString(arg472) + if err474 != nil { Usage() return } - factory408 := thrift.NewTJSONProtocolFactory() - jsProt409 := factory408.GetProtocol(mbTrans406) + factory475 := thrift.NewTJSONProtocolFactory() + jsProt476 := factory475.GetProtocol(mbTrans473) argvalue1 := aurora.NewResourceAggregate() - err410 := argvalue1.Read(context.Background(), jsProt409) - if err410 != nil { + err477 := argvalue1.Read(context.Background(), jsProt476) + if err477 != nil { Usage() return } @@ -263,19 +264,19 @@ func main() { fmt.Fprintln(os.Stderr, "QueryRecovery requires 1 args") flag.Usage() } - arg413 := flag.Arg(1) - mbTrans414 := thrift.NewTMemoryBufferLen(len(arg413)) - defer mbTrans414.Close() - _, err415 := mbTrans414.WriteString(arg413) - if err415 != nil { + arg480 := flag.Arg(1) + mbTrans481 := thrift.NewTMemoryBufferLen(len(arg480)) + defer mbTrans481.Close() + _, err482 := mbTrans481.WriteString(arg480) + if err482 != nil { Usage() return } - factory416 := thrift.NewTJSONProtocolFactory() - jsProt417 := factory416.GetProtocol(mbTrans414) + factory483 := thrift.NewTJSONProtocolFactory() + jsProt484 := factory483.GetProtocol(mbTrans481) argvalue0 := aurora.NewTaskQuery() - err418 := argvalue0.Read(context.Background(), jsProt417) - if err418 != nil { + err485 := argvalue0.Read(context.Background(), jsProt484) + if err485 != nil { Usage() return } @@ -288,19 +289,19 @@ func main() { fmt.Fprintln(os.Stderr, "DeleteRecoveryTasks requires 1 args") flag.Usage() } - arg419 := flag.Arg(1) - mbTrans420 := thrift.NewTMemoryBufferLen(len(arg419)) - defer mbTrans420.Close() - _, err421 := mbTrans420.WriteString(arg419) - if err421 != nil { + arg486 := flag.Arg(1) + mbTrans487 := thrift.NewTMemoryBufferLen(len(arg486)) + defer mbTrans487.Close() + _, err488 := mbTrans487.WriteString(arg486) + if err488 != nil { Usage() return } - factory422 := thrift.NewTJSONProtocolFactory() - jsProt423 := factory422.GetProtocol(mbTrans420) + factory489 := thrift.NewTJSONProtocolFactory() + jsProt490 := factory489.GetProtocol(mbTrans487) argvalue0 := aurora.NewTaskQuery() - err424 := argvalue0.Read(context.Background(), jsProt423) - if err424 != nil { + err491 := argvalue0.Read(context.Background(), jsProt490) + if err491 != nil { Usage() return } @@ -329,19 +330,19 @@ func main() { fmt.Fprintln(os.Stderr, "StartMaintenance requires 1 args") flag.Usage() } - arg425 := flag.Arg(1) - mbTrans426 := thrift.NewTMemoryBufferLen(len(arg425)) - defer mbTrans426.Close() - _, err427 := mbTrans426.WriteString(arg425) - if err427 != nil { + arg492 := flag.Arg(1) + mbTrans493 := thrift.NewTMemoryBufferLen(len(arg492)) + defer mbTrans493.Close() + _, err494 := mbTrans493.WriteString(arg492) + if err494 != nil { Usage() return } - factory428 := thrift.NewTJSONProtocolFactory() - jsProt429 := factory428.GetProtocol(mbTrans426) + factory495 := thrift.NewTJSONProtocolFactory() + jsProt496 := factory495.GetProtocol(mbTrans493) argvalue0 := aurora.NewHosts() - err430 := argvalue0.Read(context.Background(), jsProt429) - if err430 != nil { + err497 := argvalue0.Read(context.Background(), jsProt496) + if err497 != nil { Usage() return } @@ -354,19 +355,19 @@ func main() { fmt.Fprintln(os.Stderr, "DrainHosts requires 1 args") flag.Usage() } - arg431 := flag.Arg(1) - mbTrans432 := thrift.NewTMemoryBufferLen(len(arg431)) - defer mbTrans432.Close() - _, err433 := mbTrans432.WriteString(arg431) - if err433 != nil { + arg498 := flag.Arg(1) + mbTrans499 := thrift.NewTMemoryBufferLen(len(arg498)) + defer mbTrans499.Close() + _, err500 := mbTrans499.WriteString(arg498) + if err500 != nil { Usage() return } - factory434 := thrift.NewTJSONProtocolFactory() - jsProt435 := factory434.GetProtocol(mbTrans432) + factory501 := thrift.NewTJSONProtocolFactory() + jsProt502 := factory501.GetProtocol(mbTrans499) argvalue0 := aurora.NewHosts() - err436 := argvalue0.Read(context.Background(), jsProt435) - if err436 != nil { + err503 := argvalue0.Read(context.Background(), jsProt502) + if err503 != nil { Usage() return } @@ -379,19 +380,19 @@ func main() { fmt.Fprintln(os.Stderr, "MaintenanceStatus requires 1 args") flag.Usage() } - arg437 := flag.Arg(1) - mbTrans438 := thrift.NewTMemoryBufferLen(len(arg437)) - defer mbTrans438.Close() - _, err439 := mbTrans438.WriteString(arg437) - if err439 != nil { + arg504 := flag.Arg(1) + mbTrans505 := thrift.NewTMemoryBufferLen(len(arg504)) + defer mbTrans505.Close() + _, err506 := mbTrans505.WriteString(arg504) + if err506 != nil { Usage() return } - factory440 := thrift.NewTJSONProtocolFactory() - jsProt441 := factory440.GetProtocol(mbTrans438) + factory507 := thrift.NewTJSONProtocolFactory() + jsProt508 := factory507.GetProtocol(mbTrans505) argvalue0 := aurora.NewHosts() - err442 := argvalue0.Read(context.Background(), jsProt441) - if err442 != nil { + err509 := argvalue0.Read(context.Background(), jsProt508) + if err509 != nil { Usage() return } @@ -404,19 +405,19 @@ func main() { fmt.Fprintln(os.Stderr, "EndMaintenance requires 1 args") flag.Usage() } - arg443 := flag.Arg(1) - mbTrans444 := thrift.NewTMemoryBufferLen(len(arg443)) - defer mbTrans444.Close() - _, err445 := mbTrans444.WriteString(arg443) - if err445 != nil { + arg510 := flag.Arg(1) + mbTrans511 := thrift.NewTMemoryBufferLen(len(arg510)) + defer mbTrans511.Close() + _, err512 := mbTrans511.WriteString(arg510) + if err512 != nil { Usage() return } - factory446 := thrift.NewTJSONProtocolFactory() - jsProt447 := factory446.GetProtocol(mbTrans444) + factory513 := thrift.NewTJSONProtocolFactory() + jsProt514 := factory513.GetProtocol(mbTrans511) argvalue0 := aurora.NewHosts() - err448 := argvalue0.Read(context.Background(), jsProt447) - if err448 != nil { + err515 := argvalue0.Read(context.Background(), jsProt514) + if err515 != nil { Usage() return } @@ -429,42 +430,42 @@ func main() { fmt.Fprintln(os.Stderr, "SlaDrainHosts requires 3 args") flag.Usage() } - arg449 := flag.Arg(1) - mbTrans450 := thrift.NewTMemoryBufferLen(len(arg449)) - defer mbTrans450.Close() - _, err451 := mbTrans450.WriteString(arg449) - if err451 != nil { + arg516 := flag.Arg(1) + mbTrans517 := thrift.NewTMemoryBufferLen(len(arg516)) + defer mbTrans517.Close() + _, err518 := mbTrans517.WriteString(arg516) + if err518 != nil { Usage() return } - factory452 := thrift.NewTJSONProtocolFactory() - jsProt453 := factory452.GetProtocol(mbTrans450) + factory519 := thrift.NewTJSONProtocolFactory() + jsProt520 := factory519.GetProtocol(mbTrans517) argvalue0 := aurora.NewHosts() - err454 := argvalue0.Read(context.Background(), jsProt453) - if err454 != nil { + err521 := argvalue0.Read(context.Background(), jsProt520) + if err521 != nil { Usage() return } value0 := argvalue0 - arg455 := flag.Arg(2) - mbTrans456 := thrift.NewTMemoryBufferLen(len(arg455)) - defer mbTrans456.Close() - _, err457 := mbTrans456.WriteString(arg455) - if err457 != nil { + arg522 := flag.Arg(2) + mbTrans523 := thrift.NewTMemoryBufferLen(len(arg522)) + defer mbTrans523.Close() + _, err524 := mbTrans523.WriteString(arg522) + if err524 != nil { Usage() return } - factory458 := thrift.NewTJSONProtocolFactory() - jsProt459 := factory458.GetProtocol(mbTrans456) + factory525 := thrift.NewTJSONProtocolFactory() + jsProt526 := factory525.GetProtocol(mbTrans523) argvalue1 := aurora.NewSlaPolicy() - err460 := argvalue1.Read(context.Background(), jsProt459) - if err460 != nil { + err527 := argvalue1.Read(context.Background(), jsProt526) + if err527 != nil { Usage() return } value1 := argvalue1 - argvalue2, err461 := (strconv.ParseInt(flag.Arg(3), 10, 64)) - if err461 != nil { + argvalue2, err528 := (strconv.ParseInt(flag.Arg(3), 10, 64)) + if err528 != nil { Usage() return } @@ -485,19 +486,19 @@ func main() { fmt.Fprintln(os.Stderr, "TriggerExplicitTaskReconciliation requires 1 args") flag.Usage() } - arg462 := flag.Arg(1) - mbTrans463 := thrift.NewTMemoryBufferLen(len(arg462)) - defer mbTrans463.Close() - _, err464 := mbTrans463.WriteString(arg462) - if err464 != nil { + arg529 := flag.Arg(1) + mbTrans530 := thrift.NewTMemoryBufferLen(len(arg529)) + defer mbTrans530.Close() + _, err531 := mbTrans530.WriteString(arg529) + if err531 != nil { Usage() return } - factory465 := thrift.NewTJSONProtocolFactory() - jsProt466 := factory465.GetProtocol(mbTrans463) + factory532 := thrift.NewTJSONProtocolFactory() + jsProt533 := factory532.GetProtocol(mbTrans530) argvalue0 := aurora.NewExplicitReconciliationSettings() - err467 := argvalue0.Read(context.Background(), jsProt466) - if err467 != nil { + err534 := argvalue0.Read(context.Background(), jsProt533) + if err534 != nil { Usage() return } @@ -518,19 +519,19 @@ func main() { fmt.Fprintln(os.Stderr, "PruneTasks requires 1 args") flag.Usage() } - arg468 := flag.Arg(1) - mbTrans469 := thrift.NewTMemoryBufferLen(len(arg468)) - defer mbTrans469.Close() - _, err470 := mbTrans469.WriteString(arg468) - if err470 != nil { + arg535 := flag.Arg(1) + mbTrans536 := thrift.NewTMemoryBufferLen(len(arg535)) + defer mbTrans536.Close() + _, err537 := mbTrans536.WriteString(arg535) + if err537 != nil { Usage() return } - factory471 := thrift.NewTJSONProtocolFactory() - jsProt472 := factory471.GetProtocol(mbTrans469) + factory538 := thrift.NewTJSONProtocolFactory() + jsProt539 := factory538.GetProtocol(mbTrans536) argvalue0 := aurora.NewTaskQuery() - err473 := argvalue0.Read(context.Background(), jsProt472) - if err473 != nil { + err540 := argvalue0.Read(context.Background(), jsProt539) + if err540 != nil { Usage() return } @@ -543,19 +544,19 @@ func main() { fmt.Fprintln(os.Stderr, "CreateJob requires 1 args") flag.Usage() } - arg474 := flag.Arg(1) - mbTrans475 := thrift.NewTMemoryBufferLen(len(arg474)) - defer mbTrans475.Close() - _, err476 := mbTrans475.WriteString(arg474) - if err476 != nil { + arg541 := flag.Arg(1) + mbTrans542 := thrift.NewTMemoryBufferLen(len(arg541)) + defer mbTrans542.Close() + _, err543 := mbTrans542.WriteString(arg541) + if err543 != nil { Usage() return } - factory477 := thrift.NewTJSONProtocolFactory() - jsProt478 := factory477.GetProtocol(mbTrans475) + factory544 := thrift.NewTJSONProtocolFactory() + jsProt545 := factory544.GetProtocol(mbTrans542) argvalue0 := aurora.NewJobConfiguration() - err479 := argvalue0.Read(context.Background(), jsProt478) - if err479 != nil { + err546 := argvalue0.Read(context.Background(), jsProt545) + if err546 != nil { Usage() return } @@ -568,19 +569,19 @@ func main() { fmt.Fprintln(os.Stderr, "ScheduleCronJob requires 1 args") flag.Usage() } - arg480 := flag.Arg(1) - mbTrans481 := thrift.NewTMemoryBufferLen(len(arg480)) - defer mbTrans481.Close() - _, err482 := mbTrans481.WriteString(arg480) - if err482 != nil { + arg547 := flag.Arg(1) + mbTrans548 := thrift.NewTMemoryBufferLen(len(arg547)) + defer mbTrans548.Close() + _, err549 := mbTrans548.WriteString(arg547) + if err549 != nil { Usage() return } - factory483 := thrift.NewTJSONProtocolFactory() - jsProt484 := factory483.GetProtocol(mbTrans481) + factory550 := thrift.NewTJSONProtocolFactory() + jsProt551 := factory550.GetProtocol(mbTrans548) argvalue0 := aurora.NewJobConfiguration() - err485 := argvalue0.Read(context.Background(), jsProt484) - if err485 != nil { + err552 := argvalue0.Read(context.Background(), jsProt551) + if err552 != nil { Usage() return } @@ -593,19 +594,19 @@ func main() { fmt.Fprintln(os.Stderr, "DescheduleCronJob requires 1 args") flag.Usage() } - arg486 := flag.Arg(1) - mbTrans487 := thrift.NewTMemoryBufferLen(len(arg486)) - defer mbTrans487.Close() - _, err488 := mbTrans487.WriteString(arg486) - if err488 != nil { + arg553 := flag.Arg(1) + mbTrans554 := thrift.NewTMemoryBufferLen(len(arg553)) + defer mbTrans554.Close() + _, err555 := mbTrans554.WriteString(arg553) + if err555 != nil { Usage() return } - factory489 := thrift.NewTJSONProtocolFactory() - jsProt490 := factory489.GetProtocol(mbTrans487) + factory556 := thrift.NewTJSONProtocolFactory() + jsProt557 := factory556.GetProtocol(mbTrans554) argvalue0 := aurora.NewJobKey() - err491 := argvalue0.Read(context.Background(), jsProt490) - if err491 != nil { + err558 := argvalue0.Read(context.Background(), jsProt557) + if err558 != nil { Usage() return } @@ -618,19 +619,19 @@ func main() { fmt.Fprintln(os.Stderr, "StartCronJob requires 1 args") flag.Usage() } - arg492 := flag.Arg(1) - mbTrans493 := thrift.NewTMemoryBufferLen(len(arg492)) - defer mbTrans493.Close() - _, err494 := mbTrans493.WriteString(arg492) - if err494 != nil { + arg559 := flag.Arg(1) + mbTrans560 := thrift.NewTMemoryBufferLen(len(arg559)) + defer mbTrans560.Close() + _, err561 := mbTrans560.WriteString(arg559) + if err561 != nil { Usage() return } - factory495 := thrift.NewTJSONProtocolFactory() - jsProt496 := factory495.GetProtocol(mbTrans493) + factory562 := thrift.NewTJSONProtocolFactory() + jsProt563 := factory562.GetProtocol(mbTrans560) argvalue0 := aurora.NewJobKey() - err497 := argvalue0.Read(context.Background(), jsProt496) - if err497 != nil { + err564 := argvalue0.Read(context.Background(), jsProt563) + if err564 != nil { Usage() return } @@ -643,36 +644,36 @@ func main() { fmt.Fprintln(os.Stderr, "RestartShards requires 2 args") flag.Usage() } - arg498 := flag.Arg(1) - mbTrans499 := thrift.NewTMemoryBufferLen(len(arg498)) - defer mbTrans499.Close() - _, err500 := mbTrans499.WriteString(arg498) - if err500 != nil { + arg565 := flag.Arg(1) + mbTrans566 := thrift.NewTMemoryBufferLen(len(arg565)) + defer mbTrans566.Close() + _, err567 := mbTrans566.WriteString(arg565) + if err567 != nil { Usage() return } - factory501 := thrift.NewTJSONProtocolFactory() - jsProt502 := factory501.GetProtocol(mbTrans499) + factory568 := thrift.NewTJSONProtocolFactory() + jsProt569 := factory568.GetProtocol(mbTrans566) argvalue0 := aurora.NewJobKey() - err503 := argvalue0.Read(context.Background(), jsProt502) - if err503 != nil { + err570 := argvalue0.Read(context.Background(), jsProt569) + if err570 != nil { Usage() return } value0 := argvalue0 - arg504 := flag.Arg(2) - mbTrans505 := thrift.NewTMemoryBufferLen(len(arg504)) - defer mbTrans505.Close() - _, err506 := mbTrans505.WriteString(arg504) - if err506 != nil { + arg571 := flag.Arg(2) + mbTrans572 := thrift.NewTMemoryBufferLen(len(arg571)) + defer mbTrans572.Close() + _, err573 := mbTrans572.WriteString(arg571) + if err573 != nil { Usage() return } - factory507 := thrift.NewTJSONProtocolFactory() - jsProt508 := factory507.GetProtocol(mbTrans505) + factory574 := thrift.NewTJSONProtocolFactory() + jsProt575 := factory574.GetProtocol(mbTrans572) containerStruct1 := aurora.NewAuroraSchedulerManagerRestartShardsArgs() - err509 := containerStruct1.ReadField2(context.Background(), jsProt508) - if err509 != nil { + err576 := containerStruct1.ReadField2(context.Background(), jsProt575) + if err576 != nil { Usage() return } @@ -681,41 +682,101 @@ func main() { fmt.Print(client.RestartShards(context.Background(), value0, value1)) fmt.Print("\n") break + case "slaRestartShards": + if flag.NArg() - 1 != 3 { + fmt.Fprintln(os.Stderr, "SlaRestartShards requires 3 args") + flag.Usage() + } + arg577 := flag.Arg(1) + mbTrans578 := thrift.NewTMemoryBufferLen(len(arg577)) + defer mbTrans578.Close() + _, err579 := mbTrans578.WriteString(arg577) + if err579 != nil { + Usage() + return + } + factory580 := thrift.NewTJSONProtocolFactory() + jsProt581 := factory580.GetProtocol(mbTrans578) + argvalue0 := aurora.NewJobKey() + err582 := argvalue0.Read(context.Background(), jsProt581) + if err582 != nil { + Usage() + return + } + value0 := argvalue0 + arg583 := flag.Arg(2) + mbTrans584 := thrift.NewTMemoryBufferLen(len(arg583)) + defer mbTrans584.Close() + _, err585 := mbTrans584.WriteString(arg583) + if err585 != nil { + Usage() + return + } + factory586 := thrift.NewTJSONProtocolFactory() + jsProt587 := factory586.GetProtocol(mbTrans584) + containerStruct1 := aurora.NewAuroraSchedulerManagerSlaRestartShardsArgs() + err588 := containerStruct1.ReadField2(context.Background(), jsProt587) + if err588 != nil { + Usage() + return + } + argvalue1 := containerStruct1.ShardIds + value1 := argvalue1 + arg589 := flag.Arg(3) + mbTrans590 := thrift.NewTMemoryBufferLen(len(arg589)) + defer mbTrans590.Close() + _, err591 := mbTrans590.WriteString(arg589) + if err591 != nil { + Usage() + return + } + factory592 := thrift.NewTJSONProtocolFactory() + jsProt593 := factory592.GetProtocol(mbTrans590) + argvalue2 := aurora.NewSlaPolicy() + err594 := argvalue2.Read(context.Background(), jsProt593) + if err594 != nil { + Usage() + return + } + value2 := argvalue2 + fmt.Print(client.SlaRestartShards(context.Background(), value0, value1, value2)) + fmt.Print("\n") + break case "killTasks": if flag.NArg() - 1 != 3 { fmt.Fprintln(os.Stderr, "KillTasks requires 3 args") flag.Usage() } - arg510 := flag.Arg(1) - mbTrans511 := thrift.NewTMemoryBufferLen(len(arg510)) - defer mbTrans511.Close() - _, err512 := mbTrans511.WriteString(arg510) - if err512 != nil { + arg595 := flag.Arg(1) + mbTrans596 := thrift.NewTMemoryBufferLen(len(arg595)) + defer mbTrans596.Close() + _, err597 := mbTrans596.WriteString(arg595) + if err597 != nil { Usage() return } - factory513 := thrift.NewTJSONProtocolFactory() - jsProt514 := factory513.GetProtocol(mbTrans511) + factory598 := thrift.NewTJSONProtocolFactory() + jsProt599 := factory598.GetProtocol(mbTrans596) argvalue0 := aurora.NewJobKey() - err515 := argvalue0.Read(context.Background(), jsProt514) - if err515 != nil { + err600 := argvalue0.Read(context.Background(), jsProt599) + if err600 != nil { Usage() return } value0 := argvalue0 - arg516 := flag.Arg(2) - mbTrans517 := thrift.NewTMemoryBufferLen(len(arg516)) - defer mbTrans517.Close() - _, err518 := mbTrans517.WriteString(arg516) - if err518 != nil { + arg601 := flag.Arg(2) + mbTrans602 := thrift.NewTMemoryBufferLen(len(arg601)) + defer mbTrans602.Close() + _, err603 := mbTrans602.WriteString(arg601) + if err603 != nil { Usage() return } - factory519 := thrift.NewTJSONProtocolFactory() - jsProt520 := factory519.GetProtocol(mbTrans517) + factory604 := thrift.NewTJSONProtocolFactory() + jsProt605 := factory604.GetProtocol(mbTrans602) containerStruct1 := aurora.NewAuroraSchedulerManagerKillTasksArgs() - err521 := containerStruct1.ReadField2(context.Background(), jsProt520) - if err521 != nil { + err606 := containerStruct1.ReadField2(context.Background(), jsProt605) + if err606 != nil { Usage() return } @@ -731,25 +792,25 @@ func main() { fmt.Fprintln(os.Stderr, "AddInstances requires 2 args") flag.Usage() } - arg523 := flag.Arg(1) - mbTrans524 := thrift.NewTMemoryBufferLen(len(arg523)) - defer mbTrans524.Close() - _, err525 := mbTrans524.WriteString(arg523) - if err525 != nil { + arg608 := flag.Arg(1) + mbTrans609 := thrift.NewTMemoryBufferLen(len(arg608)) + defer mbTrans609.Close() + _, err610 := mbTrans609.WriteString(arg608) + if err610 != nil { Usage() return } - factory526 := thrift.NewTJSONProtocolFactory() - jsProt527 := factory526.GetProtocol(mbTrans524) + factory611 := thrift.NewTJSONProtocolFactory() + jsProt612 := factory611.GetProtocol(mbTrans609) argvalue0 := aurora.NewInstanceKey() - err528 := argvalue0.Read(context.Background(), jsProt527) - if err528 != nil { + err613 := argvalue0.Read(context.Background(), jsProt612) + if err613 != nil { Usage() return } value0 := argvalue0 - tmp1, err529 := (strconv.Atoi(flag.Arg(2))) - if err529 != nil { + tmp1, err614 := (strconv.Atoi(flag.Arg(2))) + if err614 != nil { Usage() return } @@ -763,19 +824,19 @@ func main() { fmt.Fprintln(os.Stderr, "ReplaceCronTemplate requires 1 args") flag.Usage() } - arg530 := flag.Arg(1) - mbTrans531 := thrift.NewTMemoryBufferLen(len(arg530)) - defer mbTrans531.Close() - _, err532 := mbTrans531.WriteString(arg530) - if err532 != nil { + arg615 := flag.Arg(1) + mbTrans616 := thrift.NewTMemoryBufferLen(len(arg615)) + defer mbTrans616.Close() + _, err617 := mbTrans616.WriteString(arg615) + if err617 != nil { Usage() return } - factory533 := thrift.NewTJSONProtocolFactory() - jsProt534 := factory533.GetProtocol(mbTrans531) + factory618 := thrift.NewTJSONProtocolFactory() + jsProt619 := factory618.GetProtocol(mbTrans616) argvalue0 := aurora.NewJobConfiguration() - err535 := argvalue0.Read(context.Background(), jsProt534) - if err535 != nil { + err620 := argvalue0.Read(context.Background(), jsProt619) + if err620 != nil { Usage() return } @@ -788,19 +849,19 @@ func main() { fmt.Fprintln(os.Stderr, "StartJobUpdate requires 2 args") flag.Usage() } - arg536 := flag.Arg(1) - mbTrans537 := thrift.NewTMemoryBufferLen(len(arg536)) - defer mbTrans537.Close() - _, err538 := mbTrans537.WriteString(arg536) - if err538 != nil { + arg621 := flag.Arg(1) + mbTrans622 := thrift.NewTMemoryBufferLen(len(arg621)) + defer mbTrans622.Close() + _, err623 := mbTrans622.WriteString(arg621) + if err623 != nil { Usage() return } - factory539 := thrift.NewTJSONProtocolFactory() - jsProt540 := factory539.GetProtocol(mbTrans537) + factory624 := thrift.NewTJSONProtocolFactory() + jsProt625 := factory624.GetProtocol(mbTrans622) argvalue0 := aurora.NewJobUpdateRequest() - err541 := argvalue0.Read(context.Background(), jsProt540) - if err541 != nil { + err626 := argvalue0.Read(context.Background(), jsProt625) + if err626 != nil { Usage() return } @@ -815,19 +876,19 @@ func main() { fmt.Fprintln(os.Stderr, "PauseJobUpdate requires 2 args") flag.Usage() } - arg543 := flag.Arg(1) - mbTrans544 := thrift.NewTMemoryBufferLen(len(arg543)) - defer mbTrans544.Close() - _, err545 := mbTrans544.WriteString(arg543) - if err545 != nil { + arg628 := flag.Arg(1) + mbTrans629 := thrift.NewTMemoryBufferLen(len(arg628)) + defer mbTrans629.Close() + _, err630 := mbTrans629.WriteString(arg628) + if err630 != nil { Usage() return } - factory546 := thrift.NewTJSONProtocolFactory() - jsProt547 := factory546.GetProtocol(mbTrans544) + factory631 := thrift.NewTJSONProtocolFactory() + jsProt632 := factory631.GetProtocol(mbTrans629) argvalue0 := aurora.NewJobUpdateKey() - err548 := argvalue0.Read(context.Background(), jsProt547) - if err548 != nil { + err633 := argvalue0.Read(context.Background(), jsProt632) + if err633 != nil { Usage() return } @@ -842,19 +903,19 @@ func main() { fmt.Fprintln(os.Stderr, "ResumeJobUpdate requires 2 args") flag.Usage() } - arg550 := flag.Arg(1) - mbTrans551 := thrift.NewTMemoryBufferLen(len(arg550)) - defer mbTrans551.Close() - _, err552 := mbTrans551.WriteString(arg550) - if err552 != nil { + arg635 := flag.Arg(1) + mbTrans636 := thrift.NewTMemoryBufferLen(len(arg635)) + defer mbTrans636.Close() + _, err637 := mbTrans636.WriteString(arg635) + if err637 != nil { Usage() return } - factory553 := thrift.NewTJSONProtocolFactory() - jsProt554 := factory553.GetProtocol(mbTrans551) + factory638 := thrift.NewTJSONProtocolFactory() + jsProt639 := factory638.GetProtocol(mbTrans636) argvalue0 := aurora.NewJobUpdateKey() - err555 := argvalue0.Read(context.Background(), jsProt554) - if err555 != nil { + err640 := argvalue0.Read(context.Background(), jsProt639) + if err640 != nil { Usage() return } @@ -869,19 +930,19 @@ func main() { fmt.Fprintln(os.Stderr, "AbortJobUpdate requires 2 args") flag.Usage() } - arg557 := flag.Arg(1) - mbTrans558 := thrift.NewTMemoryBufferLen(len(arg557)) - defer mbTrans558.Close() - _, err559 := mbTrans558.WriteString(arg557) - if err559 != nil { + arg642 := flag.Arg(1) + mbTrans643 := thrift.NewTMemoryBufferLen(len(arg642)) + defer mbTrans643.Close() + _, err644 := mbTrans643.WriteString(arg642) + if err644 != nil { Usage() return } - factory560 := thrift.NewTJSONProtocolFactory() - jsProt561 := factory560.GetProtocol(mbTrans558) + factory645 := thrift.NewTJSONProtocolFactory() + jsProt646 := factory645.GetProtocol(mbTrans643) argvalue0 := aurora.NewJobUpdateKey() - err562 := argvalue0.Read(context.Background(), jsProt561) - if err562 != nil { + err647 := argvalue0.Read(context.Background(), jsProt646) + if err647 != nil { Usage() return } @@ -896,19 +957,19 @@ func main() { fmt.Fprintln(os.Stderr, "RollbackJobUpdate requires 2 args") flag.Usage() } - arg564 := flag.Arg(1) - mbTrans565 := thrift.NewTMemoryBufferLen(len(arg564)) - defer mbTrans565.Close() - _, err566 := mbTrans565.WriteString(arg564) - if err566 != nil { + arg649 := flag.Arg(1) + mbTrans650 := thrift.NewTMemoryBufferLen(len(arg649)) + defer mbTrans650.Close() + _, err651 := mbTrans650.WriteString(arg649) + if err651 != nil { Usage() return } - factory567 := thrift.NewTJSONProtocolFactory() - jsProt568 := factory567.GetProtocol(mbTrans565) + factory652 := thrift.NewTJSONProtocolFactory() + jsProt653 := factory652.GetProtocol(mbTrans650) argvalue0 := aurora.NewJobUpdateKey() - err569 := argvalue0.Read(context.Background(), jsProt568) - if err569 != nil { + err654 := argvalue0.Read(context.Background(), jsProt653) + if err654 != nil { Usage() return } @@ -923,19 +984,19 @@ func main() { fmt.Fprintln(os.Stderr, "PulseJobUpdate requires 1 args") flag.Usage() } - arg571 := flag.Arg(1) - mbTrans572 := thrift.NewTMemoryBufferLen(len(arg571)) - defer mbTrans572.Close() - _, err573 := mbTrans572.WriteString(arg571) - if err573 != nil { + arg656 := flag.Arg(1) + mbTrans657 := thrift.NewTMemoryBufferLen(len(arg656)) + defer mbTrans657.Close() + _, err658 := mbTrans657.WriteString(arg656) + if err658 != nil { Usage() return } - factory574 := thrift.NewTJSONProtocolFactory() - jsProt575 := factory574.GetProtocol(mbTrans572) + factory659 := thrift.NewTJSONProtocolFactory() + jsProt660 := factory659.GetProtocol(mbTrans657) argvalue0 := aurora.NewJobUpdateKey() - err576 := argvalue0.Read(context.Background(), jsProt575) - if err576 != nil { + err661 := argvalue0.Read(context.Background(), jsProt660) + if err661 != nil { Usage() return } @@ -966,19 +1027,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetTasksStatus requires 1 args") flag.Usage() } - arg578 := flag.Arg(1) - mbTrans579 := thrift.NewTMemoryBufferLen(len(arg578)) - defer mbTrans579.Close() - _, err580 := mbTrans579.WriteString(arg578) - if err580 != nil { + arg663 := flag.Arg(1) + mbTrans664 := thrift.NewTMemoryBufferLen(len(arg663)) + defer mbTrans664.Close() + _, err665 := mbTrans664.WriteString(arg663) + if err665 != nil { Usage() return } - factory581 := thrift.NewTJSONProtocolFactory() - jsProt582 := factory581.GetProtocol(mbTrans579) + factory666 := thrift.NewTJSONProtocolFactory() + jsProt667 := factory666.GetProtocol(mbTrans664) argvalue0 := aurora.NewTaskQuery() - err583 := argvalue0.Read(context.Background(), jsProt582) - if err583 != nil { + err668 := argvalue0.Read(context.Background(), jsProt667) + if err668 != nil { Usage() return } @@ -991,19 +1052,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetTasksWithoutConfigs requires 1 args") flag.Usage() } - arg584 := flag.Arg(1) - mbTrans585 := thrift.NewTMemoryBufferLen(len(arg584)) - defer mbTrans585.Close() - _, err586 := mbTrans585.WriteString(arg584) - if err586 != nil { + arg669 := flag.Arg(1) + mbTrans670 := thrift.NewTMemoryBufferLen(len(arg669)) + defer mbTrans670.Close() + _, err671 := mbTrans670.WriteString(arg669) + if err671 != nil { Usage() return } - factory587 := thrift.NewTJSONProtocolFactory() - jsProt588 := factory587.GetProtocol(mbTrans585) + factory672 := thrift.NewTJSONProtocolFactory() + jsProt673 := factory672.GetProtocol(mbTrans670) argvalue0 := aurora.NewTaskQuery() - err589 := argvalue0.Read(context.Background(), jsProt588) - if err589 != nil { + err674 := argvalue0.Read(context.Background(), jsProt673) + if err674 != nil { Usage() return } @@ -1016,19 +1077,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetPendingReason requires 1 args") flag.Usage() } - arg590 := flag.Arg(1) - mbTrans591 := thrift.NewTMemoryBufferLen(len(arg590)) - defer mbTrans591.Close() - _, err592 := mbTrans591.WriteString(arg590) - if err592 != nil { + arg675 := flag.Arg(1) + mbTrans676 := thrift.NewTMemoryBufferLen(len(arg675)) + defer mbTrans676.Close() + _, err677 := mbTrans676.WriteString(arg675) + if err677 != nil { Usage() return } - factory593 := thrift.NewTJSONProtocolFactory() - jsProt594 := factory593.GetProtocol(mbTrans591) + factory678 := thrift.NewTJSONProtocolFactory() + jsProt679 := factory678.GetProtocol(mbTrans676) argvalue0 := aurora.NewTaskQuery() - err595 := argvalue0.Read(context.Background(), jsProt594) - if err595 != nil { + err680 := argvalue0.Read(context.Background(), jsProt679) + if err680 != nil { Usage() return } @@ -1041,19 +1102,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetConfigSummary requires 1 args") flag.Usage() } - arg596 := flag.Arg(1) - mbTrans597 := thrift.NewTMemoryBufferLen(len(arg596)) - defer mbTrans597.Close() - _, err598 := mbTrans597.WriteString(arg596) - if err598 != nil { + arg681 := flag.Arg(1) + mbTrans682 := thrift.NewTMemoryBufferLen(len(arg681)) + defer mbTrans682.Close() + _, err683 := mbTrans682.WriteString(arg681) + if err683 != nil { Usage() return } - factory599 := thrift.NewTJSONProtocolFactory() - jsProt600 := factory599.GetProtocol(mbTrans597) + factory684 := thrift.NewTJSONProtocolFactory() + jsProt685 := factory684.GetProtocol(mbTrans682) argvalue0 := aurora.NewJobKey() - err601 := argvalue0.Read(context.Background(), jsProt600) - if err601 != nil { + err686 := argvalue0.Read(context.Background(), jsProt685) + if err686 != nil { Usage() return } @@ -1086,19 +1147,19 @@ func main() { fmt.Fprintln(os.Stderr, "PopulateJobConfig requires 1 args") flag.Usage() } - arg604 := flag.Arg(1) - mbTrans605 := thrift.NewTMemoryBufferLen(len(arg604)) - defer mbTrans605.Close() - _, err606 := mbTrans605.WriteString(arg604) - if err606 != nil { + arg689 := flag.Arg(1) + mbTrans690 := thrift.NewTMemoryBufferLen(len(arg689)) + defer mbTrans690.Close() + _, err691 := mbTrans690.WriteString(arg689) + if err691 != nil { Usage() return } - factory607 := thrift.NewTJSONProtocolFactory() - jsProt608 := factory607.GetProtocol(mbTrans605) + factory692 := thrift.NewTJSONProtocolFactory() + jsProt693 := factory692.GetProtocol(mbTrans690) argvalue0 := aurora.NewJobConfiguration() - err609 := argvalue0.Read(context.Background(), jsProt608) - if err609 != nil { + err694 := argvalue0.Read(context.Background(), jsProt693) + if err694 != nil { Usage() return } @@ -1111,19 +1172,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateSummaries requires 1 args") flag.Usage() } - arg610 := flag.Arg(1) - mbTrans611 := thrift.NewTMemoryBufferLen(len(arg610)) - defer mbTrans611.Close() - _, err612 := mbTrans611.WriteString(arg610) - if err612 != nil { + arg695 := flag.Arg(1) + mbTrans696 := thrift.NewTMemoryBufferLen(len(arg695)) + defer mbTrans696.Close() + _, err697 := mbTrans696.WriteString(arg695) + if err697 != nil { Usage() return } - factory613 := thrift.NewTJSONProtocolFactory() - jsProt614 := factory613.GetProtocol(mbTrans611) + factory698 := thrift.NewTJSONProtocolFactory() + jsProt699 := factory698.GetProtocol(mbTrans696) argvalue0 := aurora.NewJobUpdateQuery() - err615 := argvalue0.Read(context.Background(), jsProt614) - if err615 != nil { + err700 := argvalue0.Read(context.Background(), jsProt699) + if err700 != nil { Usage() return } @@ -1136,19 +1197,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateDetails requires 1 args") flag.Usage() } - arg616 := flag.Arg(1) - mbTrans617 := thrift.NewTMemoryBufferLen(len(arg616)) - defer mbTrans617.Close() - _, err618 := mbTrans617.WriteString(arg616) - if err618 != nil { + arg701 := flag.Arg(1) + mbTrans702 := thrift.NewTMemoryBufferLen(len(arg701)) + defer mbTrans702.Close() + _, err703 := mbTrans702.WriteString(arg701) + if err703 != nil { Usage() return } - factory619 := thrift.NewTJSONProtocolFactory() - jsProt620 := factory619.GetProtocol(mbTrans617) + factory704 := thrift.NewTJSONProtocolFactory() + jsProt705 := factory704.GetProtocol(mbTrans702) argvalue0 := aurora.NewJobUpdateQuery() - err621 := argvalue0.Read(context.Background(), jsProt620) - if err621 != nil { + err706 := argvalue0.Read(context.Background(), jsProt705) + if err706 != nil { Usage() return } @@ -1161,19 +1222,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateDiff requires 1 args") flag.Usage() } - arg622 := flag.Arg(1) - mbTrans623 := thrift.NewTMemoryBufferLen(len(arg622)) - defer mbTrans623.Close() - _, err624 := mbTrans623.WriteString(arg622) - if err624 != nil { + arg707 := flag.Arg(1) + mbTrans708 := thrift.NewTMemoryBufferLen(len(arg707)) + defer mbTrans708.Close() + _, err709 := mbTrans708.WriteString(arg707) + if err709 != nil { Usage() return } - factory625 := thrift.NewTJSONProtocolFactory() - jsProt626 := factory625.GetProtocol(mbTrans623) + factory710 := thrift.NewTJSONProtocolFactory() + jsProt711 := factory710.GetProtocol(mbTrans708) argvalue0 := aurora.NewJobUpdateRequest() - err627 := argvalue0.Read(context.Background(), jsProt626) - if err627 != nil { + err712 := argvalue0.Read(context.Background(), jsProt711) + if err712 != nil { Usage() return } diff --git a/gen-go/apache/aurora/aurora_scheduler_manager-remote/aurora_scheduler_manager-remote.go b/gen-go/apache/aurora/aurora_scheduler_manager-remote/aurora_scheduler_manager-remote.go index b4715d5..5bcb59b 100755 --- a/gen-go/apache/aurora/aurora_scheduler_manager-remote/aurora_scheduler_manager-remote.go +++ b/gen-go/apache/aurora/aurora_scheduler_manager-remote/aurora_scheduler_manager-remote.go @@ -1,4 +1,4 @@ -// Code generated by Thrift Compiler (0.14.0). DO NOT EDIT. +// Code generated by Thrift Compiler (0.14.1). DO NOT EDIT. package main @@ -13,7 +13,7 @@ import ( "strconv" "strings" "github.com/apache/thrift/lib/go/thrift" - "apache/aurora" + "aurora" ) var _ = aurora.GoUnusedProtection__ @@ -27,6 +27,7 @@ func Usage() { fmt.Fprintln(os.Stderr, " Response descheduleCronJob(JobKey job)") fmt.Fprintln(os.Stderr, " Response startCronJob(JobKey job)") fmt.Fprintln(os.Stderr, " Response restartShards(JobKey job, shardIds)") + fmt.Fprintln(os.Stderr, " Response slaRestartShards(JobKey job, shardIds, SlaPolicy slaPolicy)") fmt.Fprintln(os.Stderr, " Response killTasks(JobKey job, instances, string message)") fmt.Fprintln(os.Stderr, " Response addInstances(InstanceKey key, i32 count)") fmt.Fprintln(os.Stderr, " Response replaceCronTemplate(JobConfiguration config)") @@ -175,19 +176,19 @@ func main() { fmt.Fprintln(os.Stderr, "CreateJob requires 1 args") flag.Usage() } - arg213 := flag.Arg(1) - mbTrans214 := thrift.NewTMemoryBufferLen(len(arg213)) - defer mbTrans214.Close() - _, err215 := mbTrans214.WriteString(arg213) - if err215 != nil { + arg244 := flag.Arg(1) + mbTrans245 := thrift.NewTMemoryBufferLen(len(arg244)) + defer mbTrans245.Close() + _, err246 := mbTrans245.WriteString(arg244) + if err246 != nil { Usage() return } - factory216 := thrift.NewTJSONProtocolFactory() - jsProt217 := factory216.GetProtocol(mbTrans214) + factory247 := thrift.NewTJSONProtocolFactory() + jsProt248 := factory247.GetProtocol(mbTrans245) argvalue0 := aurora.NewJobConfiguration() - err218 := argvalue0.Read(context.Background(), jsProt217) - if err218 != nil { + err249 := argvalue0.Read(context.Background(), jsProt248) + if err249 != nil { Usage() return } @@ -200,19 +201,19 @@ func main() { fmt.Fprintln(os.Stderr, "ScheduleCronJob requires 1 args") flag.Usage() } - arg219 := flag.Arg(1) - mbTrans220 := thrift.NewTMemoryBufferLen(len(arg219)) - defer mbTrans220.Close() - _, err221 := mbTrans220.WriteString(arg219) - if err221 != nil { + arg250 := flag.Arg(1) + mbTrans251 := thrift.NewTMemoryBufferLen(len(arg250)) + defer mbTrans251.Close() + _, err252 := mbTrans251.WriteString(arg250) + if err252 != nil { Usage() return } - factory222 := thrift.NewTJSONProtocolFactory() - jsProt223 := factory222.GetProtocol(mbTrans220) + factory253 := thrift.NewTJSONProtocolFactory() + jsProt254 := factory253.GetProtocol(mbTrans251) argvalue0 := aurora.NewJobConfiguration() - err224 := argvalue0.Read(context.Background(), jsProt223) - if err224 != nil { + err255 := argvalue0.Read(context.Background(), jsProt254) + if err255 != nil { Usage() return } @@ -225,19 +226,19 @@ func main() { fmt.Fprintln(os.Stderr, "DescheduleCronJob requires 1 args") flag.Usage() } - arg225 := flag.Arg(1) - mbTrans226 := thrift.NewTMemoryBufferLen(len(arg225)) - defer mbTrans226.Close() - _, err227 := mbTrans226.WriteString(arg225) - if err227 != nil { + arg256 := flag.Arg(1) + mbTrans257 := thrift.NewTMemoryBufferLen(len(arg256)) + defer mbTrans257.Close() + _, err258 := mbTrans257.WriteString(arg256) + if err258 != nil { Usage() return } - factory228 := thrift.NewTJSONProtocolFactory() - jsProt229 := factory228.GetProtocol(mbTrans226) + factory259 := thrift.NewTJSONProtocolFactory() + jsProt260 := factory259.GetProtocol(mbTrans257) argvalue0 := aurora.NewJobKey() - err230 := argvalue0.Read(context.Background(), jsProt229) - if err230 != nil { + err261 := argvalue0.Read(context.Background(), jsProt260) + if err261 != nil { Usage() return } @@ -250,19 +251,19 @@ func main() { fmt.Fprintln(os.Stderr, "StartCronJob requires 1 args") flag.Usage() } - arg231 := flag.Arg(1) - mbTrans232 := thrift.NewTMemoryBufferLen(len(arg231)) - defer mbTrans232.Close() - _, err233 := mbTrans232.WriteString(arg231) - if err233 != nil { + arg262 := flag.Arg(1) + mbTrans263 := thrift.NewTMemoryBufferLen(len(arg262)) + defer mbTrans263.Close() + _, err264 := mbTrans263.WriteString(arg262) + if err264 != nil { Usage() return } - factory234 := thrift.NewTJSONProtocolFactory() - jsProt235 := factory234.GetProtocol(mbTrans232) + factory265 := thrift.NewTJSONProtocolFactory() + jsProt266 := factory265.GetProtocol(mbTrans263) argvalue0 := aurora.NewJobKey() - err236 := argvalue0.Read(context.Background(), jsProt235) - if err236 != nil { + err267 := argvalue0.Read(context.Background(), jsProt266) + if err267 != nil { Usage() return } @@ -275,36 +276,36 @@ func main() { fmt.Fprintln(os.Stderr, "RestartShards requires 2 args") flag.Usage() } - arg237 := flag.Arg(1) - mbTrans238 := thrift.NewTMemoryBufferLen(len(arg237)) - defer mbTrans238.Close() - _, err239 := mbTrans238.WriteString(arg237) - if err239 != nil { + arg268 := flag.Arg(1) + mbTrans269 := thrift.NewTMemoryBufferLen(len(arg268)) + defer mbTrans269.Close() + _, err270 := mbTrans269.WriteString(arg268) + if err270 != nil { Usage() return } - factory240 := thrift.NewTJSONProtocolFactory() - jsProt241 := factory240.GetProtocol(mbTrans238) + factory271 := thrift.NewTJSONProtocolFactory() + jsProt272 := factory271.GetProtocol(mbTrans269) argvalue0 := aurora.NewJobKey() - err242 := argvalue0.Read(context.Background(), jsProt241) - if err242 != nil { + err273 := argvalue0.Read(context.Background(), jsProt272) + if err273 != nil { Usage() return } value0 := argvalue0 - arg243 := flag.Arg(2) - mbTrans244 := thrift.NewTMemoryBufferLen(len(arg243)) - defer mbTrans244.Close() - _, err245 := mbTrans244.WriteString(arg243) - if err245 != nil { + arg274 := flag.Arg(2) + mbTrans275 := thrift.NewTMemoryBufferLen(len(arg274)) + defer mbTrans275.Close() + _, err276 := mbTrans275.WriteString(arg274) + if err276 != nil { Usage() return } - factory246 := thrift.NewTJSONProtocolFactory() - jsProt247 := factory246.GetProtocol(mbTrans244) + factory277 := thrift.NewTJSONProtocolFactory() + jsProt278 := factory277.GetProtocol(mbTrans275) containerStruct1 := aurora.NewAuroraSchedulerManagerRestartShardsArgs() - err248 := containerStruct1.ReadField2(context.Background(), jsProt247) - if err248 != nil { + err279 := containerStruct1.ReadField2(context.Background(), jsProt278) + if err279 != nil { Usage() return } @@ -313,41 +314,101 @@ func main() { fmt.Print(client.RestartShards(context.Background(), value0, value1)) fmt.Print("\n") break + case "slaRestartShards": + if flag.NArg() - 1 != 3 { + fmt.Fprintln(os.Stderr, "SlaRestartShards requires 3 args") + flag.Usage() + } + arg280 := flag.Arg(1) + mbTrans281 := thrift.NewTMemoryBufferLen(len(arg280)) + defer mbTrans281.Close() + _, err282 := mbTrans281.WriteString(arg280) + if err282 != nil { + Usage() + return + } + factory283 := thrift.NewTJSONProtocolFactory() + jsProt284 := factory283.GetProtocol(mbTrans281) + argvalue0 := aurora.NewJobKey() + err285 := argvalue0.Read(context.Background(), jsProt284) + if err285 != nil { + Usage() + return + } + value0 := argvalue0 + arg286 := flag.Arg(2) + mbTrans287 := thrift.NewTMemoryBufferLen(len(arg286)) + defer mbTrans287.Close() + _, err288 := mbTrans287.WriteString(arg286) + if err288 != nil { + Usage() + return + } + factory289 := thrift.NewTJSONProtocolFactory() + jsProt290 := factory289.GetProtocol(mbTrans287) + containerStruct1 := aurora.NewAuroraSchedulerManagerSlaRestartShardsArgs() + err291 := containerStruct1.ReadField2(context.Background(), jsProt290) + if err291 != nil { + Usage() + return + } + argvalue1 := containerStruct1.ShardIds + value1 := argvalue1 + arg292 := flag.Arg(3) + mbTrans293 := thrift.NewTMemoryBufferLen(len(arg292)) + defer mbTrans293.Close() + _, err294 := mbTrans293.WriteString(arg292) + if err294 != nil { + Usage() + return + } + factory295 := thrift.NewTJSONProtocolFactory() + jsProt296 := factory295.GetProtocol(mbTrans293) + argvalue2 := aurora.NewSlaPolicy() + err297 := argvalue2.Read(context.Background(), jsProt296) + if err297 != nil { + Usage() + return + } + value2 := argvalue2 + fmt.Print(client.SlaRestartShards(context.Background(), value0, value1, value2)) + fmt.Print("\n") + break case "killTasks": if flag.NArg() - 1 != 3 { fmt.Fprintln(os.Stderr, "KillTasks requires 3 args") flag.Usage() } - arg249 := flag.Arg(1) - mbTrans250 := thrift.NewTMemoryBufferLen(len(arg249)) - defer mbTrans250.Close() - _, err251 := mbTrans250.WriteString(arg249) - if err251 != nil { + arg298 := flag.Arg(1) + mbTrans299 := thrift.NewTMemoryBufferLen(len(arg298)) + defer mbTrans299.Close() + _, err300 := mbTrans299.WriteString(arg298) + if err300 != nil { Usage() return } - factory252 := thrift.NewTJSONProtocolFactory() - jsProt253 := factory252.GetProtocol(mbTrans250) + factory301 := thrift.NewTJSONProtocolFactory() + jsProt302 := factory301.GetProtocol(mbTrans299) argvalue0 := aurora.NewJobKey() - err254 := argvalue0.Read(context.Background(), jsProt253) - if err254 != nil { + err303 := argvalue0.Read(context.Background(), jsProt302) + if err303 != nil { Usage() return } value0 := argvalue0 - arg255 := flag.Arg(2) - mbTrans256 := thrift.NewTMemoryBufferLen(len(arg255)) - defer mbTrans256.Close() - _, err257 := mbTrans256.WriteString(arg255) - if err257 != nil { + arg304 := flag.Arg(2) + mbTrans305 := thrift.NewTMemoryBufferLen(len(arg304)) + defer mbTrans305.Close() + _, err306 := mbTrans305.WriteString(arg304) + if err306 != nil { Usage() return } - factory258 := thrift.NewTJSONProtocolFactory() - jsProt259 := factory258.GetProtocol(mbTrans256) + factory307 := thrift.NewTJSONProtocolFactory() + jsProt308 := factory307.GetProtocol(mbTrans305) containerStruct1 := aurora.NewAuroraSchedulerManagerKillTasksArgs() - err260 := containerStruct1.ReadField2(context.Background(), jsProt259) - if err260 != nil { + err309 := containerStruct1.ReadField2(context.Background(), jsProt308) + if err309 != nil { Usage() return } @@ -363,25 +424,25 @@ func main() { fmt.Fprintln(os.Stderr, "AddInstances requires 2 args") flag.Usage() } - arg262 := flag.Arg(1) - mbTrans263 := thrift.NewTMemoryBufferLen(len(arg262)) - defer mbTrans263.Close() - _, err264 := mbTrans263.WriteString(arg262) - if err264 != nil { + arg311 := flag.Arg(1) + mbTrans312 := thrift.NewTMemoryBufferLen(len(arg311)) + defer mbTrans312.Close() + _, err313 := mbTrans312.WriteString(arg311) + if err313 != nil { Usage() return } - factory265 := thrift.NewTJSONProtocolFactory() - jsProt266 := factory265.GetProtocol(mbTrans263) + factory314 := thrift.NewTJSONProtocolFactory() + jsProt315 := factory314.GetProtocol(mbTrans312) argvalue0 := aurora.NewInstanceKey() - err267 := argvalue0.Read(context.Background(), jsProt266) - if err267 != nil { + err316 := argvalue0.Read(context.Background(), jsProt315) + if err316 != nil { Usage() return } value0 := argvalue0 - tmp1, err268 := (strconv.Atoi(flag.Arg(2))) - if err268 != nil { + tmp1, err317 := (strconv.Atoi(flag.Arg(2))) + if err317 != nil { Usage() return } @@ -395,19 +456,19 @@ func main() { fmt.Fprintln(os.Stderr, "ReplaceCronTemplate requires 1 args") flag.Usage() } - arg269 := flag.Arg(1) - mbTrans270 := thrift.NewTMemoryBufferLen(len(arg269)) - defer mbTrans270.Close() - _, err271 := mbTrans270.WriteString(arg269) - if err271 != nil { + arg318 := flag.Arg(1) + mbTrans319 := thrift.NewTMemoryBufferLen(len(arg318)) + defer mbTrans319.Close() + _, err320 := mbTrans319.WriteString(arg318) + if err320 != nil { Usage() return } - factory272 := thrift.NewTJSONProtocolFactory() - jsProt273 := factory272.GetProtocol(mbTrans270) + factory321 := thrift.NewTJSONProtocolFactory() + jsProt322 := factory321.GetProtocol(mbTrans319) argvalue0 := aurora.NewJobConfiguration() - err274 := argvalue0.Read(context.Background(), jsProt273) - if err274 != nil { + err323 := argvalue0.Read(context.Background(), jsProt322) + if err323 != nil { Usage() return } @@ -420,19 +481,19 @@ func main() { fmt.Fprintln(os.Stderr, "StartJobUpdate requires 2 args") flag.Usage() } - arg275 := flag.Arg(1) - mbTrans276 := thrift.NewTMemoryBufferLen(len(arg275)) - defer mbTrans276.Close() - _, err277 := mbTrans276.WriteString(arg275) - if err277 != nil { + arg324 := flag.Arg(1) + mbTrans325 := thrift.NewTMemoryBufferLen(len(arg324)) + defer mbTrans325.Close() + _, err326 := mbTrans325.WriteString(arg324) + if err326 != nil { Usage() return } - factory278 := thrift.NewTJSONProtocolFactory() - jsProt279 := factory278.GetProtocol(mbTrans276) + factory327 := thrift.NewTJSONProtocolFactory() + jsProt328 := factory327.GetProtocol(mbTrans325) argvalue0 := aurora.NewJobUpdateRequest() - err280 := argvalue0.Read(context.Background(), jsProt279) - if err280 != nil { + err329 := argvalue0.Read(context.Background(), jsProt328) + if err329 != nil { Usage() return } @@ -447,19 +508,19 @@ func main() { fmt.Fprintln(os.Stderr, "PauseJobUpdate requires 2 args") flag.Usage() } - arg282 := flag.Arg(1) - mbTrans283 := thrift.NewTMemoryBufferLen(len(arg282)) - defer mbTrans283.Close() - _, err284 := mbTrans283.WriteString(arg282) - if err284 != nil { + arg331 := flag.Arg(1) + mbTrans332 := thrift.NewTMemoryBufferLen(len(arg331)) + defer mbTrans332.Close() + _, err333 := mbTrans332.WriteString(arg331) + if err333 != nil { Usage() return } - factory285 := thrift.NewTJSONProtocolFactory() - jsProt286 := factory285.GetProtocol(mbTrans283) + factory334 := thrift.NewTJSONProtocolFactory() + jsProt335 := factory334.GetProtocol(mbTrans332) argvalue0 := aurora.NewJobUpdateKey() - err287 := argvalue0.Read(context.Background(), jsProt286) - if err287 != nil { + err336 := argvalue0.Read(context.Background(), jsProt335) + if err336 != nil { Usage() return } @@ -474,19 +535,19 @@ func main() { fmt.Fprintln(os.Stderr, "ResumeJobUpdate requires 2 args") flag.Usage() } - arg289 := flag.Arg(1) - mbTrans290 := thrift.NewTMemoryBufferLen(len(arg289)) - defer mbTrans290.Close() - _, err291 := mbTrans290.WriteString(arg289) - if err291 != nil { + arg338 := flag.Arg(1) + mbTrans339 := thrift.NewTMemoryBufferLen(len(arg338)) + defer mbTrans339.Close() + _, err340 := mbTrans339.WriteString(arg338) + if err340 != nil { Usage() return } - factory292 := thrift.NewTJSONProtocolFactory() - jsProt293 := factory292.GetProtocol(mbTrans290) + factory341 := thrift.NewTJSONProtocolFactory() + jsProt342 := factory341.GetProtocol(mbTrans339) argvalue0 := aurora.NewJobUpdateKey() - err294 := argvalue0.Read(context.Background(), jsProt293) - if err294 != nil { + err343 := argvalue0.Read(context.Background(), jsProt342) + if err343 != nil { Usage() return } @@ -501,19 +562,19 @@ func main() { fmt.Fprintln(os.Stderr, "AbortJobUpdate requires 2 args") flag.Usage() } - arg296 := flag.Arg(1) - mbTrans297 := thrift.NewTMemoryBufferLen(len(arg296)) - defer mbTrans297.Close() - _, err298 := mbTrans297.WriteString(arg296) - if err298 != nil { + arg345 := flag.Arg(1) + mbTrans346 := thrift.NewTMemoryBufferLen(len(arg345)) + defer mbTrans346.Close() + _, err347 := mbTrans346.WriteString(arg345) + if err347 != nil { Usage() return } - factory299 := thrift.NewTJSONProtocolFactory() - jsProt300 := factory299.GetProtocol(mbTrans297) + factory348 := thrift.NewTJSONProtocolFactory() + jsProt349 := factory348.GetProtocol(mbTrans346) argvalue0 := aurora.NewJobUpdateKey() - err301 := argvalue0.Read(context.Background(), jsProt300) - if err301 != nil { + err350 := argvalue0.Read(context.Background(), jsProt349) + if err350 != nil { Usage() return } @@ -528,19 +589,19 @@ func main() { fmt.Fprintln(os.Stderr, "RollbackJobUpdate requires 2 args") flag.Usage() } - arg303 := flag.Arg(1) - mbTrans304 := thrift.NewTMemoryBufferLen(len(arg303)) - defer mbTrans304.Close() - _, err305 := mbTrans304.WriteString(arg303) - if err305 != nil { + arg352 := flag.Arg(1) + mbTrans353 := thrift.NewTMemoryBufferLen(len(arg352)) + defer mbTrans353.Close() + _, err354 := mbTrans353.WriteString(arg352) + if err354 != nil { Usage() return } - factory306 := thrift.NewTJSONProtocolFactory() - jsProt307 := factory306.GetProtocol(mbTrans304) + factory355 := thrift.NewTJSONProtocolFactory() + jsProt356 := factory355.GetProtocol(mbTrans353) argvalue0 := aurora.NewJobUpdateKey() - err308 := argvalue0.Read(context.Background(), jsProt307) - if err308 != nil { + err357 := argvalue0.Read(context.Background(), jsProt356) + if err357 != nil { Usage() return } @@ -555,19 +616,19 @@ func main() { fmt.Fprintln(os.Stderr, "PulseJobUpdate requires 1 args") flag.Usage() } - arg310 := flag.Arg(1) - mbTrans311 := thrift.NewTMemoryBufferLen(len(arg310)) - defer mbTrans311.Close() - _, err312 := mbTrans311.WriteString(arg310) - if err312 != nil { + arg359 := flag.Arg(1) + mbTrans360 := thrift.NewTMemoryBufferLen(len(arg359)) + defer mbTrans360.Close() + _, err361 := mbTrans360.WriteString(arg359) + if err361 != nil { Usage() return } - factory313 := thrift.NewTJSONProtocolFactory() - jsProt314 := factory313.GetProtocol(mbTrans311) + factory362 := thrift.NewTJSONProtocolFactory() + jsProt363 := factory362.GetProtocol(mbTrans360) argvalue0 := aurora.NewJobUpdateKey() - err315 := argvalue0.Read(context.Background(), jsProt314) - if err315 != nil { + err364 := argvalue0.Read(context.Background(), jsProt363) + if err364 != nil { Usage() return } @@ -598,19 +659,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetTasksStatus requires 1 args") flag.Usage() } - arg317 := flag.Arg(1) - mbTrans318 := thrift.NewTMemoryBufferLen(len(arg317)) - defer mbTrans318.Close() - _, err319 := mbTrans318.WriteString(arg317) - if err319 != nil { + arg366 := flag.Arg(1) + mbTrans367 := thrift.NewTMemoryBufferLen(len(arg366)) + defer mbTrans367.Close() + _, err368 := mbTrans367.WriteString(arg366) + if err368 != nil { Usage() return } - factory320 := thrift.NewTJSONProtocolFactory() - jsProt321 := factory320.GetProtocol(mbTrans318) + factory369 := thrift.NewTJSONProtocolFactory() + jsProt370 := factory369.GetProtocol(mbTrans367) argvalue0 := aurora.NewTaskQuery() - err322 := argvalue0.Read(context.Background(), jsProt321) - if err322 != nil { + err371 := argvalue0.Read(context.Background(), jsProt370) + if err371 != nil { Usage() return } @@ -623,19 +684,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetTasksWithoutConfigs requires 1 args") flag.Usage() } - arg323 := flag.Arg(1) - mbTrans324 := thrift.NewTMemoryBufferLen(len(arg323)) - defer mbTrans324.Close() - _, err325 := mbTrans324.WriteString(arg323) - if err325 != nil { + arg372 := flag.Arg(1) + mbTrans373 := thrift.NewTMemoryBufferLen(len(arg372)) + defer mbTrans373.Close() + _, err374 := mbTrans373.WriteString(arg372) + if err374 != nil { Usage() return } - factory326 := thrift.NewTJSONProtocolFactory() - jsProt327 := factory326.GetProtocol(mbTrans324) + factory375 := thrift.NewTJSONProtocolFactory() + jsProt376 := factory375.GetProtocol(mbTrans373) argvalue0 := aurora.NewTaskQuery() - err328 := argvalue0.Read(context.Background(), jsProt327) - if err328 != nil { + err377 := argvalue0.Read(context.Background(), jsProt376) + if err377 != nil { Usage() return } @@ -648,19 +709,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetPendingReason requires 1 args") flag.Usage() } - arg329 := flag.Arg(1) - mbTrans330 := thrift.NewTMemoryBufferLen(len(arg329)) - defer mbTrans330.Close() - _, err331 := mbTrans330.WriteString(arg329) - if err331 != nil { + arg378 := flag.Arg(1) + mbTrans379 := thrift.NewTMemoryBufferLen(len(arg378)) + defer mbTrans379.Close() + _, err380 := mbTrans379.WriteString(arg378) + if err380 != nil { Usage() return } - factory332 := thrift.NewTJSONProtocolFactory() - jsProt333 := factory332.GetProtocol(mbTrans330) + factory381 := thrift.NewTJSONProtocolFactory() + jsProt382 := factory381.GetProtocol(mbTrans379) argvalue0 := aurora.NewTaskQuery() - err334 := argvalue0.Read(context.Background(), jsProt333) - if err334 != nil { + err383 := argvalue0.Read(context.Background(), jsProt382) + if err383 != nil { Usage() return } @@ -673,19 +734,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetConfigSummary requires 1 args") flag.Usage() } - arg335 := flag.Arg(1) - mbTrans336 := thrift.NewTMemoryBufferLen(len(arg335)) - defer mbTrans336.Close() - _, err337 := mbTrans336.WriteString(arg335) - if err337 != nil { + arg384 := flag.Arg(1) + mbTrans385 := thrift.NewTMemoryBufferLen(len(arg384)) + defer mbTrans385.Close() + _, err386 := mbTrans385.WriteString(arg384) + if err386 != nil { Usage() return } - factory338 := thrift.NewTJSONProtocolFactory() - jsProt339 := factory338.GetProtocol(mbTrans336) + factory387 := thrift.NewTJSONProtocolFactory() + jsProt388 := factory387.GetProtocol(mbTrans385) argvalue0 := aurora.NewJobKey() - err340 := argvalue0.Read(context.Background(), jsProt339) - if err340 != nil { + err389 := argvalue0.Read(context.Background(), jsProt388) + if err389 != nil { Usage() return } @@ -718,19 +779,19 @@ func main() { fmt.Fprintln(os.Stderr, "PopulateJobConfig requires 1 args") flag.Usage() } - arg343 := flag.Arg(1) - mbTrans344 := thrift.NewTMemoryBufferLen(len(arg343)) - defer mbTrans344.Close() - _, err345 := mbTrans344.WriteString(arg343) - if err345 != nil { + arg392 := flag.Arg(1) + mbTrans393 := thrift.NewTMemoryBufferLen(len(arg392)) + defer mbTrans393.Close() + _, err394 := mbTrans393.WriteString(arg392) + if err394 != nil { Usage() return } - factory346 := thrift.NewTJSONProtocolFactory() - jsProt347 := factory346.GetProtocol(mbTrans344) + factory395 := thrift.NewTJSONProtocolFactory() + jsProt396 := factory395.GetProtocol(mbTrans393) argvalue0 := aurora.NewJobConfiguration() - err348 := argvalue0.Read(context.Background(), jsProt347) - if err348 != nil { + err397 := argvalue0.Read(context.Background(), jsProt396) + if err397 != nil { Usage() return } @@ -743,19 +804,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateSummaries requires 1 args") flag.Usage() } - arg349 := flag.Arg(1) - mbTrans350 := thrift.NewTMemoryBufferLen(len(arg349)) - defer mbTrans350.Close() - _, err351 := mbTrans350.WriteString(arg349) - if err351 != nil { + arg398 := flag.Arg(1) + mbTrans399 := thrift.NewTMemoryBufferLen(len(arg398)) + defer mbTrans399.Close() + _, err400 := mbTrans399.WriteString(arg398) + if err400 != nil { Usage() return } - factory352 := thrift.NewTJSONProtocolFactory() - jsProt353 := factory352.GetProtocol(mbTrans350) + factory401 := thrift.NewTJSONProtocolFactory() + jsProt402 := factory401.GetProtocol(mbTrans399) argvalue0 := aurora.NewJobUpdateQuery() - err354 := argvalue0.Read(context.Background(), jsProt353) - if err354 != nil { + err403 := argvalue0.Read(context.Background(), jsProt402) + if err403 != nil { Usage() return } @@ -768,19 +829,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateDetails requires 1 args") flag.Usage() } - arg355 := flag.Arg(1) - mbTrans356 := thrift.NewTMemoryBufferLen(len(arg355)) - defer mbTrans356.Close() - _, err357 := mbTrans356.WriteString(arg355) - if err357 != nil { + arg404 := flag.Arg(1) + mbTrans405 := thrift.NewTMemoryBufferLen(len(arg404)) + defer mbTrans405.Close() + _, err406 := mbTrans405.WriteString(arg404) + if err406 != nil { Usage() return } - factory358 := thrift.NewTJSONProtocolFactory() - jsProt359 := factory358.GetProtocol(mbTrans356) + factory407 := thrift.NewTJSONProtocolFactory() + jsProt408 := factory407.GetProtocol(mbTrans405) argvalue0 := aurora.NewJobUpdateQuery() - err360 := argvalue0.Read(context.Background(), jsProt359) - if err360 != nil { + err409 := argvalue0.Read(context.Background(), jsProt408) + if err409 != nil { Usage() return } @@ -793,19 +854,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateDiff requires 1 args") flag.Usage() } - arg361 := flag.Arg(1) - mbTrans362 := thrift.NewTMemoryBufferLen(len(arg361)) - defer mbTrans362.Close() - _, err363 := mbTrans362.WriteString(arg361) - if err363 != nil { + arg410 := flag.Arg(1) + mbTrans411 := thrift.NewTMemoryBufferLen(len(arg410)) + defer mbTrans411.Close() + _, err412 := mbTrans411.WriteString(arg410) + if err412 != nil { Usage() return } - factory364 := thrift.NewTJSONProtocolFactory() - jsProt365 := factory364.GetProtocol(mbTrans362) + factory413 := thrift.NewTJSONProtocolFactory() + jsProt414 := factory413.GetProtocol(mbTrans411) argvalue0 := aurora.NewJobUpdateRequest() - err366 := argvalue0.Read(context.Background(), jsProt365) - if err366 != nil { + err415 := argvalue0.Read(context.Background(), jsProt414) + if err415 != nil { Usage() return } diff --git a/gen-go/apache/aurora/read_only_scheduler-remote/read_only_scheduler-remote.go b/gen-go/apache/aurora/read_only_scheduler-remote/read_only_scheduler-remote.go index e802328..f306258 100755 --- a/gen-go/apache/aurora/read_only_scheduler-remote/read_only_scheduler-remote.go +++ b/gen-go/apache/aurora/read_only_scheduler-remote/read_only_scheduler-remote.go @@ -1,4 +1,4 @@ -// Code generated by Thrift Compiler (0.14.0). DO NOT EDIT. +// Code generated by Thrift Compiler (0.14.1). DO NOT EDIT. package main @@ -13,7 +13,7 @@ import ( "strconv" "strings" "github.com/apache/thrift/lib/go/thrift" - "apache/aurora" + "aurora" ) var _ = aurora.GoUnusedProtection__ @@ -179,19 +179,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetTasksStatus requires 1 args") flag.Usage() } - arg132 := flag.Arg(1) - mbTrans133 := thrift.NewTMemoryBufferLen(len(arg132)) - defer mbTrans133.Close() - _, err134 := mbTrans133.WriteString(arg132) - if err134 != nil { + arg145 := flag.Arg(1) + mbTrans146 := thrift.NewTMemoryBufferLen(len(arg145)) + defer mbTrans146.Close() + _, err147 := mbTrans146.WriteString(arg145) + if err147 != nil { Usage() return } - factory135 := thrift.NewTJSONProtocolFactory() - jsProt136 := factory135.GetProtocol(mbTrans133) + factory148 := thrift.NewTJSONProtocolFactory() + jsProt149 := factory148.GetProtocol(mbTrans146) argvalue0 := aurora.NewTaskQuery() - err137 := argvalue0.Read(context.Background(), jsProt136) - if err137 != nil { + err150 := argvalue0.Read(context.Background(), jsProt149) + if err150 != nil { Usage() return } @@ -204,19 +204,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetTasksWithoutConfigs requires 1 args") flag.Usage() } - arg138 := flag.Arg(1) - mbTrans139 := thrift.NewTMemoryBufferLen(len(arg138)) - defer mbTrans139.Close() - _, err140 := mbTrans139.WriteString(arg138) - if err140 != nil { + arg151 := flag.Arg(1) + mbTrans152 := thrift.NewTMemoryBufferLen(len(arg151)) + defer mbTrans152.Close() + _, err153 := mbTrans152.WriteString(arg151) + if err153 != nil { Usage() return } - factory141 := thrift.NewTJSONProtocolFactory() - jsProt142 := factory141.GetProtocol(mbTrans139) + factory154 := thrift.NewTJSONProtocolFactory() + jsProt155 := factory154.GetProtocol(mbTrans152) argvalue0 := aurora.NewTaskQuery() - err143 := argvalue0.Read(context.Background(), jsProt142) - if err143 != nil { + err156 := argvalue0.Read(context.Background(), jsProt155) + if err156 != nil { Usage() return } @@ -229,19 +229,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetPendingReason requires 1 args") flag.Usage() } - arg144 := flag.Arg(1) - mbTrans145 := thrift.NewTMemoryBufferLen(len(arg144)) - defer mbTrans145.Close() - _, err146 := mbTrans145.WriteString(arg144) - if err146 != nil { + arg157 := flag.Arg(1) + mbTrans158 := thrift.NewTMemoryBufferLen(len(arg157)) + defer mbTrans158.Close() + _, err159 := mbTrans158.WriteString(arg157) + if err159 != nil { Usage() return } - factory147 := thrift.NewTJSONProtocolFactory() - jsProt148 := factory147.GetProtocol(mbTrans145) + factory160 := thrift.NewTJSONProtocolFactory() + jsProt161 := factory160.GetProtocol(mbTrans158) argvalue0 := aurora.NewTaskQuery() - err149 := argvalue0.Read(context.Background(), jsProt148) - if err149 != nil { + err162 := argvalue0.Read(context.Background(), jsProt161) + if err162 != nil { Usage() return } @@ -254,19 +254,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetConfigSummary requires 1 args") flag.Usage() } - arg150 := flag.Arg(1) - mbTrans151 := thrift.NewTMemoryBufferLen(len(arg150)) - defer mbTrans151.Close() - _, err152 := mbTrans151.WriteString(arg150) - if err152 != nil { + arg163 := flag.Arg(1) + mbTrans164 := thrift.NewTMemoryBufferLen(len(arg163)) + defer mbTrans164.Close() + _, err165 := mbTrans164.WriteString(arg163) + if err165 != nil { Usage() return } - factory153 := thrift.NewTJSONProtocolFactory() - jsProt154 := factory153.GetProtocol(mbTrans151) + factory166 := thrift.NewTJSONProtocolFactory() + jsProt167 := factory166.GetProtocol(mbTrans164) argvalue0 := aurora.NewJobKey() - err155 := argvalue0.Read(context.Background(), jsProt154) - if err155 != nil { + err168 := argvalue0.Read(context.Background(), jsProt167) + if err168 != nil { Usage() return } @@ -299,19 +299,19 @@ func main() { fmt.Fprintln(os.Stderr, "PopulateJobConfig requires 1 args") flag.Usage() } - arg158 := flag.Arg(1) - mbTrans159 := thrift.NewTMemoryBufferLen(len(arg158)) - defer mbTrans159.Close() - _, err160 := mbTrans159.WriteString(arg158) - if err160 != nil { + arg171 := flag.Arg(1) + mbTrans172 := thrift.NewTMemoryBufferLen(len(arg171)) + defer mbTrans172.Close() + _, err173 := mbTrans172.WriteString(arg171) + if err173 != nil { Usage() return } - factory161 := thrift.NewTJSONProtocolFactory() - jsProt162 := factory161.GetProtocol(mbTrans159) + factory174 := thrift.NewTJSONProtocolFactory() + jsProt175 := factory174.GetProtocol(mbTrans172) argvalue0 := aurora.NewJobConfiguration() - err163 := argvalue0.Read(context.Background(), jsProt162) - if err163 != nil { + err176 := argvalue0.Read(context.Background(), jsProt175) + if err176 != nil { Usage() return } @@ -324,19 +324,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateSummaries requires 1 args") flag.Usage() } - arg164 := flag.Arg(1) - mbTrans165 := thrift.NewTMemoryBufferLen(len(arg164)) - defer mbTrans165.Close() - _, err166 := mbTrans165.WriteString(arg164) - if err166 != nil { + arg177 := flag.Arg(1) + mbTrans178 := thrift.NewTMemoryBufferLen(len(arg177)) + defer mbTrans178.Close() + _, err179 := mbTrans178.WriteString(arg177) + if err179 != nil { Usage() return } - factory167 := thrift.NewTJSONProtocolFactory() - jsProt168 := factory167.GetProtocol(mbTrans165) + factory180 := thrift.NewTJSONProtocolFactory() + jsProt181 := factory180.GetProtocol(mbTrans178) argvalue0 := aurora.NewJobUpdateQuery() - err169 := argvalue0.Read(context.Background(), jsProt168) - if err169 != nil { + err182 := argvalue0.Read(context.Background(), jsProt181) + if err182 != nil { Usage() return } @@ -349,19 +349,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateDetails requires 1 args") flag.Usage() } - arg170 := flag.Arg(1) - mbTrans171 := thrift.NewTMemoryBufferLen(len(arg170)) - defer mbTrans171.Close() - _, err172 := mbTrans171.WriteString(arg170) - if err172 != nil { + arg183 := flag.Arg(1) + mbTrans184 := thrift.NewTMemoryBufferLen(len(arg183)) + defer mbTrans184.Close() + _, err185 := mbTrans184.WriteString(arg183) + if err185 != nil { Usage() return } - factory173 := thrift.NewTJSONProtocolFactory() - jsProt174 := factory173.GetProtocol(mbTrans171) + factory186 := thrift.NewTJSONProtocolFactory() + jsProt187 := factory186.GetProtocol(mbTrans184) argvalue0 := aurora.NewJobUpdateQuery() - err175 := argvalue0.Read(context.Background(), jsProt174) - if err175 != nil { + err188 := argvalue0.Read(context.Background(), jsProt187) + if err188 != nil { Usage() return } @@ -374,19 +374,19 @@ func main() { fmt.Fprintln(os.Stderr, "GetJobUpdateDiff requires 1 args") flag.Usage() } - arg176 := flag.Arg(1) - mbTrans177 := thrift.NewTMemoryBufferLen(len(arg176)) - defer mbTrans177.Close() - _, err178 := mbTrans177.WriteString(arg176) - if err178 != nil { + arg189 := flag.Arg(1) + mbTrans190 := thrift.NewTMemoryBufferLen(len(arg189)) + defer mbTrans190.Close() + _, err191 := mbTrans190.WriteString(arg189) + if err191 != nil { Usage() return } - factory179 := thrift.NewTJSONProtocolFactory() - jsProt180 := factory179.GetProtocol(mbTrans177) + factory192 := thrift.NewTJSONProtocolFactory() + jsProt193 := factory192.GetProtocol(mbTrans190) argvalue0 := aurora.NewJobUpdateRequest() - err181 := argvalue0.Read(context.Background(), jsProt180) - if err181 != nil { + err194 := argvalue0.Read(context.Background(), jsProt193) + if err194 != nil { Usage() return } diff --git a/realis.go b/realis.go index a7c536e..c14db59 100644 --- a/realis.go +++ b/realis.go @@ -522,6 +522,20 @@ func (c *Client) RestartInstances(key aurora.JobKey, instances ...int32) error { return nil } +// Restarts specific instances with SLA specified +func (c *Client) SlaRestartInstances(key aurora.JobKey, policy *aurora.SlaPolicy, instances ...int32) error { + c.logger.DebugPrintf("SlaRestartInstances Thrift Payload: %+v %v\n", key, instances) + + _, retryErr := c.thriftCallWithRetries(false, func() (*aurora.Response, error) { + return c.client.SlaRestartShards(context.TODO(), &key, instances, policy) + }) + + if retryErr != nil { + return errors.Wrap(retryErr, "error sending SlaRestartInstances command to Aurora Scheduler") + } + return nil +} + // Restarts all active tasks under a job configuration. func (c *Client) RestartJob(key aurora.JobKey) error {