Renaming RealisClient to Client to avoid stuttering. Moving monitors under Client. Making configuration object private. Deleted legacy code to generate configuration object.
This commit is contained in:
parent
c1be2fe62b
commit
76300782ba
7 changed files with 253 additions and 256 deletions
|
@ -75,8 +75,7 @@ func main() {
|
||||||
|
|
||||||
var job *realis.AuroraJob
|
var job *realis.AuroraJob
|
||||||
var err error
|
var err error
|
||||||
var monitor *realis.Monitor
|
var r *realis.Client
|
||||||
var r *realis.RealisClient
|
|
||||||
|
|
||||||
clientOptions := []realis.ClientOption{
|
clientOptions := []realis.ClientOption{
|
||||||
realis.BasicAuth(username, password),
|
realis.BasicAuth(username, password),
|
||||||
|
@ -91,7 +90,7 @@ func main() {
|
||||||
realis.Debug(),
|
realis.Debug(),
|
||||||
}
|
}
|
||||||
|
|
||||||
//check if zkUrl is available.
|
// Check if zkUrl is available.
|
||||||
if zkUrl != "" {
|
if zkUrl != "" {
|
||||||
fmt.Println("zkUrl: ", zkUrl)
|
fmt.Println("zkUrl: ", zkUrl)
|
||||||
clientOptions = append(clientOptions, realis.ZKUrl(zkUrl))
|
clientOptions = append(clientOptions, realis.ZKUrl(zkUrl))
|
||||||
|
@ -107,11 +106,10 @@ func main() {
|
||||||
clientOptions = append(clientOptions, realis.ClientCerts(clientKey, clientCert))
|
clientOptions = append(clientOptions, realis.ClientCerts(clientKey, clientCert))
|
||||||
}
|
}
|
||||||
|
|
||||||
r, err = realis.NewRealisClient(clientOptions...)
|
r, err = realis.NewClient(clientOptions...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatalln(err)
|
log.Fatalln(err)
|
||||||
}
|
}
|
||||||
monitor = &realis.Monitor{r}
|
|
||||||
defer r.Close()
|
defer r.Close()
|
||||||
|
|
||||||
switch executor {
|
switch executor {
|
||||||
|
@ -171,7 +169,7 @@ func main() {
|
||||||
log.Fatalln(err)
|
log.Fatalln(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ok, mErr := monitor.Instances(job.JobKey(), job.GetInstanceCount(), 5*time.Second, 50*time.Second); !ok || mErr != nil {
|
if ok, mErr := r.InstancesMonitor(job.JobKey(), job.GetInstanceCount(), 5*time.Second, 50*time.Second); !ok || mErr != nil {
|
||||||
err := r.KillJob(job.JobKey())
|
err := r.KillJob(job.JobKey())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatalln(err)
|
log.Fatalln(err)
|
||||||
|
@ -190,7 +188,7 @@ func main() {
|
||||||
}
|
}
|
||||||
fmt.Println(result.String())
|
fmt.Println(result.String())
|
||||||
|
|
||||||
if ok, mErr := monitor.JobUpdate(*result.GetKey(), 5*time.Second, 180*time.Second); !ok || mErr != nil {
|
if ok, mErr := r.JobUpdateMonitor(*result.GetKey(), 5*time.Second, 180*time.Second); !ok || mErr != nil {
|
||||||
err := r.AbortJobUpdate(*result.GetKey(), "Monitor timed out")
|
err := r.AbortJobUpdate(*result.GetKey(), "Monitor timed out")
|
||||||
err = r.KillJob(job.JobKey())
|
err = r.KillJob(job.JobKey())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -208,7 +206,7 @@ func main() {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ok, err := monitor.Instances(job.JobKey(), job.GetInstanceCount(), 10*time.Second, 300*time.Second); !ok || err != nil {
|
if ok, err := r.InstancesMonitor(job.JobKey(), job.GetInstanceCount(), 10*time.Second, 300*time.Second); !ok || err != nil {
|
||||||
err := r.KillJob(job.JobKey())
|
err := r.KillJob(job.JobKey())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
|
@ -224,7 +222,7 @@ func main() {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ok, err := monitor.Instances(job.JobKey(), job.GetInstanceCount(), 10*time.Second, 300*time.Second); !ok || err != nil {
|
if ok, err := r.InstancesMonitor(job.JobKey(), job.GetInstanceCount(), 10*time.Second, 300*time.Second); !ok || err != nil {
|
||||||
err := r.KillJob(job.JobKey())
|
err := r.KillJob(job.JobKey())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
|
@ -263,7 +261,7 @@ func main() {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ok, err := monitor.Instances(job.JobKey(), 0, 5*time.Second, 50*time.Second); !ok || err != nil {
|
if ok, err := r.InstancesMonitor(job.JobKey(), 0, 5*time.Second, 50*time.Second); !ok || err != nil {
|
||||||
log.Fatal("Unable to kill all instances of job")
|
log.Fatal("Unable to kill all instances of job")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -316,7 +314,7 @@ func main() {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ok, err := monitor.Instances(job.JobKey(), int32(currInstances+numOfInstances), 5*time.Second, 50*time.Second); !ok || err != nil {
|
if ok, err := r.InstancesMonitor(job.JobKey(), int32(currInstances+numOfInstances), 5*time.Second, 50*time.Second); !ok || err != nil {
|
||||||
fmt.Println("Flexing up failed")
|
fmt.Println("Flexing up failed")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -337,7 +335,7 @@ func main() {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if ok, err := monitor.Instances(job.JobKey(), int32(currInstances-numOfInstances), 5*time.Second, 100*time.Second); !ok || err != nil {
|
if ok, err := r.InstancesMonitor(job.JobKey(), int32(currInstances-numOfInstances), 5*time.Second, 100*time.Second); !ok || err != nil {
|
||||||
fmt.Println("flexDown failed")
|
fmt.Println("flexDown failed")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -363,7 +361,10 @@ func main() {
|
||||||
}
|
}
|
||||||
|
|
||||||
jobUpdateKey := result.GetKey()
|
jobUpdateKey := result.GetKey()
|
||||||
monitor.JobUpdate(*jobUpdateKey, 5*time.Second, 6*time.Minute)
|
_, err = r.JobUpdateMonitor(*jobUpdateKey, 5*time.Second, 6*time.Minute)
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
case "pauseJobUpdate":
|
case "pauseJobUpdate":
|
||||||
err := r.PauseJobUpdate(&aurora.JobUpdateKey{
|
err := r.PauseJobUpdate(&aurora.JobUpdateKey{
|
||||||
|
@ -508,7 +509,7 @@ func main() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
hostResult, err := monitor.HostMaintenance(
|
hostResult, err := r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
||||||
5*time.Second,
|
5*time.Second,
|
||||||
|
@ -537,7 +538,7 @@ func main() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
hostResult, err := monitor.HostMaintenance(
|
hostResult, err := r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
||||||
5*time.Second,
|
5*time.Second,
|
||||||
|
@ -563,7 +564,7 @@ func main() {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
hostResult, err := monitor.HostMaintenance(
|
hostResult, err := r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
||||||
5*time.Second,
|
5*time.Second,
|
||||||
|
|
|
@ -125,7 +125,7 @@ func init() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func CreateRealisClient(config *Config) (*realis.RealisClient, error) {
|
func CreateRealisClient(config *Config) (*realis.Client, error) {
|
||||||
var transportOption realis.ClientOption
|
var transportOption realis.ClientOption
|
||||||
// Configuring transport protocol. If not transport is provided, then using JSON as the
|
// Configuring transport protocol. If not transport is provided, then using JSON as the
|
||||||
// default transport protocol.
|
// default transport protocol.
|
||||||
|
@ -157,7 +157,7 @@ func CreateRealisClient(config *Config) (*realis.RealisClient, error) {
|
||||||
clientOptions = append(clientOptions, realis.Debug())
|
clientOptions = append(clientOptions, realis.Debug())
|
||||||
}
|
}
|
||||||
|
|
||||||
return realis.NewRealisClient(clientOptions...)
|
return realis.NewClient(clientOptions...)
|
||||||
}
|
}
|
||||||
|
|
||||||
func main() {
|
func main() {
|
||||||
|
@ -165,7 +165,6 @@ func main() {
|
||||||
fmt.Println(clientCreationErr)
|
fmt.Println(clientCreationErr)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
} else {
|
} else {
|
||||||
monitor := &realis.Monitor{Client: r}
|
|
||||||
defer r.Close()
|
defer r.Close()
|
||||||
uris := job.URIs
|
uris := job.URIs
|
||||||
labels := job.Labels
|
labels := job.Labels
|
||||||
|
@ -209,7 +208,7 @@ func main() {
|
||||||
fmt.Println("Error creating Aurora job: ", jobCreationErr)
|
fmt.Println("Error creating Aurora job: ", jobCreationErr)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
} else {
|
} else {
|
||||||
if ok, monitorErr := monitor.Instances(auroraJob.JobKey(), auroraJob.GetInstanceCount(), 5, 50); !ok || monitorErr != nil {
|
if ok, monitorErr := r.InstancesMonitor(auroraJob.JobKey(), auroraJob.GetInstanceCount(), 5, 50); !ok || monitorErr != nil {
|
||||||
if jobErr := r.KillJob(auroraJob.JobKey()); jobErr !=
|
if jobErr := r.KillJob(auroraJob.JobKey()); jobErr !=
|
||||||
nil {
|
nil {
|
||||||
fmt.Println(jobErr)
|
fmt.Println(jobErr)
|
||||||
|
|
4
job.go
4
job.go
|
@ -32,7 +32,7 @@ func NewJob() *AuroraJob {
|
||||||
|
|
||||||
jobKey := &aurora.JobKey{}
|
jobKey := &aurora.JobKey{}
|
||||||
|
|
||||||
// Task Config
|
// Task clientConfig
|
||||||
taskConfig := &aurora.TaskConfig{
|
taskConfig := &aurora.TaskConfig{
|
||||||
Job: jobKey,
|
Job: jobKey,
|
||||||
MesosFetcherUris: make([]*aurora.MesosFetcherURI, 0),
|
MesosFetcherUris: make([]*aurora.MesosFetcherURI, 0),
|
||||||
|
@ -42,7 +42,7 @@ func NewJob() *AuroraJob {
|
||||||
Container: NewMesosContainer().Build(),
|
Container: NewMesosContainer().Build(),
|
||||||
}
|
}
|
||||||
|
|
||||||
// AuroraJob Config
|
// AuroraJob clientConfig
|
||||||
jobConfig := &aurora.JobConfiguration{
|
jobConfig := &aurora.JobConfiguration{
|
||||||
Key: jobKey,
|
Key: jobKey,
|
||||||
TaskConfig: taskConfig,
|
TaskConfig: taskConfig,
|
||||||
|
|
12
monitors.go
12
monitors.go
|
@ -29,7 +29,7 @@ const (
|
||||||
)
|
)
|
||||||
|
|
||||||
// Polls the scheduler every certain amount of time to see if the update has succeeded
|
// Polls the scheduler every certain amount of time to see if the update has succeeded
|
||||||
func (c *RealisClient) JobUpdateMonitor(updateKey aurora.JobUpdateKey, interval, timeout time.Duration) (bool, error) {
|
func (c *Client) JobUpdateMonitor(updateKey aurora.JobUpdateKey, interval, timeout time.Duration) (bool, error) {
|
||||||
if interval < 1*time.Second {
|
if interval < 1*time.Second {
|
||||||
interval = interval * time.Second
|
interval = interval * time.Second
|
||||||
}
|
}
|
||||||
|
@ -50,13 +50,13 @@ func (c *RealisClient) JobUpdateMonitor(updateKey aurora.JobUpdateKey, interval,
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-ticker.C:
|
case <-ticker.C:
|
||||||
updateDetail, cliErr := m.JobUpdateDetails(updateQ)
|
updateDetail, cliErr := c.JobUpdateDetails(updateQ)
|
||||||
if cliErr != nil {
|
if cliErr != nil {
|
||||||
return false, cliErr
|
return false, cliErr
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(updateDetail) == 0 {
|
if len(updateDetail) == 0 {
|
||||||
m.RealisConfig().logger.Println("No update found")
|
c.RealisConfig().logger.Println("No update found")
|
||||||
return false, errors.New("No update found for " + updateKey.String())
|
return false, errors.New("No update found for " + updateKey.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -88,14 +88,14 @@ func (c *RealisClient) JobUpdateMonitor(updateKey aurora.JobUpdateKey, interval,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Monitor a AuroraJob until all instances enter one of the LiveStates
|
// Monitor a AuroraJob until all instances enter one of the LiveStates
|
||||||
func (c *RealisClient) InstancesMonitor(key *aurora.JobKey, instances int32, interval, timeout time.Duration) (bool, error) {
|
func (c *Client) InstancesMonitor(key *aurora.JobKey, instances int32, interval, timeout time.Duration) (bool, error) {
|
||||||
return c.ScheduleStatusMonitor(key, instances, aurora.LIVE_STATES, interval, timeout)
|
return c.ScheduleStatusMonitor(key, instances, aurora.LIVE_STATES, interval, timeout)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Monitor a AuroraJob until all instances enter a desired status.
|
// Monitor a AuroraJob until all instances enter a desired status.
|
||||||
// Defaults sets of desired statuses provided by the thrift API include:
|
// Defaults sets of desired statuses provided by the thrift API include:
|
||||||
// ActiveStates, SlaveAssignedStates, LiveStates, and TerminalStates
|
// ActiveStates, SlaveAssignedStates, LiveStates, and TerminalStates
|
||||||
func (c *RealisClient) ScheduleStatusMonitor(key *aurora.JobKey, instanceCount int32, desiredStatuses []aurora.ScheduleStatus, interval, timeout time.Duration) (bool, error) {
|
func (c *Client) ScheduleStatusMonitor(key *aurora.JobKey, instanceCount int32, desiredStatuses []aurora.ScheduleStatus, interval, timeout time.Duration) (bool, error) {
|
||||||
if interval < 1*time.Second {
|
if interval < 1*time.Second {
|
||||||
interval = interval * time.Second
|
interval = interval * time.Second
|
||||||
}
|
}
|
||||||
|
@ -131,7 +131,7 @@ func (c *RealisClient) ScheduleStatusMonitor(key *aurora.JobKey, instanceCount i
|
||||||
|
|
||||||
// Monitor host status until all hosts match the status provided. Returns a map where the value is true if the host
|
// Monitor host status until all hosts match the status provided. Returns a map where the value is true if the host
|
||||||
// is in one of the desired mode(s) or false if it is not as of the time when the monitor exited.
|
// is in one of the desired mode(s) or false if it is not as of the time when the monitor exited.
|
||||||
func (c *RealisClient) HostMaintenanceMonitor(hosts []string, modes []aurora.MaintenanceMode, interval, timeout time.Duration) (map[string]bool, error) {
|
func (c *Client) HostMaintenanceMonitor(hosts []string, modes []aurora.MaintenanceMode, interval, timeout time.Duration) (map[string]bool, error) {
|
||||||
if interval < 1*time.Second {
|
if interval < 1*time.Second {
|
||||||
interval = interval * time.Second
|
interval = interval * time.Second
|
||||||
}
|
}
|
||||||
|
|
380
realis.go
380
realis.go
|
@ -39,8 +39,8 @@ import (
|
||||||
|
|
||||||
const VERSION = "2.0.0"
|
const VERSION = "2.0.0"
|
||||||
|
|
||||||
type RealisClient struct {
|
type Client struct {
|
||||||
config *RealisConfig
|
config *clientConfig
|
||||||
client *aurora.AuroraSchedulerManagerClient
|
client *aurora.AuroraSchedulerManagerClient
|
||||||
readonlyClient *aurora.ReadOnlySchedulerClient
|
readonlyClient *aurora.ReadOnlySchedulerClient
|
||||||
adminClient *aurora.AuroraAdminClient
|
adminClient *aurora.AuroraAdminClient
|
||||||
|
@ -50,7 +50,7 @@ type RealisClient struct {
|
||||||
transport thrift.TTransport
|
transport thrift.TTransport
|
||||||
}
|
}
|
||||||
|
|
||||||
type RealisConfig struct {
|
type clientConfig struct {
|
||||||
username, password string
|
username, password string
|
||||||
url string
|
url string
|
||||||
timeout time.Duration
|
timeout time.Duration
|
||||||
|
@ -75,30 +75,30 @@ var defaultBackoff = Backoff{
|
||||||
Jitter: 0.1,
|
Jitter: 0.1,
|
||||||
}
|
}
|
||||||
|
|
||||||
type ClientOption func(*RealisConfig)
|
type ClientOption func(*clientConfig)
|
||||||
|
|
||||||
// Config sets for options in RealisConfig.
|
// clientConfig sets for options in clientConfig.
|
||||||
func BasicAuth(username, password string) ClientOption {
|
func BasicAuth(username, password string) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.username = username
|
config.username = username
|
||||||
config.password = password
|
config.password = password
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func SchedulerUrl(url string) ClientOption {
|
func SchedulerUrl(url string) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.url = url
|
config.url = url
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func Timeout(timeout time.Duration) ClientOption {
|
func Timeout(timeout time.Duration) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.timeout = timeout
|
config.timeout = timeout
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func ZKCluster(cluster *Cluster) ClientOption {
|
func ZKCluster(cluster *Cluster) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.cluster = cluster
|
config.cluster = cluster
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -107,7 +107,7 @@ func ZKUrl(url string) ClientOption {
|
||||||
|
|
||||||
opts := []ZKOpt{ZKEndpoints(strings.Split(url, ",")...), ZKPath("/aurora/scheduler")}
|
opts := []ZKOpt{ZKEndpoints(strings.Split(url, ",")...), ZKPath("/aurora/scheduler")}
|
||||||
|
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
if config.zkOptions == nil {
|
if config.zkOptions == nil {
|
||||||
config.zkOptions = opts
|
config.zkOptions = opts
|
||||||
} else {
|
} else {
|
||||||
|
@ -117,37 +117,37 @@ func ZKUrl(url string) ClientOption {
|
||||||
}
|
}
|
||||||
|
|
||||||
func ThriftJSON() ClientOption {
|
func ThriftJSON() ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.jsonTransport = true
|
config.jsonTransport = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func ThriftBinary() ClientOption {
|
func ThriftBinary() ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.binTransport = true
|
config.binTransport = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func BackOff(b Backoff) ClientOption {
|
func BackOff(b Backoff) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.backoff = b
|
config.backoff = b
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func InsecureSkipVerify(InsecureSkipVerify bool) ClientOption {
|
func InsecureSkipVerify(InsecureSkipVerify bool) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.InsecureSkipVerify = InsecureSkipVerify
|
config.InsecureSkipVerify = InsecureSkipVerify
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func CertsPath(certspath string) ClientOption {
|
func CertsPath(certspath string) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.certsPath = certspath
|
config.certsPath = certspath
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func ClientCerts(clientKey, clientCert string) ClientOption {
|
func ClientCerts(clientKey, clientCert string) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.clientKey, config.clientCert = clientKey, clientCert
|
config.clientKey, config.clientCert = clientKey, clientCert
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -155,26 +155,26 @@ func ClientCerts(clientKey, clientCert string) ClientOption {
|
||||||
// Use this option if you'd like to override default settings for connecting to Zookeeper.
|
// Use this option if you'd like to override default settings for connecting to Zookeeper.
|
||||||
// See zk.go for what is possible to set as an option.
|
// See zk.go for what is possible to set as an option.
|
||||||
func ZookeeperOptions(opts ...ZKOpt) ClientOption {
|
func ZookeeperOptions(opts ...ZKOpt) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.zkOptions = opts
|
config.zkOptions = opts
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Using the word set to avoid name collision with Interface.
|
// Using the word set to avoid name collision with Interface.
|
||||||
func SetLogger(l Logger) ClientOption {
|
func SetLogger(l Logger) ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.logger = &LevelLogger{l, false}
|
config.logger = &LevelLogger{l, false}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Enable debug statements.
|
// Enable debug statements.
|
||||||
func Debug() ClientOption {
|
func Debug() ClientOption {
|
||||||
return func(config *RealisConfig) {
|
return func(config *clientConfig) {
|
||||||
config.debug = true
|
config.debug = true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func newTJSONTransport(url string, timeout time.Duration, config *RealisConfig) (thrift.TTransport, error) {
|
func newTJSONTransport(url string, timeout time.Duration, config *clientConfig) (thrift.TTransport, error) {
|
||||||
trans, err := defaultTTransport(url, timeout, config)
|
trans, err := defaultTTransport(url, timeout, config)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errors.Wrap(err, "Error creating realis")
|
return nil, errors.Wrap(err, "Error creating realis")
|
||||||
|
@ -185,7 +185,7 @@ func newTJSONTransport(url string, timeout time.Duration, config *RealisConfig)
|
||||||
return trans, err
|
return trans, err
|
||||||
}
|
}
|
||||||
|
|
||||||
func newTBinTransport(url string, timeout time.Duration, config *RealisConfig) (thrift.TTransport, error) {
|
func newTBinTransport(url string, timeout time.Duration, config *clientConfig) (thrift.TTransport, error) {
|
||||||
trans, err := defaultTTransport(url, timeout, config)
|
trans, err := defaultTTransport(url, timeout, config)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errors.Wrap(err, "Error creating realis")
|
return nil, errors.Wrap(err, "Error creating realis")
|
||||||
|
@ -199,12 +199,12 @@ func newTBinTransport(url string, timeout time.Duration, config *RealisConfig) (
|
||||||
return trans, err
|
return trans, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// This client implementation of the realis interface uses a retry mechanism for all Thrift Calls.
|
// This client implementation uses a retry mechanism for all Thrift Calls.
|
||||||
// It will retry all calls which result in a temporary failure as well as calls that fail due to an EOF
|
// It will retry all calls which result in a temporary failure as well as calls that fail due to an EOF
|
||||||
// being returned by the http client. Most permanent failures are now being caught by the thriftCallWithRetries
|
// being returned by the http client. Most permanent failures are now being caught by the thriftCallWithRetries
|
||||||
// function and not being retried but there may be corner cases not yet handled.
|
// function and not being retried but there may be corner cases not yet handled.
|
||||||
func NewRealisClient(options ...ClientOption) (*RealisClient, error) {
|
func NewClient(options ...ClientOption) (*Client, error) {
|
||||||
config := &RealisConfig{}
|
config := &clientConfig{}
|
||||||
|
|
||||||
// Default configs
|
// Default configs
|
||||||
config.timeout = 10 * time.Second
|
config.timeout = 10 * time.Second
|
||||||
|
@ -234,7 +234,7 @@ func NewRealisClient(options ...ClientOption) (*RealisClient, error) {
|
||||||
// Note, by this point, a LevelLogger should have been created.
|
// Note, by this point, a LevelLogger should have been created.
|
||||||
config.logger.EnableDebug(config.debug)
|
config.logger.EnableDebug(config.debug)
|
||||||
|
|
||||||
config.logger.DebugPrintln("Number of options applied to config: ", len(options))
|
config.logger.DebugPrintln("Number of options applied to clientConfig: ", len(options))
|
||||||
|
|
||||||
// Set default Transport to JSON if needed.
|
// Set default Transport to JSON if needed.
|
||||||
if !config.jsonTransport && !config.binTransport {
|
if !config.jsonTransport && !config.binTransport {
|
||||||
|
@ -284,7 +284,7 @@ func NewRealisClient(options ...ClientOption) (*RealisClient, error) {
|
||||||
config.protoFactory = thrift.NewTBinaryProtocolFactoryDefault()
|
config.protoFactory = thrift.NewTBinaryProtocolFactoryDefault()
|
||||||
}
|
}
|
||||||
|
|
||||||
config.logger.Printf("gorealis config url: %+v\n", url)
|
config.logger.Printf("gorealis clientConfig url: %+v\n", url)
|
||||||
|
|
||||||
// Adding Basic Authentication.
|
// Adding Basic Authentication.
|
||||||
if config.username != "" && config.password != "" {
|
if config.username != "" && config.password != "" {
|
||||||
|
@ -292,7 +292,7 @@ func NewRealisClient(options ...ClientOption) (*RealisClient, error) {
|
||||||
httpTrans.SetHeader("Authorization", "Basic "+basicAuth(config.username, config.password))
|
httpTrans.SetHeader("Authorization", "Basic "+basicAuth(config.username, config.password))
|
||||||
}
|
}
|
||||||
|
|
||||||
return &RealisClient{
|
return &Client{
|
||||||
config: config,
|
config: config,
|
||||||
client: aurora.NewAuroraSchedulerManagerClientFactory(config.transport, config.protoFactory),
|
client: aurora.NewAuroraSchedulerManagerClientFactory(config.transport, config.protoFactory),
|
||||||
readonlyClient: aurora.NewReadOnlySchedulerClientFactory(config.transport, config.protoFactory),
|
readonlyClient: aurora.NewReadOnlySchedulerClientFactory(config.transport, config.protoFactory),
|
||||||
|
@ -321,7 +321,7 @@ func GetCerts(certPath string) (*x509.CertPool, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Creates a default Thrift Transport object for communications in gorealis using an HTTP Post Client
|
// Creates a default Thrift Transport object for communications in gorealis using an HTTP Post Client
|
||||||
func defaultTTransport(url string, timeout time.Duration, config *RealisConfig) (thrift.TTransport, error) {
|
func defaultTTransport(url string, timeout time.Duration, config *clientConfig) (thrift.TTransport, error) {
|
||||||
var transport http.Transport
|
var transport http.Transport
|
||||||
|
|
||||||
jar, err := cookiejar.New(nil)
|
jar, err := cookiejar.New(nil)
|
||||||
|
@ -375,29 +375,29 @@ func defaultTTransport(url string, timeout time.Duration, config *RealisConfig)
|
||||||
|
|
||||||
// Create a default configuration of the transport layer, requires a URL to test connection with.
|
// Create a default configuration of the transport layer, requires a URL to test connection with.
|
||||||
// Uses HTTP Post as transport layer and Thrift JSON as the wire protocol by default.
|
// Uses HTTP Post as transport layer and Thrift JSON as the wire protocol by default.
|
||||||
func newDefaultConfig(url string, timeout time.Duration, config *RealisConfig) (*RealisConfig, error) {
|
func newDefaultConfig(url string, timeout time.Duration, config *clientConfig) (*clientConfig, error) {
|
||||||
return newTJSONConfig(url, timeout, config)
|
return newTJSONConfig(url, timeout, config)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Creates a realis config object using HTTP Post and Thrift JSON protocol to communicate with Aurora.
|
// Creates a realis clientConfig object using HTTP Post and Thrift JSON protocol to communicate with Aurora.
|
||||||
func newTJSONConfig(url string, timeout time.Duration, config *RealisConfig) (*RealisConfig, error) {
|
func newTJSONConfig(url string, timeout time.Duration, config *clientConfig) (*clientConfig, error) {
|
||||||
trans, err := defaultTTransport(url, timeout, config)
|
trans, err := defaultTTransport(url, timeout, config)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errors.Wrap(err, "Error creating realis config")
|
return nil, errors.Wrap(err, "Error creating realis clientConfig")
|
||||||
}
|
}
|
||||||
|
|
||||||
httpTrans := (trans).(*thrift.THttpClient)
|
httpTrans := (trans).(*thrift.THttpClient)
|
||||||
httpTrans.SetHeader("Content-Type", "application/x-thrift")
|
httpTrans.SetHeader("Content-Type", "application/x-thrift")
|
||||||
httpTrans.SetHeader("User-Agent", "gorealis v"+VERSION)
|
httpTrans.SetHeader("User-Agent", "gorealis v"+VERSION)
|
||||||
|
|
||||||
return &RealisConfig{transport: trans, protoFactory: thrift.NewTJSONProtocolFactory()}, nil
|
return &clientConfig{transport: trans, protoFactory: thrift.NewTJSONProtocolFactory()}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Creates a realis config config using HTTP Post and Thrift Binary protocol to communicate with Aurora.
|
// Creates a realis clientConfig clientConfig using HTTP Post and Thrift Binary protocol to communicate with Aurora.
|
||||||
func newTBinaryConfig(url string, timeout time.Duration, config *RealisConfig) (*RealisConfig, error) {
|
func newTBinaryConfig(url string, timeout time.Duration, config *clientConfig) (*clientConfig, error) {
|
||||||
trans, err := defaultTTransport(url, timeout, config)
|
trans, err := defaultTTransport(url, timeout, config)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errors.Wrap(err, "Error creating realis config")
|
return nil, errors.Wrap(err, "Error creating realis clientConfig")
|
||||||
}
|
}
|
||||||
|
|
||||||
httpTrans := (trans).(*thrift.THttpClient)
|
httpTrans := (trans).(*thrift.THttpClient)
|
||||||
|
@ -407,7 +407,7 @@ func newTBinaryConfig(url string, timeout time.Duration, config *RealisConfig) (
|
||||||
httpTrans.SetHeader("Content-Type", "application/vnd.apache.thrift.binary")
|
httpTrans.SetHeader("Content-Type", "application/vnd.apache.thrift.binary")
|
||||||
httpTrans.SetHeader("User-Agent", "gorealis v"+VERSION)
|
httpTrans.SetHeader("User-Agent", "gorealis v"+VERSION)
|
||||||
|
|
||||||
return &RealisConfig{transport: trans, protoFactory: thrift.NewTBinaryProtocolFactoryDefault()}, nil
|
return &clientConfig{transport: trans, protoFactory: thrift.NewTBinaryProtocolFactoryDefault()}, nil
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -416,16 +416,16 @@ func basicAuth(username, password string) string {
|
||||||
return base64.StdEncoding.EncodeToString([]byte(auth))
|
return base64.StdEncoding.EncodeToString([]byte(auth))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) ReestablishConn() error {
|
func (c *Client) ReestablishConn() error {
|
||||||
// Close existing connection
|
// Close existing connection
|
||||||
r.logger.Println("Re-establishing Connection to Aurora")
|
c.logger.Println("Re-establishing Connection to Aurora")
|
||||||
r.Close()
|
c.Close()
|
||||||
|
|
||||||
r.lock.Lock()
|
c.lock.Lock()
|
||||||
defer r.lock.Unlock()
|
defer c.lock.Unlock()
|
||||||
|
|
||||||
// Recreate connection from scratch using original options
|
// Recreate connection from scratch using original options
|
||||||
newRealis, err := NewRealisClient(r.config.options...)
|
newRealis, err := NewClient(c.config.options...)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// This could be a temporary network hiccup
|
// This could be a temporary network hiccup
|
||||||
return NewTemporaryError(err)
|
return NewTemporaryError(err)
|
||||||
|
@ -433,26 +433,26 @@ func (r *RealisClient) ReestablishConn() error {
|
||||||
|
|
||||||
// If we are able to successfully re-connect, make receiver
|
// If we are able to successfully re-connect, make receiver
|
||||||
// point to newly established connections.
|
// point to newly established connections.
|
||||||
r.config = newRealis.config
|
c.config = newRealis.config
|
||||||
r.client = newRealis.client
|
c.client = newRealis.client
|
||||||
r.readonlyClient = newRealis.readonlyClient
|
c.readonlyClient = newRealis.readonlyClient
|
||||||
r.adminClient = newRealis.adminClient
|
c.adminClient = newRealis.adminClient
|
||||||
r.logger = newRealis.logger
|
c.logger = newRealis.logger
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Releases resources associated with the realis client.
|
// Releases resources associated with the realis client.
|
||||||
func (r *RealisClient) Close() {
|
func (c *Client) Close() {
|
||||||
|
|
||||||
r.lock.Lock()
|
c.lock.Lock()
|
||||||
defer r.lock.Unlock()
|
defer c.lock.Unlock()
|
||||||
|
|
||||||
r.transport.Close()
|
c.transport.Close()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Uses predefined set of states to retrieve a set of active jobs in Apache Aurora.
|
// Uses predefined set of states to retrieve a set of active jobs in Apache Aurora.
|
||||||
func (r *RealisClient) GetInstanceIds(key *aurora.JobKey, states []aurora.ScheduleStatus) ([]int32, error) {
|
func (c *Client) GetInstanceIds(key *aurora.JobKey, states []aurora.ScheduleStatus) ([]int32, error) {
|
||||||
taskQ := &aurora.TaskQuery{
|
taskQ := &aurora.TaskQuery{
|
||||||
Role: &key.Role,
|
Role: &key.Role,
|
||||||
Environment: &key.Environment,
|
Environment: &key.Environment,
|
||||||
|
@ -460,10 +460,10 @@ func (r *RealisClient) GetInstanceIds(key *aurora.JobKey, states []aurora.Schedu
|
||||||
Statuses: states,
|
Statuses: states,
|
||||||
}
|
}
|
||||||
|
|
||||||
r.logger.DebugPrintf("GetTasksWithoutConfigs Thrift Payload: %+v\n", taskQ)
|
c.logger.DebugPrintf("GetTasksWithoutConfigs Thrift Payload: %+v\n", taskQ)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.GetTasksWithoutConfigs(nil, taskQ)
|
return c.client.GetTasksWithoutConfigs(nil, taskQ)
|
||||||
})
|
})
|
||||||
|
|
||||||
// If we encountered an error we couldn't recover from by retrying, return an error to the user
|
// If we encountered an error we couldn't recover from by retrying, return an error to the user
|
||||||
|
@ -481,11 +481,11 @@ func (r *RealisClient) GetInstanceIds(key *aurora.JobKey, states []aurora.Schedu
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) GetJobUpdateSummaries(jobUpdateQuery *aurora.JobUpdateQuery) (*aurora.GetJobUpdateSummariesResult_, error) {
|
func (c *Client) GetJobUpdateSummaries(jobUpdateQuery *aurora.JobUpdateQuery) (*aurora.GetJobUpdateSummariesResult_, error) {
|
||||||
r.logger.DebugPrintf("GetJobUpdateSummaries Thrift Payload: %+v\n", jobUpdateQuery)
|
c.logger.DebugPrintf("GetJobUpdateSummaries Thrift Payload: %+v\n", jobUpdateQuery)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.readonlyClient.GetJobUpdateSummaries(nil, jobUpdateQuery)
|
return c.readonlyClient.GetJobUpdateSummaries(nil, jobUpdateQuery)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -495,12 +495,12 @@ func (r *RealisClient) GetJobUpdateSummaries(jobUpdateQuery *aurora.JobUpdateQue
|
||||||
return resp.GetResult_().GetGetJobUpdateSummariesResult_(), nil
|
return resp.GetResult_().GetGetJobUpdateSummariesResult_(), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) GetJobs(role string) (*aurora.GetJobsResult_, error) {
|
func (c *Client) GetJobs(role string) (*aurora.GetJobsResult_, error) {
|
||||||
|
|
||||||
var result *aurora.GetJobsResult_
|
var result *aurora.GetJobsResult_
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.readonlyClient.GetJobs(nil, role)
|
return c.readonlyClient.GetJobs(nil, role)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -516,11 +516,11 @@ func (r *RealisClient) GetJobs(role string) (*aurora.GetJobsResult_, error) {
|
||||||
|
|
||||||
// Kill specific instances of a job. Returns true, nil if a task was actually killed as a result of this API call.
|
// Kill specific instances of a job. Returns true, nil if a task was actually killed as a result of this API call.
|
||||||
// Returns false, nil if no tasks were killed as a result of this call but there was no error making the call.
|
// Returns false, nil if no tasks were killed as a result of this call but there was no error making the call.
|
||||||
func (r *RealisClient) KillInstances(key *aurora.JobKey, instances ...int32) (bool, error) {
|
func (c *Client) KillInstances(key *aurora.JobKey, instances ...int32) (bool, error) {
|
||||||
r.logger.DebugPrintf("KillTasks Thrift Payload: %+v %v\n", key, instances)
|
c.logger.DebugPrintf("KillTasks Thrift Payload: %+v %v\n", key, instances)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.KillTasks(nil, key, instances, "")
|
return c.client.KillTasks(nil, key, instances, "")
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -528,7 +528,7 @@ func (r *RealisClient) KillInstances(key *aurora.JobKey, instances ...int32) (bo
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(resp.GetDetails()) > 0 {
|
if len(resp.GetDetails()) > 0 {
|
||||||
r.logger.Println("KillTasks was called but no tasks killed as a result.")
|
c.logger.Println("KillTasks was called but no tasks killed as a result.")
|
||||||
return false, nil
|
return false, nil
|
||||||
} else {
|
} else {
|
||||||
return true, nil
|
return true, nil
|
||||||
|
@ -536,18 +536,18 @@ func (r *RealisClient) KillInstances(key *aurora.JobKey, instances ...int32) (bo
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) RealisConfig() *RealisConfig {
|
func (c *Client) RealisConfig() *clientConfig {
|
||||||
return r.config
|
return c.config
|
||||||
}
|
}
|
||||||
|
|
||||||
// Sends a kill message to the scheduler for all active tasks under a job.
|
// Sends a kill message to the scheduler for all active tasks under a job.
|
||||||
func (r *RealisClient) KillJob(key *aurora.JobKey) error {
|
func (c *Client) KillJob(key *aurora.JobKey) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("KillTasks Thrift Payload: %+v\n", key)
|
c.logger.DebugPrintf("KillTasks Thrift Payload: %+v\n", key)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
// Giving the KillTasks thrift call an empty set tells the Aurora scheduler to kill all active shards
|
// Giving the KillTasks thrift call an empty set tells the Aurora scheduler to kill all active shards
|
||||||
return r.client.KillTasks(nil, key, nil, "")
|
return c.client.KillTasks(nil, key, nil, "")
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -560,12 +560,12 @@ func (r *RealisClient) KillJob(key *aurora.JobKey) error {
|
||||||
// Although this API is able to create service jobs, it is better to use CreateService instead
|
// Although this API is able to create service jobs, it is better to use CreateService instead
|
||||||
// as that API uses the update thrift call which has a few extra features available.
|
// as that API uses the update thrift call which has a few extra features available.
|
||||||
// Use this API to create ad-hoc jobs.
|
// Use this API to create ad-hoc jobs.
|
||||||
func (r *RealisClient) CreateJob(auroraJob *AuroraJob) error {
|
func (c *Client) CreateJob(auroraJob *AuroraJob) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("CreateJob Thrift Payload: %+v\n", auroraJob.JobConfig())
|
c.logger.DebugPrintf("CreateJob Thrift Payload: %+v\n", auroraJob.JobConfig())
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.CreateJob(nil, auroraJob.JobConfig())
|
return c.client.CreateJob(nil, auroraJob.JobConfig())
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -576,12 +576,12 @@ func (r *RealisClient) CreateJob(auroraJob *AuroraJob) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// This API uses an update thrift call to create the services giving a few more robust features.
|
// This API uses an update thrift call to create the services giving a few more robust features.
|
||||||
func (r *RealisClient) CreateService(auroraJob *AuroraJob, settings *aurora.JobUpdateSettings) (*aurora.StartJobUpdateResult_, error) {
|
func (c *Client) CreateService(auroraJob *AuroraJob, settings *aurora.JobUpdateSettings) (*aurora.StartJobUpdateResult_, error) {
|
||||||
// Create a new job update object and ship it to the StartJobUpdate api
|
// Create a new job update object and ship it to the StartJobUpdate api
|
||||||
update := NewUpdateJob(auroraJob.TaskConfig(), settings)
|
update := NewUpdateJob(auroraJob.TaskConfig(), settings)
|
||||||
update.InstanceCount(auroraJob.GetInstanceCount())
|
update.InstanceCount(auroraJob.GetInstanceCount())
|
||||||
|
|
||||||
updateResult, err := r.StartJobUpdate(update, "")
|
updateResult, err := c.StartJobUpdate(update, "")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errors.Wrap(err, "unable to create service")
|
return nil, errors.Wrap(err, "unable to create service")
|
||||||
}
|
}
|
||||||
|
@ -593,11 +593,11 @@ func (r *RealisClient) CreateService(auroraJob *AuroraJob, settings *aurora.JobU
|
||||||
return nil, errors.New("results object is nil")
|
return nil, errors.New("results object is nil")
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) ScheduleCronJob(auroraJob *AuroraJob) error {
|
func (c *Client) ScheduleCronJob(auroraJob *AuroraJob) error {
|
||||||
r.logger.DebugPrintf("ScheduleCronJob Thrift Payload: %+v\n", auroraJob.JobConfig())
|
c.logger.DebugPrintf("ScheduleCronJob Thrift Payload: %+v\n", auroraJob.JobConfig())
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.ScheduleCronJob(nil, auroraJob.JobConfig())
|
return c.client.ScheduleCronJob(nil, auroraJob.JobConfig())
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -606,12 +606,12 @@ func (r *RealisClient) ScheduleCronJob(auroraJob *AuroraJob) error {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) DescheduleCronJob(key *aurora.JobKey) error {
|
func (c *Client) DescheduleCronJob(key *aurora.JobKey) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("DescheduleCronJob Thrift Payload: %+v\n", key)
|
c.logger.DebugPrintf("DescheduleCronJob Thrift Payload: %+v\n", key)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.DescheduleCronJob(nil, key)
|
return c.client.DescheduleCronJob(nil, key)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -622,12 +622,12 @@ func (r *RealisClient) DescheduleCronJob(key *aurora.JobKey) error {
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) StartCronJob(key *aurora.JobKey) error {
|
func (c *Client) StartCronJob(key *aurora.JobKey) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("StartCronJob Thrift Payload: %+v\n", key)
|
c.logger.DebugPrintf("StartCronJob Thrift Payload: %+v\n", key)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.StartCronJob(nil, key)
|
return c.client.StartCronJob(nil, key)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -638,11 +638,11 @@ func (r *RealisClient) StartCronJob(key *aurora.JobKey) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Restarts specific instances specified
|
// Restarts specific instances specified
|
||||||
func (r *RealisClient) RestartInstances(key *aurora.JobKey, instances ...int32) error {
|
func (c *Client) RestartInstances(key *aurora.JobKey, instances ...int32) error {
|
||||||
r.logger.DebugPrintf("RestartShards Thrift Payload: %+v %v\n", key, instances)
|
c.logger.DebugPrintf("RestartShards Thrift Payload: %+v %v\n", key, instances)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.RestartShards(nil, key, instances)
|
return c.client.RestartShards(nil, key, instances)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -652,18 +652,18 @@ func (r *RealisClient) RestartInstances(key *aurora.JobKey, instances ...int32)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Restarts all active tasks under a job configuration.
|
// Restarts all active tasks under a job configuration.
|
||||||
func (r *RealisClient) RestartJob(key *aurora.JobKey) error {
|
func (c *Client) RestartJob(key *aurora.JobKey) error {
|
||||||
|
|
||||||
instanceIds, err1 := r.GetInstanceIds(key, aurora.ACTIVE_STATES)
|
instanceIds, err1 := c.GetInstanceIds(key, aurora.ACTIVE_STATES)
|
||||||
if err1 != nil {
|
if err1 != nil {
|
||||||
return errors.Wrap(err1, "Could not retrieve relevant task instance IDs")
|
return errors.Wrap(err1, "Could not retrieve relevant task instance IDs")
|
||||||
}
|
}
|
||||||
|
|
||||||
r.logger.DebugPrintf("RestartShards Thrift Payload: %+v %v\n", key, instanceIds)
|
c.logger.DebugPrintf("RestartShards Thrift Payload: %+v %v\n", key, instanceIds)
|
||||||
|
|
||||||
if len(instanceIds) > 0 {
|
if len(instanceIds) > 0 {
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.RestartShards(nil, key, instanceIds)
|
return c.client.RestartShards(nil, key, instanceIds)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -677,12 +677,12 @@ func (r *RealisClient) RestartJob(key *aurora.JobKey) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Update all tasks under a job configuration. Currently gorealis doesn't support for canary deployments.
|
// Update all tasks under a job configuration. Currently gorealis doesn't support for canary deployments.
|
||||||
func (r *RealisClient) StartJobUpdate(updateJob *UpdateJob, message string) (*aurora.StartJobUpdateResult_, error) {
|
func (c *Client) StartJobUpdate(updateJob *UpdateJob, message string) (*aurora.StartJobUpdateResult_, error) {
|
||||||
|
|
||||||
r.logger.DebugPrintf("StartJobUpdate Thrift Payload: %+v %v\n", updateJob, message)
|
c.logger.DebugPrintf("StartJobUpdate Thrift Payload: %+v %v\n", updateJob, message)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.StartJobUpdate(nil, updateJob.req, message)
|
return c.client.StartJobUpdate(nil, updateJob.req, message)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -692,12 +692,12 @@ func (r *RealisClient) StartJobUpdate(updateJob *UpdateJob, message string) (*au
|
||||||
}
|
}
|
||||||
|
|
||||||
// Abort AuroraJob Update on Aurora. Requires the updateId which can be obtained on the Aurora web UI.
|
// Abort AuroraJob Update on Aurora. Requires the updateId which can be obtained on the Aurora web UI.
|
||||||
func (r *RealisClient) AbortJobUpdate(updateKey aurora.JobUpdateKey, message string) error {
|
func (c *Client) AbortJobUpdate(updateKey aurora.JobUpdateKey, message string) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("AbortJobUpdate Thrift Payload: %+v %v\n", updateKey, message)
|
c.logger.DebugPrintf("AbortJobUpdate Thrift Payload: %+v %v\n", updateKey, message)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.AbortJobUpdate(nil, &updateKey, message)
|
return c.client.AbortJobUpdate(nil, &updateKey, message)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -707,12 +707,12 @@ func (r *RealisClient) AbortJobUpdate(updateKey aurora.JobUpdateKey, message str
|
||||||
}
|
}
|
||||||
|
|
||||||
// Pause AuroraJob Update. UpdateID is returned from StartJobUpdate or the Aurora web UI.
|
// Pause AuroraJob Update. UpdateID is returned from StartJobUpdate or the Aurora web UI.
|
||||||
func (r *RealisClient) PauseJobUpdate(updateKey *aurora.JobUpdateKey, message string) error {
|
func (c *Client) PauseJobUpdate(updateKey *aurora.JobUpdateKey, message string) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("PauseJobUpdate Thrift Payload: %+v %v\n", updateKey, message)
|
c.logger.DebugPrintf("PauseJobUpdate Thrift Payload: %+v %v\n", updateKey, message)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.PauseJobUpdate(nil, updateKey, message)
|
return c.client.PauseJobUpdate(nil, updateKey, message)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -723,12 +723,12 @@ func (r *RealisClient) PauseJobUpdate(updateKey *aurora.JobUpdateKey, message st
|
||||||
}
|
}
|
||||||
|
|
||||||
// Resume Paused AuroraJob Update. UpdateID is returned from StartJobUpdate or the Aurora web UI.
|
// Resume Paused AuroraJob Update. UpdateID is returned from StartJobUpdate or the Aurora web UI.
|
||||||
func (r *RealisClient) ResumeJobUpdate(updateKey *aurora.JobUpdateKey, message string) error {
|
func (c *Client) ResumeJobUpdate(updateKey *aurora.JobUpdateKey, message string) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("ResumeJobUpdate Thrift Payload: %+v %v\n", updateKey, message)
|
c.logger.DebugPrintf("ResumeJobUpdate Thrift Payload: %+v %v\n", updateKey, message)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.ResumeJobUpdate(nil, updateKey, message)
|
return c.client.ResumeJobUpdate(nil, updateKey, message)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -739,12 +739,12 @@ func (r *RealisClient) ResumeJobUpdate(updateKey *aurora.JobUpdateKey, message s
|
||||||
}
|
}
|
||||||
|
|
||||||
// Pulse AuroraJob Update on Aurora. UpdateID is returned from StartJobUpdate or the Aurora web UI.
|
// Pulse AuroraJob Update on Aurora. UpdateID is returned from StartJobUpdate or the Aurora web UI.
|
||||||
func (r *RealisClient) PulseJobUpdate(updateKey *aurora.JobUpdateKey) (aurora.JobUpdatePulseStatus, error) {
|
func (c *Client) PulseJobUpdate(updateKey *aurora.JobUpdateKey) (aurora.JobUpdatePulseStatus, error) {
|
||||||
|
|
||||||
r.logger.DebugPrintf("PulseJobUpdate Thrift Payload: %+v\n", updateKey)
|
c.logger.DebugPrintf("PulseJobUpdate Thrift Payload: %+v\n", updateKey)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.PulseJobUpdate(nil, updateKey)
|
return c.client.PulseJobUpdate(nil, updateKey)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -761,12 +761,12 @@ func (r *RealisClient) PulseJobUpdate(updateKey *aurora.JobUpdateKey) (aurora.Jo
|
||||||
|
|
||||||
// Scale up the number of instances under a job configuration using the configuration for specific
|
// Scale up the number of instances under a job configuration using the configuration for specific
|
||||||
// instance to scale up.
|
// instance to scale up.
|
||||||
func (r *RealisClient) AddInstances(instKey aurora.InstanceKey, count int32) error {
|
func (c *Client) AddInstances(instKey aurora.InstanceKey, count int32) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("AddInstances Thrift Payload: %+v %v\n", instKey, count)
|
c.logger.DebugPrintf("AddInstances Thrift Payload: %+v %v\n", instKey, count)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.AddInstances(nil, &instKey, count)
|
return c.client.AddInstances(nil, &instKey, count)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -779,8 +779,8 @@ func (r *RealisClient) AddInstances(instKey aurora.InstanceKey, count int32) err
|
||||||
// Scale down the number of instances under a job configuration using the configuration of a specific instance
|
// Scale down the number of instances under a job configuration using the configuration of a specific instance
|
||||||
// Instances with a higher instance ID will be removed first. For example, if our instance ID list is [0,1,2,3]
|
// Instances with a higher instance ID will be removed first. For example, if our instance ID list is [0,1,2,3]
|
||||||
// and we want to remove 2 instances, 2 and 3 will always be picked.
|
// and we want to remove 2 instances, 2 and 3 will always be picked.
|
||||||
func (r *RealisClient) RemoveInstances(key *aurora.JobKey, count int) error {
|
func (c *Client) RemoveInstances(key *aurora.JobKey, count int) error {
|
||||||
instanceIds, err := r.GetInstanceIds(key, aurora.ACTIVE_STATES)
|
instanceIds, err := c.GetInstanceIds(key, aurora.ACTIVE_STATES)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return errors.Wrap(err, "RemoveInstances: Could not retrieve relevant instance IDs")
|
return errors.Wrap(err, "RemoveInstances: Could not retrieve relevant instance IDs")
|
||||||
}
|
}
|
||||||
|
@ -796,7 +796,7 @@ func (r *RealisClient) RemoveInstances(key *aurora.JobKey, count int) error {
|
||||||
|
|
||||||
// Get the last count instance ids to kill
|
// Get the last count instance ids to kill
|
||||||
instanceIds = instanceIds[:count]
|
instanceIds = instanceIds[:count]
|
||||||
killed, err := r.KillInstances(key, instanceIds...)
|
killed, err := c.KillInstances(key, instanceIds...)
|
||||||
|
|
||||||
if !killed {
|
if !killed {
|
||||||
return errors.New("Flex down was not able to reduce the number of instances running.")
|
return errors.New("Flex down was not able to reduce the number of instances running.")
|
||||||
|
@ -806,12 +806,12 @@ func (r *RealisClient) RemoveInstances(key *aurora.JobKey, count int) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get information about task including a fully hydrated task configuration object
|
// Get information about task including a fully hydrated task configuration object
|
||||||
func (r *RealisClient) GetTaskStatus(query *aurora.TaskQuery) ([]*aurora.ScheduledTask, error) {
|
func (c *Client) GetTaskStatus(query *aurora.TaskQuery) ([]*aurora.ScheduledTask, error) {
|
||||||
|
|
||||||
r.logger.DebugPrintf("GetTasksStatus Thrift Payload: %+v\n", query)
|
c.logger.DebugPrintf("GetTasksStatus Thrift Payload: %+v\n", query)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.GetTasksStatus(nil, query)
|
return c.client.GetTasksStatus(nil, query)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -822,12 +822,12 @@ func (r *RealisClient) GetTaskStatus(query *aurora.TaskQuery) ([]*aurora.Schedul
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get pending reason
|
// Get pending reason
|
||||||
func (r *RealisClient) GetPendingReason(query *aurora.TaskQuery) ([]*aurora.PendingReason, error) {
|
func (c *Client) GetPendingReason(query *aurora.TaskQuery) ([]*aurora.PendingReason, error) {
|
||||||
|
|
||||||
r.logger.DebugPrintf("GetPendingReason Thrift Payload: %+v\n", query)
|
c.logger.DebugPrintf("GetPendingReason Thrift Payload: %+v\n", query)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.GetPendingReason(nil, query)
|
return c.client.GetPendingReason(nil, query)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -844,12 +844,12 @@ func (r *RealisClient) GetPendingReason(query *aurora.TaskQuery) ([]*aurora.Pend
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get information about task including without a task configuration object
|
// Get information about task including without a task configuration object
|
||||||
func (r *RealisClient) GetTasksWithoutConfigs(query *aurora.TaskQuery) ([]*aurora.ScheduledTask, error) {
|
func (c *Client) GetTasksWithoutConfigs(query *aurora.TaskQuery) ([]*aurora.ScheduledTask, error) {
|
||||||
|
|
||||||
r.logger.DebugPrintf("GetTasksWithoutConfigs Thrift Payload: %+v\n", query)
|
c.logger.DebugPrintf("GetTasksWithoutConfigs Thrift Payload: %+v\n", query)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.GetTasksWithoutConfigs(nil, query)
|
return c.client.GetTasksWithoutConfigs(nil, query)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -861,7 +861,7 @@ func (r *RealisClient) GetTasksWithoutConfigs(query *aurora.TaskQuery) ([]*auror
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get the task configuration from the aurora scheduler for a job
|
// Get the task configuration from the aurora scheduler for a job
|
||||||
func (r *RealisClient) FetchTaskConfig(instKey aurora.InstanceKey) (*aurora.TaskConfig, error) {
|
func (c *Client) FetchTaskConfig(instKey aurora.InstanceKey) (*aurora.TaskConfig, error) {
|
||||||
|
|
||||||
ids := []int32{instKey.GetInstanceId()}
|
ids := []int32{instKey.GetInstanceId()}
|
||||||
|
|
||||||
|
@ -873,10 +873,10 @@ func (r *RealisClient) FetchTaskConfig(instKey aurora.InstanceKey) (*aurora.Task
|
||||||
Statuses: aurora.ACTIVE_STATES,
|
Statuses: aurora.ACTIVE_STATES,
|
||||||
}
|
}
|
||||||
|
|
||||||
r.logger.DebugPrintf("GetTasksStatus Thrift Payload: %+v\n", taskQ)
|
c.logger.DebugPrintf("GetTasksStatus Thrift Payload: %+v\n", taskQ)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.GetTasksStatus(nil, taskQ)
|
return c.client.GetTasksStatus(nil, taskQ)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -897,12 +897,12 @@ func (r *RealisClient) FetchTaskConfig(instKey aurora.InstanceKey) (*aurora.Task
|
||||||
return tasks[0].AssignedTask.Task, nil
|
return tasks[0].AssignedTask.Task, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) JobUpdateDetails(updateQuery aurora.JobUpdateQuery) ([]*aurora.JobUpdateDetails, error) {
|
func (c *Client) JobUpdateDetails(updateQuery aurora.JobUpdateQuery) ([]*aurora.JobUpdateDetails, error) {
|
||||||
|
|
||||||
r.logger.DebugPrintf("GetJobUpdateDetails Thrift Payload: %+v\n", updateQuery)
|
c.logger.DebugPrintf("GetJobUpdateDetails Thrift Payload: %+v\n", updateQuery)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.GetJobUpdateDetails(nil, &updateQuery)
|
return c.client.GetJobUpdateDetails(nil, &updateQuery)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -916,12 +916,12 @@ func (r *RealisClient) JobUpdateDetails(updateQuery aurora.JobUpdateQuery) ([]*a
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) RollbackJobUpdate(key aurora.JobUpdateKey, message string) error {
|
func (c *Client) RollbackJobUpdate(key aurora.JobUpdateKey, message string) error {
|
||||||
|
|
||||||
r.logger.DebugPrintf("RollbackJobUpdate Thrift Payload: %+v %v\n", key, message)
|
c.logger.DebugPrintf("RollbackJobUpdate Thrift Payload: %+v %v\n", key, message)
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.client.RollbackJobUpdate(nil, &key, message)
|
return c.client.RollbackJobUpdate(nil, &key, message)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -937,7 +937,7 @@ func (r *RealisClient) RollbackJobUpdate(key aurora.JobUpdateKey, message string
|
||||||
// Set a list of nodes to DRAINING. This means nothing will be able to be scheduled on them and any existing
|
// Set a list of nodes to DRAINING. This means nothing will be able to be scheduled on them and any existing
|
||||||
// tasks will be killed and re-scheduled elsewhere in the cluster. Tasks from DRAINING nodes are not guaranteed
|
// tasks will be killed and re-scheduled elsewhere in the cluster. Tasks from DRAINING nodes are not guaranteed
|
||||||
// to return to running unless there is enough capacity in the cluster to run them.
|
// to return to running unless there is enough capacity in the cluster to run them.
|
||||||
func (r *RealisClient) DrainHosts(hosts ...string) ([]*aurora.HostStatus, error) {
|
func (c *Client) DrainHosts(hosts ...string) ([]*aurora.HostStatus, error) {
|
||||||
|
|
||||||
if len(hosts) == 0 {
|
if len(hosts) == 0 {
|
||||||
return nil, errors.New("no hosts provided to drain")
|
return nil, errors.New("no hosts provided to drain")
|
||||||
|
@ -946,10 +946,10 @@ func (r *RealisClient) DrainHosts(hosts ...string) ([]*aurora.HostStatus, error)
|
||||||
drainList := aurora.NewHosts()
|
drainList := aurora.NewHosts()
|
||||||
drainList.HostNames = hosts
|
drainList.HostNames = hosts
|
||||||
|
|
||||||
r.logger.DebugPrintf("DrainHosts Thrift Payload: %v\n", drainList)
|
c.logger.DebugPrintf("DrainHosts Thrift Payload: %v\n", drainList)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.DrainHosts(nil, drainList)
|
return c.adminClient.DrainHosts(nil, drainList)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -966,7 +966,7 @@ func (r *RealisClient) DrainHosts(hosts ...string) ([]*aurora.HostStatus, error)
|
||||||
// Start SLA Aware Drain.
|
// Start SLA Aware Drain.
|
||||||
// defaultSlaPolicy is the fallback SlaPolicy to use if a task does not have an SlaPolicy.
|
// defaultSlaPolicy is the fallback SlaPolicy to use if a task does not have an SlaPolicy.
|
||||||
// After timeoutSecs, tasks will be forcefully drained without checking SLA.
|
// After timeoutSecs, tasks will be forcefully drained without checking SLA.
|
||||||
func (r *RealisClient) SLADrainHosts(policy *aurora.SlaPolicy, timeout int64, hosts ...string) ([]*aurora.HostStatus, error) {
|
func (c *Client) SLADrainHosts(policy *aurora.SlaPolicy, timeout int64, hosts ...string) ([]*aurora.HostStatus, error) {
|
||||||
|
|
||||||
if len(hosts) == 0 {
|
if len(hosts) == 0 {
|
||||||
return nil, errors.New("no hosts provided to drain")
|
return nil, errors.New("no hosts provided to drain")
|
||||||
|
@ -975,10 +975,10 @@ func (r *RealisClient) SLADrainHosts(policy *aurora.SlaPolicy, timeout int64, ho
|
||||||
drainList := aurora.NewHosts()
|
drainList := aurora.NewHosts()
|
||||||
drainList.HostNames = hosts
|
drainList.HostNames = hosts
|
||||||
|
|
||||||
r.logger.DebugPrintf("SLADrainHosts Thrift Payload: %v\n", drainList)
|
c.logger.DebugPrintf("SLADrainHosts Thrift Payload: %v\n", drainList)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.SlaDrainHosts(nil, drainList, policy, timeout)
|
return c.adminClient.SlaDrainHosts(nil, drainList, policy, timeout)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -992,7 +992,7 @@ func (r *RealisClient) SLADrainHosts(policy *aurora.SlaPolicy, timeout int64, ho
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) StartMaintenance(hosts ...string) ([]*aurora.HostStatus, error) {
|
func (c *Client) StartMaintenance(hosts ...string) ([]*aurora.HostStatus, error) {
|
||||||
|
|
||||||
if len(hosts) == 0 {
|
if len(hosts) == 0 {
|
||||||
return nil, errors.New("no hosts provided to start maintenance on")
|
return nil, errors.New("no hosts provided to start maintenance on")
|
||||||
|
@ -1001,10 +1001,10 @@ func (r *RealisClient) StartMaintenance(hosts ...string) ([]*aurora.HostStatus,
|
||||||
hostList := aurora.NewHosts()
|
hostList := aurora.NewHosts()
|
||||||
hostList.HostNames = hosts
|
hostList.HostNames = hosts
|
||||||
|
|
||||||
r.logger.DebugPrintf("StartMaintenance Thrift Payload: %v\n", hostList)
|
c.logger.DebugPrintf("StartMaintenance Thrift Payload: %v\n", hostList)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.StartMaintenance(nil, hostList)
|
return c.adminClient.StartMaintenance(nil, hostList)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1018,7 +1018,7 @@ func (r *RealisClient) StartMaintenance(hosts ...string) ([]*aurora.HostStatus,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) EndMaintenance(hosts ...string) ([]*aurora.HostStatus, error) {
|
func (c *Client) EndMaintenance(hosts ...string) ([]*aurora.HostStatus, error) {
|
||||||
|
|
||||||
if len(hosts) == 0 {
|
if len(hosts) == 0 {
|
||||||
return nil, errors.New("no hosts provided to end maintenance on")
|
return nil, errors.New("no hosts provided to end maintenance on")
|
||||||
|
@ -1027,10 +1027,10 @@ func (r *RealisClient) EndMaintenance(hosts ...string) ([]*aurora.HostStatus, er
|
||||||
hostList := aurora.NewHosts()
|
hostList := aurora.NewHosts()
|
||||||
hostList.HostNames = hosts
|
hostList.HostNames = hosts
|
||||||
|
|
||||||
r.logger.DebugPrintf("EndMaintenance Thrift Payload: %v\n", hostList)
|
c.logger.DebugPrintf("EndMaintenance Thrift Payload: %v\n", hostList)
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.EndMaintenance(nil, hostList)
|
return c.adminClient.EndMaintenance(nil, hostList)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1045,7 +1045,7 @@ func (r *RealisClient) EndMaintenance(hosts ...string) ([]*aurora.HostStatus, er
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RealisClient) MaintenanceStatus(hosts ...string) (*aurora.MaintenanceStatusResult_, error) {
|
func (c *Client) MaintenanceStatus(hosts ...string) (*aurora.MaintenanceStatusResult_, error) {
|
||||||
|
|
||||||
var result *aurora.MaintenanceStatusResult_
|
var result *aurora.MaintenanceStatusResult_
|
||||||
|
|
||||||
|
@ -1056,12 +1056,12 @@ func (r *RealisClient) MaintenanceStatus(hosts ...string) (*aurora.MaintenanceSt
|
||||||
hostList := aurora.NewHosts()
|
hostList := aurora.NewHosts()
|
||||||
hostList.HostNames = hosts
|
hostList.HostNames = hosts
|
||||||
|
|
||||||
r.logger.DebugPrintf("MaintenanceStatus Thrift Payload: %v\n", hostList)
|
c.logger.DebugPrintf("MaintenanceStatus Thrift Payload: %v\n", hostList)
|
||||||
|
|
||||||
// Make thrift call. If we encounter an error sending the call, attempt to reconnect
|
// Make thrift call. If we encounter an error sending the call, attempt to reconnect
|
||||||
// and continue trying to resend command until we run out of retries.
|
// and continue trying to resend command until we run out of retries.
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.MaintenanceStatus(nil, hostList)
|
return c.adminClient.MaintenanceStatus(nil, hostList)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1077,7 +1077,7 @@ func (r *RealisClient) MaintenanceStatus(hosts ...string) (*aurora.MaintenanceSt
|
||||||
|
|
||||||
// SetQuota sets a quota aggregate for the given role
|
// SetQuota sets a quota aggregate for the given role
|
||||||
// TODO(zircote) Currently investigating an error that is returned from thrift calls that include resources for `NamedPort` and `NumGpu`
|
// TODO(zircote) Currently investigating an error that is returned from thrift calls that include resources for `NamedPort` and `NumGpu`
|
||||||
func (r *RealisClient) SetQuota(role string, cpu *float64, ramMb *int64, diskMb *int64) error {
|
func (c *Client) SetQuota(role string, cpu *float64, ramMb *int64, diskMb *int64) error {
|
||||||
ramResource := aurora.NewResource()
|
ramResource := aurora.NewResource()
|
||||||
ramResource.RamMb = ramMb
|
ramResource.RamMb = ramMb
|
||||||
cpuResource := aurora.NewResource()
|
cpuResource := aurora.NewResource()
|
||||||
|
@ -1088,8 +1088,8 @@ func (r *RealisClient) SetQuota(role string, cpu *float64, ramMb *int64, diskMb
|
||||||
quota := aurora.NewResourceAggregate()
|
quota := aurora.NewResourceAggregate()
|
||||||
quota.Resources = []*aurora.Resource{ramResource, cpuResource, diskResource}
|
quota.Resources = []*aurora.Resource{ramResource, cpuResource, diskResource}
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.SetQuota(nil, role, quota)
|
return c.adminClient.SetQuota(nil, role, quota)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1100,10 +1100,10 @@ func (r *RealisClient) SetQuota(role string, cpu *float64, ramMb *int64, diskMb
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetQuota returns the resource aggregate for the given role
|
// GetQuota returns the resource aggregate for the given role
|
||||||
func (r *RealisClient) GetQuota(role string) (*aurora.GetQuotaResult_, error) {
|
func (c *Client) GetQuota(role string) (*aurora.GetQuotaResult_, error) {
|
||||||
|
|
||||||
resp, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
resp, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.GetQuota(nil, role)
|
return c.adminClient.GetQuota(nil, role)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1118,10 +1118,10 @@ func (r *RealisClient) GetQuota(role string) (*aurora.GetQuotaResult_, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Force Aurora Scheduler to perform a snapshot and write to Mesos log
|
// Force Aurora Scheduler to perform a snapshot and write to Mesos log
|
||||||
func (r *RealisClient) Snapshot() error {
|
func (c *Client) Snapshot() error {
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.Snapshot(nil)
|
return c.adminClient.Snapshot(nil)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1132,10 +1132,10 @@ func (r *RealisClient) Snapshot() error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Force Aurora Scheduler to write backup file to a file in the backup directory
|
// Force Aurora Scheduler to write backup file to a file in the backup directory
|
||||||
func (r *RealisClient) PerformBackup() error {
|
func (c *Client) PerformBackup() error {
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.PerformBackup(nil)
|
return c.adminClient.PerformBackup(nil)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1146,10 +1146,10 @@ func (r *RealisClient) PerformBackup() error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Force an Implicit reconciliation between Mesos and Aurora
|
// Force an Implicit reconciliation between Mesos and Aurora
|
||||||
func (r *RealisClient) ForceImplicitTaskReconciliation() error {
|
func (c *Client) ForceImplicitTaskReconciliation() error {
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.TriggerImplicitTaskReconciliation(nil)
|
return c.adminClient.TriggerImplicitTaskReconciliation(nil)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
@ -1160,7 +1160,7 @@ func (r *RealisClient) ForceImplicitTaskReconciliation() error {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Force an Explicit reconciliation between Mesos and Aurora
|
// Force an Explicit reconciliation between Mesos and Aurora
|
||||||
func (r *RealisClient) ForceExplicitTaskReconciliation(batchSize *int32) error {
|
func (c *Client) ForceExplicitTaskReconciliation(batchSize *int32) error {
|
||||||
|
|
||||||
if batchSize != nil && *batchSize < 1 {
|
if batchSize != nil && *batchSize < 1 {
|
||||||
return errors.New("Invalid batch size.")
|
return errors.New("Invalid batch size.")
|
||||||
|
@ -1169,8 +1169,8 @@ func (r *RealisClient) ForceExplicitTaskReconciliation(batchSize *int32) error {
|
||||||
|
|
||||||
settings.BatchSize = batchSize
|
settings.BatchSize = batchSize
|
||||||
|
|
||||||
_, retryErr := r.thriftCallWithRetries(func() (*aurora.Response, error) {
|
_, retryErr := c.thriftCallWithRetries(func() (*aurora.Response, error) {
|
||||||
return r.adminClient.TriggerExplicitTaskReconciliation(nil, settings)
|
return c.adminClient.TriggerExplicitTaskReconciliation(nil, settings)
|
||||||
})
|
})
|
||||||
|
|
||||||
if retryErr != nil {
|
if retryErr != nil {
|
||||||
|
|
|
@ -28,15 +28,14 @@ import (
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
)
|
)
|
||||||
|
|
||||||
var r *realis.RealisClient
|
var r *realis.Client
|
||||||
var monitor *realis.Monitor
|
|
||||||
var thermosPayload []byte
|
var thermosPayload []byte
|
||||||
|
|
||||||
func TestMain(m *testing.M) {
|
func TestMain(m *testing.M) {
|
||||||
var err error
|
var err error
|
||||||
|
|
||||||
// New configuration to connect to docker container
|
// New configuration to connect to docker container
|
||||||
r, err = realis.NewRealisClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
r, err = realis.NewClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
||||||
realis.BasicAuth("aurora", "secret"),
|
realis.BasicAuth("aurora", "secret"),
|
||||||
realis.Timeout(20*time.Second))
|
realis.Timeout(20*time.Second))
|
||||||
|
|
||||||
|
@ -47,9 +46,6 @@ func TestMain(m *testing.M) {
|
||||||
|
|
||||||
defer r.Close()
|
defer r.Close()
|
||||||
|
|
||||||
// Create monitor
|
|
||||||
monitor = &realis.Monitor{Client: r}
|
|
||||||
|
|
||||||
thermosPayload, err = ioutil.ReadFile("examples/thermos_payload.json")
|
thermosPayload, err = ioutil.ReadFile("examples/thermos_payload.json")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Println("Error reading thermos payload file: ", err)
|
fmt.Println("Error reading thermos payload file: ", err)
|
||||||
|
@ -67,7 +63,7 @@ func TestNonExistentEndpoint(t *testing.T) {
|
||||||
Jitter: 0.1}
|
Jitter: 0.1}
|
||||||
|
|
||||||
// Attempt to connect to a bad endpoint
|
// Attempt to connect to a bad endpoint
|
||||||
r, err := realis.NewRealisClient(realis.SchedulerUrl("http://192.168.33.7:8081/doesntexist/"),
|
r, err := realis.NewClient(realis.SchedulerUrl("http://192.168.33.7:8081/doesntexist/"),
|
||||||
realis.Timeout(200*time.Millisecond),
|
realis.Timeout(200*time.Millisecond),
|
||||||
realis.BackOff(backoff),
|
realis.BackOff(backoff),
|
||||||
)
|
)
|
||||||
|
@ -90,7 +86,7 @@ func TestNonExistentEndpoint(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestThriftBinary(t *testing.T) {
|
func TestThriftBinary(t *testing.T) {
|
||||||
r, err := realis.NewRealisClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
r, err := realis.NewClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
||||||
realis.BasicAuth("aurora", "secret"),
|
realis.BasicAuth("aurora", "secret"),
|
||||||
realis.Timeout(20*time.Second),
|
realis.Timeout(20*time.Second),
|
||||||
realis.ThriftBinary())
|
realis.ThriftBinary())
|
||||||
|
@ -112,7 +108,7 @@ func TestThriftBinary(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestThriftJSON(t *testing.T) {
|
func TestThriftJSON(t *testing.T) {
|
||||||
r, err := realis.NewRealisClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
r, err := realis.NewClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
||||||
realis.BasicAuth("aurora", "secret"),
|
realis.BasicAuth("aurora", "secret"),
|
||||||
realis.Timeout(20*time.Second),
|
realis.Timeout(20*time.Second),
|
||||||
realis.ThriftJSON())
|
realis.ThriftJSON())
|
||||||
|
@ -134,7 +130,7 @@ func TestThriftJSON(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestNoopLogger(t *testing.T) {
|
func TestNoopLogger(t *testing.T) {
|
||||||
r, err := realis.NewRealisClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
r, err := realis.NewClient(realis.SchedulerUrl("http://192.168.33.7:8081"),
|
||||||
realis.BasicAuth("aurora", "secret"),
|
realis.BasicAuth("aurora", "secret"),
|
||||||
realis.SetLogger(realis.NoopLogger{}))
|
realis.SetLogger(realis.NoopLogger{}))
|
||||||
|
|
||||||
|
@ -196,7 +192,7 @@ func TestRealisClient_CreateJob_Thermos(t *testing.T) {
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
// Test Instances Monitor
|
// Test Instances Monitor
|
||||||
success, err := monitor.Instances(job.JobKey(), job.GetInstanceCount(), 1*time.Second, 50*time.Second)
|
success, err := r.InstancesMonitor(job.JobKey(), job.GetInstanceCount(), 1*time.Second, 50*time.Second)
|
||||||
assert.True(t, success)
|
assert.True(t, success)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
|
@ -223,7 +219,7 @@ func TestRealisClient_CreateJob_Thermos(t *testing.T) {
|
||||||
err := r.KillJob(job.JobKey())
|
err := r.KillJob(job.JobKey())
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
success, err := monitor.Instances(job.JobKey(), 0, 1*time.Second, 50*time.Second)
|
success, err := r.InstancesMonitor(job.JobKey(), 0, 1*time.Second, 50*time.Second)
|
||||||
assert.True(t, success)
|
assert.True(t, success)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
})
|
})
|
||||||
|
@ -352,8 +348,9 @@ func TestRealisClient_CreateService_WithPulse_Thermos(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
t.Run("TestRealisClient_KillJob_Thermos", func(t *testing.T) {
|
t.Run("TestRealisClient_KillJob_Thermos", func(t *testing.T) {
|
||||||
r.AbortJobUpdate(*updateDetails[0].GetUpdate().GetSummary().GetKey(), "")
|
err := r.AbortJobUpdate(*updateDetails[0].GetUpdate().GetSummary().GetKey(), "")
|
||||||
err := r.KillJob(job.JobKey())
|
assert.NoError(t, err)
|
||||||
|
err = r.KillJob(job.JobKey())
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -386,7 +383,7 @@ func TestRealisClient_CreateService(t *testing.T) {
|
||||||
var ok bool
|
var ok bool
|
||||||
var mErr error
|
var mErr error
|
||||||
|
|
||||||
if ok, mErr = monitor.JobUpdate(*result.GetKey(), 5*time.Second, 180*time.Second); !ok || mErr != nil {
|
if ok, mErr = r.JobUpdateMonitor(*result.GetKey(), 5*time.Second, 180*time.Second); !ok || mErr != nil {
|
||||||
// Update may already be in a terminal state so don't check for error
|
// Update may already be in a terminal state so don't check for error
|
||||||
err := r.AbortJobUpdate(*result.GetKey(), "Monitor timed out.")
|
err := r.AbortJobUpdate(*result.GetKey(), "Monitor timed out.")
|
||||||
|
|
||||||
|
@ -466,7 +463,7 @@ func TestRealisClient_StartMaintenance(t *testing.T) {
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
hostResults, err := monitor.HostMaintenance(
|
hostResults, err := r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_SCHEDULED},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_SCHEDULED},
|
||||||
1*time.Second,
|
1*time.Second,
|
||||||
|
@ -478,7 +475,7 @@ func TestRealisClient_StartMaintenance(t *testing.T) {
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
_, err = monitor.HostMaintenance(
|
_, err = r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
||||||
5*time.Second,
|
5*time.Second,
|
||||||
|
@ -492,7 +489,7 @@ func TestRealisClient_DrainHosts(t *testing.T) {
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
hostResults, err := monitor.HostMaintenance(
|
hostResults, err := r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
||||||
1*time.Second,
|
1*time.Second,
|
||||||
|
@ -502,7 +499,7 @@ func TestRealisClient_DrainHosts(t *testing.T) {
|
||||||
|
|
||||||
t.Run("TestRealisClient_MonitorNontransitioned", func(t *testing.T) {
|
t.Run("TestRealisClient_MonitorNontransitioned", func(t *testing.T) {
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
hostResults, err := monitor.HostMaintenance(
|
hostResults, err := r.HostMaintenanceMonitor(
|
||||||
append(hosts, "IMAGINARY_HOST"),
|
append(hosts, "IMAGINARY_HOST"),
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
||||||
1*time.Second,
|
1*time.Second,
|
||||||
|
@ -518,7 +515,7 @@ func TestRealisClient_DrainHosts(t *testing.T) {
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
_, err = monitor.HostMaintenance(
|
_, err = r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
||||||
5*time.Second,
|
5*time.Second,
|
||||||
|
@ -539,7 +536,7 @@ func TestRealisClient_SLADrainHosts(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
hostResults, err := monitor.HostMaintenance(
|
hostResults, err := r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_DRAINED, aurora.MaintenanceMode_DRAINING},
|
||||||
1*time.Second,
|
1*time.Second,
|
||||||
|
@ -551,7 +548,7 @@ func TestRealisClient_SLADrainHosts(t *testing.T) {
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
// Monitor change to DRAINING and DRAINED mode
|
// Monitor change to DRAINING and DRAINED mode
|
||||||
_, err = monitor.HostMaintenance(
|
_, err = r.HostMaintenanceMonitor(
|
||||||
hosts,
|
hosts,
|
||||||
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
[]aurora.MaintenanceMode{aurora.MaintenanceMode_NONE},
|
||||||
5*time.Second,
|
5*time.Second,
|
||||||
|
@ -587,7 +584,7 @@ func TestRealisClient_SessionThreadSafety(t *testing.T) {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
|
|
||||||
// Test Schedule status monitor for terminal state and timing out after 30 seconds
|
// Test Schedule status monitor for terminal state and timing out after 30 seconds
|
||||||
success, err := monitor.ScheduleStatus(job.JobKey(), job.GetInstanceCount(), aurora.LIVE_STATES, 1, 30)
|
success, err := r.ScheduleStatusMonitor(job.JobKey(), job.GetInstanceCount(), aurora.LIVE_STATES, 1, 30)
|
||||||
assert.False(t, success)
|
assert.False(t, success)
|
||||||
assert.Error(t, err)
|
assert.Error(t, err)
|
||||||
|
|
||||||
|
|
30
retry.go
30
retry.go
|
@ -116,12 +116,12 @@ func ExponentialBackoff(backoff Backoff, logger Logger, condition ConditionFunc)
|
||||||
type auroraThriftCall func() (resp *aurora.Response, err error)
|
type auroraThriftCall func() (resp *aurora.Response, err error)
|
||||||
|
|
||||||
// Duplicates the functionality of ExponentialBackoff but is specifically targeted towards ThriftCalls.
|
// Duplicates the functionality of ExponentialBackoff but is specifically targeted towards ThriftCalls.
|
||||||
func (r *RealisClient) thriftCallWithRetries(thriftCall auroraThriftCall) (*aurora.Response, error) {
|
func (c *Client) thriftCallWithRetries(thriftCall auroraThriftCall) (*aurora.Response, error) {
|
||||||
var resp *aurora.Response
|
var resp *aurora.Response
|
||||||
var clientErr error
|
var clientErr error
|
||||||
var curStep int
|
var curStep int
|
||||||
|
|
||||||
backoff := r.config.backoff
|
backoff := c.config.backoff
|
||||||
duration := backoff.Duration
|
duration := backoff.Duration
|
||||||
|
|
||||||
for curStep = 0; curStep < backoff.Steps; curStep++ {
|
for curStep = 0; curStep < backoff.Steps; curStep++ {
|
||||||
|
@ -133,7 +133,7 @@ func (r *RealisClient) thriftCallWithRetries(thriftCall auroraThriftCall) (*auro
|
||||||
adjusted = Jitter(duration, backoff.Jitter)
|
adjusted = Jitter(duration, backoff.Jitter)
|
||||||
}
|
}
|
||||||
|
|
||||||
r.logger.Printf("A retriable error occurred during thrift call, backing off for %v before retry %v\n", adjusted, curStep)
|
c.logger.Printf("A retriable error occurred during thrift call, backing off for %v before retry %v\n", adjusted, curStep)
|
||||||
|
|
||||||
time.Sleep(adjusted)
|
time.Sleep(adjusted)
|
||||||
duration = time.Duration(float64(duration) * backoff.Factor)
|
duration = time.Duration(float64(duration) * backoff.Factor)
|
||||||
|
@ -143,12 +143,12 @@ func (r *RealisClient) thriftCallWithRetries(thriftCall auroraThriftCall) (*auro
|
||||||
// Placing this in an anonymous function in order to create a new, short-lived stack allowing unlock
|
// Placing this in an anonymous function in order to create a new, short-lived stack allowing unlock
|
||||||
// to be run in case of a panic inside of thriftCall.
|
// to be run in case of a panic inside of thriftCall.
|
||||||
func() {
|
func() {
|
||||||
r.lock.Lock()
|
c.lock.Lock()
|
||||||
defer r.lock.Unlock()
|
defer c.lock.Unlock()
|
||||||
|
|
||||||
resp, clientErr = thriftCall()
|
resp, clientErr = thriftCall()
|
||||||
|
|
||||||
r.logger.DebugPrintf("Aurora Thrift Call ended resp: %v clientErr: %v\n", resp, clientErr)
|
c.logger.DebugPrintf("Aurora Thrift Call ended resp: %v clientErr: %v\n", resp, clientErr)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
// Check if our thrift call is returning an error. This is a retriable event as we don't know
|
// Check if our thrift call is returning an error. This is a retriable event as we don't know
|
||||||
|
@ -156,12 +156,12 @@ func (r *RealisClient) thriftCallWithRetries(thriftCall auroraThriftCall) (*auro
|
||||||
if clientErr != nil {
|
if clientErr != nil {
|
||||||
|
|
||||||
// Print out the error to the user
|
// Print out the error to the user
|
||||||
r.logger.Printf("Client Error: %v\n", clientErr)
|
c.logger.Printf("Client Error: %v\n", clientErr)
|
||||||
|
|
||||||
// Determine if error is a temporary URL error by going up the stack
|
// Determine if error is a temporary URL error by going up the stack
|
||||||
e, ok := clientErr.(thrift.TTransportException)
|
e, ok := clientErr.(thrift.TTransportException)
|
||||||
if ok {
|
if ok {
|
||||||
r.logger.DebugPrint("Encountered a transport exception")
|
c.logger.DebugPrint("Encountered a transport exception")
|
||||||
|
|
||||||
e, ok := e.Err().(*url.Error)
|
e, ok := e.Err().(*url.Error)
|
||||||
if ok {
|
if ok {
|
||||||
|
@ -176,7 +176,7 @@ func (r *RealisClient) thriftCallWithRetries(thriftCall auroraThriftCall) (*auro
|
||||||
|
|
||||||
// In the future, reestablish connection should be able to check if it is actually possible
|
// In the future, reestablish connection should be able to check if it is actually possible
|
||||||
// to make a thrift call to Aurora. For now, a reconnect should always lead to a retry.
|
// to make a thrift call to Aurora. For now, a reconnect should always lead to a retry.
|
||||||
r.ReestablishConn()
|
c.ReestablishConn()
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
|
|
||||||
|
@ -195,31 +195,31 @@ func (r *RealisClient) thriftCallWithRetries(thriftCall auroraThriftCall) (*auro
|
||||||
|
|
||||||
// If the response code is transient, continue retrying
|
// If the response code is transient, continue retrying
|
||||||
case aurora.ResponseCode_ERROR_TRANSIENT:
|
case aurora.ResponseCode_ERROR_TRANSIENT:
|
||||||
r.logger.Println("Aurora replied with Transient error code, retrying")
|
c.logger.Println("Aurora replied with Transient error code, retrying")
|
||||||
continue
|
continue
|
||||||
|
|
||||||
// Failure scenarios, these indicate a bad payload or a bad config. Stop retrying.
|
// Failure scenarios, these indicate a bad payload or a bad clientConfig. Stop retrying.
|
||||||
case aurora.ResponseCode_INVALID_REQUEST,
|
case aurora.ResponseCode_INVALID_REQUEST,
|
||||||
aurora.ResponseCode_ERROR,
|
aurora.ResponseCode_ERROR,
|
||||||
aurora.ResponseCode_AUTH_FAILED,
|
aurora.ResponseCode_AUTH_FAILED,
|
||||||
aurora.ResponseCode_JOB_UPDATING_ERROR:
|
aurora.ResponseCode_JOB_UPDATING_ERROR:
|
||||||
r.logger.Printf("Terminal Response Code %v from Aurora, won't retry\n", resp.GetResponseCode().String())
|
c.logger.Printf("Terminal Response Code %v from Aurora, won't retry\n", resp.GetResponseCode().String())
|
||||||
return resp, errors.New(response.CombineMessage(resp))
|
return resp, errors.New(response.CombineMessage(resp))
|
||||||
|
|
||||||
// The only case that should fall down to here is a WARNING response code.
|
// The only case that should fall down to here is a WARNING response code.
|
||||||
// It is currently not used as a response in the scheduler so it is unknown how to handle it.
|
// It is currently not used as a response in the scheduler so it is unknown how to handle it.
|
||||||
default:
|
default:
|
||||||
r.logger.DebugPrintf("unhandled response code %v received from Aurora\n", responseCode)
|
c.logger.DebugPrintf("unhandled response code %v received from Aurora\n", responseCode)
|
||||||
return nil, errors.Errorf("unhandled response code from Aurora %v\n", responseCode.String())
|
return nil, errors.Errorf("unhandled response code from Aurora %v\n", responseCode.String())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
r.logger.DebugPrintf("it took %v retries to complete this operation\n", curStep)
|
c.logger.DebugPrintf("it took %v retries to complete this operation\n", curStep)
|
||||||
|
|
||||||
if curStep > 1 {
|
if curStep > 1 {
|
||||||
r.config.logger.Printf("retried this thrift call %d time(s)", curStep)
|
c.config.logger.Printf("retried this thrift call %d time(s)", curStep)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Provide more information to the user wherever possible.
|
// Provide more information to the user wherever possible.
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue