From 8eda0c68b1a4089d637fc020eb7a737c951798b5 Mon Sep 17 00:00:00 2001 From: Pradyumna Kaushik Date: Mon, 28 Nov 2016 22:38:02 -0500 Subject: [PATCH] Removed TODO comment in ResourceOffers(...), that wasn't necessary anymore. --- schedulers/proactiveclusterwidecappingfcfs.go | 3 --- 1 file changed, 3 deletions(-) diff --git a/schedulers/proactiveclusterwidecappingfcfs.go b/schedulers/proactiveclusterwidecappingfcfs.go index 49094cd..37fa19a 100644 --- a/schedulers/proactiveclusterwidecappingfcfs.go +++ b/schedulers/proactiveclusterwidecappingfcfs.go @@ -234,7 +234,6 @@ func (s *ProactiveClusterwideCapFCFS) stopRecapping() { } } -// TODO: Need to reduce the time complexity: looping over offers twice (Possible to do it just once?). func (s *ProactiveClusterwideCapFCFS) ResourceOffers(driver sched.SchedulerDriver, offers []*mesos.Offer) { log.Printf("Received %d resource offers", len(offers)) @@ -272,8 +271,6 @@ func (s *ProactiveClusterwideCapFCFS) ResourceOffers(driver sched.SchedulerDrive 3. fcfsCurrentCapValue is updated with the determined cluster wide cap. Cluster wide capping is currently performed at regular intervals of time. - TODO: We can choose to cap the cluster only if the clusterwide cap varies more than the current clusterwide cap. - Although this sounds like a better approach, it only works when the resource requirements of neighbouring tasks are similar. */ taken := false