2017-02-15 19:22:56 -05:00
package pcp
import (
2017-09-26 00:18:35 -04:00
"bitbucket.org/sunybingcloud/elektron/constants"
2017-09-26 00:26:01 -04:00
"bitbucket.org/sunybingcloud/elektron/pcp"
2017-09-26 13:17:47 -04:00
"bitbucket.org/sunybingcloud/elektron/rapl"
2017-09-26 00:18:35 -04:00
"bitbucket.org/sunybingcloud/elektron/utilities"
2017-02-15 19:22:56 -05:00
"bufio"
"container/ring"
"log"
2017-02-20 20:55:06 -05:00
"math"
2017-02-15 19:22:56 -05:00
"os"
"os/exec"
"sort"
"strconv"
"strings"
"syscall"
2017-02-20 20:55:06 -05:00
"time"
2017-02-15 19:22:56 -05:00
)
func round ( num float64 ) int {
return int ( math . Floor ( num + math . Copysign ( 0.5 , num ) ) )
}
func getNextCapValue ( curCapValue float64 , precision int ) float64 {
2017-02-20 20:55:06 -05:00
curCapValue /= 2.0
output := math . Pow ( 10 , float64 ( precision ) )
return float64 ( round ( curCapValue * output ) ) / output
}
2017-02-15 19:22:56 -05:00
func StartPCPLogAndProgressiveExtremaCap ( quit chan struct { } , logging * bool , prefix string , hiThreshold , loThreshold float64 ) {
2017-02-20 20:55:06 -05:00
log . Println ( "Inside Log and Progressive Extrema" )
2017-02-15 19:22:56 -05:00
const pcpCommand string = "pmdumptext -m -l -f '' -t 1.0 -d , -c config"
cmd := exec . Command ( "sh" , "-c" , pcpCommand )
cmd . SysProcAttr = & syscall . SysProcAttr { Setpgid : true }
if hiThreshold < loThreshold {
2017-02-20 20:55:06 -05:00
log . Println ( "High threshold is lower than low threshold!" )
2017-02-15 19:22:56 -05:00
}
logFile , err := os . Create ( "./" + prefix + ".pcplog" )
if err != nil {
log . Fatal ( err )
}
defer logFile . Close ( )
pipe , err := cmd . StdoutPipe ( )
if err != nil {
log . Fatal ( err )
}
//cmd.Stdout = stdout
scanner := bufio . NewScanner ( pipe )
go func ( logging * bool , hiThreshold , loThreshold float64 ) {
// Get names of the columns
scanner . Scan ( )
// Write to logfile
logFile . WriteString ( scanner . Text ( ) + "\n" )
headers := strings . Split ( scanner . Text ( ) , "," )
powerIndexes := make ( [ ] int , 0 , 0 )
powerHistories := make ( map [ string ] * ring . Ring )
indexToHost := make ( map [ int ] string )
for i , hostMetric := range headers {
metricSplit := strings . Split ( hostMetric , ":" )
//log.Printf("%d Host %s: Metric: %s\n", i, split[0], split[1])
if strings . Contains ( metricSplit [ 1 ] , "RAPL_ENERGY_PKG" ) ||
strings . Contains ( metricSplit [ 1 ] , "RAPL_ENERGY_DRAM" ) {
//fmt.Println("Index: ", i)
powerIndexes = append ( powerIndexes , i )
indexToHost [ i ] = metricSplit [ 0 ]
// Only create one ring per host
if _ , ok := powerHistories [ metricSplit [ 0 ] ] ; ! ok {
powerHistories [ metricSplit [ 0 ] ] = ring . New ( 20 ) // Two PKGS, two DRAM per node, 20 = 5 seconds of tracking
}
}
}
// Throw away first set of results
scanner . Scan ( )
2017-02-22 20:09:04 -05:00
// To keep track of the capped states of the capped victims
2017-02-15 19:22:56 -05:00
cappedVictims := make ( map [ string ] float64 )
2017-02-22 20:09:04 -05:00
// TODO: Come with a better name for this.
2017-02-15 19:22:56 -05:00
orderCapped := make ( [ ] string , 0 , 8 )
2017-02-22 20:09:04 -05:00
// TODO: Change this to a priority queue ordered by the cap value. This will get rid of the sorting performed in the code.
// Parallel data structure to orderCapped to keep track of the uncapped states of the uncapped victims
orderCappedVictims := make ( map [ string ] float64 )
2017-02-15 19:22:56 -05:00
clusterPowerHist := ring . New ( 5 )
seconds := 0
for scanner . Scan ( ) {
if * logging {
log . Println ( "Logging PCP..." )
split := strings . Split ( scanner . Text ( ) , "," )
logFile . WriteString ( scanner . Text ( ) + "\n" )
totalPower := 0.0
for _ , powerIndex := range powerIndexes {
power , _ := strconv . ParseFloat ( split [ powerIndex ] , 64 )
host := indexToHost [ powerIndex ]
powerHistories [ host ] . Value = power
powerHistories [ host ] = powerHistories [ host ] . Next ( )
2017-09-26 00:26:01 -04:00
log . Printf ( "Host: %s, Power: %f" , indexToHost [ powerIndex ] , ( power * pcp . RAPLUnits ) )
2017-02-15 19:22:56 -05:00
totalPower += power
}
2017-09-26 00:26:01 -04:00
clusterPower := totalPower * pcp . RAPLUnits
2017-02-15 19:22:56 -05:00
clusterPowerHist . Value = clusterPower
clusterPowerHist = clusterPowerHist . Next ( )
2017-09-26 00:26:01 -04:00
clusterMean := pcp . AverageClusterPowerHistory ( clusterPowerHist )
2017-02-15 19:22:56 -05:00
log . Printf ( "Total power: %f, %d Sec Avg: %f" , clusterPower , clusterPowerHist . Len ( ) , clusterMean )
if clusterMean >= hiThreshold {
2017-02-20 20:55:06 -05:00
log . Println ( "Need to cap a node" )
2017-02-22 20:09:04 -05:00
log . Printf ( "Cap values of capped victims: %v" , cappedVictims )
log . Printf ( "Cap values of victims to uncap: %v" , orderCappedVictims )
2017-02-15 19:22:56 -05:00
// Create statics for all victims and choose one to cap
2017-09-26 00:26:01 -04:00
victims := make ( [ ] pcp . Victim , 0 , 8 )
2017-02-15 19:22:56 -05:00
// TODO: Just keep track of the largest to reduce fron nlogn to n
for name , history := range powerHistories {
2017-09-26 00:26:01 -04:00
histMean := pcp . AverageNodePowerHistory ( history )
2017-02-15 19:22:56 -05:00
// Consider doing mean calculations using go routines if we need to speed up
2017-09-26 00:26:01 -04:00
victims = append ( victims , pcp . Victim { Watts : histMean , Host : name } )
2017-02-15 19:22:56 -05:00
}
2017-09-26 00:26:01 -04:00
sort . Sort ( pcp . VictimSorter ( victims ) ) // Sort by average wattage
2017-02-15 19:22:56 -05:00
2017-02-20 20:55:06 -05:00
// Finding the best victim to cap in a round robin manner
newVictimFound := false
2017-02-22 20:09:04 -05:00
alreadyCappedHosts := [ ] string { } // Host-names of victims that are already capped
2017-02-15 19:22:56 -05:00
for i := 0 ; i < len ( victims ) ; i ++ {
2017-02-22 20:09:04 -05:00
// Try to pick a victim that hasn't been capped yet
2017-02-20 20:55:06 -05:00
if _ , ok := cappedVictims [ victims [ i ] . Host ] ; ! ok {
2017-02-22 20:09:04 -05:00
// If this victim can't be capped further, then we move on to find another victim
2017-02-20 20:55:06 -05:00
if _ , ok := orderCappedVictims [ victims [ i ] . Host ] ; ok {
continue
}
2017-02-22 20:09:04 -05:00
// Need to cap this victim
2017-02-20 20:55:06 -05:00
if err := rapl . Cap ( victims [ i ] . Host , "rapl" , 50.0 ) ; err != nil {
log . Printf ( "Error capping host %s" , victims [ i ] . Host )
2017-02-15 19:22:56 -05:00
} else {
2017-02-20 20:55:06 -05:00
log . Printf ( "Capped host[%s] at %f" , victims [ i ] . Host , 50.0 )
2017-02-22 20:09:04 -05:00
// Keeping track of this victim and it's cap value
2017-02-20 20:55:06 -05:00
cappedVictims [ victims [ i ] . Host ] = 50.0
newVictimFound = true
// This node can be uncapped and hence adding to orderCapped
2017-02-15 19:22:56 -05:00
orderCapped = append ( orderCapped , victims [ i ] . Host )
2017-02-20 20:55:06 -05:00
orderCappedVictims [ victims [ i ] . Host ] = 50.0
2017-02-22 20:09:04 -05:00
break // Breaking only on successful cap
2017-02-20 20:55:06 -05:00
}
} else {
alreadyCappedHosts = append ( alreadyCappedHosts , victims [ i ] . Host )
}
}
2017-02-22 20:09:04 -05:00
// If no new victim found, then we need to cap the best victim among the ones that are already capped
2017-02-20 20:55:06 -05:00
if ! newVictimFound {
2017-03-09 19:17:06 -05:00
canCapAlreadyCappedVictim := false
2017-02-20 20:55:06 -05:00
for i := 0 ; i < len ( alreadyCappedHosts ) ; i ++ {
2017-02-21 21:05:47 -05:00
// If already capped then the host must be present in orderCappedVictims
capValue := orderCappedVictims [ alreadyCappedHosts [ i ] ]
2017-02-22 20:09:04 -05:00
// If capValue is greater than the threshold then cap, else continue
2017-03-09 19:20:13 -05:00
if capValue > constants . LowerCapLimit {
2017-02-21 21:05:47 -05:00
newCapValue := getNextCapValue ( capValue , 2 )
2017-02-20 20:55:06 -05:00
if err := rapl . Cap ( alreadyCappedHosts [ i ] , "rapl" , newCapValue ) ; err != nil {
2017-02-22 20:09:04 -05:00
log . Printf ( "Error capping host[%s]" , alreadyCappedHosts [ i ] )
2017-02-20 20:55:06 -05:00
} else {
2017-02-22 20:09:04 -05:00
// Successful cap
2017-02-20 20:55:06 -05:00
log . Printf ( "Capped host[%s] at %f" , alreadyCappedHosts [ i ] , newCapValue )
// Checking whether this victim can be capped further
2017-03-09 19:20:13 -05:00
if newCapValue <= constants . LowerCapLimit {
2017-02-22 20:09:04 -05:00
// Deleting victim from cappedVictims
2017-02-20 20:55:06 -05:00
delete ( cappedVictims , alreadyCappedHosts [ i ] )
2017-02-22 20:09:04 -05:00
// Updating the cap value in orderCappedVictims
2017-02-21 21:05:47 -05:00
orderCappedVictims [ alreadyCappedHosts [ i ] ] = newCapValue
2017-02-20 20:55:06 -05:00
} else {
2017-02-22 20:09:04 -05:00
// Updating the cap value
2017-02-20 20:55:06 -05:00
cappedVictims [ alreadyCappedHosts [ i ] ] = newCapValue
2017-02-21 21:05:47 -05:00
orderCappedVictims [ alreadyCappedHosts [ i ] ] = newCapValue
2017-02-20 20:55:06 -05:00
}
2017-03-09 19:17:06 -05:00
canCapAlreadyCappedVictim = true
2017-02-22 20:09:04 -05:00
break // Breaking only on successful cap.
2017-02-20 20:55:06 -05:00
}
2017-02-21 21:05:47 -05:00
} else {
// Do nothing
// Continue to find another victim to cap.
2017-02-22 20:09:04 -05:00
// If cannot find any victim, then all nodes have been capped to the maximum and we stop capping at this point.
2017-02-15 19:22:56 -05:00
}
}
2017-03-23 22:13:29 -04:00
if ! canCapAlreadyCappedVictim {
2017-03-09 19:17:06 -05:00
log . Println ( "No Victim left to cap." )
}
2017-02-15 19:22:56 -05:00
}
} else if clusterMean < loThreshold {
2017-02-20 20:55:06 -05:00
log . Println ( "Need to uncap a node" )
2017-02-22 20:09:04 -05:00
log . Printf ( "Cap values of capped victims: %v" , cappedVictims )
log . Printf ( "Cap values of victims to uncap: %v" , orderCappedVictims )
2017-02-15 19:22:56 -05:00
if len ( orderCapped ) > 0 {
2017-02-22 20:09:04 -05:00
// We pick the host that is capped the most to uncap
orderCappedToSort := utilities . GetPairList ( orderCappedVictims )
sort . Sort ( orderCappedToSort ) // Sorted hosts in non-decreasing order of capped states
hostToUncap := orderCappedToSort [ 0 ] . Key
2017-03-09 19:20:13 -05:00
// Uncapping the host.
// This is a floating point operation and might suffer from precision loss.
2017-02-22 20:09:04 -05:00
newUncapValue := orderCappedVictims [ hostToUncap ] * 2.0
if err := rapl . Cap ( hostToUncap , "rapl" , newUncapValue ) ; err != nil {
log . Printf ( "Error uncapping host[%s]" , hostToUncap )
2017-02-20 20:55:06 -05:00
} else {
2017-02-21 21:05:47 -05:00
// Successful uncap
2017-02-22 20:09:04 -05:00
log . Printf ( "Uncapped host[%s] to %f" , hostToUncap , newUncapValue )
// Can we uncap this host further. If not, then we remove its entry from orderCapped
2017-02-21 21:05:47 -05:00
if newUncapValue >= 100.0 { // can compare using ==
2017-02-22 20:09:04 -05:00
// Deleting entry from orderCapped
for i , victimHost := range orderCapped {
if victimHost == hostToUncap {
orderCapped = append ( orderCapped [ : i ] , orderCapped [ i + 1 : ] ... )
break // We are done removing host from orderCapped
}
}
// Removing entry for host from the parallel data structure
delete ( orderCappedVictims , hostToUncap )
// Removing entry from cappedVictims as this host is no longer capped
delete ( cappedVictims , hostToUncap )
2017-03-09 19:20:13 -05:00
} else if newUncapValue > constants . LowerCapLimit { // this check is unnecessary and can be converted to 'else'
2017-02-22 20:09:04 -05:00
// Updating the cap value
orderCappedVictims [ hostToUncap ] = newUncapValue
cappedVictims [ hostToUncap ] = newUncapValue
2017-02-20 20:55:06 -05:00
}
2017-02-15 19:22:56 -05:00
}
2017-02-20 20:55:06 -05:00
} else {
2017-02-21 21:05:47 -05:00
log . Println ( "No host staged for Uncapping" )
2017-02-15 19:22:56 -05:00
}
}
}
seconds ++
}
} ( logging , hiThreshold , loThreshold )
2017-02-20 20:55:06 -05:00
log . Println ( "PCP logging started" )
if err := cmd . Start ( ) ; err != nil {
log . Fatal ( err )
}
pgid , err := syscall . Getpgid ( cmd . Process . Pid )
select {
case <- quit :
log . Println ( "Stopping PCP logging in 5 seconds" )
time . Sleep ( 5 * time . Second )
// http://stackoverflow.com/questions/22470193/why-wont-go-kill-a-child-process-correctly
// kill process and all children processes
syscall . Kill ( - pgid , 15 )
return
}
2017-02-15 19:22:56 -05:00
}