2015-06-30 07:49:13 +08:00
|
|
|
// +build linux
|
|
|
|
|
2015-06-22 10:31:12 +08:00
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
|
|
|
"encoding/json"
|
2016-05-10 13:58:09 +08:00
|
|
|
"fmt"
|
2015-06-22 10:31:12 +08:00
|
|
|
"os"
|
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/Sirupsen/logrus"
|
|
|
|
"github.com/opencontainers/runc/libcontainer"
|
2016-04-26 07:15:48 +08:00
|
|
|
"github.com/opencontainers/runc/libcontainer/cgroups"
|
2016-06-07 02:45:46 +08:00
|
|
|
"github.com/urfave/cli"
|
2015-06-22 10:31:12 +08:00
|
|
|
)
|
|
|
|
|
|
|
|
// event struct for encoding the event data to json.
|
|
|
|
type event struct {
|
|
|
|
Type string `json:"type"`
|
|
|
|
ID string `json:"id"`
|
|
|
|
Data interface{} `json:"data,omitempty"`
|
|
|
|
}
|
|
|
|
|
2016-04-26 07:15:48 +08:00
|
|
|
// stats is the runc specific stats structure for stability when encoding and decoding stats.
|
|
|
|
type stats struct {
|
|
|
|
Cpu cpu `json:"cpu"`
|
|
|
|
Memory memory `json:"memory"`
|
|
|
|
Pids pids `json:"pids"`
|
|
|
|
Blkio blkio `json:"blkio"`
|
|
|
|
Hugetlb map[string]hugetlb `json:"hugetlb"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type hugetlb struct {
|
|
|
|
Usage uint64 `json:"usage,omitempty"`
|
|
|
|
Max uint64 `json:"max,omitempty"`
|
|
|
|
Failcnt uint64 `json:"failcnt"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type blkioEntry struct {
|
|
|
|
Major uint64 `json:"major,omitempty"`
|
|
|
|
Minor uint64 `json:"minor,omitempty"`
|
|
|
|
Op string `json:"op,omitempty"`
|
|
|
|
Value uint64 `json:"value,omitempty"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type blkio struct {
|
|
|
|
IoServiceBytesRecursive []blkioEntry `json:"ioServiceBytesRecursive,omitempty"`
|
|
|
|
IoServicedRecursive []blkioEntry `json:"ioServicedRecursive,omitempty"`
|
|
|
|
IoQueuedRecursive []blkioEntry `json:"ioQueueRecursive,omitempty"`
|
|
|
|
IoServiceTimeRecursive []blkioEntry `json:"ioServiceTimeRecursive,omitempty"`
|
|
|
|
IoWaitTimeRecursive []blkioEntry `json:"ioWaitTimeRecursive,omitempty"`
|
|
|
|
IoMergedRecursive []blkioEntry `json:"ioMergedRecursive,omitempty"`
|
|
|
|
IoTimeRecursive []blkioEntry `json:"ioTimeRecursive,omitempty"`
|
|
|
|
SectorsRecursive []blkioEntry `json:"sectorsRecursive,omitempty"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type pids struct {
|
|
|
|
Current uint64 `json:"current,omitempty"`
|
|
|
|
Limit uint64 `json:"limit,omitempty"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type throttling struct {
|
|
|
|
Periods uint64 `json:"periods,omitempty"`
|
|
|
|
ThrottledPeriods uint64 `json:"throttledPeriods,omitempty"`
|
|
|
|
ThrottledTime uint64 `json:"throttledTime,omitempty"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type cpuUsage struct {
|
|
|
|
// Units: nanoseconds.
|
|
|
|
Total uint64 `json:"total,omitempty"`
|
|
|
|
Percpu []uint64 `json:"percpu,omitempty"`
|
|
|
|
Kernel uint64 `json:"kernel"`
|
|
|
|
User uint64 `json:"user"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type cpu struct {
|
|
|
|
Usage cpuUsage `json:"usage,omitempty"`
|
|
|
|
Throttling throttling `json:"throttling,omitempty"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type memoryEntry struct {
|
|
|
|
Limit uint64 `json:"limit"`
|
|
|
|
Usage uint64 `json:"usage,omitempty"`
|
|
|
|
Max uint64 `json:"max,omitempty"`
|
|
|
|
Failcnt uint64 `json:"failcnt"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type memory struct {
|
|
|
|
Cache uint64 `json:"cache,omitempty"`
|
|
|
|
Usage memoryEntry `json:"usage,omitempty"`
|
|
|
|
Swap memoryEntry `json:"swap,omitempty"`
|
|
|
|
Kernel memoryEntry `json:"kernel,omitempty"`
|
|
|
|
KernelTCP memoryEntry `json:"kernelTCP,omitempty"`
|
|
|
|
Raw map[string]uint64 `json:"raw,omitempty"`
|
|
|
|
}
|
|
|
|
|
2015-06-22 10:31:12 +08:00
|
|
|
var eventsCommand = cli.Command{
|
|
|
|
Name: "events",
|
2016-04-26 07:15:48 +08:00
|
|
|
Usage: "display container events such as OOM notifications, cpu, memory, and IO usage statistics",
|
2016-02-11 01:30:06 +08:00
|
|
|
ArgsUsage: `<container-id>
|
|
|
|
|
|
|
|
Where "<container-id>" is the name for the instance of the container.`,
|
|
|
|
Description: `The events command displays information about the container. By default the
|
|
|
|
information is displayed once every 5 seconds.`,
|
2015-06-22 10:31:12 +08:00
|
|
|
Flags: []cli.Flag{
|
|
|
|
cli.DurationFlag{Name: "interval", Value: 5 * time.Second, Usage: "set the stats collection interval"},
|
|
|
|
cli.BoolFlag{Name: "stats", Usage: "display the container's stats then exit"},
|
|
|
|
},
|
2016-05-10 13:58:09 +08:00
|
|
|
Action: func(context *cli.Context) error {
|
2015-06-22 10:31:12 +08:00
|
|
|
container, err := getContainer(context)
|
|
|
|
if err != nil {
|
2016-05-10 13:58:09 +08:00
|
|
|
return err
|
2015-06-22 10:31:12 +08:00
|
|
|
}
|
2016-06-02 15:29:42 +08:00
|
|
|
duration := context.Duration("interval")
|
|
|
|
if duration <= 0 {
|
|
|
|
return fmt.Errorf("duration interval must be greater than 0")
|
|
|
|
}
|
2016-04-20 23:03:39 +08:00
|
|
|
status, err := container.Status()
|
|
|
|
if err != nil {
|
2016-05-10 13:58:09 +08:00
|
|
|
return err
|
2016-04-20 23:03:39 +08:00
|
|
|
}
|
2016-05-14 08:01:12 +08:00
|
|
|
if status == libcontainer.Stopped {
|
2016-05-10 13:58:09 +08:00
|
|
|
return fmt.Errorf("container with id %s is not running", container.ID())
|
2016-04-20 23:03:39 +08:00
|
|
|
}
|
2015-06-22 10:31:12 +08:00
|
|
|
var (
|
|
|
|
stats = make(chan *libcontainer.Stats, 1)
|
|
|
|
events = make(chan *event, 1024)
|
|
|
|
group = &sync.WaitGroup{}
|
|
|
|
)
|
|
|
|
group.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer group.Done()
|
|
|
|
enc := json.NewEncoder(os.Stdout)
|
|
|
|
for e := range events {
|
|
|
|
if err := enc.Encode(e); err != nil {
|
|
|
|
logrus.Error(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
if context.Bool("stats") {
|
|
|
|
s, err := container.Stats()
|
|
|
|
if err != nil {
|
2016-05-10 13:58:09 +08:00
|
|
|
return err
|
2015-06-22 10:31:12 +08:00
|
|
|
}
|
2016-04-26 07:15:48 +08:00
|
|
|
events <- &event{Type: "stats", ID: container.ID(), Data: convertLibcontainerStats(s)}
|
2015-06-22 10:31:12 +08:00
|
|
|
close(events)
|
|
|
|
group.Wait()
|
2016-05-10 13:58:09 +08:00
|
|
|
return nil
|
2015-06-22 10:31:12 +08:00
|
|
|
}
|
|
|
|
go func() {
|
2015-06-30 07:49:13 +08:00
|
|
|
for range time.Tick(context.Duration("interval")) {
|
2015-06-22 10:31:12 +08:00
|
|
|
s, err := container.Stats()
|
|
|
|
if err != nil {
|
|
|
|
logrus.Error(err)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
stats <- s
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
n, err := container.NotifyOOM()
|
|
|
|
if err != nil {
|
2016-05-10 13:58:09 +08:00
|
|
|
return err
|
2015-06-22 10:31:12 +08:00
|
|
|
}
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case _, ok := <-n:
|
|
|
|
if ok {
|
|
|
|
// this means an oom event was received, if it is !ok then
|
|
|
|
// the channel was closed because the container stopped and
|
|
|
|
// the cgroups no longer exist.
|
|
|
|
events <- &event{Type: "oom", ID: container.ID()}
|
|
|
|
} else {
|
|
|
|
n = nil
|
|
|
|
}
|
|
|
|
case s := <-stats:
|
2016-04-26 07:15:48 +08:00
|
|
|
events <- &event{Type: "stats", ID: container.ID(), Data: convertLibcontainerStats(s)}
|
2015-06-22 10:31:12 +08:00
|
|
|
}
|
|
|
|
if n == nil {
|
|
|
|
close(events)
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
group.Wait()
|
2016-05-10 13:58:09 +08:00
|
|
|
return nil
|
2015-06-22 10:31:12 +08:00
|
|
|
},
|
|
|
|
}
|
2016-04-26 07:15:48 +08:00
|
|
|
|
|
|
|
func convertLibcontainerStats(ls *libcontainer.Stats) *stats {
|
|
|
|
cg := ls.CgroupStats
|
|
|
|
if cg == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
var s stats
|
|
|
|
s.Pids.Current = cg.PidsStats.Current
|
|
|
|
s.Pids.Limit = cg.PidsStats.Limit
|
|
|
|
|
|
|
|
s.Cpu.Usage.Kernel = cg.CpuStats.CpuUsage.UsageInKernelmode
|
|
|
|
s.Cpu.Usage.User = cg.CpuStats.CpuUsage.UsageInUsermode
|
|
|
|
s.Cpu.Usage.Total = cg.CpuStats.CpuUsage.TotalUsage
|
|
|
|
s.Cpu.Usage.Percpu = cg.CpuStats.CpuUsage.PercpuUsage
|
|
|
|
s.Cpu.Throttling.Periods = cg.CpuStats.ThrottlingData.Periods
|
|
|
|
s.Cpu.Throttling.ThrottledPeriods = cg.CpuStats.ThrottlingData.ThrottledPeriods
|
|
|
|
s.Cpu.Throttling.ThrottledTime = cg.CpuStats.ThrottlingData.ThrottledTime
|
|
|
|
|
|
|
|
s.Memory.Cache = cg.MemoryStats.Cache
|
|
|
|
s.Memory.Kernel = convertMemoryEntry(cg.MemoryStats.KernelUsage)
|
|
|
|
s.Memory.KernelTCP = convertMemoryEntry(cg.MemoryStats.KernelTCPUsage)
|
|
|
|
s.Memory.Swap = convertMemoryEntry(cg.MemoryStats.SwapUsage)
|
|
|
|
s.Memory.Usage = convertMemoryEntry(cg.MemoryStats.Usage)
|
|
|
|
s.Memory.Raw = cg.MemoryStats.Stats
|
|
|
|
|
|
|
|
s.Blkio.IoServiceBytesRecursive = convertBlkioEntry(cg.BlkioStats.IoServiceBytesRecursive)
|
|
|
|
s.Blkio.IoServicedRecursive = convertBlkioEntry(cg.BlkioStats.IoServicedRecursive)
|
|
|
|
s.Blkio.IoQueuedRecursive = convertBlkioEntry(cg.BlkioStats.IoQueuedRecursive)
|
|
|
|
s.Blkio.IoServiceTimeRecursive = convertBlkioEntry(cg.BlkioStats.IoServiceTimeRecursive)
|
|
|
|
s.Blkio.IoWaitTimeRecursive = convertBlkioEntry(cg.BlkioStats.IoWaitTimeRecursive)
|
|
|
|
s.Blkio.IoMergedRecursive = convertBlkioEntry(cg.BlkioStats.IoMergedRecursive)
|
|
|
|
s.Blkio.IoTimeRecursive = convertBlkioEntry(cg.BlkioStats.IoTimeRecursive)
|
|
|
|
s.Blkio.SectorsRecursive = convertBlkioEntry(cg.BlkioStats.SectorsRecursive)
|
|
|
|
|
|
|
|
s.Hugetlb = make(map[string]hugetlb)
|
|
|
|
for k, v := range cg.HugetlbStats {
|
|
|
|
s.Hugetlb[k] = convertHugtlb(v)
|
|
|
|
}
|
|
|
|
return &s
|
|
|
|
}
|
|
|
|
|
|
|
|
func convertHugtlb(c cgroups.HugetlbStats) hugetlb {
|
|
|
|
return hugetlb{
|
|
|
|
Usage: c.Usage,
|
|
|
|
Max: c.MaxUsage,
|
|
|
|
Failcnt: c.Failcnt,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func convertMemoryEntry(c cgroups.MemoryData) memoryEntry {
|
|
|
|
return memoryEntry{
|
|
|
|
Limit: c.Limit,
|
|
|
|
Usage: c.Usage,
|
|
|
|
Max: c.MaxUsage,
|
|
|
|
Failcnt: c.Failcnt,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func convertBlkioEntry(c []cgroups.BlkioStatEntry) []blkioEntry {
|
|
|
|
var out []blkioEntry
|
|
|
|
for _, e := range c {
|
|
|
|
out = append(out, blkioEntry{
|
|
|
|
Major: e.Major,
|
|
|
|
Minor: e.Minor,
|
|
|
|
Op: e.Op,
|
|
|
|
Value: e.Value,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
return out
|
|
|
|
}
|