mirror of
https://github.com/containers/podman.git
synced 2025-09-28 17:25:31 +08:00

When you run podman stats, the first interval always shows the wrong cpu usage. To calculate cpu percentage we get the cpu time from the cgroup and compare this against the system time between two stats. Since the first time we do not have a previous stats an empty struct is used instead. Thus we do not use the actual running time of the container but the current unix timestamp (time since Jan 1 1970). To fix this we make sure that the previous stats time is set to the container start time, when it is empty. [NO NEW TESTS NEEDED] No idea how I could create a test which would have a predictable cpu usage. See the linked bugzilla for a reproducer. Fixes https://bugzilla.redhat.com/show_bug.cgi?id=2066145 Signed-off-by: Paul Holzinger <pholzing@redhat.com>
238 lines
6.4 KiB
Go
238 lines
6.4 KiB
Go
package compat
|
|
|
|
import (
|
|
"encoding/json"
|
|
"net/http"
|
|
"time"
|
|
|
|
"github.com/containers/common/pkg/cgroups"
|
|
"github.com/containers/podman/v4/libpod"
|
|
"github.com/containers/podman/v4/libpod/define"
|
|
"github.com/containers/podman/v4/pkg/api/handlers/utils"
|
|
api "github.com/containers/podman/v4/pkg/api/types"
|
|
docker "github.com/docker/docker/api/types"
|
|
"github.com/gorilla/schema"
|
|
"github.com/pkg/errors"
|
|
"github.com/sirupsen/logrus"
|
|
)
|
|
|
|
const DefaultStatsPeriod = 5 * time.Second
|
|
|
|
func StatsContainer(w http.ResponseWriter, r *http.Request) {
|
|
runtime := r.Context().Value(api.RuntimeKey).(*libpod.Runtime)
|
|
decoder := r.Context().Value(api.DecoderKey).(*schema.Decoder)
|
|
|
|
query := struct {
|
|
Stream bool `schema:"stream"`
|
|
OneShot bool `schema:"one-shot"` // added schema for one shot
|
|
}{
|
|
Stream: true,
|
|
}
|
|
if err := decoder.Decode(&query, r.URL.Query()); err != nil {
|
|
utils.Error(w, http.StatusBadRequest, errors.Wrapf(err, "failed to parse parameters for %s", r.URL.String()))
|
|
return
|
|
}
|
|
if query.Stream && query.OneShot { // mismatch. one-shot can only be passed with stream=false
|
|
utils.Error(w, http.StatusBadRequest, define.ErrInvalidArg)
|
|
return
|
|
}
|
|
|
|
name := utils.GetName(r)
|
|
ctnr, err := runtime.LookupContainer(name)
|
|
if err != nil {
|
|
utils.ContainerNotFound(w, name, err)
|
|
return
|
|
}
|
|
|
|
// If the container isn't running, then let's not bother and return
|
|
// immediately.
|
|
state, err := ctnr.State()
|
|
if err != nil {
|
|
utils.InternalServerError(w, err)
|
|
return
|
|
}
|
|
if state != define.ContainerStateRunning {
|
|
utils.Error(w, http.StatusConflict, define.ErrCtrStateInvalid)
|
|
return
|
|
}
|
|
|
|
stats, err := ctnr.GetContainerStats(nil)
|
|
if err != nil {
|
|
utils.InternalServerError(w, errors.Wrapf(err, "failed to obtain Container %s stats", name))
|
|
return
|
|
}
|
|
|
|
coder := json.NewEncoder(w)
|
|
// Write header and content type.
|
|
w.WriteHeader(http.StatusOK)
|
|
w.Header().Set("Content-Type", "application/json")
|
|
if flusher, ok := w.(http.Flusher); ok {
|
|
flusher.Flush()
|
|
}
|
|
|
|
// Setup JSON encoder for streaming.
|
|
coder.SetEscapeHTML(true)
|
|
var preRead time.Time
|
|
var preCPUStats CPUStats
|
|
if query.Stream {
|
|
preRead = time.Now()
|
|
systemUsage, _ := cgroups.GetSystemCPUUsage()
|
|
preCPUStats = CPUStats{
|
|
CPUUsage: docker.CPUUsage{
|
|
TotalUsage: stats.CPUNano,
|
|
PercpuUsage: stats.PerCPU,
|
|
UsageInKernelmode: stats.CPUSystemNano,
|
|
UsageInUsermode: stats.CPUNano - stats.CPUSystemNano,
|
|
},
|
|
CPU: stats.CPU,
|
|
SystemUsage: systemUsage,
|
|
OnlineCPUs: 0,
|
|
ThrottlingData: docker.ThrottlingData{},
|
|
}
|
|
}
|
|
|
|
streamLabel: // A label to flatten the scope
|
|
select {
|
|
case <-r.Context().Done():
|
|
logrus.Debugf("Client connection (container stats) cancelled")
|
|
|
|
default:
|
|
// Container stats
|
|
stats, err = ctnr.GetContainerStats(stats)
|
|
if err != nil {
|
|
logrus.Errorf("Unable to get container stats: %v", err)
|
|
return
|
|
}
|
|
inspect, err := ctnr.Inspect(false)
|
|
if err != nil {
|
|
logrus.Errorf("Unable to inspect container: %v", err)
|
|
return
|
|
}
|
|
// Cgroup stats
|
|
cgroupPath, err := ctnr.CgroupPath()
|
|
if err != nil {
|
|
logrus.Errorf("Unable to get cgroup path of container: %v", err)
|
|
return
|
|
}
|
|
cgroup, err := cgroups.Load(cgroupPath)
|
|
if err != nil {
|
|
logrus.Errorf("Unable to load cgroup: %v", err)
|
|
return
|
|
}
|
|
cgroupStat, err := cgroup.Stat()
|
|
if err != nil {
|
|
logrus.Errorf("Unable to get cgroup stats: %v", err)
|
|
return
|
|
}
|
|
|
|
// FIXME: network inspection does not yet work entirely
|
|
net := make(map[string]docker.NetworkStats)
|
|
networkName := inspect.NetworkSettings.EndpointID
|
|
if networkName == "" {
|
|
networkName = "network"
|
|
}
|
|
net[networkName] = docker.NetworkStats{
|
|
RxBytes: stats.NetInput,
|
|
RxPackets: 0,
|
|
RxErrors: 0,
|
|
RxDropped: 0,
|
|
TxBytes: stats.NetOutput,
|
|
TxPackets: 0,
|
|
TxErrors: 0,
|
|
TxDropped: 0,
|
|
EndpointID: inspect.NetworkSettings.EndpointID,
|
|
InstanceID: "",
|
|
}
|
|
|
|
systemUsage, _ := cgroups.GetSystemCPUUsage()
|
|
s := StatsJSON{
|
|
Stats: Stats{
|
|
Read: time.Now(),
|
|
PreRead: preRead,
|
|
PidsStats: docker.PidsStats{
|
|
Current: cgroupStat.Pids.Current,
|
|
Limit: 0,
|
|
},
|
|
BlkioStats: docker.BlkioStats{
|
|
IoServiceBytesRecursive: toBlkioStatEntry(cgroupStat.Blkio.IoServiceBytesRecursive),
|
|
IoServicedRecursive: nil,
|
|
IoQueuedRecursive: nil,
|
|
IoServiceTimeRecursive: nil,
|
|
IoWaitTimeRecursive: nil,
|
|
IoMergedRecursive: nil,
|
|
IoTimeRecursive: nil,
|
|
SectorsRecursive: nil,
|
|
},
|
|
CPUStats: CPUStats{
|
|
CPUUsage: docker.CPUUsage{
|
|
TotalUsage: cgroupStat.CPU.Usage.Total,
|
|
PercpuUsage: cgroupStat.CPU.Usage.PerCPU,
|
|
UsageInKernelmode: cgroupStat.CPU.Usage.Kernel,
|
|
UsageInUsermode: cgroupStat.CPU.Usage.Total - cgroupStat.CPU.Usage.Kernel,
|
|
},
|
|
CPU: stats.CPU,
|
|
SystemUsage: systemUsage,
|
|
OnlineCPUs: uint32(len(cgroupStat.CPU.Usage.PerCPU)),
|
|
ThrottlingData: docker.ThrottlingData{
|
|
Periods: 0,
|
|
ThrottledPeriods: 0,
|
|
ThrottledTime: 0,
|
|
},
|
|
},
|
|
PreCPUStats: preCPUStats,
|
|
MemoryStats: docker.MemoryStats{
|
|
Usage: cgroupStat.Memory.Usage.Usage,
|
|
MaxUsage: cgroupStat.Memory.Usage.Limit,
|
|
Stats: nil,
|
|
Failcnt: 0,
|
|
Limit: cgroupStat.Memory.Usage.Limit,
|
|
Commit: 0,
|
|
CommitPeak: 0,
|
|
PrivateWorkingSet: 0,
|
|
},
|
|
},
|
|
Name: stats.Name,
|
|
ID: stats.ContainerID,
|
|
Networks: net,
|
|
}
|
|
|
|
if err := coder.Encode(s); err != nil {
|
|
logrus.Errorf("Unable to encode stats: %v", err)
|
|
return
|
|
}
|
|
if flusher, ok := w.(http.Flusher); ok {
|
|
flusher.Flush()
|
|
}
|
|
|
|
if !query.Stream || query.OneShot {
|
|
return
|
|
}
|
|
|
|
preRead = s.Read
|
|
bits, err := json.Marshal(s.CPUStats)
|
|
if err != nil {
|
|
logrus.Errorf("Unable to marshal cpu stats: %q", err)
|
|
}
|
|
if err := json.Unmarshal(bits, &preCPUStats); err != nil {
|
|
logrus.Errorf("Unable to unmarshal previous stats: %q", err)
|
|
}
|
|
|
|
time.Sleep(DefaultStatsPeriod)
|
|
goto streamLabel
|
|
}
|
|
}
|
|
|
|
func toBlkioStatEntry(entries []cgroups.BlkIOEntry) []docker.BlkioStatEntry {
|
|
results := make([]docker.BlkioStatEntry, len(entries))
|
|
for i, e := range entries {
|
|
bits, err := json.Marshal(e)
|
|
if err != nil {
|
|
logrus.Errorf("Unable to marshal blkio stats: %q", err)
|
|
}
|
|
if err := json.Unmarshal(bits, &results[i]); err != nil {
|
|
logrus.Errorf("Unable to unmarshal blkio stats: %q", err)
|
|
}
|
|
}
|
|
return results
|
|
}
|