mirror of
https://github.com/containers/podman.git
synced 2025-05-17 15:18:43 +08:00

unfortunately the papr CI system cannot test ubuntu as a VM; therefore, this PR still keeps travis. but it does include fixes that will be required for running on modern versions of ubuntu. Signed-off-by: baude <bbaude@redhat.com>
295 lines
8.9 KiB
Go
295 lines
8.9 KiB
Go
// +build linux
|
|
|
|
package libpod
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"path"
|
|
"path/filepath"
|
|
"strings"
|
|
|
|
"github.com/containerd/cgroups"
|
|
"github.com/pkg/errors"
|
|
"github.com/sirupsen/logrus"
|
|
)
|
|
|
|
// NewPod makes a new, empty pod
|
|
func (r *Runtime) NewPod(ctx context.Context, options ...PodCreateOption) (*Pod, error) {
|
|
r.lock.Lock()
|
|
defer r.lock.Unlock()
|
|
|
|
if !r.valid {
|
|
return nil, ErrRuntimeStopped
|
|
}
|
|
|
|
pod, err := newPod(r.lockDir, r)
|
|
if err != nil {
|
|
return nil, errors.Wrapf(err, "error creating pod")
|
|
}
|
|
|
|
// Set default namespace to runtime's namespace
|
|
// Do so before options run so they can override it
|
|
if r.config.Namespace != "" {
|
|
pod.config.Namespace = r.config.Namespace
|
|
}
|
|
|
|
for _, option := range options {
|
|
if err := option(pod); err != nil {
|
|
return nil, errors.Wrapf(err, "error running pod create option")
|
|
}
|
|
}
|
|
|
|
if pod.config.Name == "" {
|
|
name, err := r.generateName()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
pod.config.Name = name
|
|
}
|
|
|
|
pod.valid = true
|
|
|
|
// Check CGroup parent sanity, and set it if it was not set
|
|
switch r.config.CgroupManager {
|
|
case CgroupfsCgroupsManager:
|
|
if pod.config.CgroupParent == "" {
|
|
pod.config.CgroupParent = CgroupfsDefaultCgroupParent
|
|
} else if strings.HasSuffix(path.Base(pod.config.CgroupParent), ".slice") {
|
|
return nil, errors.Wrapf(ErrInvalidArg, "systemd slice received as cgroup parent when using cgroupfs")
|
|
}
|
|
// If we are set to use pod cgroups, set the cgroup parent that
|
|
// all containers in the pod will share
|
|
// No need to create it with cgroupfs - the first container to
|
|
// launch should do it for us
|
|
if pod.config.UsePodCgroup {
|
|
pod.state.CgroupPath = filepath.Join(pod.config.CgroupParent, pod.ID())
|
|
}
|
|
case SystemdCgroupsManager:
|
|
if pod.config.CgroupParent == "" {
|
|
pod.config.CgroupParent = SystemdDefaultCgroupParent
|
|
} else if len(pod.config.CgroupParent) < 6 || !strings.HasSuffix(path.Base(pod.config.CgroupParent), ".slice") {
|
|
return nil, errors.Wrapf(ErrInvalidArg, "did not receive systemd slice as cgroup parent when using systemd to manage cgroups")
|
|
}
|
|
// If we are set to use pod cgroups, set the cgroup parent that
|
|
// all containers in the pod will share
|
|
if pod.config.UsePodCgroup {
|
|
cgroupPath, err := systemdSliceFromPath(pod.config.CgroupParent, fmt.Sprintf("libpod_pod_%s", pod.ID()))
|
|
if err != nil {
|
|
return nil, errors.Wrapf(err, "unable to create pod cgroup for pod %s", pod.ID())
|
|
}
|
|
pod.state.CgroupPath = cgroupPath
|
|
}
|
|
default:
|
|
return nil, errors.Wrapf(ErrInvalidArg, "unsupported CGroup manager: %s - cannot validate cgroup parent", r.config.CgroupManager)
|
|
}
|
|
|
|
if pod.config.UsePodCgroup {
|
|
logrus.Debugf("Got pod cgroup as %s", pod.state.CgroupPath)
|
|
}
|
|
if pod.HasInfraContainer() != pod.SharesNamespaces() {
|
|
return nil, errors.Errorf("Pods must have an infra container to share namespaces")
|
|
}
|
|
|
|
if err := r.state.AddPod(pod); err != nil {
|
|
return nil, errors.Wrapf(err, "error adding pod to state")
|
|
}
|
|
|
|
if pod.HasInfraContainer() {
|
|
ctr, err := r.createInfraContainer(ctx, pod)
|
|
if err != nil {
|
|
// Tear down pod, as it is assumed a the pod will contain
|
|
// a pause container, and it does not.
|
|
if err2 := r.removePod(ctx, pod, true, true); err2 != nil {
|
|
logrus.Errorf("Error removing pod after pause container creation failure: %v", err2)
|
|
}
|
|
return nil, errors.Wrapf(err, "error adding Infra Container")
|
|
}
|
|
pod.state.InfraContainerID = ctr.ID()
|
|
if err := pod.save(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Once the pod infra container has been created, we start it
|
|
if err := ctr.Start(ctx); err != nil {
|
|
// If the infra container does not start, we need to tear the pod down.
|
|
if err2 := r.removePod(ctx, pod, true, true); err2 != nil {
|
|
logrus.Errorf("Error removing pod after infra container failed to start: %v", err2)
|
|
}
|
|
return nil, errors.Wrapf(err, "error starting Infra Container")
|
|
}
|
|
}
|
|
|
|
return pod, nil
|
|
}
|
|
|
|
func (r *Runtime) removePod(ctx context.Context, p *Pod, removeCtrs, force bool) error {
|
|
if err := p.updatePod(); err != nil {
|
|
return err
|
|
}
|
|
|
|
ctrs, err := r.state.PodContainers(p)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
numCtrs := len(ctrs)
|
|
|
|
// If the only container in the pod is the pause container, remove the pod and container unconditionally.
|
|
pauseCtrID := p.state.InfraContainerID
|
|
if numCtrs == 1 && ctrs[0].ID() == pauseCtrID {
|
|
removeCtrs = true
|
|
force = true
|
|
}
|
|
if !removeCtrs && numCtrs > 0 {
|
|
return errors.Wrapf(ErrCtrExists, "pod %s contains containers and cannot be removed", p.ID())
|
|
}
|
|
|
|
// Go through and lock all containers so we can operate on them all at once
|
|
dependencies := make(map[string][]string)
|
|
for _, ctr := range ctrs {
|
|
ctr.lock.Lock()
|
|
defer ctr.lock.Unlock()
|
|
|
|
// Sync all containers
|
|
if err := ctr.syncContainer(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Check if the container is in a good state to be removed
|
|
if ctr.state.State == ContainerStatePaused {
|
|
return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains paused container %s, cannot remove", p.ID(), ctr.ID())
|
|
}
|
|
|
|
if ctr.state.State == ContainerStateUnknown {
|
|
return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains container %s with invalid state", p.ID(), ctr.ID())
|
|
}
|
|
|
|
// If the container is running and force is not set we can't do anything
|
|
if ctr.state.State == ContainerStateRunning && !force {
|
|
return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains container %s which is running", p.ID(), ctr.ID())
|
|
}
|
|
|
|
// If the container has active exec sessions and force is not set we can't do anything
|
|
if len(ctr.state.ExecSessions) != 0 && !force {
|
|
return errors.Wrapf(ErrCtrStateInvalid, "pod %s contains container %s which has active exec sessions", p.ID(), ctr.ID())
|
|
}
|
|
|
|
deps, err := r.state.ContainerInUse(ctr)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
dependencies[ctr.ID()] = deps
|
|
}
|
|
|
|
// Check if containers have dependencies
|
|
// If they do, and the dependencies are not in the pod, error
|
|
for ctr, deps := range dependencies {
|
|
for _, dep := range deps {
|
|
if _, ok := dependencies[dep]; !ok {
|
|
return errors.Wrapf(ErrCtrExists, "container %s depends on container %s not in pod %s", ctr, dep, p.ID())
|
|
}
|
|
}
|
|
}
|
|
|
|
// First loop through all containers and stop them
|
|
// Do not remove in this loop to ensure that we don't remove unless all
|
|
// containers are in a good state
|
|
if force {
|
|
for _, ctr := range ctrs {
|
|
// If force is set and the container is running, stop it now
|
|
if ctr.state.State == ContainerStateRunning {
|
|
if err := r.ociRuntime.stopContainer(ctr, ctr.StopTimeout()); err != nil {
|
|
return errors.Wrapf(err, "error stopping container %s to remove pod %s", ctr.ID(), p.ID())
|
|
}
|
|
|
|
// Sync again to pick up stopped state
|
|
if err := ctr.syncContainer(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
// If the container has active exec sessions, stop them now
|
|
if len(ctr.state.ExecSessions) != 0 {
|
|
if err := r.ociRuntime.execStopContainer(ctr, ctr.StopTimeout()); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Start removing containers
|
|
// We can remove containers even if they have dependencies now
|
|
// As we have guaranteed their dependencies are in the pod
|
|
for _, ctr := range ctrs {
|
|
// Clean up network namespace, cgroups, mounts
|
|
if err := ctr.cleanup(ctx); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Stop container's storage
|
|
if err := ctr.teardownStorage(); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Delete the container from runtime (only if we are not
|
|
// ContainerStateConfigured)
|
|
if ctr.state.State != ContainerStateConfigured &&
|
|
ctr.state.State != ContainerStateExited {
|
|
if err := ctr.delete(ctx); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
|
|
// Remove containers from the state
|
|
if err := r.state.RemovePodContainers(p); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Mark containers invalid
|
|
for _, ctr := range ctrs {
|
|
ctr.valid = false
|
|
}
|
|
|
|
// Remove pod cgroup, if present
|
|
if p.state.CgroupPath != "" {
|
|
logrus.Debugf("Removing pod cgroup %s", p.state.CgroupPath)
|
|
|
|
switch p.runtime.config.CgroupManager {
|
|
case SystemdCgroupsManager:
|
|
if err := deleteSystemdCgroup(p.state.CgroupPath); err != nil {
|
|
// The pod is already almost gone.
|
|
// No point in hard-failing if we fail
|
|
// this bit of cleanup.
|
|
logrus.Errorf("Error deleting pod %s cgroup %s: %v", p.ID(), p.state.CgroupPath, err)
|
|
}
|
|
case CgroupfsCgroupsManager:
|
|
// Delete the cgroupfs cgroup
|
|
v1CGroups := GetV1CGroups(getExcludedCGroups())
|
|
cgroup, err := cgroups.Load(v1CGroups, cgroups.StaticPath(p.state.CgroupPath))
|
|
if err != nil && err != cgroups.ErrCgroupDeleted {
|
|
return err
|
|
} else if err == nil {
|
|
if err := cgroup.Delete(); err != nil {
|
|
// The pod is already almost gone.
|
|
// No point in hard-failing if we fail
|
|
// this bit of cleanup.
|
|
logrus.Errorf("Error deleting pod %s cgroup %s: %v", p.ID(), p.state.CgroupPath, err)
|
|
}
|
|
}
|
|
default:
|
|
return errors.Wrapf(ErrInvalidArg, "unknown cgroups manager %s specified", p.runtime.config.CgroupManager)
|
|
}
|
|
}
|
|
|
|
// Remove pod from state
|
|
if err := r.state.RemovePod(p); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Mark pod invalid
|
|
p.valid = false
|
|
|
|
return nil
|
|
}
|