mirror of
https://github.com/containers/podman.git
synced 2025-10-18 03:33:32 +08:00
Add container restart policy to Libpod & Podman
This initial version does not support restart count, but it works as advertised otherwise. Signed-off-by: Matthew Heon <matthew.heon@pm.me>
This commit is contained in:
@ -444,7 +444,7 @@ func getCreateFlags(c *cliconfig.PodmanCommand) {
|
||||
)
|
||||
createFlags.String(
|
||||
"restart", "",
|
||||
"Restart is not supported. Please use a systemd unit file for restart",
|
||||
"Restart policy to apply when a container exits",
|
||||
)
|
||||
createFlags.Bool(
|
||||
"rm", false,
|
||||
|
@ -279,9 +279,6 @@ func ParseCreateOpts(ctx context.Context, c *GenericCLIResults, runtime *libpod.
|
||||
blkioWeight uint16
|
||||
namespaces map[string]string
|
||||
)
|
||||
if c.IsSet("restart") {
|
||||
return nil, errors.Errorf("--restart option is not supported.\nUse systemd unit files for restarting containers")
|
||||
}
|
||||
|
||||
idmappings, err := util.ParseIDMapping(c.StringSlice("uidmap"), c.StringSlice("gidmap"), c.String("subuidname"), c.String("subgidname"))
|
||||
if err != nil {
|
||||
@ -676,21 +673,22 @@ func ParseCreateOpts(ctx context.Context, c *GenericCLIResults, runtime *libpod.
|
||||
PidsLimit: c.Int64("pids-limit"),
|
||||
Ulimit: c.StringSlice("ulimit"),
|
||||
},
|
||||
Rm: c.Bool("rm"),
|
||||
StopSignal: stopSignal,
|
||||
StopTimeout: c.Uint("stop-timeout"),
|
||||
Sysctl: sysctl,
|
||||
Systemd: systemd,
|
||||
Tmpfs: c.StringSlice("tmpfs"),
|
||||
Tty: tty,
|
||||
User: user,
|
||||
UsernsMode: usernsMode,
|
||||
MountsFlag: c.StringArray("mount"),
|
||||
Volumes: c.StringArray("volume"),
|
||||
WorkDir: workDir,
|
||||
Rootfs: rootfs,
|
||||
VolumesFrom: c.StringSlice("volumes-from"),
|
||||
Syslog: c.Bool("syslog"),
|
||||
RestartPolicy: c.String("restart"),
|
||||
Rm: c.Bool("rm"),
|
||||
StopSignal: stopSignal,
|
||||
StopTimeout: c.Uint("stop-timeout"),
|
||||
Sysctl: sysctl,
|
||||
Systemd: systemd,
|
||||
Tmpfs: c.StringSlice("tmpfs"),
|
||||
Tty: tty,
|
||||
User: user,
|
||||
UsernsMode: usernsMode,
|
||||
MountsFlag: c.StringArray("mount"),
|
||||
Volumes: c.StringArray("volume"),
|
||||
WorkDir: workDir,
|
||||
Rootfs: rootfs,
|
||||
VolumesFrom: c.StringSlice("volumes-from"),
|
||||
Syslog: c.Bool("syslog"),
|
||||
}
|
||||
|
||||
if config.Privileged {
|
||||
|
@ -181,7 +181,10 @@ type ContainerState struct {
|
||||
BindMounts map[string]string `json:"bindMounts,omitempty"`
|
||||
// StoppedByUser indicates whether the container was stopped by an
|
||||
// explicit call to the Stop() API.
|
||||
StoppedByUser bool
|
||||
StoppedByUser bool `json:"stoppedByUser,omitempty"`
|
||||
// RestartPolicyMatch indicates whether the conditions for restart
|
||||
// policy have been met.
|
||||
RestartPolicyMatch bool `json:"restartPolicyMatch,omitempty"`
|
||||
|
||||
// ExtensionStageHooks holds hooks which will be executed by libpod
|
||||
// and not delegated to the OCI runtime.
|
||||
@ -349,6 +352,17 @@ type ContainerConfig struct {
|
||||
LogPath string `json:"logPath"`
|
||||
// File containing the conmon PID
|
||||
ConmonPidFile string `json:"conmonPidFile,omitempty"`
|
||||
// RestartPolicy indicates what action the container will take upon
|
||||
// exiting naturally.
|
||||
// Allowed options are "no" (take no action), "on-failure" (restart on
|
||||
// non-zero exit code, up an a maximum of RestartRetries times),
|
||||
// and "always" (always restart the container on any exit code).
|
||||
// The empty string is treated as the default ("no")
|
||||
RestartPolicy string `json:"restart_policy,omitempty"`
|
||||
// RestartRetries indicates the number of attempts that will be made to
|
||||
// restart the container. Used only if RestartPolicy is set to
|
||||
// "on-failure".
|
||||
RestartRetries uint `json:"restart_retries,omitempty"`
|
||||
// TODO log options for log drivers
|
||||
|
||||
PostConfigureNetNS bool `json:"postConfigureNetNS"`
|
||||
@ -732,6 +746,17 @@ func (c *Container) LogPath() string {
|
||||
return c.config.LogPath
|
||||
}
|
||||
|
||||
// RestartPolicy returns the container's restart policy.
|
||||
func (c *Container) RestartPolicy() string {
|
||||
return c.config.RestartPolicy
|
||||
}
|
||||
|
||||
// RestartRetries returns the number of retries that will be attempted when
|
||||
// using the "on-failure" restart policy
|
||||
func (c *Container) RestartRetries() uint {
|
||||
return c.config.RestartRetries
|
||||
}
|
||||
|
||||
// RuntimeName returns the name of the runtime
|
||||
func (c *Container) RuntimeName() string {
|
||||
return c.runtime.ociRuntime.name
|
||||
|
@ -583,6 +583,7 @@ func (c *Container) Cleanup(ctx context.Context) error {
|
||||
if !c.batched {
|
||||
c.lock.Lock()
|
||||
defer c.lock.Unlock()
|
||||
|
||||
if err := c.syncContainer(); err != nil {
|
||||
return err
|
||||
}
|
||||
@ -593,6 +594,24 @@ func (c *Container) Cleanup(ctx context.Context) error {
|
||||
return errors.Wrapf(ErrCtrStateInvalid, "container %s is running or paused, refusing to clean up", c.ID())
|
||||
}
|
||||
|
||||
// If we have a restart policy match when updating the state, we need to
|
||||
// restart the container.
|
||||
// However, perform a full validation of restart policy first.
|
||||
if c.state.RestartPolicyMatch {
|
||||
// if restart policy is on-error and exit code is 0, we do
|
||||
// nothing.
|
||||
// TODO: if restart retries is set, handle that here.
|
||||
if c.config.RestartRetries != 0 {
|
||||
return errors.Wrapf(ErrNotImplemented, "restart retries not yet implemented")
|
||||
}
|
||||
if (c.config.RestartPolicy == "on-error" && c.state.ExitCode == 0) || c.config.RestartPolicy == "always" {
|
||||
// The container stopped. We need to restart it.
|
||||
return c.handleRestartPolicy(ctx)
|
||||
}
|
||||
}
|
||||
|
||||
// If we aren't hitting restart policy, we perform a normal cleanup
|
||||
|
||||
// Check if we have active exec sessions
|
||||
if len(c.state.ExecSessions) != 0 {
|
||||
return errors.Wrapf(ErrCtrStateInvalid, "container %s has active exec sessions, refusing to clean up", c.ID())
|
||||
|
@ -210,6 +210,43 @@ func (c *Container) handleExitFile(exitFile string, fi os.FileInfo) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// Handle container restart policy.
|
||||
// This is called when a container has exited, and was not explicitly stopped by
|
||||
// an API call to stop the container or pod it is in.
|
||||
func (c *Container) handleRestartPolicy(ctx context.Context) (err error) {
|
||||
logrus.Debugf("Restarting container %s due to restart policy %s", c.ID(), c.config.RestartPolicy)
|
||||
|
||||
// Need to check if dependencies are alive.
|
||||
if err = c.checkDependenciesAndHandleError(ctx); err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
defer func() {
|
||||
if err != nil {
|
||||
if err2 := c.cleanup(ctx); err2 != nil {
|
||||
logrus.Errorf("error cleaning up container %s: %v", c.ID(), err2)
|
||||
}
|
||||
}
|
||||
}()
|
||||
if err := c.prepare(); err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
if c.state.State == ContainerStateStopped {
|
||||
// Reinitialize the container if we need to
|
||||
if err := c.reinit(ctx); err != nil {
|
||||
return err
|
||||
}
|
||||
} else if c.state.State == ContainerStateConfigured ||
|
||||
c.state.State == ContainerStateExited {
|
||||
// Initialize the container
|
||||
if err := c.init(ctx); err != nil {
|
||||
return err
|
||||
}
|
||||
}
|
||||
return c.start()
|
||||
}
|
||||
|
||||
// Sync this container with on-disk state and runtime status
|
||||
// Should only be called with container lock held
|
||||
// This function should suffice to ensure a container's state is accurate and
|
||||
@ -230,6 +267,14 @@ func (c *Container) syncContainer() error {
|
||||
}
|
||||
// Only save back to DB if state changed
|
||||
if c.state.State != oldState {
|
||||
// Check for a restart policy match
|
||||
if c.config.RestartPolicy != "" && c.config.RestartPolicy != "no" &&
|
||||
(oldState == ContainerStateRunning || oldState == ContainerStatePaused) &&
|
||||
(c.state.State == ContainerStateStopped || c.state.State == ContainerStateExited) &&
|
||||
!c.state.StoppedByUser {
|
||||
c.state.RestartPolicyMatch = true
|
||||
}
|
||||
|
||||
if err := c.save(); err != nil {
|
||||
return err
|
||||
}
|
||||
@ -377,6 +422,7 @@ func resetState(state *ContainerState) error {
|
||||
state.NetworkStatus = nil
|
||||
state.BindMounts = make(map[string]string)
|
||||
state.StoppedByUser = false
|
||||
state.RestartPolicyMatch = false
|
||||
|
||||
return nil
|
||||
}
|
||||
@ -791,6 +837,7 @@ func (c *Container) init(ctx context.Context) error {
|
||||
c.state.Exited = false
|
||||
c.state.State = ContainerStateCreated
|
||||
c.state.StoppedByUser = false
|
||||
c.state.RestartPolicyMatch = false
|
||||
|
||||
if err := c.save(); err != nil {
|
||||
return err
|
||||
|
@ -1239,6 +1239,41 @@ func WithUseImageHosts() CtrCreateOption {
|
||||
}
|
||||
}
|
||||
|
||||
// WithRestartPolicy sets the container's restart policy. Valid values are
|
||||
// "no", "on-failure", and "always". The empty string is allowed, and will be
|
||||
// equivalent to "no".
|
||||
func WithRestartPolicy(policy string) CtrCreateOption {
|
||||
return func(ctr *Container) error {
|
||||
if ctr.valid {
|
||||
return ErrCtrFinalized
|
||||
}
|
||||
|
||||
switch policy {
|
||||
case "", "no", "on-failure", "always":
|
||||
ctr.config.RestartPolicy = policy
|
||||
default:
|
||||
return errors.Wrapf(ErrInvalidArg, "%q is not a valid restart policy", policy)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
// WithRestartRetries sets the number of retries to use when restarting a
|
||||
// container with the "on-failure" restart policy.
|
||||
// 0 is an allowed value, and indicates infinite retries.
|
||||
func WithRestartRetries(tries uint) CtrCreateOption {
|
||||
return func(ctr *Container) error {
|
||||
if ctr.valid {
|
||||
return ErrCtrFinalized
|
||||
}
|
||||
|
||||
ctr.config.RestartRetries = tries
|
||||
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
// withIsInfra sets the container to be an infra container. This means the container will be sometimes hidden
|
||||
// and expected to be the first container in the pod.
|
||||
func withIsInfra() CtrCreateOption {
|
||||
|
@ -108,6 +108,7 @@ type CreateConfig struct {
|
||||
ReadOnlyRootfs bool //read-only
|
||||
ReadOnlyTmpfs bool //read-only-tmpfs
|
||||
Resources CreateResourceConfig
|
||||
RestartPolicy string
|
||||
Rm bool //rm
|
||||
StopSignal syscall.Signal // stop-signal
|
||||
StopTimeout uint // stop-timeout
|
||||
@ -359,6 +360,26 @@ func (c *CreateConfig) getContainerCreateOptions(runtime *libpod.Runtime, pod *l
|
||||
options = append(options, libpod.WithCgroupParent(c.CgroupParent))
|
||||
}
|
||||
|
||||
if c.RestartPolicy != "" {
|
||||
if c.RestartPolicy == "unless-stopped" {
|
||||
return nil, errors.Wrapf(libpod.ErrInvalidArg, "the unless-stopped restart policy is not supported")
|
||||
}
|
||||
|
||||
split := strings.Split(c.RestartPolicy, ":")
|
||||
if len(split) > 1 {
|
||||
numTries, err := strconv.Atoi(split[1])
|
||||
if err != nil {
|
||||
return nil, errors.Wrapf(err, "%s is not a valid number of retries for restart policy", split[1])
|
||||
}
|
||||
if numTries < 0 {
|
||||
return nil, errors.Wrapf(libpod.ErrInvalidArg, "restart policy requires a positive number of retries")
|
||||
}
|
||||
options = append(options, libpod.WithRestartRetries(uint(numTries)))
|
||||
}
|
||||
|
||||
options = append(options, libpod.WithRestartPolicy(c.RestartPolicy))
|
||||
}
|
||||
|
||||
// Always use a cleanup process to clean up Podman after termination
|
||||
exitCmd, err := c.createExitCommand(runtime)
|
||||
if err != nil {
|
||||
|
Reference in New Issue
Block a user