mirror of
https://github.com/containers/podman.git
synced 2025-09-22 20:56:21 +08:00

Per discussion with Dan, it would be better to automatically handle potential runtime errors by automatically syncing if they occur. Retaining the flag for `ps` makes sense, as we won't even be calling the OCI runtime and as such won't see errors if the state desyncs, but rm can be handled automatically. The automatic desync handling code will take some additional work so we'll land this as-is (sync on ps is enough to solve most desync issues). Signed-off-by: Matthew Heon <mheon@redhat.com>
94 lines
2.3 KiB
Go
94 lines
2.3 KiB
Go
package main
|
|
|
|
import (
|
|
"fmt"
|
|
"github.com/containers/libpod/cmd/podman/libpodruntime"
|
|
"github.com/containers/libpod/cmd/podman/shared"
|
|
"github.com/pkg/errors"
|
|
"github.com/sirupsen/logrus"
|
|
"github.com/urfave/cli"
|
|
)
|
|
|
|
var (
|
|
rmFlags = []cli.Flag{
|
|
cli.BoolFlag{
|
|
Name: "all, a",
|
|
Usage: "Remove all containers",
|
|
},
|
|
cli.BoolFlag{
|
|
Name: "force, f",
|
|
Usage: "Force removal of a running container. The default is false",
|
|
},
|
|
LatestFlag,
|
|
cli.BoolFlag{
|
|
Name: "volumes, v",
|
|
Usage: "Remove the volumes associated with the container (Not implemented yet)",
|
|
},
|
|
}
|
|
rmDescription = fmt.Sprintf(`
|
|
Podman rm will remove one or more containers from the host.
|
|
The container name or ID can be used. This does not remove images.
|
|
Running containers will not be removed without the -f option.
|
|
`)
|
|
rmCommand = cli.Command{
|
|
Name: "rm",
|
|
Usage: "Remove one or more containers",
|
|
Description: rmDescription,
|
|
Flags: sortFlags(rmFlags),
|
|
Action: rmCmd,
|
|
ArgsUsage: "",
|
|
UseShortOptionHandling: true,
|
|
OnUsageError: usageErrorHandler,
|
|
}
|
|
)
|
|
|
|
// saveCmd saves the image to either docker-archive or oci
|
|
func rmCmd(c *cli.Context) error {
|
|
var (
|
|
deleteFuncs []shared.ParallelWorkerInput
|
|
)
|
|
|
|
ctx := getContext()
|
|
if err := validateFlags(c, rmFlags); err != nil {
|
|
return err
|
|
}
|
|
runtime, err := libpodruntime.GetRuntime(c)
|
|
if err != nil {
|
|
return errors.Wrapf(err, "could not get runtime")
|
|
}
|
|
defer runtime.Shutdown(false)
|
|
|
|
if err := checkAllAndLatest(c); err != nil {
|
|
return err
|
|
}
|
|
|
|
delContainers, err := getAllOrLatestContainers(c, runtime, -1, "all")
|
|
if err != nil {
|
|
if len(delContainers) == 0 {
|
|
return err
|
|
}
|
|
fmt.Println(err.Error())
|
|
}
|
|
|
|
for _, container := range delContainers {
|
|
con := container
|
|
f := func() error {
|
|
return runtime.RemoveContainer(ctx, con, c.Bool("force"))
|
|
}
|
|
|
|
deleteFuncs = append(deleteFuncs, shared.ParallelWorkerInput{
|
|
ContainerID: con.ID(),
|
|
ParallelFunc: f,
|
|
})
|
|
}
|
|
maxWorkers := shared.Parallelize("rm")
|
|
if c.GlobalIsSet("max-workers") {
|
|
maxWorkers = c.GlobalInt("max-workers")
|
|
}
|
|
logrus.Debugf("Setting maximum workers to %d", maxWorkers)
|
|
|
|
// Run the parallel funcs
|
|
deleteErrors, errCount := shared.ParallelExecuteWorkerPool(maxWorkers, deleteFuncs)
|
|
return printParallelOutput(deleteErrors, errCount)
|
|
}
|