mirror of
				https://github.com/containers/podman.git
				synced 2025-11-04 00:50:15 +08:00 
			
		
		
		
	Moving from Go module v4 to v5 prepares us for public releases. Move done using gomove [1] as with the v3 and v4 moves. [1] https://github.com/KSubedi/gomove Signed-off-by: Matt Heon <mheon@redhat.com>
		
			
				
	
	
		
			383 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			383 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
//go:build !remote
 | 
						|
 | 
						|
package libpod
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"fmt"
 | 
						|
	"strings"
 | 
						|
 | 
						|
	"github.com/containers/podman/v5/libpod/define"
 | 
						|
	"github.com/sirupsen/logrus"
 | 
						|
)
 | 
						|
 | 
						|
type containerNode struct {
 | 
						|
	id         string
 | 
						|
	container  *Container
 | 
						|
	dependsOn  []*containerNode
 | 
						|
	dependedOn []*containerNode
 | 
						|
}
 | 
						|
 | 
						|
// ContainerGraph is a dependency graph based on a set of containers.
 | 
						|
type ContainerGraph struct {
 | 
						|
	nodes              map[string]*containerNode
 | 
						|
	noDepNodes         []*containerNode
 | 
						|
	notDependedOnNodes map[string]*containerNode
 | 
						|
}
 | 
						|
 | 
						|
// DependencyMap returns the dependency graph as map with the key being a
 | 
						|
// container and the value being the containers the key depends on.
 | 
						|
func (cg *ContainerGraph) DependencyMap() (dependencies map[*Container][]*Container) {
 | 
						|
	dependencies = make(map[*Container][]*Container)
 | 
						|
	for _, node := range cg.nodes {
 | 
						|
		dependsOn := make([]*Container, len(node.dependsOn))
 | 
						|
		for i, d := range node.dependsOn {
 | 
						|
			dependsOn[i] = d.container
 | 
						|
		}
 | 
						|
		dependencies[node.container] = dependsOn
 | 
						|
	}
 | 
						|
	return dependencies
 | 
						|
}
 | 
						|
 | 
						|
// BuildContainerGraph builds a dependency graph based on the container slice.
 | 
						|
func BuildContainerGraph(ctrs []*Container) (*ContainerGraph, error) {
 | 
						|
	graph := new(ContainerGraph)
 | 
						|
	graph.nodes = make(map[string]*containerNode)
 | 
						|
	graph.notDependedOnNodes = make(map[string]*containerNode)
 | 
						|
 | 
						|
	// Start by building all nodes, with no edges
 | 
						|
	for _, ctr := range ctrs {
 | 
						|
		ctrNode := new(containerNode)
 | 
						|
		ctrNode.id = ctr.ID()
 | 
						|
		ctrNode.container = ctr
 | 
						|
 | 
						|
		graph.nodes[ctr.ID()] = ctrNode
 | 
						|
		graph.notDependedOnNodes[ctr.ID()] = ctrNode
 | 
						|
	}
 | 
						|
 | 
						|
	// Now add edges based on dependencies
 | 
						|
	for _, node := range graph.nodes {
 | 
						|
		deps := node.container.Dependencies()
 | 
						|
		for _, dep := range deps {
 | 
						|
			// Get the dep's node
 | 
						|
			depNode, ok := graph.nodes[dep]
 | 
						|
			if !ok {
 | 
						|
				return nil, fmt.Errorf("container %s depends on container %s not found in input list: %w", node.id, dep, define.ErrNoSuchCtr)
 | 
						|
			}
 | 
						|
 | 
						|
			// Add the dependent node to the node's dependencies
 | 
						|
			// And add the node to the dependent node's dependedOn
 | 
						|
			node.dependsOn = append(node.dependsOn, depNode)
 | 
						|
			depNode.dependedOn = append(depNode.dependedOn, node)
 | 
						|
 | 
						|
			// The dependency now has something depending on it
 | 
						|
			delete(graph.notDependedOnNodes, dep)
 | 
						|
		}
 | 
						|
 | 
						|
		// Maintain a list of nodes with no dependencies
 | 
						|
		// (no edges coming from them)
 | 
						|
		if len(deps) == 0 {
 | 
						|
			graph.noDepNodes = append(graph.noDepNodes, node)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Need to do cycle detection
 | 
						|
	// We cannot start or stop if there are cyclic dependencies
 | 
						|
	cycle, err := detectCycles(graph)
 | 
						|
	if err != nil {
 | 
						|
		return nil, err
 | 
						|
	} else if cycle {
 | 
						|
		return nil, fmt.Errorf("cycle found in container dependency graph: %w", define.ErrInternal)
 | 
						|
	}
 | 
						|
 | 
						|
	return graph, nil
 | 
						|
}
 | 
						|
 | 
						|
// Detect cycles in a container graph using Tarjan's strongly connected
 | 
						|
// components algorithm
 | 
						|
// Return true if a cycle is found, false otherwise
 | 
						|
func detectCycles(graph *ContainerGraph) (bool, error) {
 | 
						|
	type nodeInfo struct {
 | 
						|
		index   int
 | 
						|
		lowLink int
 | 
						|
		onStack bool
 | 
						|
	}
 | 
						|
 | 
						|
	index := 0
 | 
						|
 | 
						|
	nodes := make(map[string]*nodeInfo)
 | 
						|
	stack := make([]*containerNode, 0, len(graph.nodes))
 | 
						|
 | 
						|
	var strongConnect func(*containerNode) (bool, error)
 | 
						|
	strongConnect = func(node *containerNode) (bool, error) {
 | 
						|
		logrus.Debugf("Strongconnecting node %s", node.id)
 | 
						|
 | 
						|
		info := new(nodeInfo)
 | 
						|
		info.index = index
 | 
						|
		info.lowLink = index
 | 
						|
		index++
 | 
						|
 | 
						|
		nodes[node.id] = info
 | 
						|
 | 
						|
		stack = append(stack, node)
 | 
						|
 | 
						|
		info.onStack = true
 | 
						|
 | 
						|
		logrus.Debugf("Pushed %s onto stack", node.id)
 | 
						|
 | 
						|
		// Work through all nodes we point to
 | 
						|
		for _, successor := range node.dependsOn {
 | 
						|
			if _, ok := nodes[successor.id]; !ok {
 | 
						|
				logrus.Debugf("Recursing to successor node %s", successor.id)
 | 
						|
 | 
						|
				cycle, err := strongConnect(successor)
 | 
						|
				if err != nil {
 | 
						|
					return false, err
 | 
						|
				} else if cycle {
 | 
						|
					return true, nil
 | 
						|
				}
 | 
						|
 | 
						|
				successorInfo := nodes[successor.id]
 | 
						|
				if successorInfo.lowLink < info.lowLink {
 | 
						|
					info.lowLink = successorInfo.lowLink
 | 
						|
				}
 | 
						|
			} else {
 | 
						|
				successorInfo := nodes[successor.id]
 | 
						|
				if successorInfo.index < info.lowLink && successorInfo.onStack {
 | 
						|
					info.lowLink = successorInfo.index
 | 
						|
				}
 | 
						|
			}
 | 
						|
		}
 | 
						|
 | 
						|
		if info.lowLink == info.index {
 | 
						|
			l := len(stack)
 | 
						|
			if l == 0 {
 | 
						|
				return false, fmt.Errorf("empty stack in detectCycles: %w", define.ErrInternal)
 | 
						|
			}
 | 
						|
 | 
						|
			// Pop off the stack
 | 
						|
			topOfStack := stack[l-1]
 | 
						|
			stack = stack[:l-1]
 | 
						|
 | 
						|
			// Popped item is no longer on the stack, mark as such
 | 
						|
			topInfo, ok := nodes[topOfStack.id]
 | 
						|
			if !ok {
 | 
						|
				return false, fmt.Errorf("finding node info for %s: %w", topOfStack.id, define.ErrInternal)
 | 
						|
			}
 | 
						|
			topInfo.onStack = false
 | 
						|
 | 
						|
			logrus.Debugf("Finishing node %s. Popped %s off stack", node.id, topOfStack.id)
 | 
						|
 | 
						|
			// If the top of the stack is not us, we have found a
 | 
						|
			// cycle
 | 
						|
			if topOfStack.id != node.id {
 | 
						|
				return true, nil
 | 
						|
			}
 | 
						|
		}
 | 
						|
 | 
						|
		return false, nil
 | 
						|
	}
 | 
						|
 | 
						|
	for id, node := range graph.nodes {
 | 
						|
		if _, ok := nodes[id]; !ok {
 | 
						|
			cycle, err := strongConnect(node)
 | 
						|
			if err != nil {
 | 
						|
				return false, err
 | 
						|
			} else if cycle {
 | 
						|
				return true, nil
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	return false, nil
 | 
						|
}
 | 
						|
 | 
						|
// Visit a node on a container graph and start the container, or set an error if
 | 
						|
// a dependency failed to start. if restart is true, startNode will restart the node instead of starting it.
 | 
						|
func startNode(ctx context.Context, node *containerNode, setError bool, ctrErrors map[string]error, ctrsVisited map[string]bool, restart bool) {
 | 
						|
	// First, check if we have already visited the node
 | 
						|
	if ctrsVisited[node.id] {
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// If setError is true, a dependency of us failed
 | 
						|
	// Mark us as failed and recurse
 | 
						|
	if setError {
 | 
						|
		// Mark us as visited, and set an error
 | 
						|
		ctrsVisited[node.id] = true
 | 
						|
		ctrErrors[node.id] = fmt.Errorf("a dependency of container %s failed to start: %w", node.id, define.ErrCtrStateInvalid)
 | 
						|
 | 
						|
		// Hit anyone who depends on us, and set errors on them too
 | 
						|
		for _, successor := range node.dependedOn {
 | 
						|
			startNode(ctx, successor, true, ctrErrors, ctrsVisited, restart)
 | 
						|
		}
 | 
						|
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Have all our dependencies started?
 | 
						|
	// If not, don't visit the node yet
 | 
						|
	depsVisited := true
 | 
						|
	for _, dep := range node.dependsOn {
 | 
						|
		depsVisited = depsVisited && ctrsVisited[dep.id]
 | 
						|
	}
 | 
						|
	if !depsVisited {
 | 
						|
		// Don't visit us yet, all dependencies are not up
 | 
						|
		// We'll hit the dependencies eventually, and when we do it will
 | 
						|
		// recurse here
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Going to try to start the container, mark us as visited
 | 
						|
	ctrsVisited[node.id] = true
 | 
						|
 | 
						|
	ctrErrored := false
 | 
						|
 | 
						|
	// Check if dependencies are running
 | 
						|
	// Graph traversal means we should have started them
 | 
						|
	// But they could have died before we got here
 | 
						|
	// Does not require that the container be locked, we only need to lock
 | 
						|
	// the dependencies
 | 
						|
	depsStopped, err := node.container.checkDependenciesRunning()
 | 
						|
	if err != nil {
 | 
						|
		ctrErrors[node.id] = err
 | 
						|
		ctrErrored = true
 | 
						|
	} else if len(depsStopped) > 0 {
 | 
						|
		// Our dependencies are not running
 | 
						|
		depsList := strings.Join(depsStopped, ",")
 | 
						|
		ctrErrors[node.id] = fmt.Errorf("the following dependencies of container %s are not running: %s: %w", node.id, depsList, define.ErrCtrStateInvalid)
 | 
						|
		ctrErrored = true
 | 
						|
	}
 | 
						|
 | 
						|
	// Lock before we start
 | 
						|
	node.container.lock.Lock()
 | 
						|
 | 
						|
	// Sync the container to pick up current state
 | 
						|
	if !ctrErrored {
 | 
						|
		if err := node.container.syncContainer(); err != nil {
 | 
						|
			ctrErrored = true
 | 
						|
			ctrErrors[node.id] = err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Start the container (only if it is not running)
 | 
						|
	if !ctrErrored && len(node.container.config.InitContainerType) < 1 {
 | 
						|
		if !restart && node.container.state.State != define.ContainerStateRunning {
 | 
						|
			if err := node.container.initAndStart(ctx); err != nil {
 | 
						|
				ctrErrored = true
 | 
						|
				ctrErrors[node.id] = err
 | 
						|
			}
 | 
						|
		}
 | 
						|
		if restart && node.container.state.State != define.ContainerStatePaused && node.container.state.State != define.ContainerStateUnknown {
 | 
						|
			if err := node.container.restartWithTimeout(ctx, node.container.config.StopTimeout); err != nil {
 | 
						|
				ctrErrored = true
 | 
						|
				ctrErrors[node.id] = err
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	node.container.lock.Unlock()
 | 
						|
 | 
						|
	// Recurse to anyone who depends on us and start them
 | 
						|
	for _, successor := range node.dependedOn {
 | 
						|
		startNode(ctx, successor, ctrErrored, ctrErrors, ctrsVisited, restart)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
// Visit a node on the container graph and remove it, or set an error if it
 | 
						|
// failed to remove. Only intended for use in pod removal; do *not* use when
 | 
						|
// removing individual containers.
 | 
						|
// All containers are assumed to be *UNLOCKED* on running this function.
 | 
						|
// Container locks will be acquired as necessary.
 | 
						|
// Pod and infraID are optional. If a pod is given it must be *LOCKED*.
 | 
						|
func removeNode(ctx context.Context, node *containerNode, pod *Pod, force bool, timeout *uint, setError bool, ctrErrors map[string]error, ctrsVisited map[string]bool, ctrNamedVolumes map[string]*ContainerNamedVolume) {
 | 
						|
	// If we already visited this node, we're done.
 | 
						|
	if ctrsVisited[node.id] {
 | 
						|
		return
 | 
						|
	}
 | 
						|
 | 
						|
	// Someone who depends on us failed.
 | 
						|
	// Mark us as failed and recurse.
 | 
						|
	if setError {
 | 
						|
		ctrsVisited[node.id] = true
 | 
						|
		ctrErrors[node.id] = fmt.Errorf("a container that depends on container %s could not be removed: %w", node.id, define.ErrCtrStateInvalid)
 | 
						|
 | 
						|
		// Hit anyone who depends on us, set errors there as well.
 | 
						|
		for _, successor := range node.dependsOn {
 | 
						|
			removeNode(ctx, successor, pod, force, timeout, true, ctrErrors, ctrsVisited, ctrNamedVolumes)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Does anyone still depend on us?
 | 
						|
	// Cannot remove if true. Once all our dependencies have been removed,
 | 
						|
	// we will be removed.
 | 
						|
	for _, dep := range node.dependedOn {
 | 
						|
		// The container that depends on us hasn't been removed yet.
 | 
						|
		// OK to continue on
 | 
						|
		if ok := ctrsVisited[dep.id]; !ok {
 | 
						|
			return
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Going to try to remove the node, mark us as visited
 | 
						|
	ctrsVisited[node.id] = true
 | 
						|
 | 
						|
	ctrErrored := false
 | 
						|
 | 
						|
	// Verify that all that depend on us are gone.
 | 
						|
	// Graph traversal should guarantee this is true, but this isn't that
 | 
						|
	// expensive, and it's better to be safe.
 | 
						|
	for _, dep := range node.dependedOn {
 | 
						|
		if _, err := node.container.runtime.GetContainer(dep.id); err == nil {
 | 
						|
			ctrErrored = true
 | 
						|
			ctrErrors[node.id] = fmt.Errorf("a container that depends on container %s still exists: %w", node.id, define.ErrDepExists)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	// Lock the container
 | 
						|
	node.container.lock.Lock()
 | 
						|
 | 
						|
	// Gate all subsequent bits behind a ctrErrored check - we don't want to
 | 
						|
	// proceed if a previous step failed.
 | 
						|
	if !ctrErrored {
 | 
						|
		if err := node.container.syncContainer(); err != nil {
 | 
						|
			ctrErrored = true
 | 
						|
			ctrErrors[node.id] = err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	if !ctrErrored {
 | 
						|
		for _, vol := range node.container.config.NamedVolumes {
 | 
						|
			ctrNamedVolumes[vol.Name] = vol
 | 
						|
		}
 | 
						|
 | 
						|
		if pod != nil && pod.state.InfraContainerID == node.id {
 | 
						|
			pod.state.InfraContainerID = ""
 | 
						|
			if err := pod.save(); err != nil {
 | 
						|
				ctrErrored = true
 | 
						|
				ctrErrors[node.id] = fmt.Errorf("error removing infra container %s from pod %s: %w", node.id, pod.ID(), err)
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	if !ctrErrored {
 | 
						|
		opts := ctrRmOpts{
 | 
						|
			Force:     force,
 | 
						|
			RemovePod: true,
 | 
						|
			Timeout:   timeout,
 | 
						|
		}
 | 
						|
 | 
						|
		if _, _, err := node.container.runtime.removeContainer(ctx, node.container, opts); err != nil {
 | 
						|
			ctrErrored = true
 | 
						|
			ctrErrors[node.id] = err
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	node.container.lock.Unlock()
 | 
						|
 | 
						|
	// Recurse to anyone who we depend on and remove them
 | 
						|
	for _, successor := range node.dependsOn {
 | 
						|
		removeNode(ctx, successor, pod, force, timeout, ctrErrored, ctrErrors, ctrsVisited, ctrNamedVolumes)
 | 
						|
	}
 | 
						|
}
 |