mirror of
				https://gitcode.com/gitea/gitea.git
				synced 2025-10-25 12:26:40 +08:00 
			
		
		
		
	 18f26cfbf7
			
		
	
	18f26cfbf7
	
	
	
		
			
			Before there was a "graceful function": RunWithShutdownFns, it's mainly for some modules which doesn't support context. The old queue system doesn't work well with context, so the old queues need it. After the queue refactoring, the new queue works with context well, so, use Golang context as much as possible, the `RunWithShutdownFns` could be removed (replaced by RunWithCancel for context cancel mechanism), the related code could be simplified. This PR also fixes some legacy queue-init problems, eg: * typo : archiver: "unable to create codes indexer queue" => "unable to create repo-archive queue" * no nil check for failed queues, which causes unfriendly panic After this PR, many goroutines could have better display name:  
		
			
				
	
	
		
			282 lines
		
	
	
		
			7.9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			282 lines
		
	
	
		
			7.9 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2019 The Gitea Authors. All rights reserved.
 | |
| // SPDX-License-Identifier: MIT
 | |
| 
 | |
| //go:build !windows
 | |
| 
 | |
| package graceful
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"errors"
 | |
| 	"os"
 | |
| 	"os/signal"
 | |
| 	"runtime/pprof"
 | |
| 	"strconv"
 | |
| 	"sync"
 | |
| 	"syscall"
 | |
| 	"time"
 | |
| 
 | |
| 	"code.gitea.io/gitea/modules/graceful/releasereopen"
 | |
| 	"code.gitea.io/gitea/modules/log"
 | |
| 	"code.gitea.io/gitea/modules/process"
 | |
| 	"code.gitea.io/gitea/modules/setting"
 | |
| )
 | |
| 
 | |
| // Manager manages the graceful shutdown process
 | |
| type Manager struct {
 | |
| 	isChild                bool
 | |
| 	forked                 bool
 | |
| 	lock                   *sync.RWMutex
 | |
| 	state                  state
 | |
| 	shutdownCtx            context.Context
 | |
| 	hammerCtx              context.Context
 | |
| 	terminateCtx           context.Context
 | |
| 	managerCtx             context.Context
 | |
| 	shutdownCtxCancel      context.CancelFunc
 | |
| 	hammerCtxCancel        context.CancelFunc
 | |
| 	terminateCtxCancel     context.CancelFunc
 | |
| 	managerCtxCancel       context.CancelFunc
 | |
| 	runningServerWaitGroup sync.WaitGroup
 | |
| 	createServerWaitGroup  sync.WaitGroup
 | |
| 	terminateWaitGroup     sync.WaitGroup
 | |
| 
 | |
| 	toRunAtShutdown  []func()
 | |
| 	toRunAtTerminate []func()
 | |
| }
 | |
| 
 | |
| func newGracefulManager(ctx context.Context) *Manager {
 | |
| 	manager := &Manager{
 | |
| 		isChild: len(os.Getenv(listenFDsEnv)) > 0 && os.Getppid() > 1,
 | |
| 		lock:    &sync.RWMutex{},
 | |
| 	}
 | |
| 	manager.createServerWaitGroup.Add(numberOfServersToCreate)
 | |
| 	manager.start(ctx)
 | |
| 	return manager
 | |
| }
 | |
| 
 | |
| type systemdNotifyMsg string
 | |
| 
 | |
| const (
 | |
| 	readyMsg     systemdNotifyMsg = "READY=1"
 | |
| 	stoppingMsg  systemdNotifyMsg = "STOPPING=1"
 | |
| 	reloadingMsg systemdNotifyMsg = "RELOADING=1"
 | |
| 	watchdogMsg  systemdNotifyMsg = "WATCHDOG=1"
 | |
| )
 | |
| 
 | |
| func statusMsg(msg string) systemdNotifyMsg {
 | |
| 	return systemdNotifyMsg("STATUS=" + msg)
 | |
| }
 | |
| 
 | |
| func pidMsg() systemdNotifyMsg {
 | |
| 	return systemdNotifyMsg("MAINPID=" + strconv.Itoa(os.Getpid()))
 | |
| }
 | |
| 
 | |
| // Notify systemd of status via the notify protocol
 | |
| func (g *Manager) notify(msg systemdNotifyMsg) {
 | |
| 	conn, err := getNotifySocket()
 | |
| 	if err != nil {
 | |
| 		// the err is logged in getNotifySocket
 | |
| 		return
 | |
| 	}
 | |
| 	if conn == nil {
 | |
| 		return
 | |
| 	}
 | |
| 	defer conn.Close()
 | |
| 
 | |
| 	if _, err = conn.Write([]byte(msg)); err != nil {
 | |
| 		log.Warn("Failed to notify NOTIFY_SOCKET: %v", err)
 | |
| 		return
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (g *Manager) start(ctx context.Context) {
 | |
| 	// Make contexts
 | |
| 	g.terminateCtx, g.terminateCtxCancel = context.WithCancel(ctx)
 | |
| 	g.shutdownCtx, g.shutdownCtxCancel = context.WithCancel(ctx)
 | |
| 	g.hammerCtx, g.hammerCtxCancel = context.WithCancel(ctx)
 | |
| 	g.managerCtx, g.managerCtxCancel = context.WithCancel(ctx)
 | |
| 
 | |
| 	// Next add pprof labels to these contexts
 | |
| 	g.terminateCtx = pprof.WithLabels(g.terminateCtx, pprof.Labels("graceful-lifecycle", "with-terminate"))
 | |
| 	g.shutdownCtx = pprof.WithLabels(g.shutdownCtx, pprof.Labels("graceful-lifecycle", "with-shutdown"))
 | |
| 	g.hammerCtx = pprof.WithLabels(g.hammerCtx, pprof.Labels("graceful-lifecycle", "with-hammer"))
 | |
| 	g.managerCtx = pprof.WithLabels(g.managerCtx, pprof.Labels("graceful-lifecycle", "with-manager"))
 | |
| 
 | |
| 	// Now label this and all goroutines created by this goroutine with the graceful-lifecycle manager
 | |
| 	pprof.SetGoroutineLabels(g.managerCtx)
 | |
| 	defer pprof.SetGoroutineLabels(ctx)
 | |
| 
 | |
| 	// Set the running state & handle signals
 | |
| 	g.setState(stateRunning)
 | |
| 	g.notify(statusMsg("Starting Gitea"))
 | |
| 	g.notify(pidMsg())
 | |
| 	go g.handleSignals(g.managerCtx)
 | |
| 
 | |
| 	// Handle clean up of unused provided listeners	and delayed start-up
 | |
| 	startupDone := make(chan struct{})
 | |
| 	go func() {
 | |
| 		defer close(startupDone)
 | |
| 		// Wait till we're done getting all of the listeners and then close
 | |
| 		// the unused ones
 | |
| 		g.createServerWaitGroup.Wait()
 | |
| 		// Ignore the error here there's not much we can do with it
 | |
| 		// They're logged in the CloseProvidedListeners function
 | |
| 		_ = CloseProvidedListeners()
 | |
| 		g.notify(readyMsg)
 | |
| 	}()
 | |
| 	if setting.StartupTimeout > 0 {
 | |
| 		go func() {
 | |
| 			select {
 | |
| 			case <-startupDone:
 | |
| 				return
 | |
| 			case <-g.IsShutdown():
 | |
| 				func() {
 | |
| 					// When waitgroup counter goes negative it will panic - we don't care about this so we can just ignore it.
 | |
| 					defer func() {
 | |
| 						_ = recover()
 | |
| 					}()
 | |
| 					// Ensure that the createServerWaitGroup stops waiting
 | |
| 					for {
 | |
| 						g.createServerWaitGroup.Done()
 | |
| 					}
 | |
| 				}()
 | |
| 				return
 | |
| 			case <-time.After(setting.StartupTimeout):
 | |
| 				log.Error("Startup took too long! Shutting down")
 | |
| 				g.notify(statusMsg("Startup took too long! Shutting down"))
 | |
| 				g.notify(stoppingMsg)
 | |
| 				g.doShutdown()
 | |
| 			}
 | |
| 		}()
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (g *Manager) handleSignals(ctx context.Context) {
 | |
| 	ctx, _, finished := process.GetManager().AddTypedContext(ctx, "Graceful: HandleSignals", process.SystemProcessType, true)
 | |
| 	defer finished()
 | |
| 
 | |
| 	signalChannel := make(chan os.Signal, 1)
 | |
| 
 | |
| 	signal.Notify(
 | |
| 		signalChannel,
 | |
| 		syscall.SIGHUP,
 | |
| 		syscall.SIGUSR1,
 | |
| 		syscall.SIGUSR2,
 | |
| 		syscall.SIGINT,
 | |
| 		syscall.SIGTERM,
 | |
| 		syscall.SIGTSTP,
 | |
| 	)
 | |
| 
 | |
| 	watchdogTimeout := getWatchdogTimeout()
 | |
| 	t := &time.Ticker{}
 | |
| 	if watchdogTimeout != 0 {
 | |
| 		g.notify(watchdogMsg)
 | |
| 		t = time.NewTicker(watchdogTimeout / 2)
 | |
| 	}
 | |
| 
 | |
| 	pid := syscall.Getpid()
 | |
| 	for {
 | |
| 		select {
 | |
| 		case sig := <-signalChannel:
 | |
| 			switch sig {
 | |
| 			case syscall.SIGHUP:
 | |
| 				log.Info("PID: %d. Received SIGHUP. Attempting GracefulRestart...", pid)
 | |
| 				g.DoGracefulRestart()
 | |
| 			case syscall.SIGUSR1:
 | |
| 				log.Warn("PID %d. Received SIGUSR1. Releasing and reopening logs", pid)
 | |
| 				g.notify(statusMsg("Releasing and reopening logs"))
 | |
| 				if err := releasereopen.GetManager().ReleaseReopen(); err != nil {
 | |
| 					log.Error("Error whilst releasing and reopening logs: %v", err)
 | |
| 				}
 | |
| 			case syscall.SIGUSR2:
 | |
| 				log.Warn("PID %d. Received SIGUSR2. Hammering...", pid)
 | |
| 				g.DoImmediateHammer()
 | |
| 			case syscall.SIGINT:
 | |
| 				log.Warn("PID %d. Received SIGINT. Shutting down...", pid)
 | |
| 				g.DoGracefulShutdown()
 | |
| 			case syscall.SIGTERM:
 | |
| 				log.Warn("PID %d. Received SIGTERM. Shutting down...", pid)
 | |
| 				g.DoGracefulShutdown()
 | |
| 			case syscall.SIGTSTP:
 | |
| 				log.Info("PID %d. Received SIGTSTP.", pid)
 | |
| 			default:
 | |
| 				log.Info("PID %d. Received %v.", pid, sig)
 | |
| 			}
 | |
| 		case <-t.C:
 | |
| 			g.notify(watchdogMsg)
 | |
| 		case <-ctx.Done():
 | |
| 			log.Warn("PID: %d. Background context for manager closed - %v - Shutting down...", pid, ctx.Err())
 | |
| 			g.DoGracefulShutdown()
 | |
| 			return
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (g *Manager) doFork() error {
 | |
| 	g.lock.Lock()
 | |
| 	if g.forked {
 | |
| 		g.lock.Unlock()
 | |
| 		return errors.New("another process already forked. Ignoring this one")
 | |
| 	}
 | |
| 	g.forked = true
 | |
| 	g.lock.Unlock()
 | |
| 
 | |
| 	g.notify(reloadingMsg)
 | |
| 
 | |
| 	// We need to move the file logs to append pids
 | |
| 	setting.RestartLogsWithPIDSuffix()
 | |
| 
 | |
| 	_, err := RestartProcess()
 | |
| 
 | |
| 	return err
 | |
| }
 | |
| 
 | |
| // DoGracefulRestart causes a graceful restart
 | |
| func (g *Manager) DoGracefulRestart() {
 | |
| 	if setting.GracefulRestartable {
 | |
| 		log.Info("PID: %d. Forking...", os.Getpid())
 | |
| 		err := g.doFork()
 | |
| 		if err != nil {
 | |
| 			if err.Error() == "another process already forked. Ignoring this one" {
 | |
| 				g.DoImmediateHammer()
 | |
| 			} else {
 | |
| 				log.Error("Error whilst forking from PID: %d : %v", os.Getpid(), err)
 | |
| 			}
 | |
| 		}
 | |
| 		// doFork calls RestartProcess which starts a new Gitea process, so this parent process needs to exit
 | |
| 		// Otherwise some resources (eg: leveldb lock) will be held by this parent process and the new process will fail to start
 | |
| 		log.Info("PID: %d. Shutting down after forking ...", os.Getpid())
 | |
| 		g.doShutdown()
 | |
| 	} else {
 | |
| 		log.Info("PID: %d. Not set restartable. Shutting down...", os.Getpid())
 | |
| 		g.notify(stoppingMsg)
 | |
| 		g.doShutdown()
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // DoImmediateHammer causes an immediate hammer
 | |
| func (g *Manager) DoImmediateHammer() {
 | |
| 	g.notify(statusMsg("Sending immediate hammer"))
 | |
| 	g.doHammerTime(0 * time.Second)
 | |
| }
 | |
| 
 | |
| // DoGracefulShutdown causes a graceful shutdown
 | |
| func (g *Manager) DoGracefulShutdown() {
 | |
| 	g.lock.Lock()
 | |
| 	if !g.forked {
 | |
| 		g.lock.Unlock()
 | |
| 		g.notify(stoppingMsg)
 | |
| 	} else {
 | |
| 		g.lock.Unlock()
 | |
| 		g.notify(statusMsg("Shutting down after fork"))
 | |
| 	}
 | |
| 	g.doShutdown()
 | |
| }
 | |
| 
 | |
| // RegisterServer registers the running of a listening server, in the case of unix this means that the parent process can now die.
 | |
| // Any call to RegisterServer must be matched by a call to ServerDone
 | |
| func (g *Manager) RegisterServer() {
 | |
| 	KillParent()
 | |
| 	g.runningServerWaitGroup.Add(1)
 | |
| }
 |