mirror of
https://github.com/ipfs/kubo.git
synced 2025-09-10 09:52:20 +08:00

We now consider debugerrors harmful: we've run into cases where debugerror.Wrap() hid valuable error information (err == io.EOF?). I've removed them from the main code, but left them in some tests. Go errors are lacking, but unfortunately, this isn't the solution. It is possible that debugerros.New or debugerrors.Errorf should remain still (i.e. only remove debugerrors.Wrap) but we don't use these errors often enough to keep.
544 lines
13 KiB
Go
544 lines
13 KiB
Go
// cmd/ipfs implements the primary CLI binary for ipfs
|
|
package main
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"math/rand"
|
|
"os"
|
|
"os/signal"
|
|
"runtime"
|
|
"runtime/pprof"
|
|
"strings"
|
|
"syscall"
|
|
"time"
|
|
|
|
ma "github.com/ipfs/go-ipfs/Godeps/_workspace/src/github.com/jbenet/go-multiaddr"
|
|
manet "github.com/ipfs/go-ipfs/Godeps/_workspace/src/github.com/jbenet/go-multiaddr-net"
|
|
|
|
context "github.com/ipfs/go-ipfs/Godeps/_workspace/src/golang.org/x/net/context"
|
|
cmds "github.com/ipfs/go-ipfs/commands"
|
|
cmdsCli "github.com/ipfs/go-ipfs/commands/cli"
|
|
cmdsHttp "github.com/ipfs/go-ipfs/commands/http"
|
|
core "github.com/ipfs/go-ipfs/core"
|
|
config "github.com/ipfs/go-ipfs/repo/config"
|
|
fsrepo "github.com/ipfs/go-ipfs/repo/fsrepo"
|
|
eventlog "github.com/ipfs/go-ipfs/thirdparty/eventlog"
|
|
u "github.com/ipfs/go-ipfs/util"
|
|
)
|
|
|
|
// log is the command logger
|
|
var log = eventlog.Logger("cmd/ipfs")
|
|
|
|
// signal to output help
|
|
var errHelpRequested = errors.New("Help Requested")
|
|
|
|
const (
|
|
EnvEnableProfiling = "IPFS_PROF"
|
|
cpuProfile = "ipfs.cpuprof"
|
|
heapProfile = "ipfs.memprof"
|
|
errorFormat = "ERROR: %v\n\n"
|
|
shutdownMessage = "Received interrupt signal, shutting down..."
|
|
)
|
|
|
|
type cmdInvocation struct {
|
|
path []string
|
|
cmd *cmds.Command
|
|
req cmds.Request
|
|
node *core.IpfsNode
|
|
}
|
|
|
|
// main roadmap:
|
|
// - parse the commandline to get a cmdInvocation
|
|
// - if user requests, help, print it and exit.
|
|
// - run the command invocation
|
|
// - output the response
|
|
// - if anything fails, print error, maybe with help
|
|
func main() {
|
|
rand.Seed(time.Now().UnixNano())
|
|
runtime.GOMAXPROCS(3) // FIXME rm arbitrary choice for n
|
|
ctx := eventlog.ContextWithLoggable(context.Background(), eventlog.Uuid("session"))
|
|
var err error
|
|
var invoc cmdInvocation
|
|
defer invoc.close()
|
|
|
|
// we'll call this local helper to output errors.
|
|
// this is so we control how to print errors in one place.
|
|
printErr := func(err error) {
|
|
fmt.Fprintf(os.Stderr, "Error: %s\n", err.Error())
|
|
}
|
|
|
|
stopFunc, err := profileIfEnabled()
|
|
if err != nil {
|
|
printErr(err)
|
|
os.Exit(1)
|
|
}
|
|
defer stopFunc() // to be executed as late as possible
|
|
|
|
// this is a local helper to print out help text.
|
|
// there's some considerations that this makes easier.
|
|
printHelp := func(long bool, w io.Writer) {
|
|
helpFunc := cmdsCli.ShortHelp
|
|
if long {
|
|
helpFunc = cmdsCli.LongHelp
|
|
}
|
|
|
|
helpFunc("ipfs", Root, invoc.path, w)
|
|
}
|
|
|
|
// this is a message to tell the user how to get the help text
|
|
printMetaHelp := func(w io.Writer) {
|
|
cmdPath := strings.Join(invoc.path, " ")
|
|
fmt.Fprintf(w, "Use 'ipfs %s --help' for information about this command\n", cmdPath)
|
|
}
|
|
|
|
// parse the commandline into a command invocation
|
|
parseErr := invoc.Parse(ctx, os.Args[1:])
|
|
|
|
// BEFORE handling the parse error, if we have enough information
|
|
// AND the user requested help, print it out and exit
|
|
if invoc.req != nil {
|
|
longH, shortH, err := invoc.requestedHelp()
|
|
if err != nil {
|
|
printErr(err)
|
|
os.Exit(1)
|
|
}
|
|
if longH || shortH {
|
|
printHelp(longH, os.Stdout)
|
|
os.Exit(0)
|
|
}
|
|
}
|
|
|
|
// here we handle the cases where
|
|
// - commands with no Run func are invoked directly.
|
|
// - the main command is invoked.
|
|
if invoc.cmd == nil || invoc.cmd.Run == nil {
|
|
printHelp(false, os.Stdout)
|
|
os.Exit(0)
|
|
}
|
|
|
|
// ok now handle parse error (which means cli input was wrong,
|
|
// e.g. incorrect number of args, or nonexistent subcommand)
|
|
if parseErr != nil {
|
|
printErr(parseErr)
|
|
|
|
// this was a user error, print help.
|
|
if invoc.cmd != nil {
|
|
// we need a newline space.
|
|
fmt.Fprintf(os.Stderr, "\n")
|
|
printMetaHelp(os.Stderr)
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
|
|
// our global interrupt handler may try to stop the daemon
|
|
// before the daemon is ready to be stopped; this dirty
|
|
// workaround is for the daemon only; other commands are always
|
|
// ready to be stopped
|
|
if invoc.cmd != daemonCmd {
|
|
close(invoc.req.Context().InitDone)
|
|
}
|
|
|
|
// ok, finally, run the command invocation.
|
|
output, err := invoc.Run(ctx)
|
|
if err != nil {
|
|
printErr(err)
|
|
|
|
// if this error was a client error, print short help too.
|
|
if isClientError(err) {
|
|
printMetaHelp(os.Stderr)
|
|
}
|
|
os.Exit(1)
|
|
}
|
|
|
|
// everything went better than expected :)
|
|
io.Copy(os.Stdout, output)
|
|
}
|
|
|
|
func (i *cmdInvocation) Run(ctx context.Context) (output io.Reader, err error) {
|
|
// setup our global interrupt handler.
|
|
i.setupInterruptHandler()
|
|
|
|
// check if user wants to debug. option OR env var.
|
|
debug, _, err := i.req.Option("debug").Bool()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if debug || u.GetenvBool("DEBUG") || os.Getenv("IPFS_LOGGING") == "debug" {
|
|
u.Debug = true
|
|
u.SetDebugLogging()
|
|
}
|
|
|
|
res, err := callCommand(ctx, i.req, Root, i.cmd)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if err := res.Error(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return res.Reader()
|
|
}
|
|
|
|
func (i *cmdInvocation) constructNodeFunc(ctx context.Context) func() (*core.IpfsNode, error) {
|
|
return func() (*core.IpfsNode, error) {
|
|
if i.req == nil {
|
|
return nil, errors.New("constructing node without a request")
|
|
}
|
|
|
|
cmdctx := i.req.Context()
|
|
if cmdctx == nil {
|
|
return nil, errors.New("constructing node without a request context")
|
|
}
|
|
|
|
r, err := fsrepo.Open(i.req.Context().ConfigRoot)
|
|
if err != nil { // repo is owned by the node
|
|
return nil, err
|
|
}
|
|
|
|
// ok everything is good. set it on the invocation (for ownership)
|
|
// and return it.
|
|
n, err := core.NewIPFSNode(ctx, core.Standard(r, cmdctx.Online))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
i.node = n
|
|
return i.node, nil
|
|
}
|
|
}
|
|
|
|
func (i *cmdInvocation) close() {
|
|
// let's not forget teardown. If a node was initialized, we must close it.
|
|
// Note that this means the underlying req.Context().Node variable is exposed.
|
|
// this is gross, and should be changed when we extract out the exec Context.
|
|
if i.node != nil {
|
|
log.Info("Shutting down node...")
|
|
i.node.Close()
|
|
}
|
|
}
|
|
|
|
func (i *cmdInvocation) Parse(ctx context.Context, args []string) error {
|
|
var err error
|
|
|
|
i.req, i.cmd, i.path, err = cmdsCli.Parse(args, os.Stdin, Root)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
i.req.Context().Context = ctx
|
|
|
|
repoPath, err := getRepoPath(i.req)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
log.Debugf("config path is %s", repoPath)
|
|
|
|
// this sets up the function that will initialize the config lazily.
|
|
cmdctx := i.req.Context()
|
|
cmdctx.ConfigRoot = repoPath
|
|
cmdctx.LoadConfig = loadConfig
|
|
// this sets up the function that will initialize the node
|
|
// this is so that we can construct the node lazily.
|
|
cmdctx.ConstructNode = i.constructNodeFunc(ctx)
|
|
|
|
// if no encoding was specified by user, default to plaintext encoding
|
|
// (if command doesn't support plaintext, use JSON instead)
|
|
if !i.req.Option("encoding").Found() {
|
|
if i.req.Command().Marshalers != nil && i.req.Command().Marshalers[cmds.Text] != nil {
|
|
i.req.SetOption("encoding", cmds.Text)
|
|
} else {
|
|
i.req.SetOption("encoding", cmds.JSON)
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (i *cmdInvocation) requestedHelp() (short bool, long bool, err error) {
|
|
longHelp, _, err := i.req.Option("help").Bool()
|
|
if err != nil {
|
|
return false, false, err
|
|
}
|
|
shortHelp, _, err := i.req.Option("h").Bool()
|
|
if err != nil {
|
|
return false, false, err
|
|
}
|
|
return longHelp, shortHelp, nil
|
|
}
|
|
|
|
func callPreCommandHooks(ctx context.Context, details cmdDetails, req cmds.Request, root *cmds.Command) error {
|
|
|
|
log.Event(ctx, "callPreCommandHooks", &details)
|
|
log.Debug("Calling pre-command hooks...")
|
|
|
|
return nil
|
|
}
|
|
|
|
func callCommand(ctx context.Context, req cmds.Request, root *cmds.Command, cmd *cmds.Command) (cmds.Response, error) {
|
|
log.Info(config.EnvDir, " ", req.Context().ConfigRoot)
|
|
var res cmds.Response
|
|
|
|
details, err := commandDetails(req.Path(), root)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debug("looking for running daemon...")
|
|
useDaemon, err := commandShouldRunOnDaemon(*details, req, root)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
err = callPreCommandHooks(ctx, *details, req, root)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if cmd.PreRun != nil {
|
|
err = cmd.PreRun(req)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
if useDaemon {
|
|
|
|
cfg, err := req.Context().GetConfig()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
addr, err := ma.NewMultiaddr(cfg.Addresses.API)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Infof("Executing command on daemon running at %s", addr)
|
|
_, host, err := manet.DialArgs(addr)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
client := cmdsHttp.NewClient(host)
|
|
|
|
res, err = client.Send(req)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
} else {
|
|
log.Debug("Executing command locally")
|
|
|
|
// Okay!!!!! NOW we can call the command.
|
|
res = root.Call(req)
|
|
|
|
}
|
|
|
|
if cmd.PostRun != nil {
|
|
cmd.PostRun(req, res)
|
|
}
|
|
|
|
return res, nil
|
|
}
|
|
|
|
// commandDetails returns a command's details for the command given by |path|
|
|
// within the |root| command tree.
|
|
//
|
|
// Returns an error if the command is not found in the Command tree.
|
|
func commandDetails(path []string, root *cmds.Command) (*cmdDetails, error) {
|
|
var details cmdDetails
|
|
// find the last command in path that has a cmdDetailsMap entry
|
|
cmd := root
|
|
for _, cmp := range path {
|
|
var found bool
|
|
cmd, found = cmd.Subcommands[cmp]
|
|
if !found {
|
|
return nil, fmt.Errorf("subcommand %s should be in root", cmp)
|
|
}
|
|
|
|
if cmdDetails, found := cmdDetailsMap[cmd]; found {
|
|
details = cmdDetails
|
|
}
|
|
}
|
|
return &details, nil
|
|
}
|
|
|
|
// commandShouldRunOnDaemon determines, from commmand details, whether a
|
|
// command ought to be executed on an IPFS daemon.
|
|
//
|
|
// It returns true if the command should be executed on a daemon and false if
|
|
// it should be executed on a client. It returns an error if the command must
|
|
// NOT be executed on either.
|
|
func commandShouldRunOnDaemon(details cmdDetails, req cmds.Request, root *cmds.Command) (bool, error) {
|
|
path := req.Path()
|
|
// root command.
|
|
if len(path) < 1 {
|
|
return false, nil
|
|
}
|
|
|
|
if details.cannotRunOnClient && details.cannotRunOnDaemon {
|
|
return false, fmt.Errorf("command disabled: %s", path[0])
|
|
}
|
|
|
|
if details.doesNotUseRepo && details.canRunOnClient() {
|
|
return false, nil
|
|
}
|
|
|
|
// at this point need to know whether daemon is running. we defer
|
|
// to this point so that some commands dont open files unnecessarily.
|
|
daemonLocked := fsrepo.LockedByOtherProcess(req.Context().ConfigRoot)
|
|
|
|
if daemonLocked {
|
|
|
|
log.Info("a daemon is running...")
|
|
|
|
if details.cannotRunOnDaemon {
|
|
e := "ipfs daemon is running. please stop it to run this command"
|
|
return false, cmds.ClientError(e)
|
|
}
|
|
|
|
return true, nil
|
|
}
|
|
|
|
if details.cannotRunOnClient {
|
|
return false, cmds.ClientError("must run on the ipfs daemon")
|
|
}
|
|
|
|
return false, nil
|
|
}
|
|
|
|
func isClientError(err error) bool {
|
|
|
|
// Somewhat suprisingly, the pointer cast fails to recognize commands.Error
|
|
// passed as values, so we check both.
|
|
|
|
// cast to cmds.Error
|
|
switch e := err.(type) {
|
|
case *cmds.Error:
|
|
return e.Code == cmds.ErrClient
|
|
case cmds.Error:
|
|
return e.Code == cmds.ErrClient
|
|
}
|
|
return false
|
|
}
|
|
|
|
func getRepoPath(req cmds.Request) (string, error) {
|
|
repoOpt, found, err := req.Option("config").String()
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
if found && repoOpt != "" {
|
|
return repoOpt, nil
|
|
}
|
|
|
|
repoPath, err := fsrepo.BestKnownPath()
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
return repoPath, nil
|
|
}
|
|
|
|
func loadConfig(path string) (*config.Config, error) {
|
|
return fsrepo.ConfigAt(path)
|
|
}
|
|
|
|
// startProfiling begins CPU profiling and returns a `stop` function to be
|
|
// executed as late as possible. The stop function captures the memprofile.
|
|
func startProfiling() (func(), error) {
|
|
|
|
// start CPU profiling as early as possible
|
|
ofi, err := os.Create(cpuProfile)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
pprof.StartCPUProfile(ofi)
|
|
|
|
stopProfiling := func() {
|
|
pprof.StopCPUProfile()
|
|
defer ofi.Close() // captured by the closure
|
|
err := writeHeapProfileToFile()
|
|
if err != nil {
|
|
log.Critical(err)
|
|
}
|
|
}
|
|
return stopProfiling, nil
|
|
}
|
|
|
|
func writeHeapProfileToFile() error {
|
|
mprof, err := os.Create(heapProfile)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer mprof.Close() // _after_ writing the heap profile
|
|
return pprof.WriteHeapProfile(mprof)
|
|
}
|
|
|
|
// listen for and handle SIGTERM
|
|
func (i *cmdInvocation) setupInterruptHandler() {
|
|
|
|
ctx := i.req.Context()
|
|
sig := allInterruptSignals()
|
|
|
|
go func() {
|
|
// first time, try to shut down.
|
|
|
|
// loop because we may be
|
|
for count := 0; ; count++ {
|
|
<-sig
|
|
|
|
// if we're still initializing, cannot use `ctx.GetNode()`
|
|
select {
|
|
default: // initialization not done
|
|
fmt.Println(shutdownMessage)
|
|
os.Exit(-1)
|
|
case <-ctx.InitDone:
|
|
}
|
|
|
|
switch count {
|
|
case 0:
|
|
fmt.Println(shutdownMessage)
|
|
if ctx.Online {
|
|
go func() {
|
|
// TODO cancel the command context instead
|
|
n, err := ctx.GetNode()
|
|
if err != nil {
|
|
log.Error(err)
|
|
fmt.Println(shutdownMessage)
|
|
os.Exit(-1)
|
|
}
|
|
n.Close()
|
|
log.Info("Gracefully shut down.")
|
|
}()
|
|
} else {
|
|
os.Exit(0)
|
|
}
|
|
|
|
default:
|
|
fmt.Println("Received another interrupt before graceful shutdown, terminating...")
|
|
os.Exit(-1)
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
func allInterruptSignals() chan os.Signal {
|
|
sigc := make(chan os.Signal, 1)
|
|
signal.Notify(sigc, syscall.SIGHUP, syscall.SIGINT,
|
|
syscall.SIGTERM)
|
|
return sigc
|
|
}
|
|
|
|
func profileIfEnabled() (func(), error) {
|
|
// FIXME this is a temporary hack so profiling of asynchronous operations
|
|
// works as intended.
|
|
if os.Getenv(EnvEnableProfiling) != "" {
|
|
stopProfilingFunc, err := startProfiling() // TODO maybe change this to its own option... profiling makes it slower.
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return stopProfilingFunc, nil
|
|
}
|
|
return func() {}, nil
|
|
}
|