mirror of
https://github.com/ipfs/kubo.git
synced 2025-06-28 00:39:31 +08:00
ctx closer races #270
This commit is contained in:
@ -89,13 +89,13 @@ func newSingleConn(ctx context.Context, local, remote peer.Peer,
|
||||
log.Info("newSingleConn: %v to %v", local, remote)
|
||||
|
||||
// setup the various io goroutines
|
||||
conn.Children().Add(1)
|
||||
go func() {
|
||||
conn.Children().Add(1)
|
||||
conn.msgio.outgoing.WriteTo(maconn)
|
||||
conn.Children().Done()
|
||||
}()
|
||||
conn.Children().Add(1)
|
||||
go func() {
|
||||
conn.Children().Add(1)
|
||||
conn.msgio.incoming.ReadFrom(maconn, MaxMessageSize)
|
||||
conn.Children().Done()
|
||||
}()
|
||||
|
@ -47,7 +47,6 @@ func (l *listener) close() error {
|
||||
}
|
||||
|
||||
func (l *listener) listen() {
|
||||
l.Children().Add(1)
|
||||
defer l.Children().Done()
|
||||
|
||||
// handle at most chansize concurrent handshakes
|
||||
@ -143,6 +142,7 @@ func Listen(ctx context.Context, addr ma.Multiaddr, local peer.Peer, peers peer.
|
||||
ctx2, _ := context.WithCancel(ctx)
|
||||
l.ContextCloser = ctxc.NewContextCloser(ctx2, l.close)
|
||||
|
||||
l.Children().Add(1)
|
||||
go l.listen()
|
||||
|
||||
return l, nil
|
||||
|
@ -57,6 +57,8 @@ func NewMultiConn(ctx context.Context, local, remote peer.Peer, conns []Conn) (*
|
||||
if conns != nil && len(conns) > 0 {
|
||||
c.Add(conns...)
|
||||
}
|
||||
|
||||
c.Children().Add(1)
|
||||
go c.fanOut()
|
||||
return c, nil
|
||||
}
|
||||
@ -81,6 +83,8 @@ func (c *MultiConn) Add(conns ...Conn) {
|
||||
}
|
||||
|
||||
c.conns[c2.ID()] = c2
|
||||
c.Children().Add(1)
|
||||
c2.Children().Add(1) // yep, on the child too.
|
||||
go c.fanInSingle(c2)
|
||||
log.Infof("MultiConn: added %s", c2)
|
||||
}
|
||||
@ -134,7 +138,6 @@ func CloseConns(conns ...Conn) {
|
||||
// fanOut is the multiplexor out -- it sends outgoing messages over the
|
||||
// underlying single connections.
|
||||
func (c *MultiConn) fanOut() {
|
||||
c.Children().Add(1)
|
||||
defer c.Children().Done()
|
||||
|
||||
i := 0
|
||||
@ -165,9 +168,6 @@ func (c *MultiConn) fanOut() {
|
||||
// fanInSingle is a multiplexor in -- it receives incoming messages over the
|
||||
// underlying single connections.
|
||||
func (c *MultiConn) fanInSingle(child Conn) {
|
||||
c.Children().Add(1)
|
||||
child.Children().Add(1) // yep, on the child too.
|
||||
|
||||
// cleanup all data associated with this child Connection.
|
||||
defer func() {
|
||||
log.Infof("closing: %s", child)
|
||||
|
@ -139,6 +139,8 @@ func (s *Swarm) connSetup(c conn.Conn) (conn.Conn, error) {
|
||||
s.connsLock.Unlock()
|
||||
|
||||
// kick off reader goroutine
|
||||
s.Children().Add(1)
|
||||
mc.Children().Add(1) // child of Conn as well.
|
||||
go s.fanInSingle(mc)
|
||||
log.Debugf("added new multiconn: %s", mc)
|
||||
} else {
|
||||
@ -154,7 +156,6 @@ func (s *Swarm) connSetup(c conn.Conn) (conn.Conn, error) {
|
||||
|
||||
// Handles the unwrapping + sending of messages to the right connection.
|
||||
func (s *Swarm) fanOut() {
|
||||
s.Children().Add(1)
|
||||
defer s.Children().Done()
|
||||
|
||||
i := 0
|
||||
@ -194,9 +195,6 @@ func (s *Swarm) fanOut() {
|
||||
// Handles the receiving + wrapping of messages, per conn.
|
||||
// Consider using reflect.Select with one goroutine instead of n.
|
||||
func (s *Swarm) fanInSingle(c conn.Conn) {
|
||||
s.Children().Add(1)
|
||||
c.Children().Add(1) // child of Conn as well.
|
||||
|
||||
// cleanup all data associated with this child Connection.
|
||||
defer func() {
|
||||
// remove it from the map.
|
||||
|
@ -83,6 +83,7 @@ func NewSwarm(ctx context.Context, listenAddrs []ma.Multiaddr, local peer.Peer,
|
||||
// ContextCloser for proper child management.
|
||||
s.ContextCloser = ctxc.NewContextCloser(ctx, s.close)
|
||||
|
||||
s.Children().Add(1)
|
||||
go s.fanOut()
|
||||
return s, s.listen(listenAddrs)
|
||||
}
|
||||
|
@ -120,6 +120,7 @@ func NewContextCloser(ctx context.Context, cf CloseFunc) ContextCloser {
|
||||
closed: make(chan struct{}),
|
||||
}
|
||||
|
||||
c.Children().Add(1) // we're a child goroutine, to be waited upon.
|
||||
go c.closeOnContextDone()
|
||||
return c
|
||||
}
|
||||
@ -176,7 +177,6 @@ func (c *contextCloser) closeLogic() {
|
||||
// we need to go through the Close motions anyway. Hence all the sync
|
||||
// stuff all over the place...
|
||||
func (c *contextCloser) closeOnContextDone() {
|
||||
c.Children().Add(1) // we're a child goroutine, to be waited upon.
|
||||
<-c.Context().Done() // wait until parent (context) is done.
|
||||
c.internalClose()
|
||||
c.Children().Done()
|
||||
|
Reference in New Issue
Block a user