Fix flaky test: TestCloseConnectionWhenServerPrefaceNotReceived (#1870)
* Atomically update minConnectTimeout in test. * Refactor the flaky test. * Post review update * mend
This commit is contained in:
@ -926,7 +926,7 @@ func (cc *ClientConn) resolveNow(o resolver.ResolveNowOption) {
|
|||||||
|
|
||||||
// Close tears down the ClientConn and all underlying connections.
|
// Close tears down the ClientConn and all underlying connections.
|
||||||
func (cc *ClientConn) Close() error {
|
func (cc *ClientConn) Close() error {
|
||||||
cc.cancel()
|
defer cc.cancel()
|
||||||
|
|
||||||
cc.mu.Lock()
|
cc.mu.Lock()
|
||||||
if cc.conns == nil {
|
if cc.conns == nil {
|
||||||
|
|||||||
@ -19,9 +19,9 @@
|
|||||||
package grpc
|
package grpc
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"io"
|
|
||||||
"math"
|
"math"
|
||||||
"net"
|
"net"
|
||||||
|
"sync/atomic"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
@ -176,95 +176,66 @@ func TestCloseConnectionWhenServerPrefaceNotReceived(t *testing.T) {
|
|||||||
}()
|
}()
|
||||||
defer leakcheck.Check(t)
|
defer leakcheck.Check(t)
|
||||||
minConnectTimeout = time.Millisecond * 500
|
minConnectTimeout = time.Millisecond * 500
|
||||||
server, err := net.Listen("tcp", "localhost:0")
|
lis, err := net.Listen("tcp", "localhost:0")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Error while listening. Err: %v", err)
|
t.Fatalf("Error while listening. Err: %v", err)
|
||||||
}
|
}
|
||||||
defer server.Close()
|
var (
|
||||||
|
conn2 net.Conn
|
||||||
|
over uint32
|
||||||
|
)
|
||||||
|
defer func() {
|
||||||
|
lis.Close()
|
||||||
|
// conn2 shouldn't be closed until the client has
|
||||||
|
// observed a successful test.
|
||||||
|
if conn2 != nil {
|
||||||
|
conn2.Close()
|
||||||
|
}
|
||||||
|
}()
|
||||||
done := make(chan struct{})
|
done := make(chan struct{})
|
||||||
clientDone := make(chan struct{})
|
|
||||||
go func() { // Launch the server.
|
go func() { // Launch the server.
|
||||||
defer func() {
|
defer close(done)
|
||||||
if done != nil {
|
conn1, err := lis.Accept()
|
||||||
close(done)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
conn1, err := server.Accept()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("Error while accepting. Err: %v", err)
|
t.Errorf("Error while accepting. Err: %v", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
defer conn1.Close()
|
defer conn1.Close()
|
||||||
// Don't send server settings and make sure the connection is closed.
|
// Don't send server settings and the client should close the connection and try again.
|
||||||
time.Sleep(time.Millisecond * 1500) // Since the first backoff is for a second.
|
conn2, err = lis.Accept() // Accept a reconnection request from client.
|
||||||
conn1.SetDeadline(time.Now().Add(time.Second))
|
|
||||||
b := make([]byte, 24)
|
|
||||||
for {
|
|
||||||
// Make sure the connection was closed by client.
|
|
||||||
_, err = conn1.Read(b)
|
|
||||||
if err == nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if err != io.EOF {
|
|
||||||
t.Errorf(" conn1.Read(_) = _, %v, want _, io.EOF", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
break
|
|
||||||
}
|
|
||||||
|
|
||||||
conn2, err := server.Accept() // Accept a reconnection request from client.
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("Error while accepting. Err: %v", err)
|
t.Errorf("Error while accepting. Err: %v", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
defer conn2.Close()
|
|
||||||
framer := http2.NewFramer(conn2, conn2)
|
framer := http2.NewFramer(conn2, conn2)
|
||||||
if err := framer.WriteSettings(http2.Setting{}); err != nil {
|
if err = framer.WriteSettings(http2.Setting{}); err != nil {
|
||||||
t.Errorf("Error while writing settings. Err: %v", err)
|
t.Errorf("Error while writing settings. Err: %v", err)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
time.Sleep(time.Millisecond * 1500) // Since the first backoff is for a second.
|
b := make([]byte, 8)
|
||||||
conn2.SetDeadline(time.Now().Add(time.Millisecond * 500))
|
|
||||||
for {
|
for {
|
||||||
// Make sure the connection stays open and is closed
|
|
||||||
// only by connection timeout.
|
|
||||||
_, err = conn2.Read(b)
|
_, err = conn2.Read(b)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if nerr, ok := err.(net.Error); ok && nerr.Timeout() {
|
if atomic.LoadUint32(&over) == 1 {
|
||||||
|
// The connection stayed alive for the timer.
|
||||||
|
// Success.
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
t.Errorf("Unexpected error while reading. Err: %v, want timeout error", err)
|
t.Errorf("Unexpected error while reading. Err: %v, want timeout error", err)
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
close(done)
|
|
||||||
done = nil
|
|
||||||
<-clientDone
|
|
||||||
|
|
||||||
}()
|
}()
|
||||||
client, err := Dial(server.Addr().String(), WithInsecure())
|
client, err := Dial(lis.Addr().String(), WithInsecure())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Error while dialing. Err: %v", err)
|
t.Fatalf("Error while dialing. Err: %v", err)
|
||||||
}
|
}
|
||||||
<-done
|
time.Sleep(time.Second * 2) // Let things play out.
|
||||||
// TODO: The code from BEGIN to END should be delete once issue
|
atomic.StoreUint32(&over, 1)
|
||||||
// https://github.com/grpc/grpc-go/issues/1750 is fixed.
|
lis.Close()
|
||||||
// BEGIN
|
|
||||||
// Set underlying addrConns state to Shutdown so that no reconnect
|
|
||||||
// attempts take place and thereby resetting minConnectTimeout is
|
|
||||||
// race free.
|
|
||||||
client.mu.Lock()
|
|
||||||
addrConns := client.conns
|
|
||||||
client.mu.Unlock()
|
|
||||||
for ac := range addrConns {
|
|
||||||
ac.mu.Lock()
|
|
||||||
ac.state = connectivity.Shutdown
|
|
||||||
ac.mu.Unlock()
|
|
||||||
}
|
|
||||||
// END
|
|
||||||
client.Close()
|
client.Close()
|
||||||
close(clientDone)
|
<-done
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestBackoffWhenNoServerPrefaceReceived(t *testing.T) {
|
func TestBackoffWhenNoServerPrefaceReceived(t *testing.T) {
|
||||||
|
|||||||
Reference in New Issue
Block a user