mirror of
https://github.com/fluencelabs/tendermint
synced 2025-04-24 22:32:15 +00:00
* node: decrease retry conn timeout in test Should fix #3256 The retry timeout was set to the default, which is the same as the accept timeout, so it's no wonder this would fail. Here we decrease the retry timeout so we can try many times before the accept timeout. * p2p: increase handshake timeout in test This fails sometimes, presumably because the handshake timeout is so low (only 50ms). So increase it to 1s. Should fix #3187 * privval: fix race with ping. closes #3237 Pings happen in a go-routine and can happen concurrently with other messages. Since we use a request/response protocol, we expect to send a request and get back the corresponding response. But with pings happening concurrently, this assumption could be violated. We were using a mutex, but only a RWMutex, where the RLock was being held for sending messages - this was to allow the underlying connection to be replaced if it fails. Turns out we actually need to use a full lock (not just a read lock) to prevent multiple requests from happening concurrently. * node: fix test name. DelayedStop -> DelayedStart * autofile: Wait() method In the TestWALTruncate in consensus/wal_test.go we remove the WAL directory at the end of the test. However the wal.Stop() does not properly wait for the autofile group to finish shutting down. Hence it was possible that the group's go-routine is still running when the cleanup happens, which causes a panic since the directory disappeared. Here we add a Wait() method to properly wait until the go-routine exits so we can safely clean up. This fixes #2852.
272 lines
6.4 KiB
Go
272 lines
6.4 KiB
Go
package p2p
|
|
|
|
import (
|
|
"fmt"
|
|
"net"
|
|
"time"
|
|
|
|
"github.com/tendermint/tendermint/crypto"
|
|
"github.com/tendermint/tendermint/crypto/ed25519"
|
|
cmn "github.com/tendermint/tendermint/libs/common"
|
|
"github.com/tendermint/tendermint/libs/log"
|
|
|
|
"github.com/tendermint/tendermint/config"
|
|
"github.com/tendermint/tendermint/p2p/conn"
|
|
)
|
|
|
|
const testCh = 0x01
|
|
|
|
//------------------------------------------------
|
|
|
|
type mockNodeInfo struct {
|
|
addr *NetAddress
|
|
}
|
|
|
|
func (ni mockNodeInfo) ID() ID { return ni.addr.ID }
|
|
func (ni mockNodeInfo) NetAddress() *NetAddress { return ni.addr }
|
|
func (ni mockNodeInfo) Validate() error { return nil }
|
|
func (ni mockNodeInfo) CompatibleWith(other NodeInfo) error { return nil }
|
|
|
|
func AddPeerToSwitch(sw *Switch, peer Peer) {
|
|
sw.peers.Add(peer)
|
|
}
|
|
|
|
func CreateRandomPeer(outbound bool) *peer {
|
|
addr, netAddr := CreateRoutableAddr()
|
|
p := &peer{
|
|
peerConn: peerConn{
|
|
outbound: outbound,
|
|
},
|
|
nodeInfo: mockNodeInfo{netAddr},
|
|
mconn: &conn.MConnection{},
|
|
metrics: NopMetrics(),
|
|
}
|
|
p.SetLogger(log.TestingLogger().With("peer", addr))
|
|
return p
|
|
}
|
|
|
|
func CreateRoutableAddr() (addr string, netAddr *NetAddress) {
|
|
for {
|
|
var err error
|
|
addr = fmt.Sprintf("%X@%v.%v.%v.%v:26656", cmn.RandBytes(20), cmn.RandInt()%256, cmn.RandInt()%256, cmn.RandInt()%256, cmn.RandInt()%256)
|
|
netAddr, err = NewNetAddressString(addr)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
if netAddr.Routable() {
|
|
break
|
|
}
|
|
}
|
|
return
|
|
}
|
|
|
|
//------------------------------------------------------------------
|
|
// Connects switches via arbitrary net.Conn. Used for testing.
|
|
|
|
const TEST_HOST = "localhost"
|
|
|
|
// MakeConnectedSwitches returns n switches, connected according to the connect func.
|
|
// If connect==Connect2Switches, the switches will be fully connected.
|
|
// initSwitch defines how the i'th switch should be initialized (ie. with what reactors).
|
|
// NOTE: panics if any switch fails to start.
|
|
func MakeConnectedSwitches(cfg *config.P2PConfig, n int, initSwitch func(int, *Switch) *Switch, connect func([]*Switch, int, int)) []*Switch {
|
|
switches := make([]*Switch, n)
|
|
for i := 0; i < n; i++ {
|
|
switches[i] = MakeSwitch(cfg, i, TEST_HOST, "123.123.123", initSwitch)
|
|
}
|
|
|
|
if err := StartSwitches(switches); err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
for i := 0; i < n; i++ {
|
|
for j := i + 1; j < n; j++ {
|
|
connect(switches, i, j)
|
|
}
|
|
}
|
|
|
|
return switches
|
|
}
|
|
|
|
// Connect2Switches will connect switches i and j via net.Pipe().
|
|
// Blocks until a connection is established.
|
|
// NOTE: caller ensures i and j are within bounds.
|
|
func Connect2Switches(switches []*Switch, i, j int) {
|
|
switchI := switches[i]
|
|
switchJ := switches[j]
|
|
|
|
c1, c2 := conn.NetPipe()
|
|
|
|
doneCh := make(chan struct{})
|
|
go func() {
|
|
err := switchI.addPeerWithConnection(c1)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
doneCh <- struct{}{}
|
|
}()
|
|
go func() {
|
|
err := switchJ.addPeerWithConnection(c2)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
doneCh <- struct{}{}
|
|
}()
|
|
<-doneCh
|
|
<-doneCh
|
|
}
|
|
|
|
func (sw *Switch) addPeerWithConnection(conn net.Conn) error {
|
|
pc, err := testInboundPeerConn(conn, sw.config, sw.nodeKey.PrivKey)
|
|
if err != nil {
|
|
if err := conn.Close(); err != nil {
|
|
sw.Logger.Error("Error closing connection", "err", err)
|
|
}
|
|
return err
|
|
}
|
|
|
|
ni, err := handshake(conn, time.Second, sw.nodeInfo)
|
|
if err != nil {
|
|
if err := conn.Close(); err != nil {
|
|
sw.Logger.Error("Error closing connection", "err", err)
|
|
}
|
|
return err
|
|
}
|
|
|
|
p := newPeer(
|
|
pc,
|
|
MConnConfig(sw.config),
|
|
ni,
|
|
sw.reactorsByCh,
|
|
sw.chDescs,
|
|
sw.StopPeerForError,
|
|
)
|
|
|
|
if err = sw.addPeer(p); err != nil {
|
|
pc.CloseConn()
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// StartSwitches calls sw.Start() for each given switch.
|
|
// It returns the first encountered error.
|
|
func StartSwitches(switches []*Switch) error {
|
|
for _, s := range switches {
|
|
err := s.Start() // start switch and reactors
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func MakeSwitch(
|
|
cfg *config.P2PConfig,
|
|
i int,
|
|
network, version string,
|
|
initSwitch func(int, *Switch) *Switch,
|
|
opts ...SwitchOption,
|
|
) *Switch {
|
|
|
|
nodeKey := NodeKey{
|
|
PrivKey: ed25519.GenPrivKey(),
|
|
}
|
|
nodeInfo := testNodeInfo(nodeKey.ID(), fmt.Sprintf("node%d", i))
|
|
|
|
t := NewMultiplexTransport(nodeInfo, nodeKey, MConnConfig(cfg))
|
|
|
|
addr := nodeInfo.NetAddress()
|
|
if err := t.Listen(*addr); err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
// TODO: let the config be passed in?
|
|
sw := initSwitch(i, NewSwitch(cfg, t, opts...))
|
|
sw.SetLogger(log.TestingLogger().With("switch", i))
|
|
sw.SetNodeKey(&nodeKey)
|
|
|
|
ni := nodeInfo.(DefaultNodeInfo)
|
|
for ch := range sw.reactorsByCh {
|
|
ni.Channels = append(ni.Channels, ch)
|
|
}
|
|
nodeInfo = ni
|
|
|
|
// TODO: We need to setup reactors ahead of time so the NodeInfo is properly
|
|
// populated and we don't have to do those awkward overrides and setters.
|
|
t.nodeInfo = nodeInfo
|
|
sw.SetNodeInfo(nodeInfo)
|
|
|
|
return sw
|
|
}
|
|
|
|
func testInboundPeerConn(
|
|
conn net.Conn,
|
|
config *config.P2PConfig,
|
|
ourNodePrivKey crypto.PrivKey,
|
|
) (peerConn, error) {
|
|
return testPeerConn(conn, config, false, false, ourNodePrivKey, nil)
|
|
}
|
|
|
|
func testPeerConn(
|
|
rawConn net.Conn,
|
|
cfg *config.P2PConfig,
|
|
outbound, persistent bool,
|
|
ourNodePrivKey crypto.PrivKey,
|
|
originalAddr *NetAddress,
|
|
) (pc peerConn, err error) {
|
|
conn := rawConn
|
|
|
|
// Fuzz connection
|
|
if cfg.TestFuzz {
|
|
// so we have time to do peer handshakes and get set up
|
|
conn = FuzzConnAfterFromConfig(conn, 10*time.Second, cfg.TestFuzzConfig)
|
|
}
|
|
|
|
// Encrypt connection
|
|
conn, err = upgradeSecretConn(conn, cfg.HandshakeTimeout, ourNodePrivKey)
|
|
if err != nil {
|
|
return pc, cmn.ErrorWrap(err, "Error creating peer")
|
|
}
|
|
|
|
// Only the information we already have
|
|
return peerConn{
|
|
outbound: outbound,
|
|
persistent: persistent,
|
|
conn: conn,
|
|
originalAddr: originalAddr,
|
|
}, nil
|
|
}
|
|
|
|
//----------------------------------------------------------------
|
|
// rand node info
|
|
|
|
func testNodeInfo(id ID, name string) NodeInfo {
|
|
return testNodeInfoWithNetwork(id, name, "testing")
|
|
}
|
|
|
|
func testNodeInfoWithNetwork(id ID, name, network string) NodeInfo {
|
|
return DefaultNodeInfo{
|
|
ProtocolVersion: defaultProtocolVersion,
|
|
ID_: id,
|
|
ListenAddr: fmt.Sprintf("127.0.0.1:%d", getFreePort()),
|
|
Network: network,
|
|
Version: "1.2.3-rc0-deadbeef",
|
|
Channels: []byte{testCh},
|
|
Moniker: name,
|
|
Other: DefaultNodeInfoOther{
|
|
TxIndex: "on",
|
|
RPCAddress: fmt.Sprintf("127.0.0.1:%d", getFreePort()),
|
|
},
|
|
}
|
|
}
|
|
|
|
func getFreePort() int {
|
|
port, err := cmn.GetFreePort()
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
return port
|
|
}
|