2018-01-26 18:04:13 +00:00
|
|
|
package network
|
|
|
|
|
|
|
|
import (
|
2019-11-15 17:42:23 +00:00
|
|
|
"crypto/rand"
|
|
|
|
"encoding/binary"
|
2018-03-14 09:36:59 +00:00
|
|
|
"errors"
|
2018-01-26 18:04:13 +00:00
|
|
|
"fmt"
|
2019-09-13 17:38:34 +00:00
|
|
|
"net"
|
2019-10-29 17:51:17 +00:00
|
|
|
"strconv"
|
2018-03-14 09:36:59 +00:00
|
|
|
"sync"
|
2018-01-28 10:12:05 +00:00
|
|
|
"time"
|
2018-01-27 15:00:28 +00:00
|
|
|
|
2020-03-03 14:21:42 +00:00
|
|
|
"github.com/nspcc-dev/neo-go/pkg/consensus"
|
|
|
|
"github.com/nspcc-dev/neo-go/pkg/core"
|
|
|
|
"github.com/nspcc-dev/neo-go/pkg/core/block"
|
|
|
|
"github.com/nspcc-dev/neo-go/pkg/core/transaction"
|
|
|
|
"github.com/nspcc-dev/neo-go/pkg/network/payload"
|
|
|
|
"github.com/nspcc-dev/neo-go/pkg/util"
|
2019-11-15 10:32:40 +00:00
|
|
|
"go.uber.org/atomic"
|
2019-12-30 07:43:05 +00:00
|
|
|
"go.uber.org/zap"
|
2018-01-26 18:04:13 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
2019-10-22 14:56:03 +00:00
|
|
|
// peer numbers are arbitrary at the moment.
|
2019-11-06 12:17:20 +00:00
|
|
|
defaultMinPeers = 5
|
|
|
|
defaultAttemptConnPeers = 20
|
|
|
|
defaultMaxPeers = 100
|
|
|
|
maxBlockBatch = 200
|
|
|
|
maxAddrsToSend = 200
|
|
|
|
minPoolCount = 30
|
2018-01-26 18:04:13 +00:00
|
|
|
)
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
var (
|
2019-11-06 09:39:17 +00:00
|
|
|
errAlreadyConnected = errors.New("already connected")
|
2018-03-14 09:36:59 +00:00
|
|
|
errIdenticalID = errors.New("identical node id")
|
|
|
|
errInvalidHandshake = errors.New("invalid handshake")
|
|
|
|
errInvalidNetwork = errors.New("invalid network")
|
2019-11-06 12:17:20 +00:00
|
|
|
errMaxPeers = errors.New("max peers reached")
|
2018-03-14 09:36:59 +00:00
|
|
|
errServerShutdown = errors.New("server shutdown")
|
|
|
|
errInvalidInvType = errors.New("invalid inventory type")
|
2019-11-29 08:08:22 +00:00
|
|
|
errInvalidHashStart = errors.New("invalid requested HashStart")
|
2018-03-14 09:36:59 +00:00
|
|
|
)
|
2018-02-01 20:28:45 +00:00
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
type (
|
|
|
|
// Server represents the local Node in the network. Its transport could
|
|
|
|
// be of any kind.
|
|
|
|
Server struct {
|
2018-03-15 20:45:37 +00:00
|
|
|
// ServerConfig holds the Server configuration.
|
|
|
|
ServerConfig
|
2018-02-06 06:43:32 +00:00
|
|
|
|
2018-03-23 20:36:59 +00:00
|
|
|
// id also known as the nonce of the server.
|
2018-03-14 09:36:59 +00:00
|
|
|
id uint32
|
2018-01-26 18:04:13 +00:00
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
transport Transporter
|
|
|
|
discovery Discoverer
|
|
|
|
chain core.Blockchainer
|
2019-09-25 16:54:31 +00:00
|
|
|
bQueue *blockQueue
|
2019-11-08 15:40:21 +00:00
|
|
|
consensus consensus.Service
|
2018-01-26 18:04:13 +00:00
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
lock sync.RWMutex
|
|
|
|
peers map[Peer]bool
|
2018-01-26 20:39:34 +00:00
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
register chan Peer
|
|
|
|
unregister chan peerDrop
|
|
|
|
quit chan struct{}
|
2019-11-15 10:32:40 +00:00
|
|
|
|
2020-02-18 15:11:55 +00:00
|
|
|
transactions chan *transaction.Transaction
|
|
|
|
|
2020-02-14 17:46:05 +00:00
|
|
|
consensusStarted *atomic.Bool
|
2019-12-30 07:43:05 +00:00
|
|
|
|
|
|
|
log *zap.Logger
|
2018-03-14 09:36:59 +00:00
|
|
|
}
|
2018-03-10 12:04:06 +00:00
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
peerDrop struct {
|
|
|
|
peer Peer
|
|
|
|
reason error
|
2018-02-01 08:00:42 +00:00
|
|
|
}
|
2018-03-14 09:36:59 +00:00
|
|
|
)
|
|
|
|
|
2019-11-15 17:42:23 +00:00
|
|
|
func randomID() uint32 {
|
|
|
|
buf := make([]byte, 4)
|
|
|
|
_, _ = rand.Read(buf)
|
|
|
|
return binary.BigEndian.Uint32(buf)
|
|
|
|
}
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
// NewServer returns a new Server, initialized with the given configuration.
|
2020-01-22 08:17:51 +00:00
|
|
|
func NewServer(config ServerConfig, chain core.Blockchainer, log *zap.Logger) (*Server, error) {
|
2019-12-30 07:43:05 +00:00
|
|
|
if log == nil {
|
2020-01-22 08:17:51 +00:00
|
|
|
return nil, errors.New("logger is a required parameter")
|
2019-12-30 07:43:05 +00:00
|
|
|
}
|
|
|
|
|
2018-03-09 15:55:25 +00:00
|
|
|
s := &Server{
|
2020-02-14 17:46:05 +00:00
|
|
|
ServerConfig: config,
|
|
|
|
chain: chain,
|
|
|
|
id: randomID(),
|
|
|
|
quit: make(chan struct{}),
|
|
|
|
register: make(chan Peer),
|
|
|
|
unregister: make(chan peerDrop),
|
|
|
|
peers: make(map[Peer]bool),
|
|
|
|
consensusStarted: atomic.NewBool(false),
|
|
|
|
log: log,
|
2020-02-18 15:11:55 +00:00
|
|
|
transactions: make(chan *transaction.Transaction, 64),
|
2018-01-26 18:04:13 +00:00
|
|
|
}
|
2020-02-14 17:46:05 +00:00
|
|
|
s.bQueue = newBlockQueue(maxBlockBatch, chain, log, func(b *block.Block) {
|
2020-02-17 13:20:04 +00:00
|
|
|
if s.consensusStarted.Load() {
|
|
|
|
s.consensus.OnNewBlock()
|
|
|
|
} else {
|
|
|
|
s.tryStartConsensus()
|
|
|
|
}
|
2020-02-14 17:46:05 +00:00
|
|
|
s.relayBlock(b)
|
|
|
|
})
|
2018-01-26 18:04:13 +00:00
|
|
|
|
2019-11-15 10:32:40 +00:00
|
|
|
srv, err := consensus.NewService(consensus.Config{
|
2019-12-30 07:43:05 +00:00
|
|
|
Logger: log,
|
2019-11-29 09:27:15 +00:00
|
|
|
Broadcast: s.handleNewPayload,
|
|
|
|
RelayBlock: s.relayBlock,
|
|
|
|
Chain: chain,
|
|
|
|
RequestTx: s.requestTx,
|
|
|
|
Wallet: config.Wallet,
|
2020-01-13 14:57:40 +00:00
|
|
|
|
|
|
|
TimePerBlock: config.TimePerBlock,
|
2019-11-15 10:32:40 +00:00
|
|
|
})
|
|
|
|
if err != nil {
|
2020-01-22 08:17:51 +00:00
|
|
|
return nil, err
|
2019-11-15 10:32:40 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
s.consensus = srv
|
|
|
|
|
2020-01-13 12:22:21 +00:00
|
|
|
if s.MinPeers < 0 {
|
2019-12-30 07:43:05 +00:00
|
|
|
s.log.Info("bad MinPeers configured, using the default value",
|
|
|
|
zap.Int("configured", s.MinPeers),
|
|
|
|
zap.Int("actual", defaultMinPeers))
|
2019-11-01 10:29:54 +00:00
|
|
|
s.MinPeers = defaultMinPeers
|
|
|
|
}
|
|
|
|
|
2019-11-06 12:17:20 +00:00
|
|
|
if s.MaxPeers <= 0 {
|
2019-12-30 07:43:05 +00:00
|
|
|
s.log.Info("bad MaxPeers configured, using the default value",
|
|
|
|
zap.Int("configured", s.MaxPeers),
|
|
|
|
zap.Int("actual", defaultMaxPeers))
|
2019-11-06 12:17:20 +00:00
|
|
|
s.MaxPeers = defaultMaxPeers
|
|
|
|
}
|
|
|
|
|
|
|
|
if s.AttemptConnPeers <= 0 {
|
2019-12-30 07:43:05 +00:00
|
|
|
s.log.Info("bad AttemptConnPeers configured, using the default value",
|
|
|
|
zap.Int("configured", s.AttemptConnPeers),
|
|
|
|
zap.Int("actual", defaultAttemptConnPeers))
|
2019-11-06 12:17:20 +00:00
|
|
|
s.AttemptConnPeers = defaultAttemptConnPeers
|
|
|
|
}
|
|
|
|
|
2019-12-30 07:43:05 +00:00
|
|
|
s.transport = NewTCPTransport(s, fmt.Sprintf("%s:%d", config.Address, config.Port), s.log)
|
2018-03-14 09:36:59 +00:00
|
|
|
s.discovery = NewDefaultDiscovery(
|
|
|
|
s.DialTimeout,
|
|
|
|
s.transport,
|
|
|
|
)
|
2018-01-26 18:04:13 +00:00
|
|
|
|
2020-01-22 08:17:51 +00:00
|
|
|
return s, nil
|
2018-01-30 10:56:36 +00:00
|
|
|
}
|
|
|
|
|
2020-01-21 14:31:51 +00:00
|
|
|
// MkMsg creates a new message based on the server configured network and given
|
|
|
|
// parameters.
|
|
|
|
func (s *Server) MkMsg(cmd CommandType, p payload.Payload) *Message {
|
|
|
|
return NewMessage(s.Net, cmd, p)
|
|
|
|
}
|
|
|
|
|
2018-03-23 20:36:59 +00:00
|
|
|
// ID returns the servers ID.
|
|
|
|
func (s *Server) ID() uint32 {
|
|
|
|
return s.id
|
|
|
|
}
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
// Start will start the server and its underlying transport.
|
2018-03-23 20:36:59 +00:00
|
|
|
func (s *Server) Start(errChan chan error) {
|
2019-12-30 07:43:05 +00:00
|
|
|
s.log.Info("node started",
|
|
|
|
zap.Uint32("blockHeight", s.chain.BlockHeight()),
|
|
|
|
zap.Uint32("headerHeight", s.chain.HeaderHeight()))
|
2018-03-17 11:53:21 +00:00
|
|
|
|
2020-01-13 12:22:21 +00:00
|
|
|
s.tryStartConsensus()
|
|
|
|
|
2019-09-12 13:19:18 +00:00
|
|
|
s.discovery.BackFill(s.Seeds...)
|
2018-04-13 10:14:08 +00:00
|
|
|
|
2020-02-18 15:11:55 +00:00
|
|
|
go s.broadcastTxLoop()
|
2019-09-25 16:54:31 +00:00
|
|
|
go s.bQueue.run()
|
2018-03-14 09:36:59 +00:00
|
|
|
go s.transport.Accept()
|
2019-10-29 17:51:17 +00:00
|
|
|
setServerAndNodeVersions(s.UserAgent, strconv.FormatUint(uint64(s.id), 10))
|
2018-03-14 09:36:59 +00:00
|
|
|
s.run()
|
2018-01-31 19:11:08 +00:00
|
|
|
}
|
2018-01-30 10:56:36 +00:00
|
|
|
|
2018-03-23 20:36:59 +00:00
|
|
|
// Shutdown disconnects all peers and stops listening.
|
|
|
|
func (s *Server) Shutdown() {
|
2019-12-30 07:43:05 +00:00
|
|
|
s.log.Info("shutting down server", zap.Int("peers", s.PeerCount()))
|
2020-02-24 12:54:18 +00:00
|
|
|
s.transport.Close()
|
2020-02-24 12:39:31 +00:00
|
|
|
s.discovery.Close()
|
2020-02-24 12:54:18 +00:00
|
|
|
for p := range s.peers {
|
|
|
|
p.Disconnect(errServerShutdown)
|
|
|
|
}
|
2019-09-25 16:54:31 +00:00
|
|
|
s.bQueue.discard()
|
2018-03-23 20:36:59 +00:00
|
|
|
close(s.quit)
|
|
|
|
}
|
|
|
|
|
2018-04-09 16:58:09 +00:00
|
|
|
// UnconnectedPeers returns a list of peers that are in the discovery peer list
|
|
|
|
// but are not connected to the server.
|
2018-03-23 20:36:59 +00:00
|
|
|
func (s *Server) UnconnectedPeers() []string {
|
2020-01-10 12:16:14 +00:00
|
|
|
return s.discovery.UnconnectedPeers()
|
2018-03-23 20:36:59 +00:00
|
|
|
}
|
|
|
|
|
2018-04-09 16:58:09 +00:00
|
|
|
// BadPeers returns a list of peers the are flagged as "bad" peers.
|
2018-03-23 20:36:59 +00:00
|
|
|
func (s *Server) BadPeers() []string {
|
2020-01-10 12:16:14 +00:00
|
|
|
return s.discovery.BadPeers()
|
2018-03-23 20:36:59 +00:00
|
|
|
}
|
|
|
|
|
2020-01-10 12:13:29 +00:00
|
|
|
// ConnectedPeers returns a list of currently connected peers.
|
|
|
|
func (s *Server) ConnectedPeers() []string {
|
|
|
|
s.lock.RLock()
|
|
|
|
defer s.lock.RUnlock()
|
|
|
|
|
|
|
|
peers := make([]string, 0, len(s.peers))
|
|
|
|
for k := range s.peers {
|
|
|
|
peers = append(peers, k.PeerAddr().String())
|
|
|
|
}
|
|
|
|
|
|
|
|
return peers
|
|
|
|
}
|
|
|
|
|
2020-01-27 09:44:05 +00:00
|
|
|
// run is a goroutine that starts another goroutine to manage protocol specifics
|
|
|
|
// while itself dealing with peers management (handling connects/disconnects).
|
2018-03-14 09:36:59 +00:00
|
|
|
func (s *Server) run() {
|
2020-01-27 09:44:05 +00:00
|
|
|
go s.runProto()
|
2018-03-09 15:55:25 +00:00
|
|
|
for {
|
2019-11-06 12:17:20 +00:00
|
|
|
if s.PeerCount() < s.MinPeers {
|
|
|
|
s.discovery.RequestRemote(s.AttemptConnPeers)
|
2019-09-12 13:19:18 +00:00
|
|
|
}
|
2019-09-13 09:03:07 +00:00
|
|
|
if s.discovery.PoolCount() < minPoolCount {
|
2020-01-22 08:28:59 +00:00
|
|
|
s.broadcastHPMessage(s.MkMsg(CMDGetAddr, payload.NewNullPayload()))
|
2019-09-13 09:03:07 +00:00
|
|
|
}
|
2018-03-14 09:36:59 +00:00
|
|
|
select {
|
|
|
|
case <-s.quit:
|
|
|
|
return
|
|
|
|
case p := <-s.register:
|
2019-11-06 09:38:47 +00:00
|
|
|
s.lock.Lock()
|
2018-03-14 09:36:59 +00:00
|
|
|
s.peers[p] = true
|
2019-11-06 09:38:47 +00:00
|
|
|
s.lock.Unlock()
|
2019-11-06 12:17:20 +00:00
|
|
|
peerCount := s.PeerCount()
|
2020-01-28 16:39:12 +00:00
|
|
|
s.log.Info("new peer connected", zap.Stringer("addr", p.RemoteAddr()), zap.Int("peerCount", peerCount))
|
2019-11-06 12:17:20 +00:00
|
|
|
if peerCount > s.MaxPeers {
|
|
|
|
s.lock.RLock()
|
|
|
|
// Pick a random peer and drop connection to it.
|
|
|
|
for peer := range s.peers {
|
2020-02-24 09:39:46 +00:00
|
|
|
// It will send us unregister signal.
|
|
|
|
go peer.Disconnect(errMaxPeers)
|
2019-11-06 12:17:20 +00:00
|
|
|
break
|
|
|
|
}
|
|
|
|
s.lock.RUnlock()
|
|
|
|
}
|
2019-10-29 17:51:17 +00:00
|
|
|
updatePeersConnectedMetric(s.PeerCount())
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
case drop := <-s.unregister:
|
2019-11-06 09:38:47 +00:00
|
|
|
s.lock.Lock()
|
2019-09-13 12:36:53 +00:00
|
|
|
if s.peers[drop.peer] {
|
|
|
|
delete(s.peers, drop.peer)
|
2019-11-06 09:38:47 +00:00
|
|
|
s.lock.Unlock()
|
2019-12-30 07:43:05 +00:00
|
|
|
s.log.Warn("peer disconnected",
|
|
|
|
zap.Stringer("addr", drop.peer.RemoteAddr()),
|
|
|
|
zap.String("reason", drop.reason.Error()),
|
|
|
|
zap.Int("peerCount", s.PeerCount()))
|
2019-11-06 07:55:21 +00:00
|
|
|
addr := drop.peer.PeerAddr().String()
|
2019-11-27 08:56:56 +00:00
|
|
|
if drop.reason == errIdenticalID {
|
|
|
|
s.discovery.RegisterBadAddr(addr)
|
2020-01-28 16:34:34 +00:00
|
|
|
} else if drop.reason != errAlreadyConnected {
|
2019-11-27 08:56:56 +00:00
|
|
|
s.discovery.UnregisterConnectedAddr(addr)
|
|
|
|
s.discovery.BackFill(addr)
|
|
|
|
}
|
2019-10-29 17:51:17 +00:00
|
|
|
updatePeersConnectedMetric(s.PeerCount())
|
2019-11-06 09:38:47 +00:00
|
|
|
} else {
|
|
|
|
// else the peer is already gone, which can happen
|
|
|
|
// because we have two goroutines sending signals here
|
|
|
|
s.lock.Unlock()
|
2019-09-13 12:36:53 +00:00
|
|
|
}
|
2019-11-06 09:38:47 +00:00
|
|
|
|
2018-03-09 15:55:25 +00:00
|
|
|
}
|
2018-01-31 19:11:08 +00:00
|
|
|
}
|
2018-01-27 12:39:07 +00:00
|
|
|
}
|
|
|
|
|
2020-01-27 09:44:05 +00:00
|
|
|
// runProto is a goroutine that manages server-wide protocol events.
|
|
|
|
func (s *Server) runProto() {
|
|
|
|
pingTimer := time.NewTimer(s.PingInterval)
|
|
|
|
for {
|
|
|
|
prevHeight := s.chain.BlockHeight()
|
|
|
|
select {
|
|
|
|
case <-s.quit:
|
|
|
|
return
|
|
|
|
case <-pingTimer.C:
|
|
|
|
if s.chain.BlockHeight() == prevHeight {
|
|
|
|
// Get a copy of s.peers to avoid holding a lock while sending.
|
|
|
|
for peer := range s.Peers() {
|
|
|
|
_ = peer.SendPing(s.MkMsg(CMDPing, payload.NewPing(s.id, s.chain.HeaderHeight())))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
pingTimer.Reset(s.PingInterval)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-11-15 10:32:40 +00:00
|
|
|
func (s *Server) tryStartConsensus() {
|
2020-02-14 17:46:05 +00:00
|
|
|
if s.Wallet == nil || s.consensusStarted.Load() {
|
2019-11-15 10:32:40 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2020-02-14 17:46:05 +00:00
|
|
|
if s.IsInSync() {
|
|
|
|
s.log.Info("node reached synchronized state, starting consensus")
|
|
|
|
if s.consensusStarted.CAS(false, true) {
|
2019-11-15 10:32:40 +00:00
|
|
|
s.consensus.Start()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-23 20:36:59 +00:00
|
|
|
// Peers returns the current list of peers connected to
|
|
|
|
// the server.
|
|
|
|
func (s *Server) Peers() map[Peer]bool {
|
2019-11-15 10:32:40 +00:00
|
|
|
s.lock.RLock()
|
|
|
|
defer s.lock.RUnlock()
|
|
|
|
|
|
|
|
peers := make(map[Peer]bool, len(s.peers))
|
|
|
|
for k, v := range s.peers {
|
|
|
|
peers[k] = v
|
|
|
|
}
|
|
|
|
|
|
|
|
return peers
|
2018-03-23 20:36:59 +00:00
|
|
|
}
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
// PeerCount returns the number of current connected peers.
|
|
|
|
func (s *Server) PeerCount() int {
|
|
|
|
s.lock.RLock()
|
|
|
|
defer s.lock.RUnlock()
|
|
|
|
return len(s.peers)
|
2018-02-01 20:28:45 +00:00
|
|
|
}
|
|
|
|
|
2019-12-02 07:51:45 +00:00
|
|
|
// HandshakedPeersCount returns the number of connected peers
|
|
|
|
// which have already performed handshake.
|
|
|
|
func (s *Server) HandshakedPeersCount() int {
|
|
|
|
s.lock.RLock()
|
|
|
|
defer s.lock.RUnlock()
|
|
|
|
|
|
|
|
var count int
|
|
|
|
|
|
|
|
for p := range s.peers {
|
|
|
|
if p.Handshaked() {
|
|
|
|
count++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return count
|
|
|
|
}
|
2018-02-01 20:28:45 +00:00
|
|
|
|
2020-01-21 14:26:08 +00:00
|
|
|
// getVersionMsg returns current version message.
|
|
|
|
func (s *Server) getVersionMsg() *Message {
|
2018-03-14 09:36:59 +00:00
|
|
|
payload := payload.NewVersion(
|
|
|
|
s.id,
|
2019-11-05 12:22:07 +00:00
|
|
|
s.Port,
|
2018-03-14 09:36:59 +00:00
|
|
|
s.UserAgent,
|
|
|
|
s.chain.BlockHeight(),
|
|
|
|
s.Relay,
|
2018-03-09 15:55:25 +00:00
|
|
|
)
|
2020-01-21 14:31:51 +00:00
|
|
|
return s.MkMsg(CMDVersion, payload)
|
2018-03-14 09:36:59 +00:00
|
|
|
}
|
2018-03-09 15:55:25 +00:00
|
|
|
|
2020-02-14 17:46:05 +00:00
|
|
|
// IsInSync answers the question of whether the server is in sync with the
|
|
|
|
// network or not (at least how the server itself sees it). The server operates
|
|
|
|
// with the data that it has, the number of peers (that has to be more than
|
|
|
|
// minimum number) and height of these peers (our chain has to be not lower
|
|
|
|
// than 2/3 of our peers have). Ideally we would check for the highest of the
|
|
|
|
// peers, but the problem is that they can lie to us and send whatever height
|
|
|
|
// they want to.
|
|
|
|
func (s *Server) IsInSync() bool {
|
|
|
|
var peersNumber int
|
|
|
|
var notHigher int
|
|
|
|
|
|
|
|
if s.MinPeers == 0 {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
ourLastBlock := s.chain.BlockHeight()
|
|
|
|
|
|
|
|
s.lock.RLock()
|
|
|
|
for p := range s.peers {
|
|
|
|
if p.Handshaked() {
|
|
|
|
peersNumber++
|
|
|
|
if ourLastBlock >= p.LastBlockIndex() {
|
|
|
|
notHigher++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
s.lock.RUnlock()
|
|
|
|
|
|
|
|
// Checking bQueue would also be nice, but it can be filled with garbage
|
|
|
|
// easily at the moment.
|
|
|
|
return peersNumber >= s.MinPeers && (3*notHigher > 2*peersNumber) // && s.bQueue.length() == 0
|
|
|
|
}
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
// When a peer sends out his version we reply with verack after validating
|
|
|
|
// the version.
|
|
|
|
func (s *Server) handleVersionCmd(p Peer, version *payload.Version) error {
|
2019-09-13 12:43:22 +00:00
|
|
|
err := p.HandleVersion(version)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2018-03-14 09:36:59 +00:00
|
|
|
}
|
|
|
|
if s.id == version.Nonce {
|
|
|
|
return errIdenticalID
|
2018-01-28 13:59:32 +00:00
|
|
|
}
|
2019-11-06 09:39:17 +00:00
|
|
|
peerAddr := p.PeerAddr().String()
|
2020-01-28 16:10:13 +00:00
|
|
|
s.discovery.RegisterConnectedAddr(peerAddr)
|
2019-11-06 09:39:17 +00:00
|
|
|
s.lock.RLock()
|
|
|
|
for peer := range s.peers {
|
2020-01-28 10:54:09 +00:00
|
|
|
if p == peer {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
ver := peer.Version()
|
2019-11-06 09:39:17 +00:00
|
|
|
// Already connected, drop this connection.
|
2020-01-28 10:54:09 +00:00
|
|
|
if ver != nil && ver.Nonce == version.Nonce && peer.PeerAddr().String() == peerAddr {
|
2019-11-06 09:39:17 +00:00
|
|
|
s.lock.RUnlock()
|
|
|
|
return errAlreadyConnected
|
|
|
|
}
|
|
|
|
}
|
|
|
|
s.lock.RUnlock()
|
2020-01-21 14:31:51 +00:00
|
|
|
return p.SendVersionAck(s.MkMsg(CMDVerack, nil))
|
2018-01-28 13:59:32 +00:00
|
|
|
}
|
|
|
|
|
2019-10-22 14:56:03 +00:00
|
|
|
// handleHeadersCmd processes the headers received from its peer.
|
|
|
|
// If the headerHeight of the blockchain still smaller then the peer
|
2018-03-14 09:36:59 +00:00
|
|
|
// the server will request more headers.
|
|
|
|
// This method could best be called in a separate routine.
|
|
|
|
func (s *Server) handleHeadersCmd(p Peer, headers *payload.Headers) {
|
|
|
|
if err := s.chain.AddHeaders(headers.Hdrs...); err != nil {
|
2019-12-30 07:43:05 +00:00
|
|
|
s.log.Warn("failed processing headers", zap.Error(err))
|
2018-03-14 09:36:59 +00:00
|
|
|
return
|
|
|
|
}
|
|
|
|
// The peer will respond with a maximum of 2000 headers in one batch.
|
|
|
|
// We will ask one more batch here if needed. Eventually we will get synced
|
|
|
|
// due to the startProtocol routine that will ask headers every protoTick.
|
2020-01-17 10:17:19 +00:00
|
|
|
if s.chain.HeaderHeight() < p.LastBlockIndex() {
|
2018-03-14 09:36:59 +00:00
|
|
|
s.requestHeaders(p)
|
2018-01-28 10:12:05 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
// handleBlockCmd processes the received block received from its peer.
|
2020-01-14 12:32:07 +00:00
|
|
|
func (s *Server) handleBlockCmd(p Peer, block *block.Block) error {
|
2019-09-25 16:54:31 +00:00
|
|
|
return s.bQueue.putBlock(block)
|
2018-03-14 09:36:59 +00:00
|
|
|
}
|
2018-02-01 08:00:42 +00:00
|
|
|
|
2020-01-17 10:17:19 +00:00
|
|
|
// handlePing processes ping request.
|
|
|
|
func (s *Server) handlePing(p Peer, ping *payload.Ping) error {
|
2020-01-23 16:40:40 +00:00
|
|
|
return p.EnqueueP2PMessage(s.MkMsg(CMDPong, payload.NewPing(s.chain.BlockHeight(), s.id)))
|
2020-01-17 10:17:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// handlePing processes pong request.
|
|
|
|
func (s *Server) handlePong(p Peer, pong *payload.Ping) error {
|
2020-01-20 16:02:19 +00:00
|
|
|
err := p.HandlePong(pong)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2020-01-17 10:17:19 +00:00
|
|
|
}
|
|
|
|
if s.chain.HeaderHeight() < pong.LastBlockIndex {
|
|
|
|
return s.requestHeaders(p)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-10-22 14:56:03 +00:00
|
|
|
// handleInvCmd processes the received inventory.
|
2018-03-14 09:36:59 +00:00
|
|
|
func (s *Server) handleInvCmd(p Peer, inv *payload.Inventory) error {
|
2019-12-02 08:02:52 +00:00
|
|
|
reqHashes := make([]util.Uint256, 0)
|
|
|
|
var typExists = map[payload.InventoryType]func(util.Uint256) bool{
|
|
|
|
payload.TXType: s.chain.HasTransaction,
|
|
|
|
payload.BlockType: s.chain.HasBlock,
|
|
|
|
payload.ConsensusType: func(h util.Uint256) bool {
|
|
|
|
cp := s.consensus.GetPayload(h)
|
|
|
|
return cp != nil
|
|
|
|
},
|
|
|
|
}
|
|
|
|
if exists := typExists[inv.Type]; exists != nil {
|
|
|
|
for _, hash := range inv.Hashes {
|
|
|
|
if !exists(hash) {
|
|
|
|
reqHashes = append(reqHashes, hash)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(reqHashes) > 0 {
|
2020-01-21 14:31:51 +00:00
|
|
|
msg := s.MkMsg(CMDGetData, payload.NewInventory(inv.Type, reqHashes))
|
2020-01-16 18:16:31 +00:00
|
|
|
pkt, err := msg.Bytes()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if inv.Type == payload.ConsensusType {
|
|
|
|
return p.EnqueueHPPacket(pkt)
|
|
|
|
}
|
2020-01-23 16:40:40 +00:00
|
|
|
return p.EnqueueP2PPacket(pkt)
|
2019-12-02 08:02:52 +00:00
|
|
|
}
|
|
|
|
return nil
|
2018-03-09 15:55:25 +00:00
|
|
|
}
|
2018-02-01 08:00:42 +00:00
|
|
|
|
2019-10-24 07:18:30 +00:00
|
|
|
// handleInvCmd processes the received inventory.
|
|
|
|
func (s *Server) handleGetDataCmd(p Peer, inv *payload.Inventory) error {
|
2020-01-16 18:16:31 +00:00
|
|
|
for _, hash := range inv.Hashes {
|
|
|
|
var msg *Message
|
|
|
|
|
|
|
|
switch inv.Type {
|
|
|
|
case payload.TXType:
|
2019-10-24 07:18:30 +00:00
|
|
|
tx, _, err := s.chain.GetTransaction(hash)
|
|
|
|
if err == nil {
|
2020-01-21 14:31:51 +00:00
|
|
|
msg = s.MkMsg(CMDTX, tx)
|
2019-10-24 07:18:30 +00:00
|
|
|
}
|
2020-01-16 18:16:31 +00:00
|
|
|
case payload.BlockType:
|
2019-10-24 07:18:30 +00:00
|
|
|
b, err := s.chain.GetBlock(hash)
|
|
|
|
if err == nil {
|
2020-01-21 14:31:51 +00:00
|
|
|
msg = s.MkMsg(CMDBlock, b)
|
2019-10-24 07:18:30 +00:00
|
|
|
}
|
2020-01-16 18:16:31 +00:00
|
|
|
case payload.ConsensusType:
|
2019-11-08 15:40:21 +00:00
|
|
|
if cp := s.consensus.GetPayload(hash); cp != nil {
|
2020-01-21 14:31:51 +00:00
|
|
|
msg = s.MkMsg(CMDConsensus, cp)
|
2020-01-16 18:16:31 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
if msg != nil {
|
|
|
|
pkt, err := msg.Bytes()
|
2020-01-27 08:55:03 +00:00
|
|
|
if err == nil {
|
|
|
|
if inv.Type == payload.ConsensusType {
|
|
|
|
err = p.EnqueueHPPacket(pkt)
|
|
|
|
} else {
|
2020-01-23 16:40:40 +00:00
|
|
|
err = p.EnqueueP2PPacket(pkt)
|
2020-01-27 08:55:03 +00:00
|
|
|
}
|
|
|
|
}
|
2020-01-16 18:16:31 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-11-08 15:40:21 +00:00
|
|
|
}
|
2019-10-24 07:18:30 +00:00
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-12-25 16:40:18 +00:00
|
|
|
// handleGetBlocksCmd processes the getblocks request.
|
|
|
|
func (s *Server) handleGetBlocksCmd(p Peer, gb *payload.GetBlocks) error {
|
|
|
|
if len(gb.HashStart) < 1 {
|
|
|
|
return errInvalidHashStart
|
|
|
|
}
|
|
|
|
startHash := gb.HashStart[0]
|
|
|
|
if startHash.Equals(gb.HashStop) {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
start, err := s.chain.GetHeader(startHash)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
blockHashes := make([]util.Uint256, 0)
|
|
|
|
for i := start.Index + 1; i < start.Index+1+payload.MaxHashesCount; i++ {
|
|
|
|
hash := s.chain.GetHeaderHash(int(i))
|
|
|
|
if hash.Equals(util.Uint256{}) || hash.Equals(gb.HashStop) {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
blockHashes = append(blockHashes, hash)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(blockHashes) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
payload := payload.NewInventory(payload.BlockType, blockHashes)
|
2020-01-21 14:31:51 +00:00
|
|
|
msg := s.MkMsg(CMDInv, payload)
|
2020-01-23 16:40:40 +00:00
|
|
|
return p.EnqueueP2PMessage(msg)
|
2019-12-25 16:40:18 +00:00
|
|
|
}
|
|
|
|
|
2019-11-29 08:08:22 +00:00
|
|
|
// handleGetHeadersCmd processes the getheaders request.
|
|
|
|
func (s *Server) handleGetHeadersCmd(p Peer, gh *payload.GetBlocks) error {
|
|
|
|
if len(gh.HashStart) < 1 {
|
|
|
|
return errInvalidHashStart
|
|
|
|
}
|
|
|
|
startHash := gh.HashStart[0]
|
|
|
|
start, err := s.chain.GetHeader(startHash)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
resp := payload.Headers{}
|
2020-01-14 12:32:07 +00:00
|
|
|
resp.Hdrs = make([]*block.Header, 0, payload.MaxHeadersAllowed)
|
2019-11-29 08:08:22 +00:00
|
|
|
for i := start.Index + 1; i < start.Index+1+payload.MaxHeadersAllowed; i++ {
|
|
|
|
hash := s.chain.GetHeaderHash(int(i))
|
|
|
|
if hash.Equals(util.Uint256{}) || hash.Equals(gh.HashStop) {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
header, err := s.chain.GetHeader(hash)
|
|
|
|
if err != nil {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
resp.Hdrs = append(resp.Hdrs, header)
|
|
|
|
}
|
|
|
|
if len(resp.Hdrs) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
2020-01-21 14:31:51 +00:00
|
|
|
msg := s.MkMsg(CMDHeaders, &resp)
|
2020-01-23 16:40:40 +00:00
|
|
|
return p.EnqueueP2PMessage(msg)
|
2019-11-29 08:08:22 +00:00
|
|
|
}
|
|
|
|
|
2019-11-08 15:40:21 +00:00
|
|
|
// handleConsensusCmd processes received consensus payload.
|
|
|
|
// It never returns an error.
|
|
|
|
func (s *Server) handleConsensusCmd(cp *consensus.Payload) error {
|
|
|
|
s.consensus.OnPayload(cp)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-11-15 10:32:40 +00:00
|
|
|
// handleTxCmd processes received transaction.
|
|
|
|
// It never returns an error.
|
|
|
|
func (s *Server) handleTxCmd(tx *transaction.Transaction) error {
|
2019-11-29 08:09:54 +00:00
|
|
|
// It's OK for it to fail for various reasons like tx already existing
|
|
|
|
// in the pool.
|
2020-01-29 08:56:40 +00:00
|
|
|
if s.verifyAndPoolTX(tx) == RelaySucceed {
|
|
|
|
s.consensus.OnTransaction(tx)
|
2020-02-18 15:11:55 +00:00
|
|
|
s.broadcastTX(tx)
|
2020-01-29 08:56:40 +00:00
|
|
|
}
|
2019-11-15 10:32:40 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-09-13 09:03:07 +00:00
|
|
|
// handleAddrCmd will process received addresses.
|
|
|
|
func (s *Server) handleAddrCmd(p Peer, addrs *payload.AddressList) error {
|
|
|
|
for _, a := range addrs.Addrs {
|
|
|
|
s.discovery.BackFill(a.IPPortString())
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-09-13 17:38:34 +00:00
|
|
|
// handleGetAddrCmd sends to the peer some good addresses that we know of.
|
|
|
|
func (s *Server) handleGetAddrCmd(p Peer) error {
|
|
|
|
addrs := s.discovery.GoodPeers()
|
|
|
|
if len(addrs) > maxAddrsToSend {
|
|
|
|
addrs = addrs[:maxAddrsToSend]
|
|
|
|
}
|
|
|
|
alist := payload.NewAddressList(len(addrs))
|
|
|
|
ts := time.Now()
|
|
|
|
for i, addr := range addrs {
|
|
|
|
// we know it's a good address, so it can't fail
|
|
|
|
netaddr, _ := net.ResolveTCPAddr("tcp", addr)
|
|
|
|
alist.Addrs[i] = payload.NewAddressAndTime(netaddr, ts)
|
|
|
|
}
|
2020-01-23 16:40:40 +00:00
|
|
|
return p.EnqueueP2PMessage(s.MkMsg(CMDAddr, alist))
|
2019-09-13 17:38:34 +00:00
|
|
|
}
|
|
|
|
|
2019-10-22 14:56:03 +00:00
|
|
|
// requestHeaders sends a getheaders message to the peer.
|
2018-03-14 09:36:59 +00:00
|
|
|
// The peer will respond with headers op to a count of 2000.
|
2019-09-13 12:36:53 +00:00
|
|
|
func (s *Server) requestHeaders(p Peer) error {
|
2018-03-14 09:36:59 +00:00
|
|
|
start := []util.Uint256{s.chain.CurrentHeaderHash()}
|
|
|
|
payload := payload.NewGetBlocks(start, util.Uint256{})
|
2020-01-23 16:40:40 +00:00
|
|
|
return p.EnqueueP2PMessage(s.MkMsg(CMDGetHeaders, payload))
|
2018-03-09 15:55:25 +00:00
|
|
|
}
|
|
|
|
|
2019-10-22 14:56:03 +00:00
|
|
|
// requestBlocks sends a getdata message to the peer
|
2018-03-14 09:36:59 +00:00
|
|
|
// to sync up in blocks. A maximum of maxBlockBatch will
|
|
|
|
// send at once.
|
2019-09-13 12:36:53 +00:00
|
|
|
func (s *Server) requestBlocks(p Peer) error {
|
2018-03-14 09:36:59 +00:00
|
|
|
var (
|
2019-02-09 15:53:58 +00:00
|
|
|
hashes []util.Uint256
|
2018-03-14 09:36:59 +00:00
|
|
|
hashStart = s.chain.BlockHeight() + 1
|
|
|
|
headerHeight = s.chain.HeaderHeight()
|
|
|
|
)
|
2019-09-26 17:18:17 +00:00
|
|
|
for hashStart <= headerHeight && len(hashes) < maxBlockBatch {
|
2018-03-14 09:36:59 +00:00
|
|
|
hash := s.chain.GetHeaderHash(int(hashStart))
|
|
|
|
hashes = append(hashes, hash)
|
|
|
|
hashStart++
|
|
|
|
}
|
|
|
|
if len(hashes) > 0 {
|
|
|
|
payload := payload.NewInventory(payload.BlockType, hashes)
|
2020-01-23 16:40:40 +00:00
|
|
|
return p.EnqueueP2PMessage(s.MkMsg(CMDGetData, payload))
|
2020-01-20 16:17:44 +00:00
|
|
|
} else if s.chain.HeaderHeight() < p.LastBlockIndex() {
|
2019-09-13 12:36:53 +00:00
|
|
|
return s.requestHeaders(p)
|
2018-03-14 09:36:59 +00:00
|
|
|
}
|
2019-09-13 12:36:53 +00:00
|
|
|
return nil
|
2018-02-01 08:00:42 +00:00
|
|
|
}
|
|
|
|
|
2019-10-22 14:56:03 +00:00
|
|
|
// handleMessage processes the given message.
|
2018-04-13 10:14:08 +00:00
|
|
|
func (s *Server) handleMessage(peer Peer, msg *Message) error {
|
2020-01-28 13:40:38 +00:00
|
|
|
s.log.Debug("got msg",
|
|
|
|
zap.Stringer("addr", peer.RemoteAddr()),
|
|
|
|
zap.String("type", string(msg.CommandType())))
|
|
|
|
|
2018-03-14 09:36:59 +00:00
|
|
|
// Make sure both server and peer are operating on
|
|
|
|
// the same network.
|
|
|
|
if msg.Magic != s.Net {
|
|
|
|
return errInvalidNetwork
|
|
|
|
}
|
|
|
|
|
2019-09-13 12:43:22 +00:00
|
|
|
if peer.Handshaked() {
|
2019-10-24 10:10:10 +00:00
|
|
|
if inv, ok := msg.Payload.(*payload.Inventory); ok {
|
|
|
|
if !inv.Type.Valid() || len(inv.Hashes) == 0 {
|
|
|
|
return errInvalidInvType
|
|
|
|
}
|
|
|
|
}
|
2019-09-13 12:43:22 +00:00
|
|
|
switch msg.CommandType() {
|
|
|
|
case CMDAddr:
|
|
|
|
addrs := msg.Payload.(*payload.AddressList)
|
|
|
|
return s.handleAddrCmd(peer, addrs)
|
2019-09-13 17:38:34 +00:00
|
|
|
case CMDGetAddr:
|
|
|
|
// it has no payload
|
|
|
|
return s.handleGetAddrCmd(peer)
|
2019-12-25 16:40:18 +00:00
|
|
|
case CMDGetBlocks:
|
|
|
|
gb := msg.Payload.(*payload.GetBlocks)
|
|
|
|
return s.handleGetBlocksCmd(peer, gb)
|
2019-10-24 07:18:30 +00:00
|
|
|
case CMDGetData:
|
|
|
|
inv := msg.Payload.(*payload.Inventory)
|
|
|
|
return s.handleGetDataCmd(peer, inv)
|
2019-11-29 08:08:22 +00:00
|
|
|
case CMDGetHeaders:
|
|
|
|
gh := msg.Payload.(*payload.GetBlocks)
|
|
|
|
return s.handleGetHeadersCmd(peer, gh)
|
2019-09-13 12:43:22 +00:00
|
|
|
case CMDHeaders:
|
|
|
|
headers := msg.Payload.(*payload.Headers)
|
|
|
|
go s.handleHeadersCmd(peer, headers)
|
|
|
|
case CMDInv:
|
|
|
|
inventory := msg.Payload.(*payload.Inventory)
|
|
|
|
return s.handleInvCmd(peer, inventory)
|
|
|
|
case CMDBlock:
|
2020-01-14 12:32:07 +00:00
|
|
|
block := msg.Payload.(*block.Block)
|
2019-09-13 12:43:22 +00:00
|
|
|
return s.handleBlockCmd(peer, block)
|
2019-11-08 15:40:21 +00:00
|
|
|
case CMDConsensus:
|
|
|
|
cp := msg.Payload.(*consensus.Payload)
|
|
|
|
return s.handleConsensusCmd(cp)
|
2019-11-15 10:32:40 +00:00
|
|
|
case CMDTX:
|
|
|
|
tx := msg.Payload.(*transaction.Transaction)
|
|
|
|
return s.handleTxCmd(tx)
|
2020-01-17 10:17:19 +00:00
|
|
|
case CMDPing:
|
|
|
|
ping := msg.Payload.(*payload.Ping)
|
|
|
|
return s.handlePing(peer, ping)
|
|
|
|
case CMDPong:
|
|
|
|
pong := msg.Payload.(*payload.Ping)
|
|
|
|
return s.handlePong(peer, pong)
|
2019-09-13 12:43:22 +00:00
|
|
|
case CMDVersion, CMDVerack:
|
|
|
|
return fmt.Errorf("received '%s' after the handshake", msg.CommandType())
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
switch msg.CommandType() {
|
|
|
|
case CMDVersion:
|
|
|
|
version := msg.Payload.(*payload.Version)
|
|
|
|
return s.handleVersionCmd(peer, version)
|
|
|
|
case CMDVerack:
|
|
|
|
err := peer.HandleVersionAck()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2020-01-15 14:03:42 +00:00
|
|
|
go peer.StartProtocol()
|
2019-11-15 10:32:40 +00:00
|
|
|
|
|
|
|
s.tryStartConsensus()
|
2019-09-13 12:43:22 +00:00
|
|
|
default:
|
|
|
|
return fmt.Errorf("received '%s' during handshake", msg.CommandType())
|
2018-03-14 09:36:59 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
2018-01-26 18:04:13 +00:00
|
|
|
}
|
Implement rpc server method: sendrawtransaction (#174)
* Added new config attributes: 'SecondsPerBlock','LowPriorityThreshold'
* Added new files:
* Added new method: CompareTo
* Fixed empty Slice case
* Added new methods: LessThan, GreaterThan, Equal, CompareTo
* Added new method: InputIntersection
* Added MaxTransactionSize, GroupOutputByAssetID
* Added ned method: ScriptHash
* Added new method: IsDoubleSpend
* Refactor blockchainer, Added Feer interface, Verify and GetMemPool method
* 1) Added MemPool
2) Added new methods to satisfy the blockchainer interface: IsLowPriority, Verify, GetMemPool
* Added new methods: RelayTxn, RelayDirectly
* Fixed tests
* Implemented RPC server method sendrawtransaction
* Refactor getrawtransaction, sendrawtransaction in separate methods
* Moved 'secondsPerBlock' to config file
* Implemented Kim suggestions:
1) Fixed data race issues
2) refactor Verify method
3) Get rid of unused InputIntersection method due to refactoring Verify method
4) Fixed bug in https://github.com/CityOfZion/neo-go/pull/174#discussion_r264108135
5) minor simplications of the code
* Fixed minor issues related to
1) space
2) getter methods do not need pointer on the receiver
3) error message
4) refactoring CompareTo method in uint256.go
* Fixed small issues
* Use sync.RWMutex instead of sync.Mutex
* Refined (R)Lock/(R)Unlock
* return error instead of bool in Verify methods
2019-03-20 12:30:05 +00:00
|
|
|
|
2019-11-15 10:32:40 +00:00
|
|
|
func (s *Server) handleNewPayload(p *consensus.Payload) {
|
2020-01-22 08:01:13 +00:00
|
|
|
msg := s.MkMsg(CMDInv, payload.NewInventory(payload.ConsensusType, []util.Uint256{p.Hash()}))
|
|
|
|
// It's high priority because it directly affects consensus process,
|
|
|
|
// even though it's just an inv.
|
|
|
|
s.broadcastHPMessage(msg)
|
2019-11-15 10:32:40 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) requestTx(hashes ...util.Uint256) {
|
|
|
|
if len(hashes) == 0 {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2020-01-22 08:01:13 +00:00
|
|
|
msg := s.MkMsg(CMDGetData, payload.NewInventory(payload.TXType, hashes))
|
|
|
|
// It's high priority because it directly affects consensus process,
|
|
|
|
// even though it's getdata.
|
|
|
|
s.broadcastHPMessage(msg)
|
2019-11-15 10:32:40 +00:00
|
|
|
}
|
|
|
|
|
2020-01-22 08:01:13 +00:00
|
|
|
// iteratePeersWithSendMsg sends given message to all peers using two functions
|
|
|
|
// passed, one is to send the message and the other is to filtrate peers (the
|
|
|
|
// peer is considered invalid if it returns false).
|
|
|
|
func (s *Server) iteratePeersWithSendMsg(msg *Message, send func(Peer, []byte) error, peerOK func(Peer) bool) {
|
|
|
|
pkt, err := msg.Bytes()
|
|
|
|
if err != nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// Get a copy of s.peers to avoid holding a lock while sending.
|
2019-11-15 10:32:40 +00:00
|
|
|
for peer := range s.Peers() {
|
2020-01-22 08:01:13 +00:00
|
|
|
if peerOK != nil && !peerOK(peer) {
|
2019-11-15 10:32:40 +00:00
|
|
|
continue
|
|
|
|
}
|
2020-01-15 10:19:31 +00:00
|
|
|
// Who cares about these messages anyway?
|
2020-01-22 08:01:13 +00:00
|
|
|
_ = send(peer, pkt)
|
2019-11-15 10:32:40 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-01-22 08:01:13 +00:00
|
|
|
// broadcastMessage sends the message to all available peers.
|
|
|
|
func (s *Server) broadcastMessage(msg *Message) {
|
|
|
|
s.iteratePeersWithSendMsg(msg, Peer.EnqueuePacket, nil)
|
|
|
|
}
|
|
|
|
|
|
|
|
// broadcastHPMessage sends the high-priority message to all available peers.
|
|
|
|
func (s *Server) broadcastHPMessage(msg *Message) {
|
|
|
|
s.iteratePeersWithSendMsg(msg, Peer.EnqueueHPPacket, nil)
|
|
|
|
}
|
|
|
|
|
2019-11-29 09:27:15 +00:00
|
|
|
// relayBlock tells all the other connected nodes about the given block.
|
2020-01-14 12:32:07 +00:00
|
|
|
func (s *Server) relayBlock(b *block.Block) {
|
2020-01-22 08:01:13 +00:00
|
|
|
msg := s.MkMsg(CMDInv, payload.NewInventory(payload.BlockType, []util.Uint256{b.Hash()}))
|
2020-02-04 16:32:29 +00:00
|
|
|
// Filter out nodes that are more current (avoid spamming the network
|
|
|
|
// during initial sync).
|
|
|
|
s.iteratePeersWithSendMsg(msg, Peer.EnqueuePacket, func(p Peer) bool {
|
|
|
|
return p.Handshaked() && p.LastBlockIndex() < b.Index
|
|
|
|
})
|
2019-11-29 09:27:15 +00:00
|
|
|
}
|
|
|
|
|
2020-01-29 08:56:40 +00:00
|
|
|
// verifyAndPoolTX verifies the TX and adds it to the local mempool.
|
|
|
|
func (s *Server) verifyAndPoolTX(t *transaction.Transaction) RelayReason {
|
Implement rpc server method: sendrawtransaction (#174)
* Added new config attributes: 'SecondsPerBlock','LowPriorityThreshold'
* Added new files:
* Added new method: CompareTo
* Fixed empty Slice case
* Added new methods: LessThan, GreaterThan, Equal, CompareTo
* Added new method: InputIntersection
* Added MaxTransactionSize, GroupOutputByAssetID
* Added ned method: ScriptHash
* Added new method: IsDoubleSpend
* Refactor blockchainer, Added Feer interface, Verify and GetMemPool method
* 1) Added MemPool
2) Added new methods to satisfy the blockchainer interface: IsLowPriority, Verify, GetMemPool
* Added new methods: RelayTxn, RelayDirectly
* Fixed tests
* Implemented RPC server method sendrawtransaction
* Refactor getrawtransaction, sendrawtransaction in separate methods
* Moved 'secondsPerBlock' to config file
* Implemented Kim suggestions:
1) Fixed data race issues
2) refactor Verify method
3) Get rid of unused InputIntersection method due to refactoring Verify method
4) Fixed bug in https://github.com/CityOfZion/neo-go/pull/174#discussion_r264108135
5) minor simplications of the code
* Fixed minor issues related to
1) space
2) getter methods do not need pointer on the receiver
3) error message
4) refactoring CompareTo method in uint256.go
* Fixed small issues
* Use sync.RWMutex instead of sync.Mutex
* Refined (R)Lock/(R)Unlock
* return error instead of bool in Verify methods
2019-03-20 12:30:05 +00:00
|
|
|
if t.Type == transaction.MinerType {
|
|
|
|
return RelayInvalid
|
|
|
|
}
|
2020-02-04 15:43:21 +00:00
|
|
|
if err := s.chain.PoolTx(t); err != nil {
|
|
|
|
switch err {
|
|
|
|
case core.ErrAlreadyExists:
|
|
|
|
return RelayAlreadyExists
|
|
|
|
case core.ErrOOM:
|
|
|
|
return RelayOutOfMemory
|
2020-02-18 17:16:38 +00:00
|
|
|
case core.ErrPolicy:
|
|
|
|
return RelayPolicyFail
|
2020-02-04 15:43:21 +00:00
|
|
|
default:
|
|
|
|
return RelayInvalid
|
|
|
|
}
|
Implement rpc server method: sendrawtransaction (#174)
* Added new config attributes: 'SecondsPerBlock','LowPriorityThreshold'
* Added new files:
* Added new method: CompareTo
* Fixed empty Slice case
* Added new methods: LessThan, GreaterThan, Equal, CompareTo
* Added new method: InputIntersection
* Added MaxTransactionSize, GroupOutputByAssetID
* Added ned method: ScriptHash
* Added new method: IsDoubleSpend
* Refactor blockchainer, Added Feer interface, Verify and GetMemPool method
* 1) Added MemPool
2) Added new methods to satisfy the blockchainer interface: IsLowPriority, Verify, GetMemPool
* Added new methods: RelayTxn, RelayDirectly
* Fixed tests
* Implemented RPC server method sendrawtransaction
* Refactor getrawtransaction, sendrawtransaction in separate methods
* Moved 'secondsPerBlock' to config file
* Implemented Kim suggestions:
1) Fixed data race issues
2) refactor Verify method
3) Get rid of unused InputIntersection method due to refactoring Verify method
4) Fixed bug in https://github.com/CityOfZion/neo-go/pull/174#discussion_r264108135
5) minor simplications of the code
* Fixed minor issues related to
1) space
2) getter methods do not need pointer on the receiver
3) error message
4) refactoring CompareTo method in uint256.go
* Fixed small issues
* Use sync.RWMutex instead of sync.Mutex
* Refined (R)Lock/(R)Unlock
* return error instead of bool in Verify methods
2019-03-20 12:30:05 +00:00
|
|
|
}
|
2020-01-29 08:56:40 +00:00
|
|
|
return RelaySucceed
|
|
|
|
}
|
|
|
|
|
|
|
|
// RelayTxn a new transaction to the local node and the connected peers.
|
|
|
|
// Reference: the method OnRelay in C#: https://github.com/neo-project/neo/blob/master/neo/Network/P2P/LocalNode.cs#L159
|
|
|
|
func (s *Server) RelayTxn(t *transaction.Transaction) RelayReason {
|
|
|
|
ret := s.verifyAndPoolTX(t)
|
|
|
|
if ret == RelaySucceed {
|
|
|
|
s.broadcastTX(t)
|
|
|
|
}
|
|
|
|
return ret
|
|
|
|
}
|
Implement rpc server method: sendrawtransaction (#174)
* Added new config attributes: 'SecondsPerBlock','LowPriorityThreshold'
* Added new files:
* Added new method: CompareTo
* Fixed empty Slice case
* Added new methods: LessThan, GreaterThan, Equal, CompareTo
* Added new method: InputIntersection
* Added MaxTransactionSize, GroupOutputByAssetID
* Added ned method: ScriptHash
* Added new method: IsDoubleSpend
* Refactor blockchainer, Added Feer interface, Verify and GetMemPool method
* 1) Added MemPool
2) Added new methods to satisfy the blockchainer interface: IsLowPriority, Verify, GetMemPool
* Added new methods: RelayTxn, RelayDirectly
* Fixed tests
* Implemented RPC server method sendrawtransaction
* Refactor getrawtransaction, sendrawtransaction in separate methods
* Moved 'secondsPerBlock' to config file
* Implemented Kim suggestions:
1) Fixed data race issues
2) refactor Verify method
3) Get rid of unused InputIntersection method due to refactoring Verify method
4) Fixed bug in https://github.com/CityOfZion/neo-go/pull/174#discussion_r264108135
5) minor simplications of the code
* Fixed minor issues related to
1) space
2) getter methods do not need pointer on the receiver
3) error message
4) refactoring CompareTo method in uint256.go
* Fixed small issues
* Use sync.RWMutex instead of sync.Mutex
* Refined (R)Lock/(R)Unlock
* return error instead of bool in Verify methods
2019-03-20 12:30:05 +00:00
|
|
|
|
2020-01-29 08:56:40 +00:00
|
|
|
// broadcastTX broadcasts an inventory message about new transaction.
|
|
|
|
func (s *Server) broadcastTX(t *transaction.Transaction) {
|
2020-02-18 15:11:55 +00:00
|
|
|
select {
|
|
|
|
case s.transactions <- t:
|
|
|
|
case <-s.quit:
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) broadcastTxHashes(hs []util.Uint256) {
|
|
|
|
msg := s.MkMsg(CMDInv, payload.NewInventory(payload.TXType, hs))
|
2020-01-22 08:01:13 +00:00
|
|
|
|
|
|
|
// We need to filter out non-relaying nodes, so plain broadcast
|
|
|
|
// functions don't fit here.
|
|
|
|
s.iteratePeersWithSendMsg(msg, Peer.EnqueuePacket, func(p Peer) bool {
|
|
|
|
return p.Handshaked() && p.Version().Relay
|
|
|
|
})
|
Implement rpc server method: sendrawtransaction (#174)
* Added new config attributes: 'SecondsPerBlock','LowPriorityThreshold'
* Added new files:
* Added new method: CompareTo
* Fixed empty Slice case
* Added new methods: LessThan, GreaterThan, Equal, CompareTo
* Added new method: InputIntersection
* Added MaxTransactionSize, GroupOutputByAssetID
* Added ned method: ScriptHash
* Added new method: IsDoubleSpend
* Refactor blockchainer, Added Feer interface, Verify and GetMemPool method
* 1) Added MemPool
2) Added new methods to satisfy the blockchainer interface: IsLowPriority, Verify, GetMemPool
* Added new methods: RelayTxn, RelayDirectly
* Fixed tests
* Implemented RPC server method sendrawtransaction
* Refactor getrawtransaction, sendrawtransaction in separate methods
* Moved 'secondsPerBlock' to config file
* Implemented Kim suggestions:
1) Fixed data race issues
2) refactor Verify method
3) Get rid of unused InputIntersection method due to refactoring Verify method
4) Fixed bug in https://github.com/CityOfZion/neo-go/pull/174#discussion_r264108135
5) minor simplications of the code
* Fixed minor issues related to
1) space
2) getter methods do not need pointer on the receiver
3) error message
4) refactoring CompareTo method in uint256.go
* Fixed small issues
* Use sync.RWMutex instead of sync.Mutex
* Refined (R)Lock/(R)Unlock
* return error instead of bool in Verify methods
2019-03-20 12:30:05 +00:00
|
|
|
}
|
2020-02-18 15:11:55 +00:00
|
|
|
|
|
|
|
// broadcastTxLoop is a loop for batching and sending
|
|
|
|
// transactions hashes in an INV payload.
|
|
|
|
func (s *Server) broadcastTxLoop() {
|
|
|
|
const (
|
|
|
|
batchTime = time.Millisecond * 50
|
|
|
|
batchSize = 32
|
|
|
|
)
|
|
|
|
|
|
|
|
txs := make([]util.Uint256, 0, batchSize)
|
|
|
|
var timer *time.Timer
|
|
|
|
|
|
|
|
timerCh := func() <-chan time.Time {
|
|
|
|
if timer == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
return timer.C
|
|
|
|
}
|
|
|
|
|
|
|
|
broadcast := func() {
|
|
|
|
s.broadcastTxHashes(txs)
|
|
|
|
txs = txs[:0]
|
|
|
|
if timer != nil {
|
|
|
|
timer.Stop()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-s.quit:
|
|
|
|
loop:
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-s.transactions:
|
|
|
|
default:
|
|
|
|
break loop
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return
|
|
|
|
case <-timerCh():
|
|
|
|
if len(txs) > 0 {
|
|
|
|
broadcast()
|
|
|
|
}
|
|
|
|
case tx := <-s.transactions:
|
|
|
|
if len(txs) == 0 {
|
|
|
|
timer = time.NewTimer(batchTime)
|
|
|
|
}
|
|
|
|
|
|
|
|
txs = append(txs, tx.Hash())
|
|
|
|
if len(txs) == batchSize {
|
|
|
|
broadcast()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|