package eth import ( "bytes" "fmt" "math" "math/big" "math/rand" "sort" "sync" "time" "github.com/ethereum/go-ethereum/core/types" "github.com/ethereum/go-ethereum/ethutil" ethlogger "github.com/ethereum/go-ethereum/logger" "github.com/ethereum/go-ethereum/pow" ) var poolLogger = ethlogger.NewLogger("Blockpool") const ( blockHashesBatchSize = 256 blockBatchSize = 64 blocksRequestInterval = 500 // ms blocksRequestRepetition = 1 blockHashesRequestInterval = 500 // ms blocksRequestMaxIdleRounds = 100 blockHashesTimeout = 60 // seconds blocksTimeout = 120 // seconds ) type poolNode struct { lock sync.RWMutex hash []byte td *big.Int block *types.Block parent *poolNode peer string blockBy string } type poolEntry struct { node *poolNode section *section index int } type BlockPool struct { lock sync.RWMutex chainLock sync.RWMutex pool map[string]*poolEntry peersLock sync.RWMutex peers map[string]*peerInfo peer *peerInfo quit chan bool purgeC chan bool flushC chan bool wg sync.WaitGroup procWg sync.WaitGroup running bool // the minimal interface with blockchain hasBlock func(hash []byte) bool insertChain func(types.Blocks) error verifyPoW func(pow.Block) bool } type peerInfo struct { lock sync.RWMutex td *big.Int currentBlockHash []byte currentBlock *types.Block currentBlockC chan *types.Block parentHash []byte headSection *section headSectionC chan *section id string requestBlockHashes func([]byte) error requestBlocks func([][]byte) error peerError func(int, string, ...interface{}) sections map[string]*section quitC chan bool } // structure to store long range links on chain to skip along type section struct { lock sync.RWMutex parent *section child *section top *poolNode bottom *poolNode nodes []*poolNode controlC chan *peerInfo suicideC chan bool blockChainC chan bool forkC chan chan bool offC chan bool } func NewBlockPool(hasBlock func(hash []byte) bool, insertChain func(types.Blocks) error, verifyPoW func(pow.Block) bool, ) *BlockPool { return &BlockPool{ hasBlock: hasBlock, insertChain: insertChain, verifyPoW: verifyPoW, } } // allows restart func (self *BlockPool) Start() { self.lock.Lock() if self.running { self.lock.Unlock() return } self.running = true self.quit = make(chan bool) self.flushC = make(chan bool) self.pool = make(map[string]*poolEntry) self.lock.Unlock() self.peersLock.Lock() self.peers = make(map[string]*peerInfo) self.peersLock.Unlock() poolLogger.Infoln("Started") } func (self *BlockPool) Stop() { self.lock.Lock() if !self.running { self.lock.Unlock() return } self.running = false self.lock.Unlock() poolLogger.Infoln("Stopping...") close(self.quit) //self.wg.Wait() self.peersLock.Lock() self.peers = nil self.peer = nil self.peersLock.Unlock() self.lock.Lock() self.pool = nil self.lock.Unlock() poolLogger.Infoln("Stopped") } func (self *BlockPool) Purge() { self.lock.Lock() if !self.running { self.lock.Unlock() return } self.lock.Unlock() poolLogger.Infoln("Purging...") close(self.purgeC) self.wg.Wait() self.purgeC = make(chan bool) poolLogger.Infoln("Stopped") } func (self *BlockPool) Wait(t time.Duration) { self.lock.Lock() if !self.running { self.lock.Unlock() return } self.lock.Unlock() poolLogger.Infoln("Waiting for processes to complete...") close(self.flushC) w := make(chan bool) go func() { self.procWg.Wait() close(w) }() select { case <-w: poolLogger.Infoln("Processes complete") case <-time.After(t): poolLogger.Warnf("Timeout") } self.flushC = make(chan bool) } // AddPeer is called by the eth protocol instance running on the peer after // the status message has been received with total difficulty and current block hash // AddPeer can only be used once, RemovePeer needs to be called when the peer disconnects func (self *BlockPool) AddPeer(td *big.Int, currentBlockHash []byte, peerId string, requestBlockHashes func([]byte) error, requestBlocks func([][]byte) error, peerError func(int, string, ...interface{})) (best bool) { self.peersLock.Lock() defer self.peersLock.Unlock() peer, ok := self.peers[peerId] if ok { if bytes.Compare(peer.currentBlockHash, currentBlockHash) != 0 { poolLogger.Debugf("Update peer %v with td %v and current block %s", peerId, td, name(currentBlockHash)) peer.lock.Lock() peer.td = td peer.currentBlockHash = currentBlockHash peer.currentBlock = nil peer.parentHash = nil peer.headSection = nil peer.lock.Unlock() } } else { peer = &peerInfo{ td: td, currentBlockHash: currentBlockHash, id: peerId, //peer.Identity().Pubkey() requestBlockHashes: requestBlockHashes, requestBlocks: requestBlocks, peerError: peerError, sections: make(map[string]*section), currentBlockC: make(chan *types.Block), headSectionC: make(chan *section), } self.peers[peerId] = peer poolLogger.Debugf("add new peer %v with td %v and current block %x", peerId, td, currentBlockHash[:4]) } // check peer current head if self.hasBlock(currentBlockHash) { // peer not ahead return false } if self.peer == peer { // new block update // peer is already active best peer, request hashes poolLogger.Debugf("[%s] already the best peer. Request new head section info from %s", peerId, name(currentBlockHash)) peer.headSectionC <- nil best = true } else { currentTD := ethutil.Big0 if self.peer != nil { currentTD = self.peer.td } if td.Cmp(currentTD) > 0 { poolLogger.Debugf("peer %v promoted best peer", peerId) self.switchPeer(self.peer, peer) self.peer = peer best = true } } return } func (self *BlockPool) requestHeadSection(peer *peerInfo) { self.wg.Add(1) self.procWg.Add(1) poolLogger.Debugf("[%s] head section at [%s] requesting info", peer.id, name(peer.currentBlockHash)) go func() { var idle bool peer.lock.RLock() quitC := peer.quitC currentBlockHash := peer.currentBlockHash peer.lock.RUnlock() blockHashesRequestTimer := time.NewTimer(0) blocksRequestTimer := time.NewTimer(0) suicide := time.NewTimer(blockHashesTimeout * time.Second) blockHashesRequestTimer.Stop() defer blockHashesRequestTimer.Stop() defer blocksRequestTimer.Stop() entry := self.get(currentBlockHash) if entry != nil { entry.node.lock.RLock() currentBlock := entry.node.block entry.node.lock.RUnlock() if currentBlock != nil { peer.lock.Lock() peer.currentBlock = currentBlock peer.parentHash = currentBlock.ParentHash() poolLogger.Debugf("[%s] head block [%s] found", peer.id, name(currentBlockHash)) peer.lock.Unlock() blockHashesRequestTimer.Reset(0) blocksRequestTimer.Stop() } } LOOP: for { select { case <-self.quit: break LOOP case <-quitC: poolLogger.Debugf("[%s] head section at [%s] incomplete - quit request loop", peer.id, name(currentBlockHash)) break LOOP case headSection := <-peer.headSectionC: peer.lock.Lock() peer.headSection = headSection if headSection == nil { oldBlockHash := currentBlockHash currentBlockHash = peer.currentBlockHash poolLogger.Debugf("[%s] head section changed [%s] -> [%s]", peer.id, name(oldBlockHash), name(currentBlockHash)) if idle { idle = false suicide.Reset(blockHashesTimeout * time.Second) self.procWg.Add(1) } blocksRequestTimer.Reset(blocksRequestInterval * time.Millisecond) } else { poolLogger.DebugDetailf("[%s] head section at [%s] created", peer.id, name(currentBlockHash)) if !idle { idle = true suicide.Stop() self.procWg.Done() } } peer.lock.Unlock() blockHashesRequestTimer.Stop() case <-blockHashesRequestTimer.C: poolLogger.DebugDetailf("[%s] head section at [%s] not found, requesting block hashes", peer.id, name(currentBlockHash)) peer.requestBlockHashes(currentBlockHash) blockHashesRequestTimer.Reset(blockHashesRequestInterval * time.Millisecond) case currentBlock := <-peer.currentBlockC: peer.lock.Lock() peer.currentBlock = currentBlock peer.parentHash = currentBlock.ParentHash() poolLogger.DebugDetailf("[%s] head block [%s] found", peer.id, name(currentBlockHash)) peer.lock.Unlock() if self.hasBlock(currentBlock.ParentHash()) { if err := self.insertChain(types.Blocks([]*types.Block{currentBlock})); err != nil { peer.peerError(ErrInvalidBlock, "%v", err) } if !idle { idle = true suicide.Stop() self.procWg.Done() } } else { blockHashesRequestTimer.Reset(0) } blocksRequestTimer.Stop() case <-blocksRequestTimer.C: peer.lock.RLock() poolLogger.DebugDetailf("[%s] head block [%s] not found, requesting", peer.id, name(currentBlockHash)) peer.requestBlocks([][]byte{peer.currentBlockHash}) peer.lock.RUnlock() blocksRequestTimer.Reset(blocksRequestInterval * time.Millisecond) case <-suicide.C: peer.peerError(ErrInsufficientChainInfo, "peer failed to provide block hashes or head block for block hash %x", currentBlockHash) break LOOP } } self.wg.Done() if !idle { self.procWg.Done() } }() } // RemovePeer is called by the eth protocol when the peer disconnects func (self *BlockPool) RemovePeer(peerId string) { self.peersLock.Lock() defer self.peersLock.Unlock() peer, ok := self.peers[peerId] if !ok { return } delete(self.peers, peerId) poolLogger.Debugf("remove peer %v", peerId) // if current best peer is removed, need find a better one if self.peer == peer { var newPeer *peerInfo max := ethutil.Big0 // peer with the highest self-acclaimed TD is chosen for _, info := range self.peers { if info.td.Cmp(max) > 0 { max = info.td newPeer = info } } if newPeer != nil { poolLogger.Debugf("peer %v with td %v promoted to best peer", newPeer.id, newPeer.td) } else { poolLogger.Warnln("no peers") } self.peer = newPeer self.switchPeer(peer, newPeer) } } // Entry point for eth protocol to add block hashes received via BlockHashesMsg // only hashes from the best peer is handled // this method is always responsible to initiate further hash requests until // a known parent is reached unless cancelled by a peerChange event // this process also launches all request processes on each chain section // this function needs to run asynchronously for one peer since the message is discarded??? func (self *BlockPool) AddBlockHashes(next func() ([]byte, bool), peerId string) { // register with peer manager loop peer, best := self.getPeer(peerId) if !best { return } // peer is still the best var size, n int var hash []byte var ok, headSection bool var sec, child, parent *section var entry *poolEntry var nodes []*poolNode bestPeer := peer hash, ok = next() peer.lock.Lock() if bytes.Compare(peer.parentHash, hash) == 0 { if self.hasBlock(peer.currentBlockHash) { return } poolLogger.Debugf("adding hashes at chain head for best peer %s starting from [%s]", peerId, name(peer.currentBlockHash)) headSection = true if entry := self.get(peer.currentBlockHash); entry == nil { node := &poolNode{ hash: peer.currentBlockHash, block: peer.currentBlock, peer: peerId, blockBy: peerId, } if size == 0 { sec = newSection() } nodes = append(nodes, node) size++ n++ } else { child = entry.section } } else { poolLogger.Debugf("adding hashes for best peer %s starting from [%s]", peerId, name(hash)) } quitC := peer.quitC peer.lock.Unlock() LOOP: // iterate using next (rlp stream lazy decoder) feeding hashesC for ; ok; hash, ok = next() { n++ select { case <-self.quit: return case <-quitC: // if the peer is demoted, no more hashes taken bestPeer = nil break LOOP default: } if self.hasBlock(hash) { // check if known block connecting the downloaded chain to our blockchain poolLogger.DebugDetailf("[%s] known block", name(hash)) // mark child as absolute pool root with parent known to blockchain if sec != nil { self.connectToBlockChain(sec) } else { if child != nil { self.connectToBlockChain(child) } } break LOOP } // look up node in pool entry = self.get(hash) if entry != nil { // reached a known chain in the pool if entry.node == entry.section.bottom && n == 1 { // the first block hash received is an orphan in the pool, so rejoice and continue poolLogger.DebugDetailf("[%s] connecting child section", sectionName(entry.section)) child = entry.section continue LOOP } poolLogger.DebugDetailf("[%s] reached blockpool chain", name(hash)) parent = entry.section break LOOP } // if node for block hash does not exist, create it and index in the pool node := &poolNode{ hash: hash, peer: peerId, } if size == 0 { sec = newSection() } nodes = append(nodes, node) size++ } //for self.chainLock.Lock() poolLogger.DebugDetailf("added %v hashes sent by %s", n, peerId) if parent != nil && entry != nil && entry.node != parent.top { poolLogger.DebugDetailf("[%s] split section at fork", sectionName(parent)) parent.controlC <- nil waiter := make(chan bool) parent.forkC <- waiter chain := parent.nodes parent.nodes = chain[entry.index:] parent.top = parent.nodes[0] orphan := newSection() self.link(orphan, parent.child) self.processSection(orphan, chain[0:entry.index]) orphan.controlC <- nil close(waiter) } if size > 0 { self.processSection(sec, nodes) poolLogger.DebugDetailf("[%s]->[%s](%v)->[%s] new chain section", sectionName(parent), sectionName(sec), size, sectionName(child)) self.link(parent, sec) self.link(sec, child) } else { poolLogger.DebugDetailf("[%s]->[%s] connecting known sections", sectionName(parent), sectionName(child)) self.link(parent, child) } self.chainLock.Unlock() if parent != nil && bestPeer != nil { self.activateChain(parent, peer) poolLogger.Debugf("[%s] activate parent section [%s]", name(parent.top.hash), sectionName(parent)) } if sec != nil { peer.addSection(sec.top.hash, sec) // request next section here once, only repeat if bottom block arrives, // otherwise no way to check if it arrived peer.requestBlockHashes(sec.bottom.hash) sec.controlC <- bestPeer poolLogger.Debugf("[%s] activate new section", sectionName(sec)) } if headSection { var headSec *section switch { case sec != nil: headSec = sec case child != nil: headSec = child default: headSec = parent } peer.headSectionC <- headSec } } func name(hash []byte) (name string) { if hash == nil { name = "" } else { name = fmt.Sprintf("%x", hash[:4]) } return } func sectionName(section *section) (name string) { if section == nil { name = "" } else { name = fmt.Sprintf("%x-%x", section.bottom.hash[:4], section.top.hash[:4]) } return } // AddBlock is the entry point for the eth protocol when blockmsg is received upon requests // It has a strict interpretation of the protocol in that if the block received has not been requested, it results in an error (which can be ignored) // block is checked for PoW // only the first PoW-valid block for a hash is considered legit func (self *BlockPool) AddBlock(block *types.Block, peerId string) { hash := block.Hash() self.peersLock.Lock() peer := self.peer self.peersLock.Unlock() entry := self.get(hash) if bytes.Compare(hash, peer.currentBlockHash) == 0 { poolLogger.Debugf("add head block [%s] for peer %s", name(hash), peerId) peer.currentBlockC <- block } else { if entry == nil { poolLogger.Warnf("unrequested block [%s] by peer %s", name(hash), peerId) self.peerError(peerId, ErrUnrequestedBlock, "%x", hash) } } if entry == nil { return } node := entry.node node.lock.Lock() defer node.lock.Unlock() // check if block already present if node.block != nil { poolLogger.DebugDetailf("block [%s] already sent by %s", name(hash), node.blockBy) return } if self.hasBlock(hash) { poolLogger.DebugDetailf("block [%s] already known", name(hash)) } else { // validate block for PoW if !self.verifyPoW(block) { poolLogger.Warnf("invalid pow on block [%s %v] by peer %s", name(hash), block.Number(), peerId) self.peerError(peerId, ErrInvalidPoW, "%x", hash) return } } poolLogger.DebugDetailf("added block [%s] sent by peer %s", name(hash), peerId) node.block = block node.blockBy = peerId } func (self *BlockPool) connectToBlockChain(section *section) { select { case <-section.offC: self.addSectionToBlockChain(section) case <-section.blockChainC: default: close(section.blockChainC) } } func (self *BlockPool) addSectionToBlockChain(section *section) (rest int, err error) { var blocks types.Blocks var node *poolNode var keys []string rest = len(section.nodes) for rest > 0 { rest-- node = section.nodes[rest] node.lock.RLock() block := node.block node.lock.RUnlock() if block == nil { break } keys = append(keys, string(node.hash)) blocks = append(blocks, block) } self.lock.Lock() for _, key := range keys { delete(self.pool, key) } self.lock.Unlock() poolLogger.Infof("insert %v blocks into blockchain", len(blocks)) err = self.insertChain(blocks) if err != nil { // TODO: not clear which peer we need to address // peerError should dispatch to peer if still connected and disconnect self.peerError(node.blockBy, ErrInvalidBlock, "%v", err) poolLogger.Warnf("invalid block %x", node.hash) poolLogger.Warnf("penalise peers %v (hash), %v (block)", node.peer, node.blockBy) // penalise peer in node.blockBy // self.disconnect() } return } func (self *BlockPool) activateChain(section *section, peer *peerInfo) { poolLogger.DebugDetailf("[%s] activate known chain for peer %s", sectionName(section), peer.id) i := 0 LOOP: for section != nil { // register this section with the peer and quit if registered poolLogger.DebugDetailf("[%s] register section with peer %s", sectionName(section), peer.id) if peer.addSection(section.top.hash, section) == section { return } poolLogger.DebugDetailf("[%s] activate section process", sectionName(section)) select { case section.controlC <- peer: case <-section.offC: } i++ section = self.getParent(section) select { case <-peer.quitC: break LOOP case <-self.quit: break LOOP default: } } } // main worker thread on each section in the poolchain // - kills the section if there are blocks missing after an absolute time // - kills the section if there are maxIdleRounds of idle rounds of block requests with no response // - periodically polls the chain section for missing blocks which are then requested from peers // - registers the process controller on the peer so that if the peer is promoted as best peer the second time (after a disconnect of a better one), all active processes are switched back on unless they expire and killed () // - when turned off (if peer disconnects and new peer connects with alternative chain), no blockrequests are made but absolute expiry timer is ticking // - when turned back on it recursively calls itself on the root of the next chain section // - when exits, signals to func (self *BlockPool) processSection(sec *section, nodes []*poolNode) { for i, node := range nodes { entry := &poolEntry{node: node, section: sec, index: i} self.set(node.hash, entry) } sec.bottom = nodes[len(nodes)-1] sec.top = nodes[0] sec.nodes = nodes poolLogger.DebugDetailf("[%s] setup section process", sectionName(sec)) self.wg.Add(1) go func() { // absolute time after which sub-chain is killed if not complete (some blocks are missing) suicideTimer := time.After(blocksTimeout * time.Second) var peer, newPeer *peerInfo var blocksRequestTimer, blockHashesRequestTimer <-chan time.Time var blocksRequestTime, blockHashesRequestTime bool var blocksRequests, blockHashesRequests int var blocksRequestsComplete, blockHashesRequestsComplete bool // node channels for the section var missingC, processC, offC chan *poolNode // container for missing block hashes var hashes [][]byte var i, missing, lastMissing, depth int var idle int var init, done, same, ready bool var insertChain bool var quitC chan bool var blockChainC = sec.blockChainC var parentHash []byte LOOP: for { if insertChain { insertChain = false rest, err := self.addSectionToBlockChain(sec) if err != nil { close(sec.suicideC) continue LOOP } if rest == 0 { blocksRequestsComplete = true child := self.getChild(sec) if child != nil { self.connectToBlockChain(child) } } } if blockHashesRequestsComplete && blocksRequestsComplete { // not waiting for hashes any more poolLogger.Debugf("[%s] section complete %v blocks retrieved (%v attempts), hash requests complete on root (%v attempts)", sectionName(sec), depth, blocksRequests, blockHashesRequests) break LOOP } // otherwise suicide if no hashes coming if done { // went through all blocks in section if missing == 0 { // no missing blocks poolLogger.DebugDetailf("[%s] got all blocks. process complete (%v total blocksRequests): missing %v/%v/%v", sectionName(sec), blocksRequests, missing, lastMissing, depth) blocksRequestsComplete = true blocksRequestTimer = nil blocksRequestTime = false } else { poolLogger.DebugDetailf("[%s] section checked: missing %v/%v/%v", sectionName(sec), missing, lastMissing, depth) // some missing blocks blocksRequests++ if len(hashes) > 0 { // send block requests to peers self.requestBlocks(blocksRequests, hashes) hashes = nil } if missing == lastMissing { // idle round if same { // more than once idle++ // too many idle rounds if idle >= blocksRequestMaxIdleRounds { poolLogger.DebugDetailf("[%s] block requests had %v idle rounds (%v total attempts): missing %v/%v/%v\ngiving up...", sectionName(sec), idle, blocksRequests, missing, lastMissing, depth) close(sec.suicideC) } } else { idle = 0 } same = true } else { same = false } } lastMissing = missing ready = true done = false // save a new processC (blocks still missing) offC = missingC missingC = processC // put processC offline processC = nil } // if ready && blocksRequestTime && !blocksRequestsComplete { poolLogger.DebugDetailf("[%s] check if new blocks arrived (attempt %v): missing %v/%v/%v", sectionName(sec), blocksRequests, missing, lastMissing, depth) blocksRequestTimer = time.After(blocksRequestInterval * time.Millisecond) blocksRequestTime = false processC = offC } if blockHashesRequestTime { var parentSection = self.getParent(sec) if parentSection == nil { if parent := self.get(parentHash); parent != nil { parentSection = parent.section self.chainLock.Lock() self.link(parentSection, sec) self.chainLock.Unlock() } else { if self.hasBlock(parentHash) { insertChain = true blockHashesRequestTime = false blockHashesRequestTimer = nil blockHashesRequestsComplete = true continue LOOP } } } if parentSection != nil { // if not root of chain, switch off poolLogger.DebugDetailf("[%s] parent found, hash requests deactivated (after %v total attempts)\n", sectionName(sec), blockHashesRequests) blockHashesRequestTimer = nil blockHashesRequestsComplete = true } else { blockHashesRequests++ poolLogger.Debugf("[%s] hash request on root (%v total attempts)\n", sectionName(sec), blockHashesRequests) peer.requestBlockHashes(sec.bottom.hash) blockHashesRequestTimer = time.After(blockHashesRequestInterval * time.Millisecond) } blockHashesRequestTime = false } select { case <-self.quit: break LOOP case <-quitC: // peer quit or demoted, put section in idle mode quitC = nil go func() { sec.controlC <- nil }() case <-self.purgeC: suicideTimer = time.After(0) case <-suicideTimer: close(sec.suicideC) poolLogger.Debugf("[%s] timeout. (%v total attempts): missing %v/%v/%v", sectionName(sec), blocksRequests, missing, lastMissing, depth) case <-sec.suicideC: poolLogger.Debugf("[%s] suicide", sectionName(sec)) // first delink from child and parent under chainlock self.chainLock.Lock() self.link(nil, sec) self.link(sec, nil) self.chainLock.Unlock() // delete node entries from pool index under pool lock self.lock.Lock() for _, node := range sec.nodes { delete(self.pool, string(node.hash)) } self.lock.Unlock() break LOOP case <-blocksRequestTimer: poolLogger.DebugDetailf("[%s] block request time", sectionName(sec)) blocksRequestTime = true case <-blockHashesRequestTimer: poolLogger.DebugDetailf("[%s] hash request time", sectionName(sec)) blockHashesRequestTime = true case newPeer = <-sec.controlC: // active -> idle if peer != nil && newPeer == nil { self.procWg.Done() if init { poolLogger.Debugf("[%s] idle mode (%v total attempts): missing %v/%v/%v", sectionName(sec), blocksRequests, missing, lastMissing, depth) } blocksRequestTime = false blocksRequestTimer = nil blockHashesRequestTime = false blockHashesRequestTimer = nil if processC != nil { offC = processC processC = nil } } // idle -> active if peer == nil && newPeer != nil { self.procWg.Add(1) poolLogger.Debugf("[%s] active mode", sectionName(sec)) if !blocksRequestsComplete { blocksRequestTime = true } if !blockHashesRequestsComplete && parentHash != nil { blockHashesRequestTime = true } if !init { processC = make(chan *poolNode, blockHashesBatchSize) missingC = make(chan *poolNode, blockHashesBatchSize) i = 0 missing = 0 self.wg.Add(1) self.procWg.Add(1) depth = len(sec.nodes) lastMissing = depth // if not run at least once fully, launch iterator go func() { var node *poolNode IT: for _, node = range sec.nodes { select { case processC <- node: case <-self.quit: break IT } } close(processC) self.wg.Done() self.procWg.Done() }() } else { poolLogger.Debugf("[%s] restore earlier state", sectionName(sec)) processC = offC } } // reset quitC to current best peer if newPeer != nil { quitC = newPeer.quitC } peer = newPeer case waiter := <-sec.forkC: // this case just blocks the process until section is split at the fork <-waiter init = false done = false ready = false case node, ok := <-processC: if !ok && !init { // channel closed, first iteration finished init = true done = true processC = make(chan *poolNode, missing) poolLogger.DebugDetailf("[%s] section initalised: missing %v/%v/%v", sectionName(sec), missing, lastMissing, depth) continue LOOP } if ready { i = 0 missing = 0 ready = false } i++ // if node has no block node.lock.RLock() block := node.block node.lock.RUnlock() if block == nil { missing++ hashes = append(hashes, node.hash) if len(hashes) == blockBatchSize { poolLogger.Debugf("[%s] request %v missing blocks", sectionName(sec), len(hashes)) self.requestBlocks(blocksRequests, hashes) hashes = nil } missingC <- node } else { if i == lastMissing { if blockChainC == nil { insertChain = true } else { if parentHash == nil { parentHash = block.ParentHash() poolLogger.Debugf("[%s] found root block [%s]", sectionName(sec), name(parentHash)) blockHashesRequestTime = true } } } } if i == lastMissing && init { done = true } case <-blockChainC: // closed blockChain channel indicates that the blockpool is reached // connected to the blockchain, insert the longest chain of blocks poolLogger.Debugf("[%s] reached blockchain", sectionName(sec)) blockChainC = nil // switch off hash requests in case they were on blockHashesRequestTime = false blockHashesRequestTimer = nil blockHashesRequestsComplete = true // section root has block if len(sec.nodes) > 0 && sec.nodes[len(sec.nodes)-1].block != nil { insertChain = true } continue LOOP } // select } // for close(sec.offC) self.wg.Done() if peer != nil { self.procWg.Done() } }() return } func (self *BlockPool) peerError(peerId string, code int, format string, params ...interface{}) { self.peersLock.RLock() defer self.peersLock.RUnlock() peer, ok := self.peers[peerId] if ok { peer.peerError(code, format, params...) } } func (self *BlockPool) requestBlocks(attempts int, hashes [][]byte) { self.wg.Add(1) self.procWg.Add(1) go func() { // distribute block request among known peers self.peersLock.Lock() defer self.peersLock.Unlock() peerCount := len(self.peers) // on first attempt use the best peer if attempts == 0 { poolLogger.Debugf("request %v missing blocks from best peer %s", len(hashes), self.peer.id) self.peer.requestBlocks(hashes) return } repetitions := int(math.Min(float64(peerCount), float64(blocksRequestRepetition))) i := 0 indexes := rand.Perm(peerCount)[0:repetitions] sort.Ints(indexes) poolLogger.Debugf("request %v missing blocks from %v/%v peers: chosen %v", len(hashes), repetitions, peerCount, indexes) for _, peer := range self.peers { if i == indexes[0] { poolLogger.Debugf("request %v missing blocks [%x/%x] from peer %s", len(hashes), hashes[0][:4], hashes[len(hashes)-1][:4], peer.id) peer.requestBlocks(hashes) indexes = indexes[1:] if len(indexes) == 0 { break } } i++ } self.wg.Done() self.procWg.Done() }() } func (self *BlockPool) getPeer(peerId string) (*peerInfo, bool) { self.peersLock.RLock() defer self.peersLock.RUnlock() if self.peer != nil && self.peer.id == peerId { return self.peer, true } info, ok := self.peers[peerId] if !ok { return nil, false } return info, false } func (self *peerInfo) addSection(hash []byte, section *section) (found *section) { self.lock.Lock() defer self.lock.Unlock() key := string(hash) found = self.sections[key] poolLogger.DebugDetailf("[%s] section process stored for %s", sectionName(section), self.id) self.sections[key] = section return } func (self *BlockPool) switchPeer(oldPeer, newPeer *peerInfo) { if newPeer != nil { newPeer.quitC = make(chan bool) poolLogger.DebugDetailf("[%s] activate section processes", newPeer.id) var addSections []*section for hash, section := range newPeer.sections { // split sections get reorganised here if string(section.top.hash) != hash { addSections = append(addSections, section) if entry := self.get([]byte(hash)); entry != nil { addSections = append(addSections, entry.section) } } } for _, section := range addSections { newPeer.sections[string(section.top.hash)] = section } for hash, section := range newPeer.sections { // this will block if section process is waiting for peer lock select { case <-section.offC: poolLogger.DebugDetailf("[%s][%x] section process complete - remove", newPeer.id, hash[:4]) delete(newPeer.sections, hash) case section.controlC <- newPeer: poolLogger.DebugDetailf("[%s][%x] activates section [%s]", newPeer.id, hash[:4], sectionName(section)) } } newPeer.lock.Lock() headSection := newPeer.headSection currentBlockHash := newPeer.currentBlockHash newPeer.lock.Unlock() if headSection == nil { poolLogger.DebugDetailf("[%s] head section for [%s] not created, requesting info", newPeer.id, name(currentBlockHash)) self.requestHeadSection(newPeer) } else { if entry := self.get(currentBlockHash); entry != nil { headSection = entry.section } poolLogger.DebugDetailf("[%s] activate chain at head section [%s] for current head [%s]", newPeer.id, sectionName(headSection), name(currentBlockHash)) self.activateChain(headSection, newPeer) } } if oldPeer != nil { poolLogger.DebugDetailf("[%s] quit section processes", oldPeer.id) close(oldPeer.quitC) } } func (self *BlockPool) getParent(sec *section) *section { self.chainLock.RLock() defer self.chainLock.RUnlock() return sec.parent } func (self *BlockPool) getChild(sec *section) *section { self.chainLock.RLock() defer self.chainLock.RUnlock() return sec.child } func newSection() (sec *section) { sec = §ion{ controlC: make(chan *peerInfo), suicideC: make(chan bool), blockChainC: make(chan bool), offC: make(chan bool), forkC: make(chan chan bool), } return } // link should only be called under chainLock func (self *BlockPool) link(parent *section, child *section) { if parent != nil { exChild := parent.child parent.child = child if exChild != nil && exChild != child { poolLogger.Debugf("[%s] chain fork [%s] -> [%s]", sectionName(parent), sectionName(exChild), sectionName(child)) exChild.parent = nil } } if child != nil { exParent := child.parent if exParent != nil && exParent != parent { poolLogger.Debugf("[%s] chain reverse fork [%s] -> [%s]", sectionName(child), sectionName(exParent), sectionName(parent)) exParent.child = nil } child.parent = parent } } func (self *BlockPool) get(hash []byte) (node *poolEntry) { self.lock.RLock() defer self.lock.RUnlock() return self.pool[string(hash)] } func (self *BlockPool) set(hash []byte, node *poolEntry) { self.lock.Lock() defer self.lock.Unlock() self.pool[string(hash)] = node }