2019-03-23 05:56:03 +03:00
|
|
|
package discovery
|
|
|
|
|
|
|
|
import (
|
|
|
|
"errors"
|
|
|
|
"sync"
|
2019-07-27 04:22:37 +03:00
|
|
|
"sync/atomic"
|
2019-03-23 05:56:03 +03:00
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
|
|
"github.com/lightningnetwork/lnd/lnpeer"
|
|
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
2019-04-05 18:36:11 +03:00
|
|
|
"github.com/lightningnetwork/lnd/routing/route"
|
2019-03-23 05:56:03 +03:00
|
|
|
"github.com/lightningnetwork/lnd/ticker"
|
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
|
|
|
// DefaultSyncerRotationInterval is the default interval in which we'll
|
|
|
|
// rotate a single active syncer.
|
|
|
|
DefaultSyncerRotationInterval = 20 * time.Minute
|
|
|
|
|
|
|
|
// DefaultHistoricalSyncInterval is the default interval in which we'll
|
|
|
|
// force a historical sync to ensure we have as much of the public
|
|
|
|
// network as possible.
|
2019-04-11 05:31:48 +03:00
|
|
|
DefaultHistoricalSyncInterval = time.Hour
|
2019-03-23 05:56:03 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
// ErrSyncManagerExiting is an error returned when we attempt to
|
|
|
|
// start/stop a gossip syncer for a connected/disconnected peer, but the
|
|
|
|
// SyncManager has already been stopped.
|
|
|
|
ErrSyncManagerExiting = errors.New("sync manager exiting")
|
|
|
|
)
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
// newSyncer in an internal message we'll use within the SyncManager to signal
|
|
|
|
// that we should create a GossipSyncer for a newly connected peer.
|
|
|
|
type newSyncer struct {
|
|
|
|
// peer is the newly connected peer.
|
|
|
|
peer lnpeer.Peer
|
|
|
|
|
|
|
|
// doneChan serves as a signal to the caller that the SyncManager's
|
|
|
|
// internal state correctly reflects the stale active syncer.
|
|
|
|
doneChan chan struct{}
|
|
|
|
}
|
|
|
|
|
|
|
|
// staleSyncer is an internal message we'll use within the SyncManager to signal
|
|
|
|
// that a peer has disconnected and its GossipSyncer should be removed.
|
|
|
|
type staleSyncer struct {
|
|
|
|
// peer is the peer that has disconnected.
|
2019-04-05 18:36:11 +03:00
|
|
|
peer route.Vertex
|
2019-04-11 05:26:08 +03:00
|
|
|
|
|
|
|
// doneChan serves as a signal to the caller that the SyncManager's
|
|
|
|
// internal state correctly reflects the stale active syncer. This is
|
|
|
|
// needed to ensure we always create a new syncer for a flappy peer
|
|
|
|
// after they disconnect if they happened to be an active syncer.
|
|
|
|
doneChan chan struct{}
|
|
|
|
}
|
|
|
|
|
2019-03-23 05:56:03 +03:00
|
|
|
// SyncManagerCfg contains all of the dependencies required for the SyncManager
|
|
|
|
// to carry out its duties.
|
|
|
|
type SyncManagerCfg struct {
|
|
|
|
// ChainHash is a hash that indicates the specific network of the active
|
|
|
|
// chain.
|
|
|
|
ChainHash chainhash.Hash
|
|
|
|
|
|
|
|
// ChanSeries is an interface that provides access to a time series view
|
|
|
|
// of the current known channel graph. Each GossipSyncer enabled peer
|
|
|
|
// will utilize this in order to create and respond to channel graph
|
|
|
|
// time series queries.
|
|
|
|
ChanSeries ChannelGraphTimeSeries
|
|
|
|
|
|
|
|
// NumActiveSyncers is the number of peers for which we should have
|
|
|
|
// active syncers with. After reaching NumActiveSyncers, any future
|
|
|
|
// gossip syncers will be passive.
|
|
|
|
NumActiveSyncers int
|
|
|
|
|
|
|
|
// RotateTicker is a ticker responsible for notifying the SyncManager
|
|
|
|
// when it should rotate its active syncers. A single active syncer with
|
|
|
|
// a chansSynced state will be exchanged for a passive syncer in order
|
|
|
|
// to ensure we don't keep syncing with the same peers.
|
|
|
|
RotateTicker ticker.Ticker
|
|
|
|
|
|
|
|
// HistoricalSyncTicker is a ticker responsible for notifying the
|
|
|
|
// SyncManager when it should attempt a historical sync with a gossip
|
|
|
|
// sync peer.
|
|
|
|
HistoricalSyncTicker ticker.Ticker
|
2019-07-31 03:25:31 +03:00
|
|
|
|
|
|
|
// IgnoreHistoricalFilters will prevent syncers from replying with
|
|
|
|
// historical data when the remote peer sets a gossip_timestamp_range.
|
|
|
|
// This prevents ranges with old start times from causing us to dump the
|
|
|
|
// graph on connect.
|
|
|
|
IgnoreHistoricalFilters bool
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// SyncManager is a subsystem of the gossiper that manages the gossip syncers
|
|
|
|
// for peers currently connected. When a new peer is connected, the manager will
|
|
|
|
// create its accompanying gossip syncer and determine whether it should have an
|
|
|
|
// ActiveSync or PassiveSync sync type based on how many other gossip syncers
|
|
|
|
// are currently active. Any ActiveSync gossip syncers are started in a
|
|
|
|
// round-robin manner to ensure we're not syncing with multiple peers at the
|
|
|
|
// same time. The first GossipSyncer registered with the SyncManager will
|
|
|
|
// attempt a historical sync to ensure we have as much of the public channel
|
|
|
|
// graph as possible.
|
|
|
|
type SyncManager struct {
|
2019-07-27 04:22:37 +03:00
|
|
|
// initialHistoricalSyncCompleted serves as a barrier when initializing
|
|
|
|
// new active GossipSyncers. If 0, the initial historical sync has not
|
|
|
|
// completed, so we'll defer initializing any active GossipSyncers. If
|
|
|
|
// 1, then we can transition the GossipSyncer immediately. We set up
|
|
|
|
// this barrier to ensure we have most of the graph before attempting to
|
|
|
|
// accept new updates at tip.
|
|
|
|
//
|
|
|
|
// NOTE: This must be used atomically.
|
|
|
|
initialHistoricalSyncCompleted int32
|
|
|
|
|
2019-03-23 05:56:03 +03:00
|
|
|
start sync.Once
|
|
|
|
stop sync.Once
|
|
|
|
|
|
|
|
cfg SyncManagerCfg
|
|
|
|
|
|
|
|
// historicalSync allows us to perform an initial historical sync only
|
|
|
|
// _once_ with a peer during the SyncManager's startup.
|
|
|
|
historicalSync sync.Once
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
// newSyncers is a channel we'll use to process requests to create
|
|
|
|
// GossipSyncers for newly connected peers.
|
|
|
|
newSyncers chan *newSyncer
|
|
|
|
|
|
|
|
// staleSyncers is a channel we'll use to process requests to tear down
|
|
|
|
// GossipSyncers for disconnected peers.
|
|
|
|
staleSyncers chan *staleSyncer
|
|
|
|
|
|
|
|
// syncersMu guards the read and write access to the activeSyncers and
|
|
|
|
// inactiveSyncers maps below.
|
|
|
|
syncersMu sync.Mutex
|
|
|
|
|
2019-03-23 05:56:03 +03:00
|
|
|
// activeSyncers is the set of all syncers for which we are currently
|
|
|
|
// receiving graph updates from. The number of possible active syncers
|
|
|
|
// is bounded by NumActiveSyncers.
|
2019-04-05 18:36:11 +03:00
|
|
|
activeSyncers map[route.Vertex]*GossipSyncer
|
2019-03-23 05:56:03 +03:00
|
|
|
|
|
|
|
// inactiveSyncers is the set of all syncers for which we are not
|
|
|
|
// currently receiving new graph updates from.
|
2019-04-05 18:36:11 +03:00
|
|
|
inactiveSyncers map[route.Vertex]*GossipSyncer
|
2019-03-23 05:56:03 +03:00
|
|
|
|
|
|
|
wg sync.WaitGroup
|
|
|
|
quit chan struct{}
|
|
|
|
}
|
|
|
|
|
|
|
|
// newSyncManager constructs a new SyncManager backed by the given config.
|
|
|
|
func newSyncManager(cfg *SyncManagerCfg) *SyncManager {
|
|
|
|
return &SyncManager{
|
2019-04-11 05:26:08 +03:00
|
|
|
cfg: *cfg,
|
|
|
|
newSyncers: make(chan *newSyncer),
|
|
|
|
staleSyncers: make(chan *staleSyncer),
|
2019-03-23 05:56:03 +03:00
|
|
|
activeSyncers: make(
|
2019-04-05 18:36:11 +03:00
|
|
|
map[route.Vertex]*GossipSyncer, cfg.NumActiveSyncers,
|
2019-03-23 05:56:03 +03:00
|
|
|
),
|
2019-04-05 18:36:11 +03:00
|
|
|
inactiveSyncers: make(map[route.Vertex]*GossipSyncer),
|
2019-04-11 05:25:53 +03:00
|
|
|
quit: make(chan struct{}),
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start starts the SyncManager in order to properly carry out its duties.
|
|
|
|
func (m *SyncManager) Start() {
|
|
|
|
m.start.Do(func() {
|
2019-04-11 05:25:53 +03:00
|
|
|
m.wg.Add(1)
|
2019-03-23 05:56:03 +03:00
|
|
|
go m.syncerHandler()
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
// Stop stops the SyncManager from performing its duties.
|
|
|
|
func (m *SyncManager) Stop() {
|
|
|
|
m.stop.Do(func() {
|
|
|
|
close(m.quit)
|
|
|
|
m.wg.Wait()
|
|
|
|
|
|
|
|
for _, syncer := range m.inactiveSyncers {
|
|
|
|
syncer.Stop()
|
|
|
|
}
|
|
|
|
for _, syncer := range m.activeSyncers {
|
|
|
|
syncer.Stop()
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
// syncerHandler is the SyncManager's main event loop responsible for:
|
|
|
|
//
|
2019-04-11 05:26:08 +03:00
|
|
|
// 1. Creating and tearing down GossipSyncers for connected/disconnected peers.
|
|
|
|
|
|
|
|
// 2. Finding new peers to receive graph updates from to ensure we don't only
|
|
|
|
// receive them from the same set of peers.
|
|
|
|
|
|
|
|
// 3. Finding new peers to force a historical sync with to ensure we have as
|
|
|
|
// much of the public network as possible.
|
2019-03-23 05:56:03 +03:00
|
|
|
//
|
|
|
|
// NOTE: This must be run as a goroutine.
|
|
|
|
func (m *SyncManager) syncerHandler() {
|
|
|
|
defer m.wg.Done()
|
|
|
|
|
|
|
|
m.cfg.RotateTicker.Resume()
|
|
|
|
defer m.cfg.RotateTicker.Stop()
|
|
|
|
|
|
|
|
m.cfg.HistoricalSyncTicker.Resume()
|
|
|
|
defer m.cfg.HistoricalSyncTicker.Stop()
|
|
|
|
|
2019-04-11 05:26:56 +03:00
|
|
|
var (
|
|
|
|
// initialHistoricalSyncer is the syncer we are currently
|
|
|
|
// performing an initial historical sync with.
|
|
|
|
initialHistoricalSyncer *GossipSyncer
|
|
|
|
|
|
|
|
// initialHistoricalSyncSignal is a signal that will fire once
|
|
|
|
// the intiial historical sync has been completed. This is
|
|
|
|
// crucial to ensure that another historical sync isn't
|
|
|
|
// attempted just because the initialHistoricalSyncer was
|
|
|
|
// disconnected.
|
|
|
|
initialHistoricalSyncSignal chan struct{}
|
|
|
|
)
|
|
|
|
|
2019-03-23 05:56:03 +03:00
|
|
|
for {
|
|
|
|
select {
|
2019-04-11 05:26:08 +03:00
|
|
|
// A new peer has been connected, so we'll create its
|
|
|
|
// accompanying GossipSyncer.
|
|
|
|
case newSyncer := <-m.newSyncers:
|
|
|
|
// If we already have a syncer, then we'll exit early as
|
|
|
|
// we don't want to override it.
|
|
|
|
if _, ok := m.GossipSyncer(newSyncer.peer.PubKey()); ok {
|
|
|
|
close(newSyncer.doneChan)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
s := m.createGossipSyncer(newSyncer.peer)
|
|
|
|
|
2019-05-21 14:43:25 +03:00
|
|
|
// attemptHistoricalSync determines whether we should
|
|
|
|
// attempt an initial historical sync when a new peer
|
|
|
|
// connects.
|
|
|
|
attemptHistoricalSync := false
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
m.syncersMu.Lock()
|
|
|
|
switch {
|
2019-05-21 14:43:25 +03:00
|
|
|
// Regardless of whether the initial historical sync
|
|
|
|
// has completed, we'll re-trigger a historical sync if
|
|
|
|
// we no longer have any syncers. This might be
|
|
|
|
// necessary if we lost all our peers at one point, and
|
|
|
|
// now we finally have one again.
|
|
|
|
case len(m.activeSyncers) == 0 &&
|
|
|
|
len(m.inactiveSyncers) == 0:
|
|
|
|
|
|
|
|
attemptHistoricalSync = true
|
|
|
|
fallthrough
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
// If we've exceeded our total number of active syncers,
|
|
|
|
// we'll initialize this GossipSyncer as passive.
|
|
|
|
case len(m.activeSyncers) >= m.cfg.NumActiveSyncers:
|
2019-04-11 05:27:12 +03:00
|
|
|
fallthrough
|
|
|
|
|
2019-05-21 14:43:25 +03:00
|
|
|
// If the initial historical sync has yet to complete,
|
2019-07-27 04:22:37 +03:00
|
|
|
// then we'll declare it as passive and attempt to
|
2019-05-21 14:43:25 +03:00
|
|
|
// transition it when the initial historical sync
|
|
|
|
// completes.
|
2019-07-27 04:22:37 +03:00
|
|
|
case !m.IsGraphSynced():
|
2019-04-11 05:26:08 +03:00
|
|
|
s.setSyncType(PassiveSync)
|
|
|
|
m.inactiveSyncers[s.cfg.peerPub] = s
|
|
|
|
|
2019-04-11 05:27:12 +03:00
|
|
|
// The initial historical sync has completed, so we can
|
|
|
|
// immediately start the GossipSyncer as active.
|
2019-04-11 05:26:08 +03:00
|
|
|
default:
|
|
|
|
s.setSyncType(ActiveSync)
|
|
|
|
m.activeSyncers[s.cfg.peerPub] = s
|
|
|
|
}
|
|
|
|
m.syncersMu.Unlock()
|
|
|
|
|
|
|
|
s.Start()
|
|
|
|
|
|
|
|
// Once we create the GossipSyncer, we'll signal to the
|
|
|
|
// caller that they can proceed since the SyncManager's
|
|
|
|
// internal state has been updated.
|
|
|
|
close(newSyncer.doneChan)
|
|
|
|
|
|
|
|
// We'll force a historical sync with the first peer we
|
|
|
|
// connect to, to ensure we get as much of the graph as
|
|
|
|
// possible.
|
2019-05-21 14:43:25 +03:00
|
|
|
if !attemptHistoricalSync {
|
2019-04-11 05:26:56 +03:00
|
|
|
continue
|
2019-04-11 05:26:08 +03:00
|
|
|
}
|
2019-07-27 04:22:37 +03:00
|
|
|
m.markGraphSyncing()
|
2019-04-11 05:26:08 +03:00
|
|
|
|
2019-04-11 05:26:56 +03:00
|
|
|
log.Debugf("Attempting initial historical sync with "+
|
|
|
|
"GossipSyncer(%x)", s.cfg.peerPub)
|
|
|
|
|
|
|
|
if err := s.historicalSync(); err != nil {
|
|
|
|
log.Errorf("Unable to attempt initial "+
|
|
|
|
"historical sync with "+
|
|
|
|
"GossipSyncer(%x): %v", s.cfg.peerPub,
|
|
|
|
err)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// Once the historical sync has started, we'll get a
|
|
|
|
// keep track of the corresponding syncer to properly
|
|
|
|
// handle disconnects. We'll also use a signal to know
|
|
|
|
// when the historical sync completed.
|
|
|
|
initialHistoricalSyncer = s
|
|
|
|
initialHistoricalSyncSignal = s.ResetSyncedSignal()
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
// An existing peer has disconnected, so we'll tear down its
|
|
|
|
// corresponding GossipSyncer.
|
|
|
|
case staleSyncer := <-m.staleSyncers:
|
|
|
|
// Once the corresponding GossipSyncer has been stopped
|
|
|
|
// and removed, we'll signal to the caller that they can
|
|
|
|
// proceed since the SyncManager's internal state has
|
|
|
|
// been updated.
|
|
|
|
m.removeGossipSyncer(staleSyncer.peer)
|
|
|
|
close(staleSyncer.doneChan)
|
|
|
|
|
2019-04-11 05:26:56 +03:00
|
|
|
// If we don't have an initialHistoricalSyncer, or we do
|
|
|
|
// but it is not the peer being disconnected, then we
|
|
|
|
// have nothing left to do and can proceed.
|
|
|
|
switch {
|
|
|
|
case initialHistoricalSyncer == nil:
|
|
|
|
fallthrough
|
|
|
|
case staleSyncer.peer != initialHistoricalSyncer.cfg.peerPub:
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, our initialHistoricalSyncer corresponds to
|
|
|
|
// the peer being disconnected, so we'll have to find a
|
|
|
|
// replacement.
|
|
|
|
log.Debug("Finding replacement for intitial " +
|
|
|
|
"historical sync")
|
|
|
|
|
|
|
|
s := m.forceHistoricalSync()
|
|
|
|
if s == nil {
|
|
|
|
log.Debug("No eligible replacement found " +
|
|
|
|
"for initial historical sync")
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Debugf("Replaced initial historical "+
|
|
|
|
"GossipSyncer(%v) with GossipSyncer(%x)",
|
|
|
|
staleSyncer.peer, s.cfg.peerPub)
|
|
|
|
|
|
|
|
initialHistoricalSyncer = s
|
|
|
|
initialHistoricalSyncSignal = s.ResetSyncedSignal()
|
|
|
|
|
|
|
|
// Our initial historical sync signal has completed, so we'll
|
|
|
|
// nil all of the relevant fields as they're no longer needed.
|
|
|
|
case <-initialHistoricalSyncSignal:
|
|
|
|
initialHistoricalSyncer = nil
|
|
|
|
initialHistoricalSyncSignal = nil
|
2019-07-27 04:22:37 +03:00
|
|
|
m.markGraphSynced()
|
2019-04-11 05:27:12 +03:00
|
|
|
|
|
|
|
log.Debug("Initial historical sync completed")
|
|
|
|
|
|
|
|
// With the initial historical sync complete, we can
|
|
|
|
// begin receiving new graph updates at tip. We'll
|
|
|
|
// determine whether we can have any more active
|
|
|
|
// GossipSyncers. If we do, we'll randomly select some
|
|
|
|
// that are currently passive to transition.
|
|
|
|
m.syncersMu.Lock()
|
|
|
|
numActiveLeft := m.cfg.NumActiveSyncers - len(m.activeSyncers)
|
|
|
|
if numActiveLeft <= 0 {
|
|
|
|
m.syncersMu.Unlock()
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Debugf("Attempting to transition %v passive "+
|
|
|
|
"GossipSyncers to active", numActiveLeft)
|
|
|
|
|
|
|
|
for i := 0; i < numActiveLeft; i++ {
|
|
|
|
chooseRandomSyncer(
|
|
|
|
m.inactiveSyncers, m.transitionPassiveSyncer,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
m.syncersMu.Unlock()
|
2019-04-11 05:26:56 +03:00
|
|
|
|
2019-03-23 05:56:03 +03:00
|
|
|
// Our RotateTicker has ticked, so we'll attempt to rotate a
|
|
|
|
// single active syncer with a passive one.
|
|
|
|
case <-m.cfg.RotateTicker.Ticks():
|
|
|
|
m.rotateActiveSyncerCandidate()
|
|
|
|
|
|
|
|
// Our HistoricalSyncTicker has ticked, so we'll randomly select
|
|
|
|
// a peer and force a historical sync with them.
|
|
|
|
case <-m.cfg.HistoricalSyncTicker.Ticks():
|
2019-08-02 00:04:45 +03:00
|
|
|
s := m.forceHistoricalSync()
|
|
|
|
|
2019-08-14 04:23:05 +03:00
|
|
|
// If we don't have a syncer available or we've already
|
|
|
|
// performed our initial historical sync, then we have
|
|
|
|
// nothing left to do.
|
|
|
|
if s == nil || m.IsGraphSynced() {
|
2019-08-02 00:04:45 +03:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we'll track the peer we've performed a
|
|
|
|
// historical sync with in order to handle the case
|
|
|
|
// where our previous historical sync peer did not
|
|
|
|
// respond to our queries and we haven't ingested as
|
|
|
|
// much of the graph as we should.
|
|
|
|
initialHistoricalSyncer = s
|
|
|
|
initialHistoricalSyncSignal = s.ResetSyncedSignal()
|
2019-03-23 05:56:03 +03:00
|
|
|
|
|
|
|
case <-m.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
// createGossipSyncer creates the GossipSyncer for a newly connected peer.
|
|
|
|
func (m *SyncManager) createGossipSyncer(peer lnpeer.Peer) *GossipSyncer {
|
2019-04-05 18:36:11 +03:00
|
|
|
nodeID := route.Vertex(peer.PubKey())
|
2019-04-11 05:26:08 +03:00
|
|
|
log.Infof("Creating new GossipSyncer for peer=%x", nodeID[:])
|
|
|
|
|
|
|
|
encoding := lnwire.EncodingSortedPlain
|
|
|
|
s := newGossipSyncer(gossipSyncerCfg{
|
|
|
|
chainHash: m.cfg.ChainHash,
|
|
|
|
peerPub: nodeID,
|
|
|
|
channelSeries: m.cfg.ChanSeries,
|
|
|
|
encodingType: encoding,
|
|
|
|
chunkSize: encodingTypeToChunkSize[encoding],
|
|
|
|
batchSize: requestBatchSize,
|
|
|
|
sendToPeer: func(msgs ...lnwire.Message) error {
|
|
|
|
return peer.SendMessageLazy(false, msgs...)
|
|
|
|
},
|
2019-04-27 06:05:10 +03:00
|
|
|
sendToPeerSync: func(msgs ...lnwire.Message) error {
|
|
|
|
return peer.SendMessageLazy(true, msgs...)
|
|
|
|
},
|
2019-07-31 03:25:31 +03:00
|
|
|
ignoreHistoricalFilters: m.cfg.IgnoreHistoricalFilters,
|
2019-04-11 05:26:08 +03:00
|
|
|
})
|
|
|
|
|
|
|
|
// Gossip syncers are initialized by default in a PassiveSync type
|
|
|
|
// and chansSynced state so that they can reply to any peer queries or
|
|
|
|
// handle any sync transitions.
|
|
|
|
s.setSyncState(chansSynced)
|
|
|
|
s.setSyncType(PassiveSync)
|
|
|
|
return s
|
|
|
|
}
|
|
|
|
|
|
|
|
// removeGossipSyncer removes all internal references to the disconnected peer's
|
|
|
|
// GossipSyncer and stops it. In the event of an active GossipSyncer being
|
|
|
|
// disconnected, a passive GossipSyncer, if any, will take its place.
|
2019-04-05 18:36:11 +03:00
|
|
|
func (m *SyncManager) removeGossipSyncer(peer route.Vertex) {
|
2019-04-11 05:26:08 +03:00
|
|
|
m.syncersMu.Lock()
|
|
|
|
defer m.syncersMu.Unlock()
|
|
|
|
|
|
|
|
s, ok := m.gossipSyncer(peer)
|
|
|
|
if !ok {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Infof("Removing GossipSyncer for peer=%v", peer)
|
|
|
|
|
|
|
|
// We'll stop the GossipSyncer for the disconnected peer in a goroutine
|
|
|
|
// to prevent blocking the SyncManager.
|
|
|
|
go s.Stop()
|
|
|
|
|
|
|
|
// If it's a non-active syncer, then we can just exit now.
|
|
|
|
if _, ok := m.inactiveSyncers[peer]; ok {
|
|
|
|
delete(m.inactiveSyncers, peer)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we'll need find a new one to replace it, if any.
|
|
|
|
delete(m.activeSyncers, peer)
|
2019-04-11 05:26:24 +03:00
|
|
|
newActiveSyncer := chooseRandomSyncer(
|
|
|
|
m.inactiveSyncers, m.transitionPassiveSyncer,
|
|
|
|
)
|
2019-04-11 05:26:08 +03:00
|
|
|
if newActiveSyncer == nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Debugf("Replaced active GossipSyncer(%x) with GossipSyncer(%x)",
|
|
|
|
peer, newActiveSyncer.cfg.peerPub)
|
|
|
|
}
|
|
|
|
|
2019-03-23 05:56:03 +03:00
|
|
|
// rotateActiveSyncerCandidate rotates a single active syncer. In order to
|
|
|
|
// achieve this, the active syncer must be in a chansSynced state in order to
|
|
|
|
// process the sync transition.
|
|
|
|
func (m *SyncManager) rotateActiveSyncerCandidate() {
|
2019-04-11 05:26:08 +03:00
|
|
|
m.syncersMu.Lock()
|
|
|
|
defer m.syncersMu.Unlock()
|
|
|
|
|
2019-04-11 05:26:24 +03:00
|
|
|
// If we couldn't find an eligible active syncer to rotate, we can
|
|
|
|
// return early.
|
|
|
|
activeSyncer := chooseRandomSyncer(m.activeSyncers, nil)
|
|
|
|
if activeSyncer == nil {
|
|
|
|
log.Debug("No eligible active syncer to rotate")
|
2019-03-23 05:56:03 +03:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:26:24 +03:00
|
|
|
// Similarly, if we don't have a candidate to rotate with, we can return
|
|
|
|
// early as well.
|
|
|
|
candidate := chooseRandomSyncer(m.inactiveSyncers, nil)
|
|
|
|
if candidate == nil {
|
|
|
|
log.Debug("No eligible candidate to rotate active syncer")
|
2019-03-23 05:56:03 +03:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
// Otherwise, we'll attempt to transition each syncer to their
|
|
|
|
// respective new sync type.
|
2019-04-11 05:25:53 +03:00
|
|
|
log.Debugf("Rotating active GossipSyncer(%x) with GossipSyncer(%x)",
|
|
|
|
activeSyncer.cfg.peerPub, candidate.cfg.peerPub)
|
|
|
|
|
2019-03-23 05:56:03 +03:00
|
|
|
if err := m.transitionActiveSyncer(activeSyncer); err != nil {
|
2019-04-11 05:25:53 +03:00
|
|
|
log.Errorf("Unable to transition active GossipSyncer(%x): %v",
|
|
|
|
activeSyncer.cfg.peerPub, err)
|
2019-03-23 05:56:03 +03:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:25:53 +03:00
|
|
|
if err := m.transitionPassiveSyncer(candidate); err != nil {
|
|
|
|
log.Errorf("Unable to transition passive GossipSyncer(%x): %v",
|
|
|
|
activeSyncer.cfg.peerPub, err)
|
|
|
|
return
|
|
|
|
}
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// transitionActiveSyncer transitions an active syncer to a passive one.
|
2019-04-11 05:26:08 +03:00
|
|
|
//
|
|
|
|
// NOTE: This must be called with the syncersMu lock held.
|
2019-03-23 05:56:03 +03:00
|
|
|
func (m *SyncManager) transitionActiveSyncer(s *GossipSyncer) error {
|
|
|
|
log.Debugf("Transitioning active GossipSyncer(%x) to passive",
|
|
|
|
s.cfg.peerPub)
|
|
|
|
|
|
|
|
if err := s.ProcessSyncTransition(PassiveSync); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:25:53 +03:00
|
|
|
delete(m.activeSyncers, s.cfg.peerPub)
|
|
|
|
m.inactiveSyncers[s.cfg.peerPub] = s
|
2019-03-23 05:56:03 +03:00
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// transitionPassiveSyncer transitions a passive syncer to an active one.
|
2019-04-11 05:26:08 +03:00
|
|
|
//
|
|
|
|
// NOTE: This must be called with the syncersMu lock held.
|
2019-03-23 05:56:03 +03:00
|
|
|
func (m *SyncManager) transitionPassiveSyncer(s *GossipSyncer) error {
|
|
|
|
log.Debugf("Transitioning passive GossipSyncer(%x) to active",
|
|
|
|
s.cfg.peerPub)
|
|
|
|
|
|
|
|
if err := s.ProcessSyncTransition(ActiveSync); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:25:53 +03:00
|
|
|
delete(m.inactiveSyncers, s.cfg.peerPub)
|
2019-03-23 05:56:03 +03:00
|
|
|
m.activeSyncers[s.cfg.peerPub] = s
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// forceHistoricalSync chooses a syncer with a remote peer at random and forces
|
|
|
|
// a historical sync with it.
|
2019-04-11 05:26:56 +03:00
|
|
|
func (m *SyncManager) forceHistoricalSync() *GossipSyncer {
|
2019-04-11 05:26:08 +03:00
|
|
|
m.syncersMu.Lock()
|
|
|
|
defer m.syncersMu.Unlock()
|
2019-03-23 05:56:03 +03:00
|
|
|
|
2019-04-11 05:26:24 +03:00
|
|
|
// We'll sample from both sets of active and inactive syncers in the
|
|
|
|
// event that we don't have any inactive syncers.
|
2019-04-11 05:26:56 +03:00
|
|
|
return chooseRandomSyncer(m.gossipSyncers(), func(s *GossipSyncer) error {
|
2019-04-11 05:26:24 +03:00
|
|
|
return s.historicalSync()
|
|
|
|
})
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
|
2019-04-11 05:26:24 +03:00
|
|
|
// chooseRandomSyncer iterates through the set of syncers given and returns the
|
|
|
|
// first one which was able to successfully perform the action enclosed in the
|
|
|
|
// function closure.
|
2019-03-23 05:56:03 +03:00
|
|
|
//
|
|
|
|
// NOTE: It's possible for a nil value to be returned if there are no eligible
|
|
|
|
// candidate syncers.
|
2019-04-05 18:36:11 +03:00
|
|
|
func chooseRandomSyncer(syncers map[route.Vertex]*GossipSyncer,
|
2019-04-11 05:26:24 +03:00
|
|
|
action func(*GossipSyncer) error) *GossipSyncer {
|
2019-03-23 05:56:03 +03:00
|
|
|
|
2019-04-11 05:26:24 +03:00
|
|
|
for _, s := range syncers {
|
2019-03-23 05:56:03 +03:00
|
|
|
// Only syncers in a chansSynced state are viable for sync
|
|
|
|
// transitions, so skip any that aren't.
|
2019-04-11 05:26:24 +03:00
|
|
|
if s.syncState() != chansSynced {
|
2019-03-23 05:56:03 +03:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:26:24 +03:00
|
|
|
if action != nil {
|
|
|
|
if err := action(s); err != nil {
|
|
|
|
log.Debugf("Skipping eligible candidate "+
|
|
|
|
"GossipSyncer(%x): %v", s.cfg.peerPub,
|
|
|
|
err)
|
2019-03-23 05:56:03 +03:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
2019-04-11 05:26:24 +03:00
|
|
|
|
|
|
|
return s
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// InitSyncState is called by outside sub-systems when a connection is
|
|
|
|
// established to a new peer that understands how to perform channel range
|
|
|
|
// queries. We'll allocate a new GossipSyncer for it, and start any goroutines
|
|
|
|
// needed to handle new queries. The first GossipSyncer registered with the
|
|
|
|
// SyncManager will attempt a historical sync to ensure we have as much of the
|
|
|
|
// public channel graph as possible.
|
|
|
|
//
|
|
|
|
// TODO(wilmer): Only mark as ActiveSync if this isn't a channel peer.
|
2019-04-11 05:26:08 +03:00
|
|
|
func (m *SyncManager) InitSyncState(peer lnpeer.Peer) error {
|
|
|
|
done := make(chan struct{})
|
|
|
|
|
|
|
|
select {
|
|
|
|
case m.newSyncers <- &newSyncer{
|
|
|
|
peer: peer,
|
|
|
|
doneChan: done,
|
|
|
|
}:
|
|
|
|
case <-m.quit:
|
|
|
|
return ErrSyncManagerExiting
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
select {
|
|
|
|
case <-done:
|
|
|
|
return nil
|
|
|
|
case <-m.quit:
|
|
|
|
return ErrSyncManagerExiting
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// PruneSyncState is called by outside sub-systems once a peer that we were
|
|
|
|
// previously connected to has been disconnected. In this case we can stop the
|
|
|
|
// existing GossipSyncer assigned to the peer and free up resources.
|
2019-04-05 18:36:11 +03:00
|
|
|
func (m *SyncManager) PruneSyncState(peer route.Vertex) {
|
2019-04-11 05:26:08 +03:00
|
|
|
done := make(chan struct{})
|
|
|
|
|
|
|
|
// We avoid returning an error when the SyncManager is stopped since the
|
|
|
|
// GossipSyncer will be stopped then anyway.
|
|
|
|
select {
|
|
|
|
case m.staleSyncers <- &staleSyncer{
|
|
|
|
peer: peer,
|
|
|
|
doneChan: done,
|
|
|
|
}:
|
|
|
|
case <-m.quit:
|
2019-03-23 05:56:03 +03:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2019-04-11 05:26:08 +03:00
|
|
|
select {
|
|
|
|
case <-done:
|
|
|
|
case <-m.quit:
|
2019-03-23 05:56:03 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// GossipSyncer returns the associated gossip syncer of a peer. The boolean
|
|
|
|
// returned signals whether there exists a gossip syncer for the peer.
|
2019-04-05 18:36:11 +03:00
|
|
|
func (m *SyncManager) GossipSyncer(peer route.Vertex) (*GossipSyncer, bool) {
|
2019-04-11 05:26:08 +03:00
|
|
|
m.syncersMu.Lock()
|
|
|
|
defer m.syncersMu.Unlock()
|
2019-03-23 05:56:03 +03:00
|
|
|
return m.gossipSyncer(peer)
|
|
|
|
}
|
|
|
|
|
|
|
|
// gossipSyncer returns the associated gossip syncer of a peer. The boolean
|
|
|
|
// returned signals whether there exists a gossip syncer for the peer.
|
2019-04-05 18:36:11 +03:00
|
|
|
func (m *SyncManager) gossipSyncer(peer route.Vertex) (*GossipSyncer, bool) {
|
2019-03-23 05:56:03 +03:00
|
|
|
syncer, ok := m.inactiveSyncers[peer]
|
|
|
|
if ok {
|
|
|
|
return syncer, true
|
|
|
|
}
|
|
|
|
syncer, ok = m.activeSyncers[peer]
|
|
|
|
if ok {
|
|
|
|
return syncer, true
|
|
|
|
}
|
|
|
|
return nil, false
|
|
|
|
}
|
|
|
|
|
|
|
|
// GossipSyncers returns all of the currently initialized gossip syncers.
|
2019-04-05 18:36:11 +03:00
|
|
|
func (m *SyncManager) GossipSyncers() map[route.Vertex]*GossipSyncer {
|
2019-04-11 05:26:08 +03:00
|
|
|
m.syncersMu.Lock()
|
|
|
|
defer m.syncersMu.Unlock()
|
2019-04-11 05:26:24 +03:00
|
|
|
return m.gossipSyncers()
|
|
|
|
}
|
2019-03-23 05:56:03 +03:00
|
|
|
|
2019-04-11 05:26:24 +03:00
|
|
|
// gossipSyncers returns all of the currently initialized gossip syncers.
|
2019-04-05 18:36:11 +03:00
|
|
|
func (m *SyncManager) gossipSyncers() map[route.Vertex]*GossipSyncer {
|
2019-04-11 05:25:53 +03:00
|
|
|
numSyncers := len(m.inactiveSyncers) + len(m.activeSyncers)
|
2019-04-05 18:36:11 +03:00
|
|
|
syncers := make(map[route.Vertex]*GossipSyncer, numSyncers)
|
2019-03-23 05:56:03 +03:00
|
|
|
|
|
|
|
for _, syncer := range m.inactiveSyncers {
|
|
|
|
syncers[syncer.cfg.peerPub] = syncer
|
|
|
|
}
|
|
|
|
for _, syncer := range m.activeSyncers {
|
|
|
|
syncers[syncer.cfg.peerPub] = syncer
|
|
|
|
}
|
|
|
|
|
|
|
|
return syncers
|
|
|
|
}
|
2019-07-27 04:22:37 +03:00
|
|
|
|
|
|
|
// markGraphSynced allows us to report that the initial historical sync has
|
|
|
|
// completed.
|
|
|
|
func (m *SyncManager) markGraphSynced() {
|
|
|
|
atomic.StoreInt32(&m.initialHistoricalSyncCompleted, 1)
|
|
|
|
}
|
|
|
|
|
|
|
|
// markGraphSyncing allows us to report that the initial historical sync is
|
|
|
|
// still undergoing.
|
|
|
|
func (m *SyncManager) markGraphSyncing() {
|
|
|
|
atomic.StoreInt32(&m.initialHistoricalSyncCompleted, 0)
|
|
|
|
}
|
|
|
|
|
|
|
|
// IsGraphSynced determines whether we've completed our initial historical sync.
|
|
|
|
// The initial historical sync is done to ensure we've ingested as much of the
|
|
|
|
// public graph as possible.
|
|
|
|
func (m *SyncManager) IsGraphSynced() bool {
|
|
|
|
return atomic.LoadInt32(&m.initialHistoricalSyncCompleted) == 1
|
|
|
|
}
|