2018-01-17 06:52:22 +03:00
|
|
|
package contractcourt
|
|
|
|
|
|
|
|
import (
|
2018-09-12 02:05:26 +03:00
|
|
|
"errors"
|
2018-01-17 06:52:22 +03:00
|
|
|
"fmt"
|
|
|
|
"sync"
|
|
|
|
"sync/atomic"
|
|
|
|
|
2018-05-31 08:07:54 +03:00
|
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
|
|
"github.com/btcsuite/btcd/wire"
|
|
|
|
"github.com/btcsuite/btcutil"
|
2018-01-17 06:52:22 +03:00
|
|
|
"github.com/lightningnetwork/lnd/chainntnfs"
|
|
|
|
"github.com/lightningnetwork/lnd/channeldb"
|
|
|
|
"github.com/lightningnetwork/lnd/lnwallet"
|
|
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
2019-01-23 07:45:35 +03:00
|
|
|
"github.com/lightningnetwork/lnd/sweep"
|
2018-01-17 06:52:22 +03:00
|
|
|
)
|
|
|
|
|
2018-09-12 02:05:26 +03:00
|
|
|
// ErrChainArbExiting signals that the chain arbitrator is shutting down.
|
|
|
|
var ErrChainArbExiting = errors.New("ChainArbitrator exiting")
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
// ResolutionMsg is a message sent by resolvers to outside sub-systems once an
|
|
|
|
// outgoing contract has been fully resolved. For multi-hop contracts, if we
|
|
|
|
// resolve the outgoing contract, we'll also need to ensure that the incoming
|
|
|
|
// contract is resolved as well. We package the items required to resolve the
|
|
|
|
// incoming contracts within this message.
|
|
|
|
type ResolutionMsg struct {
|
|
|
|
// SourceChan identifies the channel that this message is being sent
|
|
|
|
// from. This is the channel's short channel ID.
|
|
|
|
SourceChan lnwire.ShortChannelID
|
|
|
|
|
|
|
|
// HtlcIndex is the index of the contract within the original
|
|
|
|
// commitment trace.
|
|
|
|
HtlcIndex uint64
|
|
|
|
|
|
|
|
// Failure will be non-nil if the incoming contract should be cancelled
|
|
|
|
// all together. This can happen if the outgoing contract was dust, if
|
|
|
|
// if the outgoing HTLC timed out.
|
|
|
|
Failure lnwire.FailureMessage
|
|
|
|
|
|
|
|
// PreImage will be non-nil if the incoming contract can successfully
|
|
|
|
// be redeemed. This can happen if we learn of the preimage from the
|
|
|
|
// outgoing HTLC on-chain.
|
|
|
|
PreImage *[32]byte
|
|
|
|
}
|
|
|
|
|
|
|
|
// ChainArbitratorConfig is a configuration struct that contains all the
|
|
|
|
// function closures and interface that required to arbitrate on-chain
|
|
|
|
// contracts for a particular chain.
|
|
|
|
type ChainArbitratorConfig struct {
|
|
|
|
// ChainHash is the chain that this arbitrator is to operate within.
|
|
|
|
ChainHash chainhash.Hash
|
|
|
|
|
|
|
|
// BroadcastDelta is the delta that we'll use to decide when to
|
|
|
|
// broadcast our commitment transaction. This value should be set
|
|
|
|
// based on our current fee estimation of the commitment transaction.
|
|
|
|
// We use this to determine when we should broadcast instead of the
|
|
|
|
// just the HTLC timeout, as we want to ensure that the commitment
|
|
|
|
// transaction is already confirmed, by the time the HTLC expires.
|
|
|
|
BroadcastDelta uint32
|
|
|
|
|
|
|
|
// NewSweepAddr is a function that returns a new address under control
|
|
|
|
// by the wallet. We'll use this to sweep any no-delay outputs as a
|
|
|
|
// result of unilateral channel closes.
|
|
|
|
//
|
|
|
|
// NOTE: This SHOULD return a p2wkh script.
|
|
|
|
NewSweepAddr func() ([]byte, error)
|
|
|
|
|
|
|
|
// PublishTx reliably broadcasts a transaction to the network. Once
|
|
|
|
// this function exits without an error, then they transaction MUST
|
|
|
|
// continually be rebroadcast if needed.
|
|
|
|
PublishTx func(*wire.MsgTx) error
|
|
|
|
|
|
|
|
// DeliverResolutionMsg is a function that will append an outgoing
|
|
|
|
// message to the "out box" for a ChannelLink. This is used to cancel
|
|
|
|
// backwards any HTLC's that are either dust, we're timing out, or
|
|
|
|
// settling on-chain to the incoming link.
|
|
|
|
DeliverResolutionMsg func(...ResolutionMsg) error
|
|
|
|
|
|
|
|
// MarkLinkInactive is a function closure that the ChainArbitrator will
|
|
|
|
// use to mark that active HTLC's shouldn't be attempt ted to be routed
|
|
|
|
// over a particular channel. This function will be called in that a
|
|
|
|
// ChannelArbitrator decides that it needs to go to chain in order to
|
|
|
|
// resolve contracts.
|
|
|
|
//
|
2018-01-20 04:17:14 +03:00
|
|
|
// TODO(roasbeef): rename, routing based
|
2018-01-17 06:52:22 +03:00
|
|
|
MarkLinkInactive func(wire.OutPoint) error
|
|
|
|
|
2018-04-18 14:41:49 +03:00
|
|
|
// ContractBreach is a function closure that the ChainArbitrator will
|
|
|
|
// use to notify the breachArbiter about a contract breach. It should
|
|
|
|
// only return a non-nil error when the breachArbiter has preserved the
|
|
|
|
// necessary breach info for this channel point, and it is safe to mark
|
|
|
|
// the channel as pending close in the database.
|
|
|
|
ContractBreach func(wire.OutPoint, *lnwallet.BreachRetribution) error
|
|
|
|
|
2018-01-20 04:17:14 +03:00
|
|
|
// IsOurAddress is a function that returns true if the passed address
|
|
|
|
// is known to the underlying wallet. Otherwise, false should be
|
|
|
|
// returned.
|
|
|
|
IsOurAddress func(btcutil.Address) bool
|
|
|
|
|
2018-04-18 05:02:04 +03:00
|
|
|
// IncubateOutput sends either an incoming HTLC, an outgoing HTLC, or
|
2018-01-17 06:52:22 +03:00
|
|
|
// both to the utxo nursery. Once this function returns, the nursery
|
|
|
|
// should have safely persisted the outputs to disk, and should start
|
|
|
|
// the process of incubation. This is used when a resolver wishes to
|
2018-02-07 06:11:11 +03:00
|
|
|
// pass off the output to the nursery as we're only waiting on an
|
2018-01-17 06:52:22 +03:00
|
|
|
// absolute/relative item block.
|
|
|
|
IncubateOutputs func(wire.OutPoint, *lnwallet.CommitOutputResolution,
|
|
|
|
*lnwallet.OutgoingHtlcResolution,
|
2018-09-05 16:43:01 +03:00
|
|
|
*lnwallet.IncomingHtlcResolution, uint32) error
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
// PreimageDB is a global store of all known pre-images. We'll use this
|
|
|
|
// to decide if we should broadcast a commitment transaction to claim
|
|
|
|
// an HTLC on-chain.
|
|
|
|
PreimageDB WitnessBeacon
|
|
|
|
|
|
|
|
// Notifier is an instance of a chain notifier we'll use to watch for
|
|
|
|
// certain on-chain events.
|
|
|
|
Notifier chainntnfs.ChainNotifier
|
|
|
|
|
|
|
|
// Signer is a signer backed by the active lnd node. This should be
|
|
|
|
// capable of producing a signature as specified by a valid
|
|
|
|
// SignDescriptor.
|
|
|
|
Signer lnwallet.Signer
|
|
|
|
|
|
|
|
// FeeEstimator will be used to return fee estimates.
|
|
|
|
FeeEstimator lnwallet.FeeEstimator
|
|
|
|
|
|
|
|
// ChainIO allows us to query the state of the current main chain.
|
|
|
|
ChainIO lnwallet.BlockChainIO
|
2018-06-14 05:43:42 +03:00
|
|
|
|
|
|
|
// DisableChannel disables a channel, resulting in it not being able to
|
|
|
|
// forward payments.
|
|
|
|
DisableChannel func(wire.OutPoint) error
|
2018-09-26 19:46:48 +03:00
|
|
|
|
|
|
|
// Sweeper allows resolvers to sweep their final outputs.
|
|
|
|
Sweeper *sweep.UtxoSweeper
|
2019-01-23 07:45:35 +03:00
|
|
|
|
|
|
|
// SettleInvoice attempts to settle an existing invoice on-chain with
|
|
|
|
// the given payment hash. ErrInvoiceNotFound is returned if an invoice
|
|
|
|
// is not found.
|
|
|
|
SettleInvoice func(chainhash.Hash, lnwire.MilliSatoshi) error
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// ChainArbitrator is a sub-system that oversees the on-chain resolution of all
|
|
|
|
// active, and channel that are in the "pending close" state. Within the
|
|
|
|
// contractcourt package, the ChainArbitrator manages a set of active
|
|
|
|
// ContractArbitrators. Each ContractArbitrators is responsible for watching
|
|
|
|
// the chain for any activity that affects the state of the channel, and also
|
|
|
|
// for monitoring each contract in order to determine if any on-chain activity is
|
|
|
|
// required. Outside sub-systems interact with the ChainArbitrator in order to
|
|
|
|
// forcibly exit a contract, update the set of live signals for each contract,
|
|
|
|
// and to receive reports on the state of contract resolution.
|
|
|
|
type ChainArbitrator struct {
|
2018-06-01 01:41:41 +03:00
|
|
|
started int32 // To be used atomically.
|
|
|
|
stopped int32 // To be used atomically.
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
sync.Mutex
|
|
|
|
|
|
|
|
// activeChannels is a map of all the active contracts that are still
|
|
|
|
// open, and not fully resolved.
|
|
|
|
activeChannels map[wire.OutPoint]*ChannelArbitrator
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// activeWatchers is a map of all the active chainWatchers for channels
|
|
|
|
// that are still considered open.
|
|
|
|
activeWatchers map[wire.OutPoint]*chainWatcher
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
// cfg is the config struct for the arbitrator that contains all
|
|
|
|
// methods and interface it needs to operate.
|
|
|
|
cfg ChainArbitratorConfig
|
|
|
|
|
|
|
|
// chanSource will be used by the ChainArbitrator to fetch all the
|
|
|
|
// active channels that it must still watch over.
|
|
|
|
chanSource *channeldb.DB
|
|
|
|
|
|
|
|
quit chan struct{}
|
|
|
|
|
|
|
|
wg sync.WaitGroup
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewChainArbitrator returns a new instance of the ChainArbitrator using the
|
|
|
|
// passed config struct, and backing persistent database.
|
|
|
|
func NewChainArbitrator(cfg ChainArbitratorConfig,
|
|
|
|
db *channeldb.DB) *ChainArbitrator {
|
|
|
|
|
|
|
|
return &ChainArbitrator{
|
|
|
|
cfg: cfg,
|
|
|
|
activeChannels: make(map[wire.OutPoint]*ChannelArbitrator),
|
2018-01-19 01:00:35 +03:00
|
|
|
activeWatchers: make(map[wire.OutPoint]*chainWatcher),
|
2018-01-17 06:52:22 +03:00
|
|
|
chanSource: db,
|
|
|
|
quit: make(chan struct{}),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// newActiveChannelArbitrator creates a new instance of an active channel
|
|
|
|
// arbitrator given the state of the target channel.
|
|
|
|
func newActiveChannelArbitrator(channel *channeldb.OpenChannel,
|
2018-01-19 01:00:35 +03:00
|
|
|
c *ChainArbitrator, chanEvents *ChainEventSubscription) (*ChannelArbitrator, error) {
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
log.Tracef("Creating ChannelArbitrator for ChannelPoint(%v)",
|
|
|
|
channel.FundingOutpoint)
|
|
|
|
|
|
|
|
// We'll start by registering for a block epoch notifications so this
|
|
|
|
// channel can keep track of the current state of the main chain.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): fetch best height (or pass in) so can ensure block
|
|
|
|
// epoch delivers all the notifications to
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): instead 1 block epoch that multi-plexes to the rest?
|
|
|
|
// * reduces the number of goroutines
|
2018-08-09 10:05:27 +03:00
|
|
|
blockEpoch, err := c.cfg.Notifier.RegisterBlockEpochNtfn(nil)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
chanPoint := channel.FundingOutpoint
|
|
|
|
|
|
|
|
// Next we'll create the matching configuration struct that contains
|
|
|
|
// all interfaces and methods the arbitrator needs to do its job.
|
|
|
|
arbCfg := ChannelArbitratorConfig{
|
|
|
|
ChanPoint: chanPoint,
|
2018-05-02 02:27:20 +03:00
|
|
|
ShortChanID: channel.ShortChanID(),
|
2018-01-17 06:52:22 +03:00
|
|
|
BlockEpochs: blockEpoch,
|
2018-03-19 17:19:19 +03:00
|
|
|
ForceCloseChan: func() (*lnwallet.LocalForceCloseSummary, error) {
|
2018-01-17 06:52:22 +03:00
|
|
|
// With the channels fetched, attempt to locate
|
|
|
|
// the target channel according to its channel
|
|
|
|
// point.
|
|
|
|
dbChannels, err := c.chanSource.FetchAllChannels()
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
var channel *channeldb.OpenChannel
|
|
|
|
for _, dbChannel := range dbChannels {
|
|
|
|
if dbChannel.FundingOutpoint == chanPoint {
|
|
|
|
channel = dbChannel
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the channel cannot be located, then we
|
|
|
|
// exit with an error to the channel.
|
|
|
|
if channel == nil {
|
|
|
|
return nil, fmt.Errorf("unable to find channel")
|
|
|
|
}
|
|
|
|
|
|
|
|
chanMachine, err := lnwallet.NewLightningChannel(
|
multi: replace per channel sigPool with global daemon level sigPool
In this commit, we remove the per channel `sigPool` within the
`lnwallet.LightningChannel` struct. With this change, we ensure that as
the number of channels grows, the number of gouroutines idling in the
sigPool stays constant. It's the case that currently on the daemon, most
channels are likely inactive, with only a hand full actually
consistently carrying out channel updates. As a result, this change
should reduce the amount of idle CPU usage, as we have less active
goroutines in select loops.
In order to make this change, the `SigPool` itself has been publicly
exported such that outside callers can make a `SigPool` and pass it into
newly created channels. Since the sig pool now lives outside the
channel, we were also able to do away with the Stop() method on the
channel all together.
Finally, the server is the sub-system that is currently responsible for
managing the `SigPool` within lnd.
2018-12-15 03:35:07 +03:00
|
|
|
c.cfg.Signer, c.cfg.PreimageDB, channel, nil,
|
|
|
|
)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := c.cfg.MarkLinkInactive(chanPoint); err != nil {
|
|
|
|
log.Errorf("unable to mark link inactive: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return chanMachine.ForceClose()
|
|
|
|
},
|
2018-04-03 15:04:30 +03:00
|
|
|
MarkCommitmentBroadcasted: channel.MarkCommitmentBroadcasted,
|
2018-08-21 13:21:15 +03:00
|
|
|
MarkChannelClosed: channel.CloseChannel,
|
2018-08-21 13:21:16 +03:00
|
|
|
IsPendingClose: false,
|
2018-04-03 15:53:51 +03:00
|
|
|
ChainArbitratorConfig: c.cfg,
|
|
|
|
ChainEvents: chanEvents,
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// The final component needed is an arbitrator log that the arbitrator
|
|
|
|
// will use to keep track of its internal state using a backed
|
|
|
|
// persistent log.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef); abstraction leak...
|
|
|
|
// * rework: adaptor method to set log scope w/ factory func
|
|
|
|
chanLog, err := newBoltArbitratorLog(
|
|
|
|
c.chanSource.DB, arbCfg, c.cfg.ChainHash, chanPoint,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
blockEpoch.Cancel()
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
arbCfg.MarkChannelResolved = func() error {
|
|
|
|
return c.resolveContract(chanPoint, chanLog)
|
|
|
|
}
|
|
|
|
|
|
|
|
return NewChannelArbitrator(
|
|
|
|
arbCfg, channel.LocalCommitment.Htlcs, chanLog,
|
|
|
|
), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// resolveContract marks a contract as fully resolved within the database.
|
|
|
|
// This is only to be done once all contracts which were live on the channel
|
|
|
|
// before hitting the chain have been resolved.
|
|
|
|
func (c *ChainArbitrator) resolveContract(chanPoint wire.OutPoint,
|
|
|
|
arbLog ArbitratorLog) error {
|
|
|
|
|
|
|
|
log.Infof("Marking ChannelPoint(%v) fully resolved", chanPoint)
|
|
|
|
|
|
|
|
// First, we'll we'll mark the channel as fully closed from the PoV of
|
|
|
|
// the channel source.
|
|
|
|
err := c.chanSource.MarkChanFullyClosed(&chanPoint)
|
|
|
|
if err != nil {
|
2018-04-13 03:13:28 +03:00
|
|
|
log.Errorf("ChainArbitrator: unable to mark ChannelPoint(%v) "+
|
|
|
|
"fully closed: %v", chanPoint, err)
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-01-20 04:13:41 +03:00
|
|
|
if arbLog != nil {
|
|
|
|
// Once this has been marked as resolved, we'll wipe the log
|
|
|
|
// that the channel arbitrator was using to store its
|
|
|
|
// persistent state. We do this after marking the channel
|
|
|
|
// resolved, as otherwise, the arbitrator would be re-created,
|
|
|
|
// and think it was starting from the default state.
|
|
|
|
if err := arbLog.WipeHistory(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
c.Lock()
|
|
|
|
delete(c.activeChannels, chanPoint)
|
2018-01-20 04:13:41 +03:00
|
|
|
|
|
|
|
chainWatcher, ok := c.activeWatchers[chanPoint]
|
|
|
|
if ok {
|
|
|
|
chainWatcher.Stop()
|
|
|
|
}
|
2018-06-19 14:25:30 +03:00
|
|
|
delete(c.activeWatchers, chanPoint)
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Unlock()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start launches all goroutines that the ChainArbitrator needs to operate.
|
|
|
|
func (c *ChainArbitrator) Start() error {
|
|
|
|
if !atomic.CompareAndSwapInt32(&c.started, 0, 1) {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Tracef("Starting ChainArbitrator")
|
|
|
|
|
|
|
|
// First, we'll fetch all the channels that are still open, in order to
|
|
|
|
// collect them within our set of active contracts.
|
|
|
|
openChannels, err := c.chanSource.FetchAllChannels()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(openChannels) > 0 {
|
|
|
|
log.Infof("Creating ChannelArbitrators for %v active channels",
|
|
|
|
len(openChannels))
|
|
|
|
}
|
|
|
|
|
|
|
|
// For each open channel, we'll configure then launch a corresponding
|
|
|
|
// ChannelArbitrator.
|
|
|
|
for _, channel := range openChannels {
|
2018-04-13 04:53:22 +03:00
|
|
|
chanPoint := channel.FundingOutpoint
|
2018-05-19 02:21:06 +03:00
|
|
|
channel := channel
|
2018-04-13 04:53:22 +03:00
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// First, we'll create an active chainWatcher for this channel
|
|
|
|
// to ensure that we detect any relevant on chain events.
|
|
|
|
chainWatcher, err := newChainWatcher(
|
2018-04-19 14:05:50 +03:00
|
|
|
chainWatcherConfig{
|
|
|
|
chanState: channel,
|
|
|
|
notifier: c.cfg.Notifier,
|
|
|
|
pCache: c.cfg.PreimageDB,
|
|
|
|
signer: c.cfg.Signer,
|
|
|
|
isOurAddr: c.cfg.IsOurAddress,
|
2018-04-18 14:41:49 +03:00
|
|
|
contractBreach: func(retInfo *lnwallet.BreachRetribution) error {
|
|
|
|
return c.cfg.ContractBreach(chanPoint, retInfo)
|
|
|
|
},
|
2018-01-20 04:17:14 +03:00
|
|
|
},
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-13 04:53:22 +03:00
|
|
|
c.activeWatchers[chanPoint] = chainWatcher
|
2018-01-19 01:00:35 +03:00
|
|
|
channelArb, err := newActiveChannelArbitrator(
|
2018-04-18 14:41:49 +03:00
|
|
|
channel, c, chainWatcher.SubscribeChannelEvents(),
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-13 04:53:22 +03:00
|
|
|
c.activeChannels[chanPoint] = channelArb
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// In addition to the channels that we know to be open, we'll also
|
|
|
|
// launch arbitrators to finishing resolving any channels that are in
|
|
|
|
// the pending close state.
|
|
|
|
closingChannels, err := c.chanSource.FetchClosedChannels(true)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(closingChannels) > 0 {
|
|
|
|
log.Infof("Creating ChannelArbitrators for %v closing channels",
|
|
|
|
len(closingChannels))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next, for each channel is the closing state, we'll launch a
|
2018-05-24 11:24:31 +03:00
|
|
|
// corresponding more restricted resolver, as we don't have to watch
|
|
|
|
// the chain any longer, only resolve the contracts on the confirmed
|
|
|
|
// commitment.
|
2018-01-17 06:52:22 +03:00
|
|
|
for _, closeChanInfo := range closingChannels {
|
2018-08-09 10:05:27 +03:00
|
|
|
blockEpoch, err := c.cfg.Notifier.RegisterBlockEpochNtfn(nil)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// We can leave off the CloseContract and ForceCloseChan
|
|
|
|
// methods as the channel is already closed at this point.
|
|
|
|
chanPoint := closeChanInfo.ChanPoint
|
|
|
|
arbCfg := ChannelArbitratorConfig{
|
|
|
|
ChanPoint: chanPoint,
|
|
|
|
ShortChanID: closeChanInfo.ShortChanID,
|
|
|
|
BlockEpochs: blockEpoch,
|
|
|
|
ChainArbitratorConfig: c.cfg,
|
2018-01-19 01:00:35 +03:00
|
|
|
ChainEvents: &ChainEventSubscription{},
|
2018-08-21 13:21:16 +03:00
|
|
|
IsPendingClose: true,
|
|
|
|
ClosingHeight: closeChanInfo.CloseHeight,
|
|
|
|
CloseType: closeChanInfo.CloseType,
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
chanLog, err := newBoltArbitratorLog(
|
|
|
|
c.chanSource.DB, arbCfg, c.cfg.ChainHash, chanPoint,
|
|
|
|
)
|
|
|
|
if err != nil {
|
2018-02-06 01:41:03 +03:00
|
|
|
blockEpoch.Cancel()
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
arbCfg.MarkChannelResolved = func() error {
|
|
|
|
return c.resolveContract(chanPoint, chanLog)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We can also leave off the set of HTLC's here as since the
|
|
|
|
// channel is already in the process of being full resolved, no
|
|
|
|
// new HTLC's we be added.
|
|
|
|
c.activeChannels[chanPoint] = NewChannelArbitrator(
|
|
|
|
arbCfg, nil, chanLog,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2018-09-12 02:05:26 +03:00
|
|
|
// Now, we'll start all chain watchers in parallel to shorten start up
|
|
|
|
// duration. In neutrino mode, this allows spend registrations to take
|
|
|
|
// advantage of batch spend reporting, instead of doing a single rescan
|
|
|
|
// per chain watcher.
|
|
|
|
//
|
|
|
|
// NOTE: After this point, we Stop the chain arb to ensure that any
|
|
|
|
// lingering goroutines are cleaned up before exiting.
|
|
|
|
watcherErrs := make(chan error, len(c.activeWatchers))
|
|
|
|
var wg sync.WaitGroup
|
2018-01-19 01:00:35 +03:00
|
|
|
for _, watcher := range c.activeWatchers {
|
2018-09-12 02:05:26 +03:00
|
|
|
wg.Add(1)
|
|
|
|
go func(w *chainWatcher) {
|
|
|
|
defer wg.Done()
|
|
|
|
select {
|
|
|
|
case watcherErrs <- w.Start():
|
|
|
|
case <-c.quit:
|
|
|
|
watcherErrs <- ErrChainArbExiting
|
|
|
|
}
|
|
|
|
}(watcher)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Once all chain watchers have been started, seal the err chan to
|
|
|
|
// signal the end of the err stream.
|
|
|
|
go func() {
|
|
|
|
wg.Wait()
|
|
|
|
close(watcherErrs)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Handle all errors returned from spawning our chain watchers. If any
|
|
|
|
// of them failed, we will stop the chain arb to shutdown any active
|
|
|
|
// goroutines.
|
|
|
|
for err := range watcherErrs {
|
|
|
|
if err != nil {
|
|
|
|
c.Stop()
|
2018-01-19 01:00:35 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
2018-09-12 02:05:26 +03:00
|
|
|
|
|
|
|
// Finally, we'll launch all the goroutines for each arbitrator so they
|
|
|
|
// can carry out their duties.
|
2018-01-17 06:52:22 +03:00
|
|
|
for _, arbitrator := range c.activeChannels {
|
|
|
|
if err := arbitrator.Start(); err != nil {
|
2018-09-12 02:05:26 +03:00
|
|
|
c.Stop()
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO(roasbeef): eventually move all breach watching here
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Stop signals the ChainArbitrator to trigger a graceful shutdown. Any active
|
|
|
|
// channel arbitrators will be signalled to exit, and this method will block
|
|
|
|
// until they've all exited.
|
|
|
|
func (c *ChainArbitrator) Stop() error {
|
|
|
|
if !atomic.CompareAndSwapInt32(&c.stopped, 0, 1) {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Infof("Stopping ChainArbitrator")
|
|
|
|
|
|
|
|
close(c.quit)
|
|
|
|
|
2018-08-04 02:43:18 +03:00
|
|
|
var (
|
|
|
|
activeWatchers = make(map[wire.OutPoint]*chainWatcher)
|
|
|
|
activeChannels = make(map[wire.OutPoint]*ChannelArbitrator)
|
|
|
|
)
|
|
|
|
|
|
|
|
// Copy the current set of active watchers and arbitrators to shutdown.
|
|
|
|
// We don't want to hold the lock when shutting down each watcher or
|
|
|
|
// arbitrator individually, as they may need to acquire this mutex.
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Lock()
|
2018-08-04 02:43:18 +03:00
|
|
|
for chanPoint, watcher := range c.activeWatchers {
|
|
|
|
activeWatchers[chanPoint] = watcher
|
|
|
|
}
|
|
|
|
for chanPoint, arbitrator := range c.activeChannels {
|
|
|
|
activeChannels[chanPoint] = arbitrator
|
|
|
|
}
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Unlock()
|
|
|
|
|
2018-08-04 02:43:18 +03:00
|
|
|
for chanPoint, watcher := range activeWatchers {
|
2018-01-19 01:00:35 +03:00
|
|
|
log.Tracef("Attempting to stop ChainWatcher(%v)",
|
|
|
|
chanPoint)
|
|
|
|
|
|
|
|
if err := watcher.Stop(); err != nil {
|
|
|
|
log.Errorf("unable to stop watcher for "+
|
|
|
|
"ChannelPoint(%v): %v", chanPoint, err)
|
|
|
|
}
|
|
|
|
}
|
2018-08-04 02:43:18 +03:00
|
|
|
for chanPoint, arbitrator := range activeChannels {
|
2018-01-17 06:52:22 +03:00
|
|
|
log.Tracef("Attempting to stop ChannelArbitrator(%v)",
|
|
|
|
chanPoint)
|
|
|
|
|
|
|
|
if err := arbitrator.Stop(); err != nil {
|
|
|
|
log.Errorf("unable to stop arbitrator for "+
|
|
|
|
"ChannelPoint(%v): %v", chanPoint, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
c.wg.Wait()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// ContractSignals wraps the two signals that affect the state of a channel
|
|
|
|
// being watched by an arbitrator. The two signals we care about are: the
|
|
|
|
// channel has a new set of HTLC's, and the remote party has just broadcast
|
|
|
|
// their version of the commitment transaction.
|
|
|
|
type ContractSignals struct {
|
|
|
|
// HtlcUpdates is a channel that once we new commitment updates takes
|
|
|
|
// place, the later set of HTLC's on the commitment transaction should
|
|
|
|
// be sent over.
|
|
|
|
HtlcUpdates chan []channeldb.HTLC
|
|
|
|
|
|
|
|
// ShortChanID is the up to date short channel ID for a contract. This
|
|
|
|
// can change either if when the contract was added it didn't yet have
|
|
|
|
// a stable identifier, or in the case of a reorg.
|
|
|
|
ShortChanID lnwire.ShortChannelID
|
|
|
|
}
|
|
|
|
|
|
|
|
// UpdateContractSignals sends a set of active, up to date contract signals to
|
|
|
|
// the ChannelArbitrator which is has been assigned to the channel infield by
|
|
|
|
// the passed channel point.
|
|
|
|
func (c *ChainArbitrator) UpdateContractSignals(chanPoint wire.OutPoint,
|
|
|
|
signals *ContractSignals) error {
|
|
|
|
|
|
|
|
log.Infof("Attempting to update ContractSignals for ChannelPoint(%v)",
|
|
|
|
chanPoint)
|
|
|
|
|
|
|
|
c.Lock()
|
|
|
|
arbitrator, ok := c.activeChannels[chanPoint]
|
|
|
|
c.Unlock()
|
|
|
|
if !ok {
|
|
|
|
return fmt.Errorf("unable to find arbitrator")
|
|
|
|
}
|
|
|
|
|
|
|
|
arbitrator.UpdateContractSignals(signals)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// forceCloseReq is a request sent from an outside sub-system to the arbitrator
|
|
|
|
// that watches a particular channel to broadcast the commitment transaction,
|
2018-01-17 06:52:22 +03:00
|
|
|
// and enter the resolution phase of the channel.
|
|
|
|
type forceCloseReq struct {
|
2018-01-19 01:00:35 +03:00
|
|
|
// errResp is a channel that will be sent upon either in the case of
|
|
|
|
// force close success (nil error), or in the case on an error.
|
2018-01-17 06:52:22 +03:00
|
|
|
//
|
|
|
|
// NOTE; This channel MUST be buffered.
|
|
|
|
errResp chan error
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// closeTx is a channel that carries the transaction which ultimately
|
2018-01-17 06:52:22 +03:00
|
|
|
// closed out the channel.
|
|
|
|
closeTx chan *wire.MsgTx
|
|
|
|
}
|
|
|
|
|
|
|
|
// ForceCloseContract attempts to force close the channel infield by the passed
|
|
|
|
// channel point. A force close will immediately terminate the contract,
|
|
|
|
// causing it to enter the resolution phase. If the force close was successful,
|
|
|
|
// then the force close transaction itself will be returned.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): just return the summary itself?
|
|
|
|
func (c *ChainArbitrator) ForceCloseContract(chanPoint wire.OutPoint) (*wire.MsgTx, error) {
|
|
|
|
c.Lock()
|
|
|
|
arbitrator, ok := c.activeChannels[chanPoint]
|
|
|
|
c.Unlock()
|
|
|
|
if !ok {
|
|
|
|
return nil, fmt.Errorf("unable to find arbitrator")
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Infof("Attempting to force close ChannelPoint(%v)", chanPoint)
|
|
|
|
|
|
|
|
errChan := make(chan error, 1)
|
|
|
|
respChan := make(chan *wire.MsgTx, 1)
|
|
|
|
|
|
|
|
// With the channel found, and the request crafted, we'll send over a
|
|
|
|
// force close request to the arbitrator that watches this channel.
|
|
|
|
select {
|
|
|
|
case arbitrator.forceCloseReqs <- &forceCloseReq{
|
|
|
|
errResp: errChan,
|
|
|
|
closeTx: respChan,
|
|
|
|
}:
|
|
|
|
case <-c.quit:
|
2018-09-12 02:05:26 +03:00
|
|
|
return nil, ErrChainArbExiting
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// We'll await two responses: the error response, and the transaction
|
2018-01-17 06:52:22 +03:00
|
|
|
// that closed out the channel.
|
|
|
|
select {
|
|
|
|
case err := <-errChan:
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
case <-c.quit:
|
2018-09-12 02:05:26 +03:00
|
|
|
return nil, ErrChainArbExiting
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
var closeTx *wire.MsgTx
|
|
|
|
select {
|
|
|
|
case closeTx = <-respChan:
|
|
|
|
case <-c.quit:
|
2018-09-12 02:05:26 +03:00
|
|
|
return nil, ErrChainArbExiting
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
2018-06-14 05:43:42 +03:00
|
|
|
// We'll attempt to disable the channel in the background to
|
|
|
|
// avoid blocking due to sending the update message to all
|
|
|
|
// active peers.
|
|
|
|
go func() {
|
|
|
|
if err := c.cfg.DisableChannel(chanPoint); err != nil {
|
|
|
|
log.Errorf("Unable to disable channel %v on "+
|
|
|
|
"close: %v", chanPoint, err)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
return closeTx, nil
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// WatchNewChannel sends the ChainArbitrator a message to create a
|
2018-01-17 06:52:22 +03:00
|
|
|
// ChannelArbitrator tasked with watching over a new channel. Once a new
|
|
|
|
// channel has finished its final funding flow, it should be registered with
|
|
|
|
// the ChainArbitrator so we can properly react to any on-chain events.
|
2018-01-19 01:00:35 +03:00
|
|
|
func (c *ChainArbitrator) WatchNewChannel(newChan *channeldb.OpenChannel) error {
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Lock()
|
|
|
|
defer c.Unlock()
|
|
|
|
|
|
|
|
log.Infof("Creating new ChannelArbitrator for ChannelPoint(%v)",
|
|
|
|
newChan.FundingOutpoint)
|
|
|
|
|
|
|
|
// If we're already watching this channel, then we'll ignore this
|
|
|
|
// request.
|
|
|
|
chanPoint := newChan.FundingOutpoint
|
|
|
|
if _, ok := c.activeChannels[chanPoint]; ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// First, also create an active chainWatcher for this channel to ensure
|
|
|
|
// that we detect any relevant on chain events.
|
|
|
|
chainWatcher, err := newChainWatcher(
|
2018-04-19 14:05:50 +03:00
|
|
|
chainWatcherConfig{
|
|
|
|
chanState: newChan,
|
|
|
|
notifier: c.cfg.Notifier,
|
|
|
|
pCache: c.cfg.PreimageDB,
|
|
|
|
signer: c.cfg.Signer,
|
|
|
|
isOurAddr: c.cfg.IsOurAddress,
|
2018-04-18 14:41:49 +03:00
|
|
|
contractBreach: func(retInfo *lnwallet.BreachRetribution) error {
|
|
|
|
return c.cfg.ContractBreach(chanPoint, retInfo)
|
|
|
|
},
|
2018-01-20 04:17:14 +03:00
|
|
|
},
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
c.activeWatchers[newChan.FundingOutpoint] = chainWatcher
|
|
|
|
|
|
|
|
// We'll also create a new channel arbitrator instance using this new
|
2018-01-17 06:52:22 +03:00
|
|
|
// channel, and our internal state.
|
2018-01-19 01:00:35 +03:00
|
|
|
channelArb, err := newActiveChannelArbitrator(
|
2018-04-18 14:41:49 +03:00
|
|
|
newChan, c, chainWatcher.SubscribeChannelEvents(),
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the arbitrator created, we'll add it to our set of active
|
|
|
|
// arbitrators, then launch it.
|
|
|
|
c.activeChannels[chanPoint] = channelArb
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
if err := channelArb.Start(); err != nil {
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
return chainWatcher.Start()
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
2018-01-19 01:02:21 +03:00
|
|
|
// SubscribeChannelEvents returns a new active subscription for the set of
|
|
|
|
// possible on-chain events for a particular channel. The struct can be used by
|
|
|
|
// callers to be notified whenever an event that changes the state of the
|
2018-01-21 07:25:54 +03:00
|
|
|
// channel on-chain occurs. If syncDispatch is true, then the sender of the
|
|
|
|
// notification will wait until an error is sent over the ProcessACK before
|
|
|
|
// modifying any database state. This allows callers to request a reliable hand
|
|
|
|
// off.
|
2018-01-19 01:02:21 +03:00
|
|
|
//
|
|
|
|
// TODO(roasbeef): can be used later to provide RPC hook for all channel
|
|
|
|
// lifetimes
|
|
|
|
func (c *ChainArbitrator) SubscribeChannelEvents(
|
2018-04-18 14:41:49 +03:00
|
|
|
chanPoint wire.OutPoint) (*ChainEventSubscription, error) {
|
2018-01-19 01:02:21 +03:00
|
|
|
|
|
|
|
// First, we'll attempt to look up the active watcher for this channel.
|
|
|
|
// If we can't find it, then we'll return an error back to the caller.
|
|
|
|
watcher, ok := c.activeWatchers[chanPoint]
|
|
|
|
if !ok {
|
|
|
|
return nil, fmt.Errorf("unable to find watcher for: %v",
|
|
|
|
chanPoint)
|
|
|
|
}
|
2018-01-17 06:52:22 +03:00
|
|
|
|
2018-01-19 01:02:21 +03:00
|
|
|
// With the watcher located, we'll request for it to create a new chain
|
|
|
|
// event subscription client.
|
2018-04-18 14:41:49 +03:00
|
|
|
return watcher.SubscribeChannelEvents(), nil
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// TODO(roasbeef): arbitration reports
|
|
|
|
// * types: contested, waiting for success conf, etc
|