2018-01-17 06:52:22 +03:00
|
|
|
package contractcourt
|
|
|
|
|
|
|
|
import (
|
2018-09-12 02:05:26 +03:00
|
|
|
"errors"
|
2018-01-17 06:52:22 +03:00
|
|
|
"fmt"
|
|
|
|
"sync"
|
|
|
|
"sync/atomic"
|
|
|
|
|
2018-05-31 08:07:54 +03:00
|
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
|
|
"github.com/btcsuite/btcd/wire"
|
|
|
|
"github.com/btcsuite/btcutil"
|
2018-01-17 06:52:22 +03:00
|
|
|
"github.com/lightningnetwork/lnd/chainntnfs"
|
|
|
|
"github.com/lightningnetwork/lnd/channeldb"
|
2019-01-16 17:47:43 +03:00
|
|
|
"github.com/lightningnetwork/lnd/input"
|
2018-01-17 06:52:22 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwallet"
|
2019-10-31 05:43:05 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwallet/chainfee"
|
2018-01-17 06:52:22 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
|
|
)
|
|
|
|
|
2018-09-12 02:05:26 +03:00
|
|
|
// ErrChainArbExiting signals that the chain arbitrator is shutting down.
|
|
|
|
var ErrChainArbExiting = errors.New("ChainArbitrator exiting")
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
// ResolutionMsg is a message sent by resolvers to outside sub-systems once an
|
|
|
|
// outgoing contract has been fully resolved. For multi-hop contracts, if we
|
|
|
|
// resolve the outgoing contract, we'll also need to ensure that the incoming
|
|
|
|
// contract is resolved as well. We package the items required to resolve the
|
|
|
|
// incoming contracts within this message.
|
|
|
|
type ResolutionMsg struct {
|
|
|
|
// SourceChan identifies the channel that this message is being sent
|
|
|
|
// from. This is the channel's short channel ID.
|
|
|
|
SourceChan lnwire.ShortChannelID
|
|
|
|
|
|
|
|
// HtlcIndex is the index of the contract within the original
|
|
|
|
// commitment trace.
|
|
|
|
HtlcIndex uint64
|
|
|
|
|
2019-10-03 18:22:43 +03:00
|
|
|
// Failure will be non-nil if the incoming contract should be canceled
|
2018-01-17 06:52:22 +03:00
|
|
|
// all together. This can happen if the outgoing contract was dust, if
|
|
|
|
// if the outgoing HTLC timed out.
|
|
|
|
Failure lnwire.FailureMessage
|
|
|
|
|
|
|
|
// PreImage will be non-nil if the incoming contract can successfully
|
|
|
|
// be redeemed. This can happen if we learn of the preimage from the
|
|
|
|
// outgoing HTLC on-chain.
|
|
|
|
PreImage *[32]byte
|
|
|
|
}
|
|
|
|
|
|
|
|
// ChainArbitratorConfig is a configuration struct that contains all the
|
|
|
|
// function closures and interface that required to arbitrate on-chain
|
|
|
|
// contracts for a particular chain.
|
|
|
|
type ChainArbitratorConfig struct {
|
|
|
|
// ChainHash is the chain that this arbitrator is to operate within.
|
|
|
|
ChainHash chainhash.Hash
|
|
|
|
|
2019-04-03 13:18:19 +03:00
|
|
|
// IncomingBroadcastDelta is the delta that we'll use to decide when to
|
|
|
|
// broadcast our commitment transaction if we have incoming htlcs. This
|
|
|
|
// value should be set based on our current fee estimation of the
|
|
|
|
// commitment transaction. We use this to determine when we should
|
|
|
|
// broadcast instead of the just the HTLC timeout, as we want to ensure
|
|
|
|
// that the commitment transaction is already confirmed, by the time the
|
|
|
|
// HTLC expires. Otherwise we may end up not settling the htlc on-chain
|
|
|
|
// because the other party managed to time it out.
|
|
|
|
IncomingBroadcastDelta uint32
|
|
|
|
|
|
|
|
// OutgoingBroadcastDelta is the delta that we'll use to decide when to
|
|
|
|
// broadcast our commitment transaction if there are active outgoing
|
|
|
|
// htlcs. This value can be lower than the incoming broadcast delta.
|
|
|
|
OutgoingBroadcastDelta uint32
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
// NewSweepAddr is a function that returns a new address under control
|
|
|
|
// by the wallet. We'll use this to sweep any no-delay outputs as a
|
|
|
|
// result of unilateral channel closes.
|
|
|
|
//
|
|
|
|
// NOTE: This SHOULD return a p2wkh script.
|
|
|
|
NewSweepAddr func() ([]byte, error)
|
|
|
|
|
|
|
|
// PublishTx reliably broadcasts a transaction to the network. Once
|
|
|
|
// this function exits without an error, then they transaction MUST
|
|
|
|
// continually be rebroadcast if needed.
|
|
|
|
PublishTx func(*wire.MsgTx) error
|
|
|
|
|
|
|
|
// DeliverResolutionMsg is a function that will append an outgoing
|
|
|
|
// message to the "out box" for a ChannelLink. This is used to cancel
|
|
|
|
// backwards any HTLC's that are either dust, we're timing out, or
|
|
|
|
// settling on-chain to the incoming link.
|
|
|
|
DeliverResolutionMsg func(...ResolutionMsg) error
|
|
|
|
|
|
|
|
// MarkLinkInactive is a function closure that the ChainArbitrator will
|
|
|
|
// use to mark that active HTLC's shouldn't be attempt ted to be routed
|
|
|
|
// over a particular channel. This function will be called in that a
|
|
|
|
// ChannelArbitrator decides that it needs to go to chain in order to
|
|
|
|
// resolve contracts.
|
|
|
|
//
|
2018-01-20 04:17:14 +03:00
|
|
|
// TODO(roasbeef): rename, routing based
|
2018-01-17 06:52:22 +03:00
|
|
|
MarkLinkInactive func(wire.OutPoint) error
|
|
|
|
|
2018-04-18 14:41:49 +03:00
|
|
|
// ContractBreach is a function closure that the ChainArbitrator will
|
|
|
|
// use to notify the breachArbiter about a contract breach. It should
|
|
|
|
// only return a non-nil error when the breachArbiter has preserved the
|
|
|
|
// necessary breach info for this channel point, and it is safe to mark
|
|
|
|
// the channel as pending close in the database.
|
|
|
|
ContractBreach func(wire.OutPoint, *lnwallet.BreachRetribution) error
|
|
|
|
|
2018-01-20 04:17:14 +03:00
|
|
|
// IsOurAddress is a function that returns true if the passed address
|
|
|
|
// is known to the underlying wallet. Otherwise, false should be
|
|
|
|
// returned.
|
|
|
|
IsOurAddress func(btcutil.Address) bool
|
|
|
|
|
2018-04-18 05:02:04 +03:00
|
|
|
// IncubateOutput sends either an incoming HTLC, an outgoing HTLC, or
|
2018-01-17 06:52:22 +03:00
|
|
|
// both to the utxo nursery. Once this function returns, the nursery
|
|
|
|
// should have safely persisted the outputs to disk, and should start
|
|
|
|
// the process of incubation. This is used when a resolver wishes to
|
2018-02-07 06:11:11 +03:00
|
|
|
// pass off the output to the nursery as we're only waiting on an
|
2018-01-17 06:52:22 +03:00
|
|
|
// absolute/relative item block.
|
2019-10-30 13:44:39 +03:00
|
|
|
IncubateOutputs func(wire.OutPoint, *lnwallet.OutgoingHtlcResolution,
|
2018-09-05 16:43:01 +03:00
|
|
|
*lnwallet.IncomingHtlcResolution, uint32) error
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
// PreimageDB is a global store of all known pre-images. We'll use this
|
|
|
|
// to decide if we should broadcast a commitment transaction to claim
|
|
|
|
// an HTLC on-chain.
|
|
|
|
PreimageDB WitnessBeacon
|
|
|
|
|
|
|
|
// Notifier is an instance of a chain notifier we'll use to watch for
|
|
|
|
// certain on-chain events.
|
|
|
|
Notifier chainntnfs.ChainNotifier
|
|
|
|
|
|
|
|
// Signer is a signer backed by the active lnd node. This should be
|
|
|
|
// capable of producing a signature as specified by a valid
|
|
|
|
// SignDescriptor.
|
2019-01-16 17:47:43 +03:00
|
|
|
Signer input.Signer
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
// FeeEstimator will be used to return fee estimates.
|
2019-10-31 05:43:05 +03:00
|
|
|
FeeEstimator chainfee.Estimator
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
// ChainIO allows us to query the state of the current main chain.
|
|
|
|
ChainIO lnwallet.BlockChainIO
|
2018-06-14 05:43:42 +03:00
|
|
|
|
|
|
|
// DisableChannel disables a channel, resulting in it not being able to
|
|
|
|
// forward payments.
|
|
|
|
DisableChannel func(wire.OutPoint) error
|
2018-09-26 19:46:48 +03:00
|
|
|
|
|
|
|
// Sweeper allows resolvers to sweep their final outputs.
|
2019-10-31 14:42:49 +03:00
|
|
|
Sweeper UtxoSweeper
|
2019-01-23 07:45:35 +03:00
|
|
|
|
2019-02-09 18:27:37 +03:00
|
|
|
// Registry is the invoice database that is used by resolvers to lookup
|
|
|
|
// preimages and settle invoices.
|
2019-04-16 11:21:02 +03:00
|
|
|
Registry Registry
|
2018-10-22 06:40:02 +03:00
|
|
|
|
|
|
|
// NotifyClosedChannel is a function closure that the ChainArbitrator
|
|
|
|
// will use to notify the ChannelNotifier about a newly closed channel.
|
|
|
|
NotifyClosedChannel func(wire.OutPoint)
|
2019-08-26 15:06:57 +03:00
|
|
|
|
|
|
|
// OnionProcessor is used to decode onion payloads for on-chain
|
|
|
|
// resolution.
|
|
|
|
OnionProcessor OnionProcessor
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// ChainArbitrator is a sub-system that oversees the on-chain resolution of all
|
|
|
|
// active, and channel that are in the "pending close" state. Within the
|
|
|
|
// contractcourt package, the ChainArbitrator manages a set of active
|
|
|
|
// ContractArbitrators. Each ContractArbitrators is responsible for watching
|
|
|
|
// the chain for any activity that affects the state of the channel, and also
|
|
|
|
// for monitoring each contract in order to determine if any on-chain activity is
|
|
|
|
// required. Outside sub-systems interact with the ChainArbitrator in order to
|
|
|
|
// forcibly exit a contract, update the set of live signals for each contract,
|
|
|
|
// and to receive reports on the state of contract resolution.
|
|
|
|
type ChainArbitrator struct {
|
2018-06-01 01:41:41 +03:00
|
|
|
started int32 // To be used atomically.
|
|
|
|
stopped int32 // To be used atomically.
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
sync.Mutex
|
|
|
|
|
|
|
|
// activeChannels is a map of all the active contracts that are still
|
|
|
|
// open, and not fully resolved.
|
|
|
|
activeChannels map[wire.OutPoint]*ChannelArbitrator
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// activeWatchers is a map of all the active chainWatchers for channels
|
|
|
|
// that are still considered open.
|
|
|
|
activeWatchers map[wire.OutPoint]*chainWatcher
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
// cfg is the config struct for the arbitrator that contains all
|
|
|
|
// methods and interface it needs to operate.
|
|
|
|
cfg ChainArbitratorConfig
|
|
|
|
|
|
|
|
// chanSource will be used by the ChainArbitrator to fetch all the
|
|
|
|
// active channels that it must still watch over.
|
|
|
|
chanSource *channeldb.DB
|
|
|
|
|
|
|
|
quit chan struct{}
|
|
|
|
|
|
|
|
wg sync.WaitGroup
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewChainArbitrator returns a new instance of the ChainArbitrator using the
|
|
|
|
// passed config struct, and backing persistent database.
|
|
|
|
func NewChainArbitrator(cfg ChainArbitratorConfig,
|
|
|
|
db *channeldb.DB) *ChainArbitrator {
|
|
|
|
|
|
|
|
return &ChainArbitrator{
|
|
|
|
cfg: cfg,
|
|
|
|
activeChannels: make(map[wire.OutPoint]*ChannelArbitrator),
|
2018-01-19 01:00:35 +03:00
|
|
|
activeWatchers: make(map[wire.OutPoint]*chainWatcher),
|
2018-01-17 06:52:22 +03:00
|
|
|
chanSource: db,
|
|
|
|
quit: make(chan struct{}),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// newActiveChannelArbitrator creates a new instance of an active channel
|
|
|
|
// arbitrator given the state of the target channel.
|
|
|
|
func newActiveChannelArbitrator(channel *channeldb.OpenChannel,
|
2018-01-19 01:00:35 +03:00
|
|
|
c *ChainArbitrator, chanEvents *ChainEventSubscription) (*ChannelArbitrator, error) {
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
log.Tracef("Creating ChannelArbitrator for ChannelPoint(%v)",
|
|
|
|
channel.FundingOutpoint)
|
|
|
|
|
|
|
|
// We'll start by registering for a block epoch notifications so this
|
|
|
|
// channel can keep track of the current state of the main chain.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): fetch best height (or pass in) so can ensure block
|
|
|
|
// epoch delivers all the notifications to
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): instead 1 block epoch that multi-plexes to the rest?
|
|
|
|
// * reduces the number of goroutines
|
2018-08-09 10:05:27 +03:00
|
|
|
blockEpoch, err := c.cfg.Notifier.RegisterBlockEpochNtfn(nil)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
chanPoint := channel.FundingOutpoint
|
|
|
|
|
|
|
|
// Next we'll create the matching configuration struct that contains
|
|
|
|
// all interfaces and methods the arbitrator needs to do its job.
|
|
|
|
arbCfg := ChannelArbitratorConfig{
|
|
|
|
ChanPoint: chanPoint,
|
2018-05-02 02:27:20 +03:00
|
|
|
ShortChanID: channel.ShortChanID(),
|
2018-01-17 06:52:22 +03:00
|
|
|
BlockEpochs: blockEpoch,
|
2018-03-19 17:19:19 +03:00
|
|
|
ForceCloseChan: func() (*lnwallet.LocalForceCloseSummary, error) {
|
2019-03-09 03:40:49 +03:00
|
|
|
// First, we mark the channel as borked, this ensure
|
|
|
|
// that no new state transitions can happen, and also
|
|
|
|
// that the link won't be loaded into the switch.
|
|
|
|
if err := channel.MarkBorked(); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the channel marked as borked, we'll now remove
|
|
|
|
// the link from the switch if its there. If the link
|
|
|
|
// is active, then this method will block until it
|
|
|
|
// exits.
|
|
|
|
if err := c.cfg.MarkLinkInactive(chanPoint); err != nil {
|
|
|
|
log.Errorf("unable to mark link inactive: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now that we know the link can't mutate the channel
|
|
|
|
// state, we'll read the channel from disk the target
|
|
|
|
// channel according to its channel point.
|
2019-01-05 01:59:04 +03:00
|
|
|
channel, err := c.chanSource.FetchChannel(chanPoint)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
2019-03-09 03:40:49 +03:00
|
|
|
// Finally, we'll force close the channel completing
|
|
|
|
// the force close workflow.
|
2018-01-17 06:52:22 +03:00
|
|
|
chanMachine, err := lnwallet.NewLightningChannel(
|
2019-04-15 15:24:43 +03:00
|
|
|
c.cfg.Signer, channel, nil,
|
multi: replace per channel sigPool with global daemon level sigPool
In this commit, we remove the per channel `sigPool` within the
`lnwallet.LightningChannel` struct. With this change, we ensure that as
the number of channels grows, the number of gouroutines idling in the
sigPool stays constant. It's the case that currently on the daemon, most
channels are likely inactive, with only a hand full actually
consistently carrying out channel updates. As a result, this change
should reduce the amount of idle CPU usage, as we have less active
goroutines in select loops.
In order to make this change, the `SigPool` itself has been publicly
exported such that outside callers can make a `SigPool` and pass it into
newly created channels. Since the sig pool now lives outside the
channel, we were also able to do away with the Stop() method on the
channel all together.
Finally, the server is the sub-system that is currently responsible for
managing the `SigPool` within lnd.
2018-12-15 03:35:07 +03:00
|
|
|
)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
return chanMachine.ForceClose()
|
|
|
|
},
|
2018-04-03 15:04:30 +03:00
|
|
|
MarkCommitmentBroadcasted: channel.MarkCommitmentBroadcasted,
|
2018-10-22 06:40:02 +03:00
|
|
|
MarkChannelClosed: func(summary *channeldb.ChannelCloseSummary) error {
|
|
|
|
if err := channel.CloseChannel(summary); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
c.cfg.NotifyClosedChannel(summary.ChanPoint)
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
IsPendingClose: false,
|
|
|
|
ChainArbitratorConfig: c.cfg,
|
|
|
|
ChainEvents: chanEvents,
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// The final component needed is an arbitrator log that the arbitrator
|
|
|
|
// will use to keep track of its internal state using a backed
|
|
|
|
// persistent log.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef); abstraction leak...
|
|
|
|
// * rework: adaptor method to set log scope w/ factory func
|
|
|
|
chanLog, err := newBoltArbitratorLog(
|
|
|
|
c.chanSource.DB, arbCfg, c.cfg.ChainHash, chanPoint,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
blockEpoch.Cancel()
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
arbCfg.MarkChannelResolved = func() error {
|
2019-11-14 03:05:49 +03:00
|
|
|
return c.ResolveContract(chanPoint)
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
multi: address lingering TODO by no longer wiping out local HTLCs on remote close
In this commit, we fix a lingering TOOD statement in the channel arb.
Before this commitment, we would simply wipe our our local HTLC set of
the HTLC set that was on the remote commitment transaction on force
close. This was incorrect as if our commitment transaction had an HTLC
that the remote commitment didn't, then we would fail to cancel that
back, and cause both channels to time out on chain.
In order to remedy this, we introduce a new `HtlcSetKey` struct to track
all 3 possible in-flight set of HTLCs: ours, theirs, and their pending.
We also we start to tack on additional data to all the unilateral close
messages we send to subscribers. This new data is the CommitSet, or the
set of valid commitments at channel closure time. This new information
will be used by the channel arb in an upcoming commit to ensure it will
cancel back HTLCs in the case of split commitment state.
Finally, we start to thread through an optional *CommitSet to the
advanceState method. This additional information will give the channel
arb addition information it needs to ensure it properly cancels back
HTLCs that are about to time out or may time out depending on which
commitment is played.
Within the htlcswitch pakage, we modify the `SignNextCommitment` method
to return the new set of pending HTLCs for the remote party's commitment
transaction and `ReceiveRevocation` to return the latest set of
commitment transactions on the remote party's commitment as well. This
is a preparatory change which is part of a larger change to address a
lingering TODO in the cnct.
Additionally, rather than just send of the set of HTLCs after the we
revoke, we'll also send of the set of HTLCs after the remote party
revokes, and we create a pending commitment state for it.
2019-05-17 03:23:26 +03:00
|
|
|
// Finally, we'll need to construct a series of htlc Sets based on all
|
|
|
|
// currently known valid commitments.
|
|
|
|
htlcSets := make(map[HtlcSetKey]htlcSet)
|
|
|
|
htlcSets[LocalHtlcSet] = newHtlcSet(channel.LocalCommitment.Htlcs)
|
|
|
|
htlcSets[RemoteHtlcSet] = newHtlcSet(channel.RemoteCommitment.Htlcs)
|
|
|
|
|
|
|
|
pendingRemoteCommitment, err := channel.RemoteCommitChainTip()
|
|
|
|
if err != nil && err != channeldb.ErrNoPendingCommit {
|
|
|
|
blockEpoch.Cancel()
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if pendingRemoteCommitment != nil {
|
|
|
|
htlcSets[RemotePendingHtlcSet] = newHtlcSet(
|
|
|
|
pendingRemoteCommitment.Commitment.Htlcs,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
return NewChannelArbitrator(
|
multi: address lingering TODO by no longer wiping out local HTLCs on remote close
In this commit, we fix a lingering TOOD statement in the channel arb.
Before this commitment, we would simply wipe our our local HTLC set of
the HTLC set that was on the remote commitment transaction on force
close. This was incorrect as if our commitment transaction had an HTLC
that the remote commitment didn't, then we would fail to cancel that
back, and cause both channels to time out on chain.
In order to remedy this, we introduce a new `HtlcSetKey` struct to track
all 3 possible in-flight set of HTLCs: ours, theirs, and their pending.
We also we start to tack on additional data to all the unilateral close
messages we send to subscribers. This new data is the CommitSet, or the
set of valid commitments at channel closure time. This new information
will be used by the channel arb in an upcoming commit to ensure it will
cancel back HTLCs in the case of split commitment state.
Finally, we start to thread through an optional *CommitSet to the
advanceState method. This additional information will give the channel
arb addition information it needs to ensure it properly cancels back
HTLCs that are about to time out or may time out depending on which
commitment is played.
Within the htlcswitch pakage, we modify the `SignNextCommitment` method
to return the new set of pending HTLCs for the remote party's commitment
transaction and `ReceiveRevocation` to return the latest set of
commitment transactions on the remote party's commitment as well. This
is a preparatory change which is part of a larger change to address a
lingering TODO in the cnct.
Additionally, rather than just send of the set of HTLCs after the we
revoke, we'll also send of the set of HTLCs after the remote party
revokes, and we create a pending commitment state for it.
2019-05-17 03:23:26 +03:00
|
|
|
arbCfg, htlcSets, chanLog,
|
2018-01-17 06:52:22 +03:00
|
|
|
), nil
|
|
|
|
}
|
|
|
|
|
2019-11-14 03:05:49 +03:00
|
|
|
// ResolveContract marks a contract as fully resolved within the database.
|
2018-01-17 06:52:22 +03:00
|
|
|
// This is only to be done once all contracts which were live on the channel
|
|
|
|
// before hitting the chain have been resolved.
|
2019-11-14 03:05:49 +03:00
|
|
|
func (c *ChainArbitrator) ResolveContract(chanPoint wire.OutPoint) error {
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
log.Infof("Marking ChannelPoint(%v) fully resolved", chanPoint)
|
|
|
|
|
|
|
|
// First, we'll we'll mark the channel as fully closed from the PoV of
|
|
|
|
// the channel source.
|
|
|
|
err := c.chanSource.MarkChanFullyClosed(&chanPoint)
|
|
|
|
if err != nil {
|
2018-04-13 03:13:28 +03:00
|
|
|
log.Errorf("ChainArbitrator: unable to mark ChannelPoint(%v) "+
|
|
|
|
"fully closed: %v", chanPoint, err)
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2019-11-14 03:05:49 +03:00
|
|
|
// Now that the channel has been marked as fully closed, we'll stop
|
|
|
|
// both the channel arbitrator and chain watcher for this channel if
|
|
|
|
// they're still active.
|
|
|
|
var arbLog ArbitratorLog
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Lock()
|
2019-11-14 03:05:49 +03:00
|
|
|
chainArb := c.activeChannels[chanPoint]
|
2018-01-17 06:52:22 +03:00
|
|
|
delete(c.activeChannels, chanPoint)
|
2018-01-20 04:13:41 +03:00
|
|
|
|
2019-11-14 03:05:49 +03:00
|
|
|
chainWatcher := c.activeWatchers[chanPoint]
|
2018-06-19 14:25:30 +03:00
|
|
|
delete(c.activeWatchers, chanPoint)
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Unlock()
|
|
|
|
|
2019-11-14 03:05:49 +03:00
|
|
|
if chainArb != nil {
|
|
|
|
arbLog = chainArb.log
|
|
|
|
|
|
|
|
if err := chainArb.Stop(); err != nil {
|
|
|
|
log.Warnf("unable to stop ChannelArbitrator(%v): %v",
|
|
|
|
chanPoint, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if chainWatcher != nil {
|
|
|
|
if err := chainWatcher.Stop(); err != nil {
|
|
|
|
log.Warnf("unable to stop ChainWatcher(%v): %v",
|
|
|
|
chanPoint, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Once this has been marked as resolved, we'll wipe the log that the
|
|
|
|
// channel arbitrator was using to store its persistent state. We do
|
|
|
|
// this after marking the channel resolved, as otherwise, the
|
|
|
|
// arbitrator would be re-created, and think it was starting from the
|
|
|
|
// default state.
|
|
|
|
if arbLog != nil {
|
|
|
|
if err := arbLog.WipeHistory(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Start launches all goroutines that the ChainArbitrator needs to operate.
|
|
|
|
func (c *ChainArbitrator) Start() error {
|
|
|
|
if !atomic.CompareAndSwapInt32(&c.started, 0, 1) {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Tracef("Starting ChainArbitrator")
|
|
|
|
|
|
|
|
// First, we'll fetch all the channels that are still open, in order to
|
|
|
|
// collect them within our set of active contracts.
|
|
|
|
openChannels, err := c.chanSource.FetchAllChannels()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(openChannels) > 0 {
|
|
|
|
log.Infof("Creating ChannelArbitrators for %v active channels",
|
|
|
|
len(openChannels))
|
|
|
|
}
|
|
|
|
|
|
|
|
// For each open channel, we'll configure then launch a corresponding
|
|
|
|
// ChannelArbitrator.
|
|
|
|
for _, channel := range openChannels {
|
2018-04-13 04:53:22 +03:00
|
|
|
chanPoint := channel.FundingOutpoint
|
2018-05-19 02:21:06 +03:00
|
|
|
channel := channel
|
2018-04-13 04:53:22 +03:00
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// First, we'll create an active chainWatcher for this channel
|
|
|
|
// to ensure that we detect any relevant on chain events.
|
|
|
|
chainWatcher, err := newChainWatcher(
|
2018-04-19 14:05:50 +03:00
|
|
|
chainWatcherConfig{
|
|
|
|
chanState: channel,
|
|
|
|
notifier: c.cfg.Notifier,
|
|
|
|
signer: c.cfg.Signer,
|
|
|
|
isOurAddr: c.cfg.IsOurAddress,
|
2018-04-18 14:41:49 +03:00
|
|
|
contractBreach: func(retInfo *lnwallet.BreachRetribution) error {
|
|
|
|
return c.cfg.ContractBreach(chanPoint, retInfo)
|
|
|
|
},
|
2019-03-13 05:19:24 +03:00
|
|
|
extractStateNumHint: lnwallet.GetStateNumHint,
|
2018-01-20 04:17:14 +03:00
|
|
|
},
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-13 04:53:22 +03:00
|
|
|
c.activeWatchers[chanPoint] = chainWatcher
|
2018-01-19 01:00:35 +03:00
|
|
|
channelArb, err := newActiveChannelArbitrator(
|
2018-04-18 14:41:49 +03:00
|
|
|
channel, c, chainWatcher.SubscribeChannelEvents(),
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-13 04:53:22 +03:00
|
|
|
c.activeChannels[chanPoint] = channelArb
|
2019-09-06 14:14:39 +03:00
|
|
|
|
2019-12-05 00:29:51 +03:00
|
|
|
// Republish any closing transactions for this channel.
|
|
|
|
err = c.publishClosingTxs(channel)
|
|
|
|
if err != nil {
|
2019-09-06 14:14:39 +03:00
|
|
|
return err
|
|
|
|
}
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// In addition to the channels that we know to be open, we'll also
|
|
|
|
// launch arbitrators to finishing resolving any channels that are in
|
|
|
|
// the pending close state.
|
|
|
|
closingChannels, err := c.chanSource.FetchClosedChannels(true)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(closingChannels) > 0 {
|
|
|
|
log.Infof("Creating ChannelArbitrators for %v closing channels",
|
|
|
|
len(closingChannels))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next, for each channel is the closing state, we'll launch a
|
2018-05-24 11:24:31 +03:00
|
|
|
// corresponding more restricted resolver, as we don't have to watch
|
|
|
|
// the chain any longer, only resolve the contracts on the confirmed
|
|
|
|
// commitment.
|
2018-01-17 06:52:22 +03:00
|
|
|
for _, closeChanInfo := range closingChannels {
|
2018-08-09 10:05:27 +03:00
|
|
|
blockEpoch, err := c.cfg.Notifier.RegisterBlockEpochNtfn(nil)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// We can leave off the CloseContract and ForceCloseChan
|
|
|
|
// methods as the channel is already closed at this point.
|
|
|
|
chanPoint := closeChanInfo.ChanPoint
|
|
|
|
arbCfg := ChannelArbitratorConfig{
|
|
|
|
ChanPoint: chanPoint,
|
|
|
|
ShortChanID: closeChanInfo.ShortChanID,
|
|
|
|
BlockEpochs: blockEpoch,
|
|
|
|
ChainArbitratorConfig: c.cfg,
|
2018-01-19 01:00:35 +03:00
|
|
|
ChainEvents: &ChainEventSubscription{},
|
2018-08-21 13:21:16 +03:00
|
|
|
IsPendingClose: true,
|
|
|
|
ClosingHeight: closeChanInfo.CloseHeight,
|
|
|
|
CloseType: closeChanInfo.CloseType,
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
chanLog, err := newBoltArbitratorLog(
|
|
|
|
c.chanSource.DB, arbCfg, c.cfg.ChainHash, chanPoint,
|
|
|
|
)
|
|
|
|
if err != nil {
|
2018-02-06 01:41:03 +03:00
|
|
|
blockEpoch.Cancel()
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
arbCfg.MarkChannelResolved = func() error {
|
2019-11-14 03:05:49 +03:00
|
|
|
return c.ResolveContract(chanPoint)
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// We can also leave off the set of HTLC's here as since the
|
|
|
|
// channel is already in the process of being full resolved, no
|
2019-05-24 06:27:04 +03:00
|
|
|
// new HTLC's will be added.
|
2018-01-17 06:52:22 +03:00
|
|
|
c.activeChannels[chanPoint] = NewChannelArbitrator(
|
|
|
|
arbCfg, nil, chanLog,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2018-09-12 02:05:26 +03:00
|
|
|
// Now, we'll start all chain watchers in parallel to shorten start up
|
|
|
|
// duration. In neutrino mode, this allows spend registrations to take
|
|
|
|
// advantage of batch spend reporting, instead of doing a single rescan
|
|
|
|
// per chain watcher.
|
|
|
|
//
|
|
|
|
// NOTE: After this point, we Stop the chain arb to ensure that any
|
|
|
|
// lingering goroutines are cleaned up before exiting.
|
|
|
|
watcherErrs := make(chan error, len(c.activeWatchers))
|
|
|
|
var wg sync.WaitGroup
|
2018-01-19 01:00:35 +03:00
|
|
|
for _, watcher := range c.activeWatchers {
|
2018-09-12 02:05:26 +03:00
|
|
|
wg.Add(1)
|
|
|
|
go func(w *chainWatcher) {
|
|
|
|
defer wg.Done()
|
|
|
|
select {
|
|
|
|
case watcherErrs <- w.Start():
|
|
|
|
case <-c.quit:
|
|
|
|
watcherErrs <- ErrChainArbExiting
|
|
|
|
}
|
|
|
|
}(watcher)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Once all chain watchers have been started, seal the err chan to
|
|
|
|
// signal the end of the err stream.
|
|
|
|
go func() {
|
|
|
|
wg.Wait()
|
|
|
|
close(watcherErrs)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Handle all errors returned from spawning our chain watchers. If any
|
|
|
|
// of them failed, we will stop the chain arb to shutdown any active
|
|
|
|
// goroutines.
|
|
|
|
for err := range watcherErrs {
|
|
|
|
if err != nil {
|
|
|
|
c.Stop()
|
2018-01-19 01:00:35 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
2018-09-12 02:05:26 +03:00
|
|
|
|
|
|
|
// Finally, we'll launch all the goroutines for each arbitrator so they
|
|
|
|
// can carry out their duties.
|
2018-01-17 06:52:22 +03:00
|
|
|
for _, arbitrator := range c.activeChannels {
|
|
|
|
if err := arbitrator.Start(); err != nil {
|
2018-09-12 02:05:26 +03:00
|
|
|
c.Stop()
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO(roasbeef): eventually move all breach watching here
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-12-05 00:29:51 +03:00
|
|
|
// publishClosingTxs will load any stored cooperative or unilater closing
|
|
|
|
// transactions and republish them. This helps ensure propagation of the
|
|
|
|
// transactions in the event that prior publications failed.
|
|
|
|
func (c *ChainArbitrator) publishClosingTxs(
|
|
|
|
channel *channeldb.OpenChannel) error {
|
|
|
|
|
|
|
|
// If the channel has had its unilateral close broadcasted already,
|
|
|
|
// republish it in case it didn't propagate.
|
|
|
|
if channel.HasChanStatus(channeldb.ChanStatusCommitBroadcasted) {
|
|
|
|
err := c.rebroadcast(
|
|
|
|
channel, channeldb.ChanStatusCommitBroadcasted,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the channel has had its cooperative close broadcasted
|
|
|
|
// already, republish it in case it didn't propagate.
|
|
|
|
if channel.HasChanStatus(channeldb.ChanStatusCoopBroadcasted) {
|
|
|
|
err := c.rebroadcast(
|
|
|
|
channel, channeldb.ChanStatusCoopBroadcasted,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// rebroadcast is a helper method which will republish the unilateral or
|
|
|
|
// cooperative close transaction or a channel in a particular state.
|
|
|
|
//
|
|
|
|
// NOTE: There is no risk to caling this method if the channel isn't in either
|
|
|
|
// CommimentBroadcasted or CoopBroadcasted, but the logs will be misleading.
|
|
|
|
func (c *ChainArbitrator) rebroadcast(channel *channeldb.OpenChannel,
|
|
|
|
state channeldb.ChannelStatus) error {
|
|
|
|
|
|
|
|
chanPoint := channel.FundingOutpoint
|
|
|
|
|
|
|
|
var (
|
|
|
|
closeTx *wire.MsgTx
|
|
|
|
kind string
|
|
|
|
err error
|
|
|
|
)
|
|
|
|
switch state {
|
|
|
|
case channeldb.ChanStatusCommitBroadcasted:
|
|
|
|
kind = "force"
|
|
|
|
closeTx, err = channel.BroadcastedCommitment()
|
|
|
|
|
|
|
|
case channeldb.ChanStatusCoopBroadcasted:
|
|
|
|
kind = "coop"
|
|
|
|
closeTx, err = channel.BroadcastedCooperative()
|
|
|
|
|
|
|
|
default:
|
|
|
|
return fmt.Errorf("unknown closing state: %v", state)
|
|
|
|
}
|
|
|
|
|
|
|
|
switch {
|
|
|
|
|
|
|
|
// This can happen for channels that had their closing tx published
|
|
|
|
// before we started storing it to disk.
|
|
|
|
case err == channeldb.ErrNoCloseTx:
|
|
|
|
log.Warnf("Channel %v is in state %v, but no %s closing tx "+
|
|
|
|
"to re-publish...", chanPoint, state, kind)
|
|
|
|
return nil
|
|
|
|
|
|
|
|
case err != nil:
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Infof("Re-publishing %s close tx(%v) for channel %v",
|
|
|
|
kind, closeTx.TxHash(), chanPoint)
|
|
|
|
|
|
|
|
err = c.cfg.PublishTx(closeTx)
|
|
|
|
if err != nil && err != lnwallet.ErrDoubleSpend {
|
|
|
|
log.Warnf("Unable to broadcast %s close tx(%v): %v",
|
|
|
|
kind, closeTx.TxHash(), err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
// Stop signals the ChainArbitrator to trigger a graceful shutdown. Any active
|
|
|
|
// channel arbitrators will be signalled to exit, and this method will block
|
|
|
|
// until they've all exited.
|
|
|
|
func (c *ChainArbitrator) Stop() error {
|
|
|
|
if !atomic.CompareAndSwapInt32(&c.stopped, 0, 1) {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Infof("Stopping ChainArbitrator")
|
|
|
|
|
|
|
|
close(c.quit)
|
|
|
|
|
2018-08-04 02:43:18 +03:00
|
|
|
var (
|
|
|
|
activeWatchers = make(map[wire.OutPoint]*chainWatcher)
|
|
|
|
activeChannels = make(map[wire.OutPoint]*ChannelArbitrator)
|
|
|
|
)
|
|
|
|
|
|
|
|
// Copy the current set of active watchers and arbitrators to shutdown.
|
|
|
|
// We don't want to hold the lock when shutting down each watcher or
|
|
|
|
// arbitrator individually, as they may need to acquire this mutex.
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Lock()
|
2018-08-04 02:43:18 +03:00
|
|
|
for chanPoint, watcher := range c.activeWatchers {
|
|
|
|
activeWatchers[chanPoint] = watcher
|
|
|
|
}
|
|
|
|
for chanPoint, arbitrator := range c.activeChannels {
|
|
|
|
activeChannels[chanPoint] = arbitrator
|
|
|
|
}
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Unlock()
|
|
|
|
|
2018-08-04 02:43:18 +03:00
|
|
|
for chanPoint, watcher := range activeWatchers {
|
2018-01-19 01:00:35 +03:00
|
|
|
log.Tracef("Attempting to stop ChainWatcher(%v)",
|
|
|
|
chanPoint)
|
|
|
|
|
|
|
|
if err := watcher.Stop(); err != nil {
|
|
|
|
log.Errorf("unable to stop watcher for "+
|
|
|
|
"ChannelPoint(%v): %v", chanPoint, err)
|
|
|
|
}
|
|
|
|
}
|
2018-08-04 02:43:18 +03:00
|
|
|
for chanPoint, arbitrator := range activeChannels {
|
2018-01-17 06:52:22 +03:00
|
|
|
log.Tracef("Attempting to stop ChannelArbitrator(%v)",
|
|
|
|
chanPoint)
|
|
|
|
|
|
|
|
if err := arbitrator.Stop(); err != nil {
|
|
|
|
log.Errorf("unable to stop arbitrator for "+
|
|
|
|
"ChannelPoint(%v): %v", chanPoint, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
c.wg.Wait()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
multi: address lingering TODO by no longer wiping out local HTLCs on remote close
In this commit, we fix a lingering TOOD statement in the channel arb.
Before this commitment, we would simply wipe our our local HTLC set of
the HTLC set that was on the remote commitment transaction on force
close. This was incorrect as if our commitment transaction had an HTLC
that the remote commitment didn't, then we would fail to cancel that
back, and cause both channels to time out on chain.
In order to remedy this, we introduce a new `HtlcSetKey` struct to track
all 3 possible in-flight set of HTLCs: ours, theirs, and their pending.
We also we start to tack on additional data to all the unilateral close
messages we send to subscribers. This new data is the CommitSet, or the
set of valid commitments at channel closure time. This new information
will be used by the channel arb in an upcoming commit to ensure it will
cancel back HTLCs in the case of split commitment state.
Finally, we start to thread through an optional *CommitSet to the
advanceState method. This additional information will give the channel
arb addition information it needs to ensure it properly cancels back
HTLCs that are about to time out or may time out depending on which
commitment is played.
Within the htlcswitch pakage, we modify the `SignNextCommitment` method
to return the new set of pending HTLCs for the remote party's commitment
transaction and `ReceiveRevocation` to return the latest set of
commitment transactions on the remote party's commitment as well. This
is a preparatory change which is part of a larger change to address a
lingering TODO in the cnct.
Additionally, rather than just send of the set of HTLCs after the we
revoke, we'll also send of the set of HTLCs after the remote party
revokes, and we create a pending commitment state for it.
2019-05-17 03:23:26 +03:00
|
|
|
// ContractUpdate is a message packages the latest set of active HTLCs on a
|
|
|
|
// commitment, and also identifies which commitment received a new set of
|
|
|
|
// HTLCs.
|
|
|
|
type ContractUpdate struct {
|
|
|
|
// HtlcKey identifies which commitment the HTLCs below are present on.
|
|
|
|
HtlcKey HtlcSetKey
|
|
|
|
|
|
|
|
// Htlcs are the of active HTLCs on the commitment identified by the
|
|
|
|
// above HtlcKey.
|
|
|
|
Htlcs []channeldb.HTLC
|
|
|
|
}
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
// ContractSignals wraps the two signals that affect the state of a channel
|
|
|
|
// being watched by an arbitrator. The two signals we care about are: the
|
|
|
|
// channel has a new set of HTLC's, and the remote party has just broadcast
|
|
|
|
// their version of the commitment transaction.
|
|
|
|
type ContractSignals struct {
|
multi: address lingering TODO by no longer wiping out local HTLCs on remote close
In this commit, we fix a lingering TOOD statement in the channel arb.
Before this commitment, we would simply wipe our our local HTLC set of
the HTLC set that was on the remote commitment transaction on force
close. This was incorrect as if our commitment transaction had an HTLC
that the remote commitment didn't, then we would fail to cancel that
back, and cause both channels to time out on chain.
In order to remedy this, we introduce a new `HtlcSetKey` struct to track
all 3 possible in-flight set of HTLCs: ours, theirs, and their pending.
We also we start to tack on additional data to all the unilateral close
messages we send to subscribers. This new data is the CommitSet, or the
set of valid commitments at channel closure time. This new information
will be used by the channel arb in an upcoming commit to ensure it will
cancel back HTLCs in the case of split commitment state.
Finally, we start to thread through an optional *CommitSet to the
advanceState method. This additional information will give the channel
arb addition information it needs to ensure it properly cancels back
HTLCs that are about to time out or may time out depending on which
commitment is played.
Within the htlcswitch pakage, we modify the `SignNextCommitment` method
to return the new set of pending HTLCs for the remote party's commitment
transaction and `ReceiveRevocation` to return the latest set of
commitment transactions on the remote party's commitment as well. This
is a preparatory change which is part of a larger change to address a
lingering TODO in the cnct.
Additionally, rather than just send of the set of HTLCs after the we
revoke, we'll also send of the set of HTLCs after the remote party
revokes, and we create a pending commitment state for it.
2019-05-17 03:23:26 +03:00
|
|
|
// HtlcUpdates is a channel that the link will use to update the
|
|
|
|
// designated channel arbitrator when the set of HTLCs on any valid
|
|
|
|
// commitment changes.
|
|
|
|
HtlcUpdates chan *ContractUpdate
|
2018-01-17 06:52:22 +03:00
|
|
|
|
|
|
|
// ShortChanID is the up to date short channel ID for a contract. This
|
|
|
|
// can change either if when the contract was added it didn't yet have
|
|
|
|
// a stable identifier, or in the case of a reorg.
|
|
|
|
ShortChanID lnwire.ShortChannelID
|
|
|
|
}
|
|
|
|
|
|
|
|
// UpdateContractSignals sends a set of active, up to date contract signals to
|
|
|
|
// the ChannelArbitrator which is has been assigned to the channel infield by
|
|
|
|
// the passed channel point.
|
|
|
|
func (c *ChainArbitrator) UpdateContractSignals(chanPoint wire.OutPoint,
|
|
|
|
signals *ContractSignals) error {
|
|
|
|
|
|
|
|
log.Infof("Attempting to update ContractSignals for ChannelPoint(%v)",
|
|
|
|
chanPoint)
|
|
|
|
|
|
|
|
c.Lock()
|
|
|
|
arbitrator, ok := c.activeChannels[chanPoint]
|
|
|
|
c.Unlock()
|
|
|
|
if !ok {
|
|
|
|
return fmt.Errorf("unable to find arbitrator")
|
|
|
|
}
|
|
|
|
|
|
|
|
arbitrator.UpdateContractSignals(signals)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-09-07 17:05:57 +03:00
|
|
|
// GetChannelArbitrator safely returns the channel arbitrator for a given
|
|
|
|
// channel outpoint.
|
|
|
|
func (c *ChainArbitrator) GetChannelArbitrator(chanPoint wire.OutPoint) (
|
|
|
|
*ChannelArbitrator, error) {
|
|
|
|
|
|
|
|
c.Lock()
|
|
|
|
arbitrator, ok := c.activeChannels[chanPoint]
|
|
|
|
c.Unlock()
|
|
|
|
if !ok {
|
|
|
|
return nil, fmt.Errorf("unable to find arbitrator")
|
|
|
|
}
|
|
|
|
|
|
|
|
return arbitrator, nil
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// forceCloseReq is a request sent from an outside sub-system to the arbitrator
|
|
|
|
// that watches a particular channel to broadcast the commitment transaction,
|
2018-01-17 06:52:22 +03:00
|
|
|
// and enter the resolution phase of the channel.
|
|
|
|
type forceCloseReq struct {
|
2018-01-19 01:00:35 +03:00
|
|
|
// errResp is a channel that will be sent upon either in the case of
|
|
|
|
// force close success (nil error), or in the case on an error.
|
2018-01-17 06:52:22 +03:00
|
|
|
//
|
|
|
|
// NOTE; This channel MUST be buffered.
|
|
|
|
errResp chan error
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// closeTx is a channel that carries the transaction which ultimately
|
2018-01-17 06:52:22 +03:00
|
|
|
// closed out the channel.
|
|
|
|
closeTx chan *wire.MsgTx
|
|
|
|
}
|
|
|
|
|
|
|
|
// ForceCloseContract attempts to force close the channel infield by the passed
|
|
|
|
// channel point. A force close will immediately terminate the contract,
|
|
|
|
// causing it to enter the resolution phase. If the force close was successful,
|
|
|
|
// then the force close transaction itself will be returned.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): just return the summary itself?
|
|
|
|
func (c *ChainArbitrator) ForceCloseContract(chanPoint wire.OutPoint) (*wire.MsgTx, error) {
|
|
|
|
c.Lock()
|
|
|
|
arbitrator, ok := c.activeChannels[chanPoint]
|
|
|
|
c.Unlock()
|
|
|
|
if !ok {
|
|
|
|
return nil, fmt.Errorf("unable to find arbitrator")
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Infof("Attempting to force close ChannelPoint(%v)", chanPoint)
|
|
|
|
|
2019-03-13 03:34:28 +03:00
|
|
|
// Before closing, we'll attempt to send a disable update for the
|
|
|
|
// channel. We do so before closing the channel as otherwise the current
|
|
|
|
// edge policy won't be retrievable from the graph.
|
|
|
|
if err := c.cfg.DisableChannel(chanPoint); err != nil {
|
|
|
|
log.Warnf("Unable to disable channel %v on "+
|
|
|
|
"close: %v", chanPoint, err)
|
|
|
|
}
|
|
|
|
|
2018-01-17 06:52:22 +03:00
|
|
|
errChan := make(chan error, 1)
|
|
|
|
respChan := make(chan *wire.MsgTx, 1)
|
|
|
|
|
|
|
|
// With the channel found, and the request crafted, we'll send over a
|
|
|
|
// force close request to the arbitrator that watches this channel.
|
|
|
|
select {
|
|
|
|
case arbitrator.forceCloseReqs <- &forceCloseReq{
|
|
|
|
errResp: errChan,
|
|
|
|
closeTx: respChan,
|
|
|
|
}:
|
|
|
|
case <-c.quit:
|
2018-09-12 02:05:26 +03:00
|
|
|
return nil, ErrChainArbExiting
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// We'll await two responses: the error response, and the transaction
|
2018-01-17 06:52:22 +03:00
|
|
|
// that closed out the channel.
|
|
|
|
select {
|
|
|
|
case err := <-errChan:
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
case <-c.quit:
|
2018-09-12 02:05:26 +03:00
|
|
|
return nil, ErrChainArbExiting
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
var closeTx *wire.MsgTx
|
|
|
|
select {
|
|
|
|
case closeTx = <-respChan:
|
|
|
|
case <-c.quit:
|
2018-09-12 02:05:26 +03:00
|
|
|
return nil, ErrChainArbExiting
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return closeTx, nil
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// WatchNewChannel sends the ChainArbitrator a message to create a
|
2018-01-17 06:52:22 +03:00
|
|
|
// ChannelArbitrator tasked with watching over a new channel. Once a new
|
|
|
|
// channel has finished its final funding flow, it should be registered with
|
|
|
|
// the ChainArbitrator so we can properly react to any on-chain events.
|
2018-01-19 01:00:35 +03:00
|
|
|
func (c *ChainArbitrator) WatchNewChannel(newChan *channeldb.OpenChannel) error {
|
2018-01-17 06:52:22 +03:00
|
|
|
c.Lock()
|
|
|
|
defer c.Unlock()
|
|
|
|
|
|
|
|
log.Infof("Creating new ChannelArbitrator for ChannelPoint(%v)",
|
|
|
|
newChan.FundingOutpoint)
|
|
|
|
|
|
|
|
// If we're already watching this channel, then we'll ignore this
|
|
|
|
// request.
|
|
|
|
chanPoint := newChan.FundingOutpoint
|
|
|
|
if _, ok := c.activeChannels[chanPoint]; ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
// First, also create an active chainWatcher for this channel to ensure
|
|
|
|
// that we detect any relevant on chain events.
|
|
|
|
chainWatcher, err := newChainWatcher(
|
2018-04-19 14:05:50 +03:00
|
|
|
chainWatcherConfig{
|
|
|
|
chanState: newChan,
|
|
|
|
notifier: c.cfg.Notifier,
|
|
|
|
signer: c.cfg.Signer,
|
|
|
|
isOurAddr: c.cfg.IsOurAddress,
|
2018-04-18 14:41:49 +03:00
|
|
|
contractBreach: func(retInfo *lnwallet.BreachRetribution) error {
|
|
|
|
return c.cfg.ContractBreach(chanPoint, retInfo)
|
|
|
|
},
|
2019-03-13 05:19:24 +03:00
|
|
|
extractStateNumHint: lnwallet.GetStateNumHint,
|
2018-01-20 04:17:14 +03:00
|
|
|
},
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
c.activeWatchers[newChan.FundingOutpoint] = chainWatcher
|
|
|
|
|
|
|
|
// We'll also create a new channel arbitrator instance using this new
|
2018-01-17 06:52:22 +03:00
|
|
|
// channel, and our internal state.
|
2018-01-19 01:00:35 +03:00
|
|
|
channelArb, err := newActiveChannelArbitrator(
|
2018-04-18 14:41:49 +03:00
|
|
|
newChan, c, chainWatcher.SubscribeChannelEvents(),
|
2018-01-19 01:00:35 +03:00
|
|
|
)
|
2018-01-17 06:52:22 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the arbitrator created, we'll add it to our set of active
|
|
|
|
// arbitrators, then launch it.
|
|
|
|
c.activeChannels[chanPoint] = channelArb
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
if err := channelArb.Start(); err != nil {
|
2018-01-17 06:52:22 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-01-19 01:00:35 +03:00
|
|
|
return chainWatcher.Start()
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
2018-01-19 01:02:21 +03:00
|
|
|
// SubscribeChannelEvents returns a new active subscription for the set of
|
|
|
|
// possible on-chain events for a particular channel. The struct can be used by
|
|
|
|
// callers to be notified whenever an event that changes the state of the
|
2018-10-22 06:40:02 +03:00
|
|
|
// channel on-chain occurs.
|
2018-01-19 01:02:21 +03:00
|
|
|
func (c *ChainArbitrator) SubscribeChannelEvents(
|
2018-04-18 14:41:49 +03:00
|
|
|
chanPoint wire.OutPoint) (*ChainEventSubscription, error) {
|
2018-01-19 01:02:21 +03:00
|
|
|
|
|
|
|
// First, we'll attempt to look up the active watcher for this channel.
|
|
|
|
// If we can't find it, then we'll return an error back to the caller.
|
|
|
|
watcher, ok := c.activeWatchers[chanPoint]
|
|
|
|
if !ok {
|
|
|
|
return nil, fmt.Errorf("unable to find watcher for: %v",
|
|
|
|
chanPoint)
|
|
|
|
}
|
2018-01-17 06:52:22 +03:00
|
|
|
|
2018-01-19 01:02:21 +03:00
|
|
|
// With the watcher located, we'll request for it to create a new chain
|
|
|
|
// event subscription client.
|
2018-04-18 14:41:49 +03:00
|
|
|
return watcher.SubscribeChannelEvents(), nil
|
2018-01-17 06:52:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// TODO(roasbeef): arbitration reports
|
|
|
|
// * types: contested, waiting for success conf, etc
|