ef28d2aaed
Previously we only penalized the outgoing connections of a failing node. This turned out not to be sufficient, because the next route sometimes went into the same failing node again to try a different outgoing connection that wasn't yet known to mission control and therefore not penalized before.
561 lines
17 KiB
Go
561 lines
17 KiB
Go
package routing
|
|
|
|
import (
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/coreos/bbolt"
|
|
"github.com/lightningnetwork/lnd/channeldb"
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
"github.com/lightningnetwork/lnd/routing/route"
|
|
)
|
|
|
|
const (
|
|
// DefaultPenaltyHalfLife is the default half-life duration. The
|
|
// half-life duration defines after how much time a penalized node or
|
|
// channel is back at 50% probability.
|
|
DefaultPenaltyHalfLife = time.Hour
|
|
|
|
// minSecondChanceInterval is the minimum time required between
|
|
// second-chance failures.
|
|
//
|
|
// If nodes return a channel policy related failure, they may get a
|
|
// second chance to forward the payment. It could be that the channel
|
|
// policy that we are aware of is not up to date. This is especially
|
|
// important in case of mobile apps that are mostly offline.
|
|
//
|
|
// However, we don't want to give nodes the option to endlessly return
|
|
// new channel updates so that we are kept busy trying to route through
|
|
// that node until the payment loop times out.
|
|
//
|
|
// Therefore we only grant a second chance to a node if the previous
|
|
// second chance is sufficiently long ago. This is what
|
|
// minSecondChanceInterval defines. If a second policy failure comes in
|
|
// within that interval, we will apply a penalty.
|
|
//
|
|
// Second chances granted are tracked on the level of node pairs. This
|
|
// means that if a node has multiple channels to the same peer, they
|
|
// will only get a single second chance to route to that peer again.
|
|
// Nodes forward non-strict, so it isn't necessary to apply a less
|
|
// restrictive channel level tracking scheme here.
|
|
minSecondChanceInterval = time.Minute
|
|
|
|
// DefaultMaxMcHistory is the default maximum history size.
|
|
DefaultMaxMcHistory = 1000
|
|
|
|
// prevSuccessProbability is the assumed probability for node pairs that
|
|
// successfully relayed the previous attempt.
|
|
prevSuccessProbability = 0.95
|
|
|
|
// DefaultAprioriWeight is the default a priori weight. See
|
|
// MissionControlConfig for further explanation.
|
|
DefaultAprioriWeight = 0.5
|
|
)
|
|
|
|
// NodeResults contains previous results from a node to its peers.
|
|
type NodeResults map[route.Vertex]TimedPairResult
|
|
|
|
// MissionControl contains state which summarizes the past attempts of HTLC
|
|
// routing by external callers when sending payments throughout the network. It
|
|
// acts as a shared memory during routing attempts with the goal to optimize the
|
|
// payment attempt success rate.
|
|
//
|
|
// Failed payment attempts are reported to mission control. These reports are
|
|
// used to track the time of the last node or channel level failure. The time
|
|
// since the last failure is used to estimate a success probability that is fed
|
|
// into the path finding process for subsequent payment attempts.
|
|
type MissionControl struct {
|
|
// lastPairResult tracks the last payment result (on a pair basis) for
|
|
// each transited node. This is a multi-layer map that allows us to look
|
|
// up the failure history of all connected channels (node pairs) for a
|
|
// particular node.
|
|
lastPairResult map[route.Vertex]NodeResults
|
|
|
|
// lastSecondChance tracks the last time a second chance was granted for
|
|
// a directed node pair.
|
|
lastSecondChance map[DirectedNodePair]time.Time
|
|
|
|
// now is expected to return the current time. It is supplied as an
|
|
// external function to enable deterministic unit tests.
|
|
now func() time.Time
|
|
|
|
cfg *MissionControlConfig
|
|
|
|
store *missionControlStore
|
|
|
|
// estimator is the probability estimator that is used with the payment
|
|
// results that mission control collects.
|
|
estimator *probabilityEstimator
|
|
|
|
sync.Mutex
|
|
|
|
// TODO(roasbeef): further counters, if vertex continually unavailable,
|
|
// add to another generation
|
|
|
|
// TODO(roasbeef): also add favorable metrics for nodes
|
|
}
|
|
|
|
// MissionControlConfig defines parameters that control mission control
|
|
// behaviour.
|
|
type MissionControlConfig struct {
|
|
// PenaltyHalfLife defines after how much time a penalized node or
|
|
// channel is back at 50% probability.
|
|
PenaltyHalfLife time.Duration
|
|
|
|
// AprioriHopProbability is the assumed success probability of a hop in
|
|
// a route when no other information is available.
|
|
AprioriHopProbability float64
|
|
|
|
// MaxMcHistory defines the maximum number of payment results that are
|
|
// held on disk.
|
|
MaxMcHistory int
|
|
|
|
// AprioriWeight is a value in the range [0, 1] that defines to what
|
|
// extent historical results should be extrapolated to untried
|
|
// connections. Setting it to one will completely ignore historical
|
|
// results and always assume the configured a priori probability for
|
|
// untried connections. A value of zero will ignore the a priori
|
|
// probability completely and only base the probability on historical
|
|
// results, unless there are none available.
|
|
AprioriWeight float64
|
|
|
|
// SelfNode is our own pubkey.
|
|
SelfNode route.Vertex
|
|
}
|
|
|
|
// TimedPairResult describes a timestamped pair result.
|
|
type TimedPairResult struct {
|
|
// FailTime is the time of the last failure.
|
|
FailTime time.Time
|
|
|
|
// FailAmt is the amount of the last failure. This amount may be pushed
|
|
// up if a later success is higher than the last failed amount.
|
|
FailAmt lnwire.MilliSatoshi
|
|
|
|
// SuccessTime is the time of the last success.
|
|
SuccessTime time.Time
|
|
|
|
// SuccessAmt is the highest amount that successfully forwarded. This
|
|
// isn't necessarily the last success amount. The value of this field
|
|
// may also be pushed down if a later failure is lower than the highest
|
|
// success amount. Because of this, SuccessAmt may not match
|
|
// SuccessTime.
|
|
SuccessAmt lnwire.MilliSatoshi
|
|
}
|
|
|
|
// MissionControlSnapshot contains a snapshot of the current state of mission
|
|
// control.
|
|
type MissionControlSnapshot struct {
|
|
// Pairs is a list of channels for which specific information is
|
|
// logged.
|
|
Pairs []MissionControlPairSnapshot
|
|
}
|
|
|
|
// MissionControlPairSnapshot contains a snapshot of the current node pair
|
|
// state in mission control.
|
|
type MissionControlPairSnapshot struct {
|
|
// Pair is the node pair of which the state is described.
|
|
Pair DirectedNodePair
|
|
|
|
// TimedPairResult contains the data for this pair.
|
|
TimedPairResult
|
|
}
|
|
|
|
// paymentResult is the information that becomes available when a payment
|
|
// attempt completes.
|
|
type paymentResult struct {
|
|
id uint64
|
|
timeFwd, timeReply time.Time
|
|
route *route.Route
|
|
success bool
|
|
failureSourceIdx *int
|
|
failure lnwire.FailureMessage
|
|
}
|
|
|
|
// NewMissionControl returns a new instance of missionControl.
|
|
func NewMissionControl(db *bbolt.DB, cfg *MissionControlConfig) (
|
|
*MissionControl, error) {
|
|
|
|
log.Debugf("Instantiating mission control with config: "+
|
|
"PenaltyHalfLife=%v, AprioriHopProbability=%v, "+
|
|
"AprioriWeight=%v", cfg.PenaltyHalfLife,
|
|
cfg.AprioriHopProbability, cfg.AprioriWeight)
|
|
|
|
store, err := newMissionControlStore(db, cfg.MaxMcHistory)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
estimator := &probabilityEstimator{
|
|
aprioriHopProbability: cfg.AprioriHopProbability,
|
|
aprioriWeight: cfg.AprioriWeight,
|
|
penaltyHalfLife: cfg.PenaltyHalfLife,
|
|
prevSuccessProbability: prevSuccessProbability,
|
|
}
|
|
|
|
mc := &MissionControl{
|
|
lastPairResult: make(map[route.Vertex]NodeResults),
|
|
lastSecondChance: make(map[DirectedNodePair]time.Time),
|
|
now: time.Now,
|
|
cfg: cfg,
|
|
store: store,
|
|
estimator: estimator,
|
|
}
|
|
|
|
if err := mc.init(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return mc, nil
|
|
}
|
|
|
|
// init initializes mission control with historical data.
|
|
func (m *MissionControl) init() error {
|
|
log.Debugf("Mission control state reconstruction started")
|
|
|
|
start := time.Now()
|
|
|
|
results, err := m.store.fetchAll()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
for _, result := range results {
|
|
m.applyPaymentResult(result)
|
|
}
|
|
|
|
log.Debugf("Mission control state reconstruction finished: "+
|
|
"n=%v, time=%v", len(results), time.Now().Sub(start))
|
|
|
|
return nil
|
|
}
|
|
|
|
// ResetHistory resets the history of MissionControl returning it to a state as
|
|
// if no payment attempts have been made.
|
|
func (m *MissionControl) ResetHistory() error {
|
|
m.Lock()
|
|
defer m.Unlock()
|
|
|
|
if err := m.store.clear(); err != nil {
|
|
return err
|
|
}
|
|
|
|
m.lastPairResult = make(map[route.Vertex]NodeResults)
|
|
m.lastSecondChance = make(map[DirectedNodePair]time.Time)
|
|
|
|
log.Debugf("Mission control history cleared")
|
|
|
|
return nil
|
|
}
|
|
|
|
// GetProbability is expected to return the success probability of a payment
|
|
// from fromNode along edge.
|
|
func (m *MissionControl) GetProbability(fromNode, toNode route.Vertex,
|
|
amt lnwire.MilliSatoshi) float64 {
|
|
|
|
m.Lock()
|
|
defer m.Unlock()
|
|
|
|
now := m.now()
|
|
results := m.lastPairResult[fromNode]
|
|
|
|
// Use a distinct probability estimation function for local channels.
|
|
if fromNode == m.cfg.SelfNode {
|
|
return m.estimator.getLocalPairProbability(now, results, toNode)
|
|
}
|
|
|
|
return m.estimator.getPairProbability(now, results, toNode, amt)
|
|
}
|
|
|
|
// setLastPairResult stores a result for a node pair.
|
|
func (m *MissionControl) setLastPairResult(fromNode, toNode route.Vertex,
|
|
timestamp time.Time, result *pairResult) {
|
|
|
|
nodePairs, ok := m.lastPairResult[fromNode]
|
|
if !ok {
|
|
nodePairs = make(NodeResults)
|
|
m.lastPairResult[fromNode] = nodePairs
|
|
}
|
|
|
|
current := nodePairs[toNode]
|
|
|
|
// Apply the new result to the existing data for this pair. If there is
|
|
// no existing data, apply it to the default values for TimedPairResult.
|
|
if result.success {
|
|
successAmt := result.amt
|
|
current.SuccessTime = timestamp
|
|
|
|
// Only update the success amount if this amount is higher. This
|
|
// prevents the success range from shrinking when there is no
|
|
// reason to do so. For example: small amount probes shouldn't
|
|
// affect a previous success for a much larger amount.
|
|
if successAmt > current.SuccessAmt {
|
|
current.SuccessAmt = successAmt
|
|
}
|
|
|
|
// If the success amount goes into the failure range, move the
|
|
// failure range up. Future attempts up to the success amount
|
|
// are likely to succeed. We don't want to clear the failure
|
|
// completely, because we haven't learnt much for amounts above
|
|
// the current success amount.
|
|
if !current.FailTime.IsZero() && successAmt >= current.FailAmt {
|
|
current.FailAmt = successAmt + 1
|
|
}
|
|
} else {
|
|
// For failures we always want to update both the amount and the
|
|
// time. Those need to relate to the same result, because the
|
|
// time is used to gradually diminish the penality for that
|
|
// specific result. Updating the timestamp but not the amount
|
|
// could cause a failure for a lower amount (a more severe
|
|
// condition) to be revived as if it just happened.
|
|
failAmt := result.amt
|
|
current.FailTime = timestamp
|
|
current.FailAmt = failAmt
|
|
|
|
switch {
|
|
// The failure amount is set to zero when the failure is
|
|
// amount-independent, meaning that the attempt would have
|
|
// failed regardless of the amount. This should also reset the
|
|
// success amount to zero.
|
|
case failAmt == 0:
|
|
current.SuccessAmt = 0
|
|
|
|
// If the failure range goes into the success range, move the
|
|
// success range down.
|
|
case failAmt <= current.SuccessAmt:
|
|
current.SuccessAmt = failAmt - 1
|
|
}
|
|
}
|
|
|
|
log.Debugf("Setting %v->%v range to [%v-%v]",
|
|
fromNode, toNode, current.SuccessAmt, current.FailAmt)
|
|
|
|
nodePairs[toNode] = current
|
|
}
|
|
|
|
// setAllFail stores a fail result for all known connections to and from the
|
|
// given node.
|
|
func (m *MissionControl) setAllFail(node route.Vertex,
|
|
timestamp time.Time) {
|
|
|
|
for fromNode, nodePairs := range m.lastPairResult {
|
|
for toNode := range nodePairs {
|
|
if fromNode == node || toNode == node {
|
|
nodePairs[toNode] = TimedPairResult{
|
|
FailTime: timestamp,
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// requestSecondChance checks whether the node fromNode can have a second chance
|
|
// at providing a channel update for its channel with toNode.
|
|
func (m *MissionControl) requestSecondChance(timestamp time.Time,
|
|
fromNode, toNode route.Vertex) bool {
|
|
|
|
// Look up previous second chance time.
|
|
pair := DirectedNodePair{
|
|
From: fromNode,
|
|
To: toNode,
|
|
}
|
|
lastSecondChance, ok := m.lastSecondChance[pair]
|
|
|
|
// If the channel hasn't already be given a second chance or its last
|
|
// second chance was long ago, we give it another chance.
|
|
if !ok || timestamp.Sub(lastSecondChance) > minSecondChanceInterval {
|
|
m.lastSecondChance[pair] = timestamp
|
|
|
|
log.Debugf("Second chance granted for %v->%v", fromNode, toNode)
|
|
|
|
return true
|
|
}
|
|
|
|
// Otherwise penalize the channel, because we don't allow channel
|
|
// updates that are that frequent. This is to prevent nodes from keeping
|
|
// us busy by continuously sending new channel updates.
|
|
log.Debugf("Second chance denied for %v->%v, remaining interval: %v",
|
|
fromNode, toNode, timestamp.Sub(lastSecondChance))
|
|
|
|
return false
|
|
}
|
|
|
|
// GetHistorySnapshot takes a snapshot from the current mission control state
|
|
// and actual probability estimates.
|
|
func (m *MissionControl) GetHistorySnapshot() *MissionControlSnapshot {
|
|
m.Lock()
|
|
defer m.Unlock()
|
|
|
|
log.Debugf("Requesting history snapshot from mission control: "+
|
|
"pair_result_count=%v", len(m.lastPairResult))
|
|
|
|
pairs := make([]MissionControlPairSnapshot, 0, len(m.lastPairResult))
|
|
|
|
for fromNode, fromPairs := range m.lastPairResult {
|
|
for toNode, result := range fromPairs {
|
|
pair := NewDirectedNodePair(fromNode, toNode)
|
|
|
|
pairSnapshot := MissionControlPairSnapshot{
|
|
Pair: pair,
|
|
TimedPairResult: result,
|
|
}
|
|
|
|
pairs = append(pairs, pairSnapshot)
|
|
}
|
|
}
|
|
|
|
snapshot := MissionControlSnapshot{
|
|
Pairs: pairs,
|
|
}
|
|
|
|
return &snapshot
|
|
}
|
|
|
|
// GetPairHistorySnapshot returns the stored history for a given node pair.
|
|
func (m *MissionControl) GetPairHistorySnapshot(
|
|
fromNode, toNode route.Vertex) TimedPairResult {
|
|
|
|
m.Lock()
|
|
defer m.Unlock()
|
|
|
|
results, ok := m.lastPairResult[fromNode]
|
|
if !ok {
|
|
return TimedPairResult{}
|
|
}
|
|
|
|
result, ok := results[toNode]
|
|
if !ok {
|
|
return TimedPairResult{}
|
|
}
|
|
|
|
return result
|
|
}
|
|
|
|
// ReportPaymentFail reports a failed payment to mission control as input for
|
|
// future probability estimates. The failureSourceIdx argument indicates the
|
|
// failure source. If it is nil, the failure source is unknown. This function
|
|
// returns a reason if this failure is a final failure. In that case no further
|
|
// payment attempts need to be made.
|
|
func (m *MissionControl) ReportPaymentFail(paymentID uint64, rt *route.Route,
|
|
failureSourceIdx *int, failure lnwire.FailureMessage) (
|
|
*channeldb.FailureReason, error) {
|
|
|
|
timestamp := m.now()
|
|
|
|
result := &paymentResult{
|
|
success: false,
|
|
timeFwd: timestamp,
|
|
timeReply: timestamp,
|
|
id: paymentID,
|
|
failureSourceIdx: failureSourceIdx,
|
|
failure: failure,
|
|
route: rt,
|
|
}
|
|
|
|
return m.processPaymentResult(result)
|
|
}
|
|
|
|
// ReportPaymentSuccess reports a successful payment to mission control as input
|
|
// for future probability estimates.
|
|
func (m *MissionControl) ReportPaymentSuccess(paymentID uint64,
|
|
rt *route.Route) error {
|
|
|
|
timestamp := m.now()
|
|
|
|
result := &paymentResult{
|
|
timeFwd: timestamp,
|
|
timeReply: timestamp,
|
|
id: paymentID,
|
|
success: true,
|
|
route: rt,
|
|
}
|
|
|
|
_, err := m.processPaymentResult(result)
|
|
return err
|
|
}
|
|
|
|
// processPaymentResult stores a payment result in the mission control store and
|
|
// updates mission control's in-memory state.
|
|
func (m *MissionControl) processPaymentResult(result *paymentResult) (
|
|
*channeldb.FailureReason, error) {
|
|
|
|
// Store complete result in database.
|
|
if err := m.store.AddResult(result); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Apply result to update mission control state.
|
|
reason := m.applyPaymentResult(result)
|
|
|
|
return reason, nil
|
|
}
|
|
|
|
// applyPaymentResult applies a payment result as input for future probability
|
|
// estimates. It returns a bool indicating whether this error is a final error
|
|
// and no further payment attempts need to be made.
|
|
func (m *MissionControl) applyPaymentResult(
|
|
result *paymentResult) *channeldb.FailureReason {
|
|
|
|
// Interpret result.
|
|
i := interpretResult(
|
|
result.route, result.success, result.failureSourceIdx,
|
|
result.failure,
|
|
)
|
|
|
|
// Update mission control state using the interpretation.
|
|
m.Lock()
|
|
defer m.Unlock()
|
|
|
|
if i.policyFailure != nil {
|
|
if m.requestSecondChance(
|
|
result.timeReply,
|
|
i.policyFailure.From, i.policyFailure.To,
|
|
) {
|
|
return nil
|
|
}
|
|
}
|
|
|
|
// If there is a node-level failure, record a failure for every tried
|
|
// connection of that node. A node-level failure can be considered as a
|
|
// failure that would have occurred with any of the node's channels.
|
|
//
|
|
// Ideally we'd also record the failure for the untried connections of
|
|
// the node. Unfortunately this would require access to the graph and
|
|
// adding this dependency and db calls does not outweigh the benefits.
|
|
//
|
|
// Untried connections will fall back to the node probability. After the
|
|
// call to setAllPairResult below, the node probability will be equal to
|
|
// the probability of the tried channels except that the a priori
|
|
// probability is mixed in too. This effect is controlled by the
|
|
// aprioriWeight parameter. If that parameter isn't set to an extreme
|
|
// and there are a few known connections, there shouldn't be much of a
|
|
// difference. The largest difference occurs when aprioriWeight is 1. In
|
|
// that case, a node-level failure would not be applied to untried
|
|
// channels.
|
|
if i.nodeFailure != nil {
|
|
log.Debugf("Reporting node failure to Mission Control: "+
|
|
"node=%v", *i.nodeFailure)
|
|
|
|
m.setAllFail(*i.nodeFailure, result.timeReply)
|
|
}
|
|
|
|
for pair, pairResult := range i.pairResults {
|
|
pairResult := pairResult
|
|
|
|
if pairResult.success {
|
|
log.Debugf("Reporting pair success to Mission "+
|
|
"Control: pair=%v", pair)
|
|
} else {
|
|
log.Debugf("Reporting pair failure to Mission "+
|
|
"Control: pair=%v, amt=%v",
|
|
pair, pairResult.amt)
|
|
}
|
|
|
|
m.setLastPairResult(
|
|
pair.From, pair.To, result.timeReply, &pairResult,
|
|
)
|
|
}
|
|
|
|
return i.finalFailureReason
|
|
}
|