routing: make routing retry behaviour consistent

Fixes the following issues:
- If the channel update of FailFeeInsufficient contains an invalid channel
  update, it is not possible to properly add to the failed channels set.
- FailAmountBelowMinimum may apply a channel update, but does not retry.
- FailIncorrectCltvExpiry immediately prunes the vertex without
  trying one more time.

In this commit, the logic for all three policy related errors is
aligned.
This commit is contained in:
Joost Jager 2018-10-24 10:45:30 +02:00
parent 6ba1144528
commit b6ce03e569
No known key found for this signature in database
GPG Key ID: AE6B0D042C8E38D9
2 changed files with 108 additions and 70 deletions

@ -339,6 +339,31 @@ func (p *paymentSession) ReportChannelFailure(e uint64) {
p.mc.Unlock() p.mc.Unlock()
} }
// ReportChannelPolicyFailure handles a failure message that relates to a
// channel policy. For these types of failures, the policy is updated and we
// want to keep it included during path finding. This function does mark the
// edge as 'policy failed once'. The next time it fails, the whole node will be
// pruned. This is to prevent nodes from keeping us busy by continuously sending
// new channel updates.
func (p *paymentSession) ReportChannelPolicyFailure(
errSource Vertex, failedChanID uint64) {
// Check to see if we've already reported a policy related failure for
// this channel. If so, then we'll prune out the vertex.
_, ok := p.errFailedPolicyChans[failedChanID]
if ok {
// TODO(joostjager): is this aggresive pruning still necessary?
// Just pruning edges may also work unless there is a huge
// number of failing channels from that node?
p.ReportVertexFailure(errSource)
return
}
// Finally, we'll record a policy failure from this node and move on.
p.errFailedPolicyChans[failedChanID] = struct{}{}
}
// RequestRoute returns a route which is likely to be capable for successfully // RequestRoute returns a route which is likely to be capable for successfully
// routing the specified HTLC payment to the target node. Initially the first // routing the specified HTLC payment to the target node. Initially the first
// set of paths returned from this method may encounter routing failure along // set of paths returned from this method may encounter routing failure along

@ -1758,6 +1758,49 @@ func (r *ChannelRouter) sendPayment(payment *LightningPayment,
"htlc=%x", errSource.SerializeCompressed(), "htlc=%x", errSource.SerializeCompressed(),
payment.PaymentHash[:]) payment.PaymentHash[:])
// Always determine chan id ourselves, because a channel
// update with id may not be available.
failedChanID, err := getFailedChannelID(route, errSource)
if err != nil {
return preImage, nil, err
}
// processChannelUpdateAndRetry is a closure that
// handles a failure message containing a channel
// update. This function always tries to apply the
// channel update and passes on the result to the
// payment session to adjust its view on the reliability
// of the network.
//
// As channel id, the locally determined channel id is
// used. It does not rely on the channel id that is part
// of the channel update message, because the remote
// node may lie to us or the update may be corrupt.
processChannelUpdateAndRetry := func(
update *lnwire.ChannelUpdate,
pubKey *btcec.PublicKey) {
// Try to apply the channel update.
updateOk := r.applyChannelUpdate(update, pubKey)
// If the update could not be applied, prune the
// edge. There is no reason to continue trying
// this channel.
//
// TODO: Could even prune the node completely?
// Or is there a valid reason for the channel
// update to fail?
if !updateOk {
paySession.ReportChannelFailure(
failedChanID,
)
}
paySession.ReportChannelPolicyFailure(
NewVertex(errSource), failedChanID,
)
}
switch onionErr := fErr.FailureMessage.(type) { switch onionErr := fErr.FailureMessage.(type) {
// If the end destination didn't know they payment // If the end destination didn't know they payment
// hash, then we'll terminate immediately. // hash, then we'll terminate immediately.
@ -1798,9 +1841,7 @@ func (r *ChannelRouter) sendPayment(payment *LightningPayment,
// that sent us this error, as it doesn't now what the // that sent us this error, as it doesn't now what the
// correct block height is. // correct block height is.
case *lnwire.FailExpiryTooSoon: case *lnwire.FailExpiryTooSoon:
update := onionErr.Update r.applyChannelUpdate(&onionErr.Update, errSource)
r.applyChannelUpdate(&update, errSource)
paySession.ReportVertexFailure(errVertex) paySession.ReportVertexFailure(errVertex)
continue continue
@ -1814,72 +1855,47 @@ func (r *ChannelRouter) sendPayment(payment *LightningPayment,
case *lnwire.FailInvalidOnionKey: case *lnwire.FailInvalidOnionKey:
return preImage, nil, sendError return preImage, nil, sendError
// If the onion error includes a channel update, and // If we get a failure due to violating the minimum
// isn't necessarily fatal, then we'll apply the update // amount, we'll apply the new minimum amount and retry
// and continue with the rest of the routes. // routing.
case *lnwire.FailAmountBelowMinimum: case *lnwire.FailAmountBelowMinimum:
update := onionErr.Update processChannelUpdateAndRetry(
r.applyChannelUpdate(&update, errSource) &onionErr.Update, errSource,
)
continue
return preImage, nil, sendError // If we get a failure due to a fee, we'll apply the
// If we get a failure due to a fee, so we'll apply the
// new fee update, and retry our attempt using the // new fee update, and retry our attempt using the
// newly updated fees. // newly updated fees.
case *lnwire.FailFeeInsufficient: case *lnwire.FailFeeInsufficient:
update := onionErr.Update processChannelUpdateAndRetry(
updateOk := r.applyChannelUpdate(&update, errSource) &onionErr.Update, errSource,
if !updateOk { )
pruneEdgeFailure(
paySession, route, errSource,
)
}
// We'll now check to see if we've already
// reported a fee related failure for this
// node. If so, then we'll actually prune out
// the vertex for now.
chanID := update.ShortChannelID.ToUint64()
_, ok := paySession.errFailedPolicyChans[chanID]
if ok {
paySession.ReportVertexFailure(errVertex)
continue
}
// Finally, we'll record a fee failure from
// this node and move on.
paySession.errFailedPolicyChans[chanID] = struct{}{}
continue continue
// If we get the failure for an intermediate node that // If we get the failure for an intermediate node that
// disagrees with our time lock values, then we'll // disagrees with our time lock values, then we'll
// prune it out for now, and continue with path // apply the new delta value and try it once more.
// finding.
case *lnwire.FailIncorrectCltvExpiry: case *lnwire.FailIncorrectCltvExpiry:
update := onionErr.Update processChannelUpdateAndRetry(
r.applyChannelUpdate(&update, errSource) &onionErr.Update, errSource,
)
paySession.ReportVertexFailure(errVertex)
continue continue
// The outgoing channel that this node was meant to // The outgoing channel that this node was meant to
// forward one is currently disabled, so we'll apply // forward one is currently disabled, so we'll apply
// the update and continue. // the update and continue.
case *lnwire.FailChannelDisabled: case *lnwire.FailChannelDisabled:
update := onionErr.Update r.applyChannelUpdate(&onionErr.Update, errSource)
r.applyChannelUpdate(&update, errSource) paySession.ReportChannelFailure(failedChanID)
pruneEdgeFailure(paySession, route, errSource)
continue continue
// It's likely that the outgoing channel didn't have // It's likely that the outgoing channel didn't have
// sufficient capacity, so we'll prune this edge for // sufficient capacity, so we'll prune this edge for
// now, and continue onwards with our path finding. // now, and continue onwards with our path finding.
case *lnwire.FailTemporaryChannelFailure: case *lnwire.FailTemporaryChannelFailure:
update := onionErr.Update r.applyChannelUpdate(onionErr.Update, errSource)
r.applyChannelUpdate(update, errSource) paySession.ReportChannelFailure(failedChanID)
pruneEdgeFailure(paySession, route, errSource)
continue continue
// If the send fail due to a node not having the // If the send fail due to a node not having the
@ -1904,7 +1920,7 @@ func (r *ChannelRouter) sendPayment(payment *LightningPayment,
// returning errors in order to attempt to black list // returning errors in order to attempt to black list
// another node. // another node.
case *lnwire.FailUnknownNextPeer: case *lnwire.FailUnknownNextPeer:
pruneEdgeFailure(paySession, route, errSource) paySession.ReportChannelFailure(failedChanID)
continue continue
// If the node wasn't able to forward for which ever // If the node wasn't able to forward for which ever
@ -1935,7 +1951,7 @@ func (r *ChannelRouter) sendPayment(payment *LightningPayment,
// we'll note this (exclude the vertex/edge), and // we'll note this (exclude the vertex/edge), and
// continue with the rest of the routes. // continue with the rest of the routes.
case *lnwire.FailPermanentChannelFailure: case *lnwire.FailPermanentChannelFailure:
pruneEdgeFailure(paySession, route, errSource) paySession.ReportChannelFailure(failedChanID)
continue continue
default: default:
@ -1947,34 +1963,31 @@ func (r *ChannelRouter) sendPayment(payment *LightningPayment,
} }
} }
// pruneEdgeFailure will attempts to prune an edge from the current available // getFailedChannelID tries to locate the failing channel given a route and the
// edges of the target payment session in response to an encountered routing // pubkey of the node that sent the error. It will assume that the error is
// error. // associated with the outgoing channel of the error node.
func pruneEdgeFailure(paySession *paymentSession, route *Route, func getFailedChannelID(route *Route, errSource *btcec.PublicKey) (
errSource *btcec.PublicKey) { uint64, error) {
// As this error indicates that the target channel was unable to carry // As this error indicates that the target channel was unable to carry
// this HTLC (for w/e reason), we'll query the index to find the // this HTLC (for w/e reason), we'll query the index to find the
// _outgoing_ channel the source of the error was meant to pass the // _outgoing_ channel the source of the error was meant to pass the
// HTLC along to. // HTLC along to.
badChan, ok := route.nextHopChannel(errSource) if badChan, ok := route.nextHopChannel(errSource); ok {
if !ok { return badChan.ChannelID, nil
// If we weren't able to find the hop *after* this node, then
// we'll attempt to disable the previous channel.
prevChan, ok := route.prevHopChannel(
errSource,
)
if !ok {
return
}
badChan = prevChan
} }
// If the channel was found, then we'll inform mission control of this // If we weren't able to find the hop *after* this node, then we'll
// failure so future attempts avoid this link temporarily. // attempt to disable the previous channel.
paySession.ReportChannelFailure(badChan.ChannelID) //
// TODO(joostjager): errSource must be the final hop then? In that case,
// certain types of errors are not expected. For example
// FailUnknownNextPeer. This could be a reason to prune the node?
if prevChan, ok := route.prevHopChannel(errSource); ok {
return prevChan.ChannelID, nil
}
return 0, fmt.Errorf("cannot find channel in route")
} }
// applyChannelUpdate validates a channel update and if valid, applies it to the // applyChannelUpdate validates a channel update and if valid, applies it to the