2017-03-21 04:25:54 +03:00
|
|
|
package routing
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"fmt"
|
2017-07-14 22:32:00 +03:00
|
|
|
"image/color"
|
2017-10-02 18:54:29 +03:00
|
|
|
"math/rand"
|
2017-10-11 05:49:09 +03:00
|
|
|
"strings"
|
2017-03-21 04:25:54 +03:00
|
|
|
"testing"
|
2017-07-14 22:32:00 +03:00
|
|
|
"time"
|
2017-03-21 04:25:54 +03:00
|
|
|
|
|
|
|
"github.com/lightningnetwork/lnd/channeldb"
|
2017-10-11 05:49:09 +03:00
|
|
|
"github.com/lightningnetwork/lnd/htlcswitch"
|
2017-03-30 04:01:28 +03:00
|
|
|
"github.com/roasbeef/btcd/wire"
|
|
|
|
|
|
|
|
"github.com/davecgh/go-spew/spew"
|
2017-06-29 16:40:45 +03:00
|
|
|
"github.com/lightningnetwork/lightning-onion"
|
2017-03-21 04:25:54 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
|
|
"github.com/roasbeef/btcd/btcec"
|
|
|
|
)
|
|
|
|
|
2018-02-28 06:09:41 +03:00
|
|
|
// defaultNumRoutes is the default value for the maximum number of routes to
|
|
|
|
// be returned by FindRoutes
|
|
|
|
const defaultNumRoutes = 10
|
|
|
|
|
2017-03-21 04:25:54 +03:00
|
|
|
type testCtx struct {
|
|
|
|
router *ChannelRouter
|
|
|
|
|
|
|
|
graph *channeldb.ChannelGraph
|
|
|
|
|
|
|
|
aliases map[string]*btcec.PublicKey
|
|
|
|
|
|
|
|
chain *mockChain
|
|
|
|
|
2017-05-11 03:22:26 +03:00
|
|
|
chainView *mockChainView
|
2017-03-21 04:25:54 +03:00
|
|
|
}
|
|
|
|
|
2017-11-07 03:00:17 +03:00
|
|
|
func (c *testCtx) RestartRouter() error {
|
|
|
|
// First, we'll reset the chainView's state as it doesn't persist the
|
|
|
|
// filter between restarts.
|
|
|
|
c.chainView.Reset()
|
|
|
|
|
|
|
|
// With the chainView reset, we'll now re-create the router itself, and
|
|
|
|
// start it.
|
|
|
|
router, err := New(Config{
|
|
|
|
Graph: c.graph,
|
|
|
|
Chain: c.chain,
|
|
|
|
ChainView: c.chainView,
|
2018-01-31 07:26:26 +03:00
|
|
|
SendToSwitch: func(_ [33]byte,
|
2017-11-07 03:00:17 +03:00
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
return [32]byte{}, nil
|
|
|
|
},
|
|
|
|
ChannelPruneExpiry: time.Hour * 24,
|
|
|
|
GraphPruneInterval: time.Hour * 2,
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("unable to create router %v", err)
|
|
|
|
}
|
|
|
|
if err := router.Start(); err != nil {
|
|
|
|
return fmt.Errorf("unable to start router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, we'll swap out the pointer in the testCtx with this fresh
|
|
|
|
// instance of the router.
|
|
|
|
c.router = router
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-11-30 03:44:08 +03:00
|
|
|
func copyPubKey(pub *btcec.PublicKey) *btcec.PublicKey {
|
|
|
|
return &btcec.PublicKey{
|
|
|
|
Curve: btcec.S256(),
|
|
|
|
X: pub.X,
|
|
|
|
Y: pub.Y,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-03-21 04:25:54 +03:00
|
|
|
func createTestCtx(startingHeight uint32, testGraph ...string) (*testCtx, func(), error) {
|
|
|
|
var (
|
|
|
|
graph *channeldb.ChannelGraph
|
|
|
|
sourceNode *channeldb.LightningNode
|
|
|
|
cleanup func()
|
|
|
|
err error
|
|
|
|
)
|
|
|
|
|
|
|
|
aliasMap := make(map[string]*btcec.PublicKey)
|
|
|
|
|
|
|
|
// If the testGraph isn't set, then we'll create an empty graph to
|
|
|
|
// start out with. Our usage of a variadic parameter allows caller to
|
|
|
|
// omit the testGraph argument all together if they wish to start with
|
|
|
|
// a blank graph.
|
|
|
|
if testGraph == nil {
|
|
|
|
// First we'll set up a test graph for usage within the test.
|
|
|
|
graph, cleanup, err = makeTestGraph()
|
|
|
|
if err != nil {
|
|
|
|
return nil, nil, fmt.Errorf("unable to create test graph: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-03-30 04:01:28 +03:00
|
|
|
sourceNode, err = createTestNode()
|
2017-03-21 04:25:54 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, nil, fmt.Errorf("unable to create source node: %v", err)
|
|
|
|
}
|
|
|
|
if err = graph.SetSourceNode(sourceNode); err != nil {
|
|
|
|
return nil, nil, fmt.Errorf("unable to set source node: %v", err)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Otherwise, we'll attempt to locate and parse out the file
|
|
|
|
// that encodes the graph that our tests should be run against.
|
|
|
|
graph, cleanup, aliasMap, err = parseTestGraph(testGraph[0])
|
|
|
|
if err != nil {
|
|
|
|
return nil, nil, fmt.Errorf("unable to create test graph: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next we'll initialize an instance of the channel router with mock
|
|
|
|
// versions of the chain and channel notifier. As we don't need to test
|
|
|
|
// any p2p functionality, the peer send and switch send messages won't
|
|
|
|
// be populated.
|
|
|
|
chain := newMockChain(startingHeight)
|
2017-11-07 03:00:51 +03:00
|
|
|
chainView := newMockChainView(chain)
|
2017-03-21 04:25:54 +03:00
|
|
|
router, err := New(Config{
|
2017-05-11 03:22:26 +03:00
|
|
|
Graph: graph,
|
|
|
|
Chain: chain,
|
|
|
|
ChainView: chainView,
|
2018-01-31 07:26:26 +03:00
|
|
|
SendToSwitch: func(_ [33]byte, _ *lnwire.UpdateAddHTLC,
|
|
|
|
_ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
|
2017-03-21 04:25:54 +03:00
|
|
|
return [32]byte{}, nil
|
|
|
|
},
|
2017-10-05 05:39:38 +03:00
|
|
|
ChannelPruneExpiry: time.Hour * 24,
|
|
|
|
GraphPruneInterval: time.Hour * 2,
|
2017-03-21 04:25:54 +03:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, nil, fmt.Errorf("unable to create router %v", err)
|
|
|
|
}
|
|
|
|
if err := router.Start(); err != nil {
|
|
|
|
return nil, nil, fmt.Errorf("unable to start router: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-11-07 03:00:51 +03:00
|
|
|
ctx := &testCtx{
|
2017-05-11 03:22:26 +03:00
|
|
|
router: router,
|
|
|
|
graph: graph,
|
|
|
|
aliases: aliasMap,
|
|
|
|
chain: chain,
|
|
|
|
chainView: chainView,
|
2017-11-07 03:00:51 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
cleanUp := func() {
|
|
|
|
ctx.router.Stop()
|
|
|
|
cleanup()
|
|
|
|
}
|
|
|
|
|
|
|
|
return ctx, cleanUp, nil
|
2017-03-21 04:25:54 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// TestFindRoutesFeeSorting asserts that routes found by the FindRoutes method
|
|
|
|
// within the channel router are properly returned in a sorted order, with the
|
|
|
|
// lowest fee route coming first.
|
|
|
|
func TestFindRoutesFeeSorting(t *testing.T) {
|
2017-06-17 01:59:20 +03:00
|
|
|
t.Parallel()
|
|
|
|
|
2017-03-21 04:25:54 +03:00
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight, basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// In this test we'd like to ensure proper integration of the various
|
|
|
|
// functions that are involved in path finding, and also route
|
|
|
|
// selection.
|
|
|
|
|
|
|
|
// Execute a query for all possible routes between roasbeef and luo ji.
|
2017-08-22 09:43:20 +03:00
|
|
|
paymentAmt := lnwire.NewMSatFromSatoshis(100)
|
2017-03-21 04:25:54 +03:00
|
|
|
target := ctx.aliases["luoji"]
|
2017-10-19 08:13:11 +03:00
|
|
|
routes, err := ctx.router.FindRoutes(target, paymentAmt,
|
2018-02-28 06:09:41 +03:00
|
|
|
defaultNumRoutes, DefaultFinalCLTVDelta)
|
2017-03-21 04:25:54 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to find any routes: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Exactly, two such paths should be found.
|
|
|
|
if len(routes) != 2 {
|
2017-12-07 05:39:50 +03:00
|
|
|
t.Fatalf("2 routes should've been selected, instead %v were: %v",
|
2017-04-14 00:42:29 +03:00
|
|
|
len(routes), spew.Sdump(routes))
|
2017-03-21 04:25:54 +03:00
|
|
|
}
|
|
|
|
|
2017-12-07 05:39:50 +03:00
|
|
|
// We shouldn't pay a fee for the fist route, but the second route
|
|
|
|
// should have a fee intact.
|
|
|
|
if routes[0].TotalFees != 0 {
|
|
|
|
t.Fatalf("incorrect fees for first route, expected 0 got: %v",
|
|
|
|
routes[0].TotalFees)
|
|
|
|
}
|
|
|
|
if routes[1].TotalFees == 0 {
|
2017-12-07 05:45:45 +03:00
|
|
|
t.Fatalf("total fees not set in second route: %v",
|
2017-12-07 05:39:50 +03:00
|
|
|
spew.Sdump(routes[0]))
|
|
|
|
}
|
|
|
|
|
2017-03-21 04:25:54 +03:00
|
|
|
// The paths should properly be ranked according to their total fee
|
|
|
|
// rate.
|
|
|
|
if routes[0].TotalFees > routes[1].TotalFees {
|
|
|
|
t.Fatalf("routes not ranked by total fee: %v",
|
|
|
|
spew.Sdump(routes))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-03-21 04:58:21 +03:00
|
|
|
// TestSendPaymentRouteFailureFallback tests that when sending a payment, if
|
|
|
|
// one of the target routes is seen as unavailable, then the next route in the
|
|
|
|
// queue is used instead. This process should continue until either a payment
|
|
|
|
// succeeds, or all routes have been exhausted.
|
|
|
|
func TestSendPaymentRouteFailureFallback(t *testing.T) {
|
2017-06-17 01:59:20 +03:00
|
|
|
t.Parallel()
|
|
|
|
|
2017-03-21 04:58:21 +03:00
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight, basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Craft a LightningPayment struct that'll send a payment from roasbeef
|
|
|
|
// to luo ji for 100 satoshis.
|
|
|
|
var payHash [32]byte
|
|
|
|
payment := LightningPayment{
|
|
|
|
Target: ctx.aliases["luoji"],
|
2017-08-22 09:43:20 +03:00
|
|
|
Amount: lnwire.NewMSatFromSatoshis(1000),
|
2017-03-21 04:58:21 +03:00
|
|
|
PaymentHash: payHash,
|
|
|
|
}
|
|
|
|
|
|
|
|
var preImage [32]byte
|
|
|
|
copy(preImage[:], bytes.Repeat([]byte{9}, 32))
|
|
|
|
|
2017-10-17 05:07:31 +03:00
|
|
|
sourceNode := ctx.router.selfNode
|
|
|
|
|
2017-03-21 04:58:21 +03:00
|
|
|
// We'll modify the SendToSwitch method that's been set within the
|
|
|
|
// router's configuration to ignore the path that has luo ji as the
|
|
|
|
// first hop. This should force the router to instead take the
|
|
|
|
// available two hop path (through satoshi).
|
2018-01-31 07:26:26 +03:00
|
|
|
ctx.router.cfg.SendToSwitch = func(n [33]byte,
|
2017-06-29 16:40:45 +03:00
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
2017-03-21 04:58:21 +03:00
|
|
|
|
2018-01-31 07:26:26 +03:00
|
|
|
if bytes.Equal(ctx.aliases["luoji"].SerializeCompressed(), n[:]) {
|
|
|
|
pub, err := sourceNode.PubKey()
|
|
|
|
if err != nil {
|
|
|
|
return preImage, err
|
|
|
|
}
|
2017-10-11 05:49:09 +03:00
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
2018-01-31 07:26:26 +03:00
|
|
|
ErrorSource: pub,
|
2017-10-17 05:07:31 +03:00
|
|
|
// TODO(roasbeef): temp node failure should be?
|
|
|
|
FailureMessage: &lnwire.FailTemporaryChannelFailure{},
|
2017-10-11 05:49:09 +03:00
|
|
|
}
|
2017-03-21 04:58:21 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return preImage, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Send off the payment request to the router, route through satoshi
|
|
|
|
// should've been selected as a fall back and succeeded correctly.
|
|
|
|
paymentPreImage, route, err := ctx.router.SendPayment(&payment)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// The route selected should have two hops
|
|
|
|
if len(route.Hops) != 2 {
|
|
|
|
t.Fatalf("incorrect route length: expected %v got %v", 2,
|
|
|
|
len(route.Hops))
|
|
|
|
}
|
|
|
|
|
|
|
|
// The preimage should match up with the once created above.
|
|
|
|
if !bytes.Equal(paymentPreImage[:], preImage[:]) {
|
|
|
|
t.Fatalf("incorrect preimage used: expected %x got %x",
|
|
|
|
preImage[:], paymentPreImage[:])
|
|
|
|
}
|
|
|
|
|
|
|
|
// The route should have satoshi as the first hop.
|
|
|
|
if route.Hops[0].Channel.Node.Alias != "satoshi" {
|
2017-10-11 05:49:09 +03:00
|
|
|
t.Fatalf("route should go through satoshi as first hop, "+
|
|
|
|
"instead passes through: %v",
|
|
|
|
route.Hops[0].Channel.Node.Alias)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-20 05:08:55 +03:00
|
|
|
// TestSendPaymentErrorRepeatedFeeInsufficient tests that if we receive
|
|
|
|
// multiple fee related errors from a channel that we're attempting to route
|
|
|
|
// through, then we'll prune the channel after the second attempt.
|
|
|
|
func TestSendPaymentErrorRepeatedFeeInsufficient(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight, basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Craft a LightningPayment struct that'll send a payment from roasbeef
|
|
|
|
// to luo ji for 100 satoshis.
|
|
|
|
var payHash [32]byte
|
|
|
|
payment := LightningPayment{
|
|
|
|
Target: ctx.aliases["luoji"],
|
|
|
|
Amount: lnwire.NewMSatFromSatoshis(1000),
|
|
|
|
PaymentHash: payHash,
|
|
|
|
}
|
|
|
|
|
|
|
|
var preImage [32]byte
|
|
|
|
copy(preImage[:], bytes.Repeat([]byte{9}, 32))
|
|
|
|
|
|
|
|
// We'll also fetch the first outgoing channel edge from roasbeef to
|
|
|
|
// luo ji. We'll obtain this as we'll need to to generate the
|
|
|
|
// FeeInsufficient error that we'll send back.
|
|
|
|
chanID := uint64(689530843)
|
|
|
|
_, _, edgeUpateToFail, err := ctx.graph.FetchChannelEdgesByID(chanID)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch chan id: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
errChanUpdate := lnwire.ChannelUpdate{
|
|
|
|
ShortChannelID: lnwire.NewShortChanIDFromInt(chanID),
|
|
|
|
Timestamp: uint32(edgeUpateToFail.LastUpdate.Unix()),
|
|
|
|
Flags: edgeUpateToFail.Flags,
|
|
|
|
TimeLockDelta: edgeUpateToFail.TimeLockDelta,
|
|
|
|
HtlcMinimumMsat: edgeUpateToFail.MinHTLC,
|
|
|
|
BaseFee: uint32(edgeUpateToFail.FeeBaseMSat),
|
|
|
|
FeeRate: uint32(edgeUpateToFail.FeeProportionalMillionths),
|
|
|
|
}
|
|
|
|
|
|
|
|
sourceNode := ctx.router.selfNode
|
|
|
|
|
|
|
|
// We'll now modify the SendToSwitch method to return an error for the
|
|
|
|
// outgoing channel to luo ji. This will be a fee related error, so it
|
|
|
|
// should only cause the edge to be pruned after the second attempt.
|
|
|
|
ctx.router.cfg.SendToSwitch = func(n [33]byte,
|
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
|
|
|
|
if bytes.Equal(ctx.aliases["luoji"].SerializeCompressed(), n[:]) {
|
|
|
|
pub, err := sourceNode.PubKey()
|
|
|
|
if err != nil {
|
|
|
|
return preImage, err
|
|
|
|
}
|
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
|
|
|
ErrorSource: pub,
|
|
|
|
|
|
|
|
// Within our error, we'll add a channel update
|
|
|
|
// which is meant to refelct he new fee
|
|
|
|
// schedule for the node/channel.
|
|
|
|
FailureMessage: &lnwire.FailFeeInsufficient{
|
|
|
|
Update: errChanUpdate,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return preImage, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Send off the payment request to the router, route through satoshi
|
|
|
|
// should've been selected as a fall back and succeeded correctly.
|
|
|
|
paymentPreImage, route, err := ctx.router.SendPayment(&payment)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// The route selected should have two hops
|
|
|
|
if len(route.Hops) != 2 {
|
|
|
|
t.Fatalf("incorrect route length: expected %v got %v", 2,
|
|
|
|
len(route.Hops))
|
|
|
|
}
|
|
|
|
|
|
|
|
// The preimage should match up with the once created above.
|
|
|
|
if !bytes.Equal(paymentPreImage[:], preImage[:]) {
|
|
|
|
t.Fatalf("incorrect preimage used: expected %x got %x",
|
|
|
|
preImage[:], paymentPreImage[:])
|
|
|
|
}
|
|
|
|
|
|
|
|
// The route should have satoshi as the first hop.
|
|
|
|
if route.Hops[0].Channel.Node.Alias != "satoshi" {
|
|
|
|
t.Fatalf("route should go through satoshi as first hop, "+
|
|
|
|
"instead passes through: %v",
|
|
|
|
route.Hops[0].Channel.Node.Alias)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-23 03:40:07 +03:00
|
|
|
// TestSendPaymentErrorNonFinalTimeLockErrors tests that if we receive either
|
|
|
|
// an ExpiryTooSoon or a IncorrectCltvExpiry error from a node, then we prune
|
|
|
|
// that node from the available graph witin a mission control session. This
|
|
|
|
// test ensures that we'll route around errors due to nodes not knowing the
|
|
|
|
// current block height.
|
|
|
|
func TestSendPaymentErrorNonFinalTimeLockErrors(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight, basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Craft a LightningPayment struct that'll send a payment from roasbeef
|
|
|
|
// to sophon for 1k satoshis.
|
|
|
|
var payHash [32]byte
|
|
|
|
payment := LightningPayment{
|
|
|
|
Target: ctx.aliases["sophon"],
|
|
|
|
Amount: lnwire.NewMSatFromSatoshis(1000),
|
|
|
|
PaymentHash: payHash,
|
|
|
|
}
|
|
|
|
|
|
|
|
var preImage [32]byte
|
|
|
|
copy(preImage[:], bytes.Repeat([]byte{9}, 32))
|
|
|
|
|
|
|
|
// We'll also fetch the first outgoing channel edge from roasbeef to
|
|
|
|
// son goku. This edge will be included in the time lock related expiry
|
|
|
|
// errors that we'll get back due to disagrements in what the current
|
|
|
|
// block height is.
|
|
|
|
chanID := uint64(3495345)
|
|
|
|
_, _, edgeUpateToFail, err := ctx.graph.FetchChannelEdgesByID(chanID)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch chan id: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
errChanUpdate := lnwire.ChannelUpdate{
|
|
|
|
ShortChannelID: lnwire.NewShortChanIDFromInt(chanID),
|
|
|
|
Timestamp: uint32(edgeUpateToFail.LastUpdate.Unix()),
|
|
|
|
Flags: edgeUpateToFail.Flags,
|
|
|
|
TimeLockDelta: edgeUpateToFail.TimeLockDelta,
|
|
|
|
HtlcMinimumMsat: edgeUpateToFail.MinHTLC,
|
|
|
|
BaseFee: uint32(edgeUpateToFail.FeeBaseMSat),
|
|
|
|
FeeRate: uint32(edgeUpateToFail.FeeProportionalMillionths),
|
|
|
|
}
|
|
|
|
|
|
|
|
// The error will be returned by Son Goku.
|
|
|
|
sourceNode := ctx.aliases["songoku"]
|
|
|
|
|
|
|
|
// We'll now modify the SendToSwitch method to return an error for the
|
|
|
|
// outgoing channel to son goku. Since this is a time lock related
|
|
|
|
// error, we should fail the payment flow all together, as Goku is the
|
|
|
|
// only channel to Sophon.
|
|
|
|
ctx.router.cfg.SendToSwitch = func(n [33]byte,
|
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
|
|
|
|
if bytes.Equal(sourceNode.SerializeCompressed(), n[:]) {
|
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
|
|
|
ErrorSource: sourceNode,
|
|
|
|
FailureMessage: &lnwire.FailExpiryTooSoon{
|
|
|
|
Update: errChanUpdate,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return preImage, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// assertExpectedPath is a helper function that asserts the returned
|
|
|
|
// route properly routes around the failure we've introduced in the
|
|
|
|
// graph.
|
|
|
|
assertExpectedPath := func(retPreImage [32]byte, route *Route) {
|
|
|
|
// The route selected should have two hops
|
|
|
|
if len(route.Hops) != 2 {
|
|
|
|
t.Fatalf("incorrect route length: expected %v got %v", 2,
|
|
|
|
len(route.Hops))
|
|
|
|
}
|
|
|
|
|
|
|
|
// The preimage should match up with the once created above.
|
|
|
|
if !bytes.Equal(retPreImage[:], preImage[:]) {
|
|
|
|
t.Fatalf("incorrect preimage used: expected %x got %x",
|
|
|
|
preImage[:], retPreImage[:])
|
|
|
|
}
|
|
|
|
|
|
|
|
// The route should have satoshi as the first hop.
|
|
|
|
if route.Hops[0].Channel.Node.Alias != "phamnuwen" {
|
|
|
|
t.Fatalf("route should go through phamnuwen as first hop, "+
|
|
|
|
"instead passes through: %v",
|
|
|
|
route.Hops[0].Channel.Node.Alias)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Send off the payment request to the router, this payment should
|
|
|
|
// suceed as we should actually go through Pham Nuwen in order to get
|
|
|
|
// to Sophon, even though he has higher fees.
|
|
|
|
paymentPreImage, route, err := ctx.router.SendPayment(&payment)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
assertExpectedPath(paymentPreImage, route)
|
|
|
|
|
|
|
|
// We'll now modify the error return an IncorrectCltvExpiry error
|
|
|
|
// instead, this should result in the same behavior of roasbeef routing
|
|
|
|
// around the faulty Son Goku node.
|
|
|
|
ctx.router.cfg.SendToSwitch = func(n [33]byte,
|
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
|
|
|
|
if bytes.Equal(sourceNode.SerializeCompressed(), n[:]) {
|
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
|
|
|
ErrorSource: sourceNode,
|
|
|
|
FailureMessage: &lnwire.FailIncorrectCltvExpiry{
|
|
|
|
Update: errChanUpdate,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return preImage, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Once again, Roasbeef should route around Goku since they disagree
|
|
|
|
// w.r.t to the block height, and instead go through Pham Nuwen.
|
|
|
|
paymentPreImage, route, err = ctx.router.SendPayment(&payment)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
assertExpectedPath(paymentPreImage, route)
|
|
|
|
}
|
|
|
|
|
2017-10-11 05:49:09 +03:00
|
|
|
// TestSendPaymentErrorPathPruning tests that the send of candidate routes
|
|
|
|
// properly gets pruned in response to ForwardingError response from the
|
|
|
|
// underlying SendToSwitch function.
|
|
|
|
func TestSendPaymentErrorPathPruning(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight, basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Craft a LightningPayment struct that'll send a payment from roasbeef
|
|
|
|
// to luo ji for 100 satoshis.
|
|
|
|
var payHash [32]byte
|
|
|
|
payment := LightningPayment{
|
|
|
|
Target: ctx.aliases["luoji"],
|
|
|
|
Amount: lnwire.NewMSatFromSatoshis(1000),
|
|
|
|
PaymentHash: payHash,
|
|
|
|
}
|
|
|
|
|
|
|
|
var preImage [32]byte
|
|
|
|
copy(preImage[:], bytes.Repeat([]byte{9}, 32))
|
|
|
|
|
2017-10-19 08:13:11 +03:00
|
|
|
sourceNode, err := ctx.graph.SourceNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch source node: %v", err)
|
|
|
|
}
|
2017-10-11 05:49:09 +03:00
|
|
|
|
2018-01-31 07:26:26 +03:00
|
|
|
sourcePub, err := sourceNode.PubKey()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch source node pub: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-10-11 05:49:09 +03:00
|
|
|
// First, we'll modify the SendToSwitch method to return an error
|
|
|
|
// indicating that the channel from roasbeef to luoji is not operable
|
|
|
|
// with an UnknownNextPeer.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): filtering should be intelligent enough so just not
|
|
|
|
// go through satoshi at all at this point.
|
2018-01-31 07:26:26 +03:00
|
|
|
ctx.router.cfg.SendToSwitch = func(n [33]byte,
|
2017-10-11 05:49:09 +03:00
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
|
2018-01-31 07:26:26 +03:00
|
|
|
if bytes.Equal(ctx.aliases["luoji"].SerializeCompressed(), n[:]) {
|
2017-10-11 05:49:09 +03:00
|
|
|
// We'll first simulate an error from the first
|
|
|
|
// outgoing link to simulate the channel from luo ji to
|
|
|
|
// roasbeef not having enough capacity.
|
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
2018-01-31 07:26:26 +03:00
|
|
|
ErrorSource: sourcePub,
|
2017-10-11 05:49:09 +03:00
|
|
|
FailureMessage: &lnwire.FailTemporaryChannelFailure{},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next, we'll create an error from satoshi to indicate
|
|
|
|
// that the luoji node is not longer online, which should
|
|
|
|
// prune out the rest of the routes.
|
2018-01-31 07:26:26 +03:00
|
|
|
if bytes.Equal(ctx.aliases["satoshi"].SerializeCompressed(), n[:]) {
|
2017-10-11 05:49:09 +03:00
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
|
|
|
ErrorSource: ctx.aliases["satoshi"],
|
|
|
|
FailureMessage: &lnwire.FailUnknownNextPeer{},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return preImage, nil
|
|
|
|
}
|
|
|
|
|
2017-10-17 05:07:31 +03:00
|
|
|
ctx.router.missionControl.ResetHistory()
|
|
|
|
|
2017-10-11 05:49:09 +03:00
|
|
|
// When we try to dispatch that payment, we should receive an error as
|
|
|
|
// both attempts should fail and cause both routes to be pruned.
|
|
|
|
_, _, err = ctx.router.SendPayment(&payment)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("payment didn't return error")
|
|
|
|
}
|
|
|
|
|
|
|
|
// The final error returned should also indicate that the peer wasn't
|
|
|
|
// online (the last error we returned).
|
|
|
|
if !strings.Contains(err.Error(), "UnknownNextPeer") {
|
|
|
|
t.Fatalf("expected UnknownNextPeer instead got: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-10-17 05:07:31 +03:00
|
|
|
ctx.router.missionControl.ResetHistory()
|
|
|
|
|
2017-10-11 05:49:09 +03:00
|
|
|
// Next, we'll modify the SendToSwitch method to indicate that luo ji
|
|
|
|
// wasn't originally online. This should also halt the send all
|
|
|
|
// together as all paths contain luoji and he can't be reached.
|
2018-01-31 07:26:26 +03:00
|
|
|
ctx.router.cfg.SendToSwitch = func(n [33]byte,
|
2017-10-11 05:49:09 +03:00
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
|
2018-01-31 07:26:26 +03:00
|
|
|
if bytes.Equal(ctx.aliases["luoji"].SerializeCompressed(), n[:]) {
|
2017-10-11 05:49:09 +03:00
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
2018-01-31 07:26:26 +03:00
|
|
|
ErrorSource: sourcePub,
|
2017-10-11 05:49:09 +03:00
|
|
|
FailureMessage: &lnwire.FailUnknownNextPeer{},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return preImage, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// The final error returned should also indicate that the peer wasn't
|
|
|
|
// online (the last error we returned).
|
|
|
|
_, _, err = ctx.router.SendPayment(&payment)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("payment didn't return error")
|
|
|
|
}
|
|
|
|
if !strings.Contains(err.Error(), "UnknownNextPeer") {
|
|
|
|
t.Fatalf("expected UnknownNextPeer instead got: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-10-17 05:07:31 +03:00
|
|
|
ctx.router.missionControl.ResetHistory()
|
|
|
|
|
2017-10-11 05:49:09 +03:00
|
|
|
// Finally, we'll modify the SendToSwitch function to indicate that the
|
|
|
|
// roasbeef -> luoji channel has insufficient capacity.
|
2018-01-31 07:26:26 +03:00
|
|
|
ctx.router.cfg.SendToSwitch = func(n [33]byte,
|
2017-10-11 05:49:09 +03:00
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
2018-01-31 07:26:26 +03:00
|
|
|
if bytes.Equal(ctx.aliases["luoji"].SerializeCompressed(), n[:]) {
|
2017-10-11 05:49:09 +03:00
|
|
|
// We'll first simulate an error from the first
|
|
|
|
// outgoing link to simulate the channel from luo ji to
|
|
|
|
// roasbeef not having enough capacity.
|
|
|
|
return [32]byte{}, &htlcswitch.ForwardingError{
|
2018-01-31 07:26:26 +03:00
|
|
|
ErrorSource: sourcePub,
|
2017-10-11 05:49:09 +03:00
|
|
|
FailureMessage: &lnwire.FailTemporaryChannelFailure{},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return preImage, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
paymentPreImage, route, err := ctx.router.SendPayment(&payment)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-10-17 05:07:31 +03:00
|
|
|
// This should succeed finally. The route selected should have two
|
|
|
|
// hops.
|
2017-10-11 05:49:09 +03:00
|
|
|
if len(route.Hops) != 2 {
|
|
|
|
t.Fatalf("incorrect route length: expected %v got %v", 2,
|
|
|
|
len(route.Hops))
|
|
|
|
}
|
|
|
|
|
|
|
|
// The preimage should match up with the once created above.
|
|
|
|
if !bytes.Equal(paymentPreImage[:], preImage[:]) {
|
|
|
|
t.Fatalf("incorrect preimage used: expected %x got %x",
|
|
|
|
preImage[:], paymentPreImage[:])
|
|
|
|
}
|
|
|
|
|
|
|
|
// The route should have satoshi as the first hop.
|
|
|
|
if route.Hops[0].Channel.Node.Alias != "satoshi" {
|
2017-03-21 04:58:21 +03:00
|
|
|
t.Fatalf("route should go through satoshi as first hop, "+
|
|
|
|
"instead passes through: %v",
|
|
|
|
route.Hops[0].Channel.Node.Alias)
|
|
|
|
}
|
|
|
|
}
|
2017-03-27 20:00:38 +03:00
|
|
|
|
|
|
|
// TestAddProof checks that we can update the channel proof after channel
|
|
|
|
// info was added to the database.
|
|
|
|
func TestAddProof(t *testing.T) {
|
2017-06-17 01:59:20 +03:00
|
|
|
t.Parallel()
|
|
|
|
|
2017-03-27 20:00:38 +03:00
|
|
|
ctx, cleanup, err := createTestCtx(0)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
defer cleanup()
|
|
|
|
|
2017-05-15 05:27:14 +03:00
|
|
|
// Before creating out edge, we'll create two new nodes within the
|
|
|
|
// network that the channel will connect.
|
|
|
|
node1, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
node2, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// In order to be able to add the edge we should have a valid funding
|
|
|
|
// UTXO within the blockchain.
|
2017-03-27 20:00:38 +03:00
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(), bitcoinKey2.SerializeCompressed(),
|
|
|
|
100, 0)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
fundingBlock := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{fundingTx},
|
|
|
|
}
|
2017-10-02 18:54:29 +03:00
|
|
|
ctx.chain.addBlock(fundingBlock, chanID.BlockHeight, chanID.BlockHeight)
|
2017-03-27 20:00:38 +03:00
|
|
|
|
|
|
|
// After utxo was recreated adding the edge without the proof.
|
|
|
|
edge := &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID.ToUint64(),
|
|
|
|
NodeKey1Bytes: node1.PubKeyBytes,
|
|
|
|
NodeKey2Bytes: node2.PubKeyBytes,
|
|
|
|
AuthProof: nil,
|
2017-03-27 20:00:38 +03:00
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(edge.BitcoinKey1Bytes[:], bitcoinKey1.SerializeCompressed())
|
|
|
|
copy(edge.BitcoinKey2Bytes[:], bitcoinKey2.SerializeCompressed())
|
2017-03-27 20:00:38 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddEdge(edge); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-05-15 05:27:14 +03:00
|
|
|
// Now we'll attempt to update the proof and check that it has been
|
|
|
|
// properly updated.
|
2017-03-27 20:00:38 +03:00
|
|
|
if err := ctx.router.AddProof(*chanID, &testAuthProof); err != nil {
|
|
|
|
t.Fatalf("unable to add proof: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
info, _, _, err := ctx.router.GetChannelByID(*chanID)
|
|
|
|
if info.AuthProof == nil {
|
|
|
|
t.Fatal("proof have been updated")
|
|
|
|
}
|
|
|
|
}
|
2017-05-15 05:27:14 +03:00
|
|
|
|
2017-07-14 22:32:00 +03:00
|
|
|
// TestIgnoreNodeAnnouncement tests that adding a node to the router that is
|
2017-10-19 08:13:11 +03:00
|
|
|
// not known from any channel announcement, leads to the announcement being
|
2017-07-14 22:32:00 +03:00
|
|
|
// ignored.
|
|
|
|
func TestIgnoreNodeAnnouncement(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight,
|
|
|
|
basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-01-31 07:26:26 +03:00
|
|
|
pub := priv1.PubKey()
|
2017-07-14 22:32:00 +03:00
|
|
|
node := &channeldb.LightningNode{
|
|
|
|
HaveNodeAnnouncement: true,
|
|
|
|
LastUpdate: time.Unix(123, 0),
|
|
|
|
Addresses: testAddrs,
|
|
|
|
Color: color.RGBA{1, 2, 3, 0},
|
|
|
|
Alias: "node11",
|
2018-01-31 07:26:26 +03:00
|
|
|
AuthSigBytes: testSig.Serialize(),
|
2017-07-14 22:32:00 +03:00
|
|
|
Features: testFeatures,
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(node.PubKeyBytes[:], pub.SerializeCompressed())
|
2017-07-14 22:32:00 +03:00
|
|
|
|
|
|
|
err = ctx.router.AddNode(node)
|
|
|
|
if !IsError(err, ErrIgnored) {
|
|
|
|
t.Fatalf("expected to get ErrIgnore, instead got: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-05-15 05:27:14 +03:00
|
|
|
// TestAddEdgeUnknownVertexes tests that if an edge is added that contains two
|
2017-07-14 22:32:00 +03:00
|
|
|
// vertexes which we don't know of, the edge should be available for use
|
|
|
|
// regardless. This is due to the fact that we don't actually need node
|
|
|
|
// announcements for the channel vertexes to be able to use the channel.
|
2017-05-15 05:27:14 +03:00
|
|
|
func TestAddEdgeUnknownVertexes(t *testing.T) {
|
2017-06-17 01:59:20 +03:00
|
|
|
t.Parallel()
|
|
|
|
|
2017-07-14 22:32:00 +03:00
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight,
|
|
|
|
basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
2017-05-15 05:27:14 +03:00
|
|
|
if err != nil {
|
2017-07-14 22:32:00 +03:00
|
|
|
t.Fatalf("unable to create router: %v", err)
|
2017-05-15 05:27:14 +03:00
|
|
|
}
|
|
|
|
|
2018-01-31 07:26:26 +03:00
|
|
|
var pub1 [33]byte
|
|
|
|
copy(pub1[:], priv1.PubKey().SerializeCompressed())
|
|
|
|
|
|
|
|
var pub2 [33]byte
|
|
|
|
copy(pub2[:], priv2.PubKey().SerializeCompressed())
|
|
|
|
|
2017-07-14 22:32:00 +03:00
|
|
|
// The two nodes we are about to add should not exist yet.
|
2018-01-31 07:26:26 +03:00
|
|
|
_, exists1, err := ctx.graph.HasLightningNode(pub1)
|
2017-07-14 22:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query graph: %v", err)
|
|
|
|
}
|
|
|
|
if exists1 {
|
|
|
|
t.Fatalf("node already existed")
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
_, exists2, err := ctx.graph.HasLightningNode(pub2)
|
2017-07-14 22:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query graph: %v", err)
|
|
|
|
}
|
|
|
|
if exists2 {
|
|
|
|
t.Fatalf("node already existed")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add the edge between the two unknown nodes to the graph, and check
|
|
|
|
// that the nodes are found after the fact.
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
2017-08-22 09:43:20 +03:00
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
10000, 500)
|
2017-05-15 05:27:14 +03:00
|
|
|
if err != nil {
|
2017-05-16 04:16:28 +03:00
|
|
|
t.Fatalf("unable to create channel edge: %v", err)
|
2017-05-15 05:27:14 +03:00
|
|
|
}
|
2017-07-14 22:32:00 +03:00
|
|
|
fundingBlock := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{fundingTx},
|
|
|
|
}
|
2017-10-02 18:54:29 +03:00
|
|
|
ctx.chain.addBlock(fundingBlock, chanID.BlockHeight, chanID.BlockHeight)
|
2017-05-15 05:27:14 +03:00
|
|
|
|
|
|
|
edge := &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID.ToUint64(),
|
|
|
|
NodeKey1Bytes: pub1,
|
|
|
|
NodeKey2Bytes: pub2,
|
|
|
|
BitcoinKey1Bytes: pub1,
|
|
|
|
BitcoinKey2Bytes: pub2,
|
|
|
|
AuthProof: nil,
|
2017-05-15 05:27:14 +03:00
|
|
|
}
|
2017-07-14 22:32:00 +03:00
|
|
|
if err := ctx.router.AddEdge(edge); err != nil {
|
|
|
|
t.Fatalf("expected to be able to add edge to the channel graph,"+
|
|
|
|
" even though the vertexes were unknown: %v.", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We must add the edge policy to be able to use the edge for route
|
|
|
|
// finding.
|
|
|
|
edgePolicy := &channeldb.ChannelEdgePolicy{
|
2018-01-31 07:26:26 +03:00
|
|
|
SigBytes: testSig.Serialize(),
|
2017-07-14 22:32:00 +03:00
|
|
|
ChannelID: edge.ChannelID,
|
|
|
|
LastUpdate: time.Now(),
|
|
|
|
TimeLockDelta: 10,
|
2017-08-22 09:43:20 +03:00
|
|
|
MinHTLC: 1,
|
|
|
|
FeeBaseMSat: 10,
|
|
|
|
FeeProportionalMillionths: 10000,
|
2017-07-14 22:32:00 +03:00
|
|
|
}
|
|
|
|
edgePolicy.Flags = 0
|
|
|
|
|
|
|
|
if err := ctx.router.UpdateEdge(edgePolicy); err != nil {
|
|
|
|
t.Fatalf("unable to update edge policy: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create edge in the other direction as well.
|
|
|
|
edgePolicy = &channeldb.ChannelEdgePolicy{
|
2018-01-31 07:26:26 +03:00
|
|
|
SigBytes: testSig.Serialize(),
|
2017-07-14 22:32:00 +03:00
|
|
|
ChannelID: edge.ChannelID,
|
|
|
|
LastUpdate: time.Now(),
|
|
|
|
TimeLockDelta: 10,
|
2017-08-22 09:43:20 +03:00
|
|
|
MinHTLC: 1,
|
|
|
|
FeeBaseMSat: 10,
|
|
|
|
FeeProportionalMillionths: 10000,
|
2017-07-14 22:32:00 +03:00
|
|
|
}
|
|
|
|
edgePolicy.Flags = 1
|
|
|
|
|
|
|
|
if err := ctx.router.UpdateEdge(edgePolicy); err != nil {
|
|
|
|
t.Fatalf("unable to update edge policy: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// After adding the edge between the two previously unknown nodes, they
|
|
|
|
// should have been added to the graph.
|
2018-01-31 07:26:26 +03:00
|
|
|
_, exists1, err = ctx.graph.HasLightningNode(pub1)
|
2017-07-14 22:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query graph: %v", err)
|
|
|
|
}
|
|
|
|
if !exists1 {
|
|
|
|
t.Fatalf("node1 was not added to the graph")
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
_, exists2, err = ctx.graph.HasLightningNode(pub2)
|
2017-07-14 22:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query graph: %v", err)
|
|
|
|
}
|
|
|
|
if !exists2 {
|
|
|
|
t.Fatalf("node2 was not added to the graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
// We will connect node1 to the rest of the test graph, and make sure
|
|
|
|
// we can find a route to node2, which will use the just added channel
|
|
|
|
// edge.
|
|
|
|
|
|
|
|
// We will connect node 1 to "sophon"
|
|
|
|
connectNode := ctx.aliases["sophon"]
|
|
|
|
if connectNode == nil {
|
|
|
|
t.Fatalf("could not find node to connect to")
|
|
|
|
}
|
|
|
|
|
|
|
|
var (
|
|
|
|
pubKey1 *btcec.PublicKey
|
|
|
|
pubKey2 *btcec.PublicKey
|
|
|
|
)
|
|
|
|
node1Bytes := priv1.PubKey().SerializeCompressed()
|
|
|
|
node2Bytes := connectNode.SerializeCompressed()
|
|
|
|
if bytes.Compare(node1Bytes, node2Bytes) == -1 {
|
|
|
|
pubKey1 = priv1.PubKey()
|
|
|
|
pubKey2 = connectNode
|
|
|
|
} else {
|
|
|
|
pubKey1 = connectNode
|
|
|
|
pubKey2 = priv1.PubKey()
|
|
|
|
}
|
|
|
|
|
|
|
|
fundingTx, _, chanID, err = createChannelEdge(ctx,
|
|
|
|
pubKey1.SerializeCompressed(), pubKey2.SerializeCompressed(),
|
|
|
|
10000, 510)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
fundingBlock = &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{fundingTx},
|
|
|
|
}
|
2017-10-02 18:54:29 +03:00
|
|
|
ctx.chain.addBlock(fundingBlock, chanID.BlockHeight, chanID.BlockHeight)
|
2017-07-14 22:32:00 +03:00
|
|
|
|
|
|
|
edge = &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID.ToUint64(),
|
|
|
|
AuthProof: nil,
|
2017-07-14 22:32:00 +03:00
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(edge.NodeKey1Bytes[:], node1Bytes)
|
|
|
|
copy(edge.NodeKey2Bytes[:], node2Bytes)
|
|
|
|
copy(edge.BitcoinKey1Bytes[:], node1Bytes)
|
|
|
|
copy(edge.BitcoinKey2Bytes[:], node2Bytes)
|
2017-07-14 22:32:00 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddEdge(edge); err != nil {
|
|
|
|
t.Fatalf("unable to add edge to the channel graph: %v.", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
edgePolicy = &channeldb.ChannelEdgePolicy{
|
2018-01-31 07:26:26 +03:00
|
|
|
SigBytes: testSig.Serialize(),
|
2017-07-14 22:32:00 +03:00
|
|
|
ChannelID: edge.ChannelID,
|
|
|
|
LastUpdate: time.Now(),
|
|
|
|
TimeLockDelta: 10,
|
2017-08-22 09:43:20 +03:00
|
|
|
MinHTLC: 1,
|
|
|
|
FeeBaseMSat: 10,
|
|
|
|
FeeProportionalMillionths: 10000,
|
2017-07-14 22:32:00 +03:00
|
|
|
}
|
|
|
|
edgePolicy.Flags = 0
|
|
|
|
|
|
|
|
if err := ctx.router.UpdateEdge(edgePolicy); err != nil {
|
|
|
|
t.Fatalf("unable to update edge policy: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
edgePolicy = &channeldb.ChannelEdgePolicy{
|
2018-01-31 07:26:26 +03:00
|
|
|
SigBytes: testSig.Serialize(),
|
2017-07-14 22:32:00 +03:00
|
|
|
ChannelID: edge.ChannelID,
|
|
|
|
LastUpdate: time.Now(),
|
|
|
|
TimeLockDelta: 10,
|
2017-08-22 09:43:20 +03:00
|
|
|
MinHTLC: 1,
|
|
|
|
FeeBaseMSat: 10,
|
|
|
|
FeeProportionalMillionths: 10000,
|
2017-07-14 22:32:00 +03:00
|
|
|
}
|
|
|
|
edgePolicy.Flags = 1
|
|
|
|
|
|
|
|
if err := ctx.router.UpdateEdge(edgePolicy); err != nil {
|
|
|
|
t.Fatalf("unable to update edge policy: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-03-23 03:39:47 +03:00
|
|
|
// We should now be able to find two routes to node 2.
|
2017-08-22 09:43:20 +03:00
|
|
|
paymentAmt := lnwire.NewMSatFromSatoshis(100)
|
2017-07-14 22:32:00 +03:00
|
|
|
targetNode := priv2.PubKey()
|
2017-10-19 08:13:11 +03:00
|
|
|
routes, err := ctx.router.FindRoutes(targetNode, paymentAmt,
|
2018-02-28 06:09:41 +03:00
|
|
|
defaultNumRoutes, DefaultFinalCLTVDelta)
|
2017-07-14 22:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to find any routes: %v", err)
|
|
|
|
}
|
2018-03-23 03:39:47 +03:00
|
|
|
if len(routes) != 2 {
|
|
|
|
t.Fatalf("expected to find 2 route, found: %v", len(routes))
|
2017-07-14 22:32:00 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Now check that we can update the node info for the partial node
|
|
|
|
// without messing up the channel graph.
|
|
|
|
n1 := &channeldb.LightningNode{
|
|
|
|
HaveNodeAnnouncement: true,
|
|
|
|
LastUpdate: time.Unix(123, 0),
|
|
|
|
Addresses: testAddrs,
|
|
|
|
Color: color.RGBA{1, 2, 3, 0},
|
|
|
|
Alias: "node11",
|
2018-01-31 07:26:26 +03:00
|
|
|
AuthSigBytes: testSig.Serialize(),
|
2017-07-14 22:32:00 +03:00
|
|
|
Features: testFeatures,
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(n1.PubKeyBytes[:], priv1.PubKey().SerializeCompressed())
|
2017-07-14 22:32:00 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddNode(n1); err != nil {
|
|
|
|
t.Fatalf("could not add node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
n2 := &channeldb.LightningNode{
|
|
|
|
HaveNodeAnnouncement: true,
|
|
|
|
LastUpdate: time.Unix(123, 0),
|
|
|
|
Addresses: testAddrs,
|
|
|
|
Color: color.RGBA{1, 2, 3, 0},
|
|
|
|
Alias: "node22",
|
2018-01-31 07:26:26 +03:00
|
|
|
AuthSigBytes: testSig.Serialize(),
|
2017-07-14 22:32:00 +03:00
|
|
|
Features: testFeatures,
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(n2.PubKeyBytes[:], priv2.PubKey().SerializeCompressed())
|
2017-07-14 22:32:00 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddNode(n2); err != nil {
|
|
|
|
t.Fatalf("could not add node: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-03-23 03:39:47 +03:00
|
|
|
// Should still be able to find the routes, and the info should be
|
2017-07-14 22:32:00 +03:00
|
|
|
// updated.
|
2017-10-19 08:13:11 +03:00
|
|
|
routes, err = ctx.router.FindRoutes(targetNode, paymentAmt,
|
2018-02-28 06:09:41 +03:00
|
|
|
defaultNumRoutes, DefaultFinalCLTVDelta)
|
2017-07-14 22:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to find any routes: %v", err)
|
|
|
|
}
|
2018-03-23 03:39:47 +03:00
|
|
|
if len(routes) != 2 {
|
|
|
|
t.Fatalf("expected to find 2 route, found: %v", len(routes))
|
2017-07-14 22:32:00 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
copy1, err := ctx.graph.FetchLightningNode(priv1.PubKey())
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if copy1.Alias != n1.Alias {
|
|
|
|
t.Fatalf("fetched node not equal to original")
|
|
|
|
}
|
|
|
|
|
|
|
|
copy2, err := ctx.graph.FetchLightningNode(priv2.PubKey())
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if copy2.Alias != n2.Alias {
|
|
|
|
t.Fatalf("fetched node not equal to original")
|
2017-05-15 05:27:14 +03:00
|
|
|
}
|
|
|
|
}
|
2017-10-02 18:54:29 +03:00
|
|
|
|
|
|
|
// TestWakeUpOnStaleBranch tests that upon startup of the ChannelRouter, if the
|
|
|
|
// the chain previously reflected in the channel graph is stale (overtaken by a
|
|
|
|
// longer chain), the channel router will prune the graph for any channels
|
|
|
|
// confirmed on the stale chain, and resync to the main chain.
|
|
|
|
func TestWakeUpOnStaleBranch(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
const chanValue = 10000
|
|
|
|
|
|
|
|
// chanID1 will not be reorged out.
|
|
|
|
var chanID1 uint64
|
|
|
|
|
|
|
|
// chanID2 will be reorged out.
|
|
|
|
var chanID2 uint64
|
|
|
|
|
|
|
|
// Create 10 common blocks, confirming chanID1.
|
|
|
|
for i := uint32(1); i <= 10; i++ {
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
height := startingBlockHeight + i
|
|
|
|
if i == 5 {
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
chanValue, height)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
block.Transactions = append(block.Transactions,
|
|
|
|
fundingTx)
|
|
|
|
chanID1 = chanID.ToUint64()
|
|
|
|
|
|
|
|
}
|
|
|
|
ctx.chain.addBlock(block, height, rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(height))
|
|
|
|
ctx.chainView.notifyBlock(block.BlockHash(), height,
|
|
|
|
[]*wire.MsgTx{})
|
|
|
|
}
|
|
|
|
|
|
|
|
// Give time to process new blocks
|
|
|
|
time.Sleep(time.Millisecond * 500)
|
|
|
|
|
|
|
|
_, forkHeight, err := ctx.chain.GetBestBlock()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to ge best block: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create 10 blocks on the minority chain, confirming chanID2.
|
|
|
|
for i := uint32(1); i <= 10; i++ {
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
height := uint32(forkHeight) + i
|
|
|
|
if i == 5 {
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
chanValue, height)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
block.Transactions = append(block.Transactions,
|
|
|
|
fundingTx)
|
|
|
|
chanID2 = chanID.ToUint64()
|
|
|
|
}
|
|
|
|
ctx.chain.addBlock(block, height, rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(height))
|
|
|
|
ctx.chainView.notifyBlock(block.BlockHash(), height,
|
|
|
|
[]*wire.MsgTx{})
|
|
|
|
}
|
|
|
|
// Give time to process new blocks
|
|
|
|
time.Sleep(time.Millisecond * 500)
|
|
|
|
|
|
|
|
// Now add the two edges to the channel graph, and check that they
|
|
|
|
// correctly show up in the database.
|
|
|
|
node1, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
|
|
}
|
|
|
|
node2, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
edge1 := &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID1,
|
|
|
|
NodeKey1Bytes: node1.PubKeyBytes,
|
|
|
|
NodeKey2Bytes: node2.PubKeyBytes,
|
2017-10-02 18:54:29 +03:00
|
|
|
AuthProof: &channeldb.ChannelAuthProof{
|
2018-01-31 07:26:26 +03:00
|
|
|
NodeSig1Bytes: testSig.Serialize(),
|
|
|
|
NodeSig2Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig1Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig2Bytes: testSig.Serialize(),
|
2017-10-02 18:54:29 +03:00
|
|
|
},
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(edge1.BitcoinKey1Bytes[:], bitcoinKey1.SerializeCompressed())
|
|
|
|
copy(edge1.BitcoinKey2Bytes[:], bitcoinKey2.SerializeCompressed())
|
2017-10-02 18:54:29 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddEdge(edge1); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
edge2 := &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID2,
|
|
|
|
NodeKey1Bytes: node1.PubKeyBytes,
|
|
|
|
NodeKey2Bytes: node2.PubKeyBytes,
|
2017-10-02 18:54:29 +03:00
|
|
|
AuthProof: &channeldb.ChannelAuthProof{
|
2018-01-31 07:26:26 +03:00
|
|
|
NodeSig1Bytes: testSig.Serialize(),
|
|
|
|
NodeSig2Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig1Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig2Bytes: testSig.Serialize(),
|
2017-10-02 18:54:29 +03:00
|
|
|
},
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(edge2.BitcoinKey1Bytes[:], bitcoinKey1.SerializeCompressed())
|
|
|
|
copy(edge2.BitcoinKey2Bytes[:], bitcoinKey2.SerializeCompressed())
|
2017-10-02 18:54:29 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddEdge(edge2); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check that the fundingTxs are in the graph db.
|
|
|
|
_, _, has, err := ctx.graph.HasChannelEdge(chanID1)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID1)
|
|
|
|
}
|
|
|
|
if !has {
|
|
|
|
t.Fatalf("could not find edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
_, _, has, err = ctx.graph.HasChannelEdge(chanID2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID2)
|
|
|
|
}
|
|
|
|
if !has {
|
|
|
|
t.Fatalf("could not find edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Stop the router, so we can reorg the chain while its offline.
|
|
|
|
if err := ctx.router.Stop(); err != nil {
|
|
|
|
t.Fatalf("unable to stop router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a 15 block fork.
|
|
|
|
for i := uint32(1); i <= 15; i++ {
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
height := uint32(forkHeight) + i
|
|
|
|
ctx.chain.addBlock(block, height, rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(height))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Give time to process new blocks.
|
|
|
|
time.Sleep(time.Millisecond * 500)
|
|
|
|
|
|
|
|
// Create new router with same graph database.
|
|
|
|
router, err := New(Config{
|
|
|
|
Graph: ctx.graph,
|
|
|
|
Chain: ctx.chain,
|
|
|
|
ChainView: ctx.chainView,
|
2018-01-31 07:26:26 +03:00
|
|
|
SendToSwitch: func(_ [33]byte,
|
2017-10-02 18:54:29 +03:00
|
|
|
_ *lnwire.UpdateAddHTLC, _ *sphinx.Circuit) ([32]byte, error) {
|
|
|
|
return [32]byte{}, nil
|
|
|
|
},
|
|
|
|
ChannelPruneExpiry: time.Hour * 24,
|
|
|
|
GraphPruneInterval: time.Hour * 2,
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// It should resync to the longer chain on startup.
|
|
|
|
if err := router.Start(); err != nil {
|
|
|
|
t.Fatalf("unable to start router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// The channel with chanID2 should not be in the database anymore,
|
|
|
|
// since it is not confirmed on the longest chain. chanID1 should
|
|
|
|
// still be.
|
|
|
|
_, _, has, err = ctx.graph.HasChannelEdge(chanID1)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID1)
|
|
|
|
}
|
|
|
|
if !has {
|
|
|
|
t.Fatalf("did not find edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
_, _, has, err = ctx.graph.HasChannelEdge(chanID2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID2)
|
|
|
|
}
|
|
|
|
if has {
|
|
|
|
t.Fatalf("found edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2017-11-07 03:00:51 +03:00
|
|
|
// TestDisconnectedBlocks checks that the router handles a reorg happening when
|
|
|
|
// it is active.
|
2017-10-02 18:54:29 +03:00
|
|
|
func TestDisconnectedBlocks(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
const chanValue = 10000
|
|
|
|
|
2017-11-07 03:00:51 +03:00
|
|
|
// chanID1 will not be reorged out, while chanID2 will be reorged out.
|
|
|
|
var chanID1, chanID2 uint64
|
2017-10-02 18:54:29 +03:00
|
|
|
|
|
|
|
// Create 10 common blocks, confirming chanID1.
|
|
|
|
for i := uint32(1); i <= 10; i++ {
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
height := startingBlockHeight + i
|
|
|
|
if i == 5 {
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
chanValue, height)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
block.Transactions = append(block.Transactions,
|
|
|
|
fundingTx)
|
|
|
|
chanID1 = chanID.ToUint64()
|
|
|
|
|
|
|
|
}
|
|
|
|
ctx.chain.addBlock(block, height, rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(height))
|
|
|
|
ctx.chainView.notifyBlock(block.BlockHash(), height,
|
|
|
|
[]*wire.MsgTx{})
|
|
|
|
}
|
|
|
|
|
|
|
|
// Give time to process new blocks
|
|
|
|
time.Sleep(time.Millisecond * 500)
|
|
|
|
|
|
|
|
_, forkHeight, err := ctx.chain.GetBestBlock()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get best block: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create 10 blocks on the minority chain, confirming chanID2.
|
|
|
|
var minorityChain []*wire.MsgBlock
|
|
|
|
for i := uint32(1); i <= 10; i++ {
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
height := uint32(forkHeight) + i
|
|
|
|
if i == 5 {
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
chanValue, height)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
block.Transactions = append(block.Transactions,
|
|
|
|
fundingTx)
|
|
|
|
chanID2 = chanID.ToUint64()
|
|
|
|
}
|
|
|
|
minorityChain = append(minorityChain, block)
|
|
|
|
ctx.chain.addBlock(block, height, rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(height))
|
|
|
|
ctx.chainView.notifyBlock(block.BlockHash(), height,
|
|
|
|
[]*wire.MsgTx{})
|
|
|
|
}
|
|
|
|
// Give time to process new blocks
|
|
|
|
time.Sleep(time.Millisecond * 500)
|
|
|
|
|
|
|
|
// Now add the two edges to the channel graph, and check that they
|
|
|
|
// correctly show up in the database.
|
|
|
|
node1, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
|
|
}
|
|
|
|
node2, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
edge1 := &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID1,
|
|
|
|
NodeKey1Bytes: node1.PubKeyBytes,
|
|
|
|
NodeKey2Bytes: node2.PubKeyBytes,
|
|
|
|
BitcoinKey1Bytes: node1.PubKeyBytes,
|
|
|
|
BitcoinKey2Bytes: node2.PubKeyBytes,
|
2017-10-02 18:54:29 +03:00
|
|
|
AuthProof: &channeldb.ChannelAuthProof{
|
2018-01-31 07:26:26 +03:00
|
|
|
NodeSig1Bytes: testSig.Serialize(),
|
|
|
|
NodeSig2Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig1Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig2Bytes: testSig.Serialize(),
|
2017-10-02 18:54:29 +03:00
|
|
|
},
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(edge1.BitcoinKey1Bytes[:], bitcoinKey1.SerializeCompressed())
|
|
|
|
copy(edge1.BitcoinKey2Bytes[:], bitcoinKey2.SerializeCompressed())
|
2017-10-02 18:54:29 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddEdge(edge1); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
edge2 := &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID2,
|
|
|
|
NodeKey1Bytes: node1.PubKeyBytes,
|
|
|
|
NodeKey2Bytes: node2.PubKeyBytes,
|
|
|
|
BitcoinKey1Bytes: node1.PubKeyBytes,
|
|
|
|
BitcoinKey2Bytes: node2.PubKeyBytes,
|
2017-10-02 18:54:29 +03:00
|
|
|
AuthProof: &channeldb.ChannelAuthProof{
|
2018-01-31 07:26:26 +03:00
|
|
|
NodeSig1Bytes: testSig.Serialize(),
|
|
|
|
NodeSig2Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig1Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig2Bytes: testSig.Serialize(),
|
2017-10-02 18:54:29 +03:00
|
|
|
},
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(edge2.BitcoinKey1Bytes[:], bitcoinKey1.SerializeCompressed())
|
|
|
|
copy(edge2.BitcoinKey2Bytes[:], bitcoinKey2.SerializeCompressed())
|
2017-10-02 18:54:29 +03:00
|
|
|
|
|
|
|
if err := ctx.router.AddEdge(edge2); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check that the fundingTxs are in the graph db.
|
|
|
|
_, _, has, err := ctx.graph.HasChannelEdge(chanID1)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID1)
|
|
|
|
}
|
|
|
|
if !has {
|
|
|
|
t.Fatalf("could not find edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
_, _, has, err = ctx.graph.HasChannelEdge(chanID2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID2)
|
|
|
|
}
|
|
|
|
if !has {
|
|
|
|
t.Fatalf("could not find edge in graph")
|
|
|
|
}
|
|
|
|
|
2017-11-07 03:00:51 +03:00
|
|
|
// Create a 15 block fork. We first let the chainView notify the router
|
|
|
|
// about stale blocks, before sending the now connected blocks. We do
|
|
|
|
// this because we expect this order from the chainview.
|
2017-10-02 18:54:29 +03:00
|
|
|
for i := len(minorityChain) - 1; i >= 0; i-- {
|
|
|
|
block := minorityChain[i]
|
|
|
|
height := uint32(forkHeight) + uint32(i) + 1
|
|
|
|
ctx.chainView.notifyStaleBlock(block.BlockHash(), height,
|
|
|
|
block.Transactions)
|
|
|
|
}
|
|
|
|
for i := uint32(1); i <= 15; i++ {
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
height := uint32(forkHeight) + i
|
|
|
|
ctx.chain.addBlock(block, height, rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(height))
|
|
|
|
ctx.chainView.notifyBlock(block.BlockHash(), height,
|
|
|
|
block.Transactions)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Give time to process new blocks
|
|
|
|
time.Sleep(time.Millisecond * 500)
|
|
|
|
|
2017-11-07 03:00:51 +03:00
|
|
|
// chanID2 should not be in the database anymore, since it is not
|
|
|
|
// confirmed on the longest chain. chanID1 should still be.
|
2017-10-02 18:54:29 +03:00
|
|
|
_, _, has, err = ctx.graph.HasChannelEdge(chanID1)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID1)
|
|
|
|
}
|
|
|
|
if !has {
|
|
|
|
t.Fatalf("did not find edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
_, _, has, err = ctx.graph.HasChannelEdge(chanID2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID2)
|
|
|
|
}
|
|
|
|
if has {
|
|
|
|
t.Fatalf("found edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
2017-11-07 03:06:11 +03:00
|
|
|
|
|
|
|
// TestChansClosedOfflinePruneGraph tests that if channels we know of are
|
|
|
|
// closed while we're offline, then once we resume operation of the
|
|
|
|
// ChannelRouter, then the channels are properly pruned.
|
|
|
|
func TestRouterChansClosedOfflinePruneGraph(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
const chanValue = 10000
|
|
|
|
|
|
|
|
// First, we'll create a channel, to be mined shortly at height 102.
|
|
|
|
block102 := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
nextHeight := startingBlockHeight + 1
|
|
|
|
fundingTx1, chanUTXO, chanID1, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
chanValue, uint32(nextHeight))
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
block102.Transactions = append(block102.Transactions, fundingTx1)
|
|
|
|
ctx.chain.addBlock(block102, uint32(nextHeight), rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(nextHeight))
|
|
|
|
ctx.chainView.notifyBlock(block102.BlockHash(), uint32(nextHeight),
|
|
|
|
[]*wire.MsgTx{})
|
|
|
|
|
|
|
|
// We'll now create the edges and nodes within the database required
|
|
|
|
// for the ChannelRouter to properly recognize the channel we added
|
|
|
|
// above.
|
|
|
|
node1, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
|
|
}
|
|
|
|
node2, err := createTestNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
|
|
}
|
|
|
|
edge1 := &channeldb.ChannelEdgeInfo{
|
2018-01-31 07:26:26 +03:00
|
|
|
ChannelID: chanID1.ToUint64(),
|
|
|
|
NodeKey1Bytes: node1.PubKeyBytes,
|
|
|
|
NodeKey2Bytes: node2.PubKeyBytes,
|
2017-11-07 03:06:11 +03:00
|
|
|
AuthProof: &channeldb.ChannelAuthProof{
|
2018-01-31 07:26:26 +03:00
|
|
|
NodeSig1Bytes: testSig.Serialize(),
|
|
|
|
NodeSig2Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig1Bytes: testSig.Serialize(),
|
|
|
|
BitcoinSig2Bytes: testSig.Serialize(),
|
2017-11-07 03:06:11 +03:00
|
|
|
},
|
|
|
|
}
|
2018-01-31 07:26:26 +03:00
|
|
|
copy(edge1.BitcoinKey1Bytes[:], bitcoinKey1.SerializeCompressed())
|
|
|
|
copy(edge1.BitcoinKey2Bytes[:], bitcoinKey2.SerializeCompressed())
|
2017-11-07 03:06:11 +03:00
|
|
|
if err := ctx.router.AddEdge(edge1); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// The router should now be aware of the channel we created above.
|
|
|
|
_, _, hasChan, err := ctx.graph.HasChannelEdge(chanID1.ToUint64())
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID1)
|
|
|
|
}
|
|
|
|
if !hasChan {
|
|
|
|
t.Fatalf("could not find edge in graph")
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the transaction included, and the router's database state
|
|
|
|
// updated, we'll now mine 5 additional blocks on top of it.
|
|
|
|
for i := 0; i < 5; i++ {
|
|
|
|
nextHeight++
|
|
|
|
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
ctx.chain.addBlock(block, uint32(nextHeight), rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(nextHeight))
|
|
|
|
ctx.chainView.notifyBlock(block.BlockHash(), uint32(nextHeight),
|
|
|
|
[]*wire.MsgTx{})
|
|
|
|
}
|
|
|
|
|
|
|
|
// At this point, our starting height should be 107.
|
|
|
|
_, chainHeight, err := ctx.chain.GetBestBlock()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get best block: %v", err)
|
|
|
|
}
|
|
|
|
if chainHeight != 107 {
|
|
|
|
t.Fatalf("incorrect chain height: expected %v, got %v",
|
2017-11-07 03:49:27 +03:00
|
|
|
107, chainHeight)
|
2017-11-07 03:06:11 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Next, we'll "shut down" the router in order to simulate downtime.
|
|
|
|
if err := ctx.router.Stop(); err != nil {
|
|
|
|
t.Fatalf("unable to shutdown router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// While the router is "offline" we'll mine 5 additional blocks, with
|
|
|
|
// the second block closing the channel we created above.
|
|
|
|
for i := 0; i < 5; i++ {
|
|
|
|
nextHeight++
|
|
|
|
|
|
|
|
block := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{},
|
|
|
|
}
|
|
|
|
|
|
|
|
if i == 2 {
|
|
|
|
// For the second block, we'll add a transaction that
|
|
|
|
// closes the channel we created above by spending the
|
|
|
|
// output.
|
|
|
|
closingTx := wire.NewMsgTx(2)
|
|
|
|
closingTx.AddTxIn(&wire.TxIn{
|
|
|
|
PreviousOutPoint: *chanUTXO,
|
|
|
|
})
|
|
|
|
block.Transactions = append(block.Transactions,
|
|
|
|
closingTx)
|
|
|
|
}
|
|
|
|
|
|
|
|
ctx.chain.addBlock(block, uint32(nextHeight), rand.Uint32())
|
|
|
|
ctx.chain.setBestBlock(int32(nextHeight))
|
|
|
|
ctx.chainView.notifyBlock(block.BlockHash(), uint32(nextHeight),
|
|
|
|
[]*wire.MsgTx{})
|
|
|
|
}
|
|
|
|
|
|
|
|
// At this point, our starting height should be 112.
|
|
|
|
_, chainHeight, err = ctx.chain.GetBestBlock()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get best block: %v", err)
|
|
|
|
}
|
|
|
|
if chainHeight != 112 {
|
|
|
|
t.Fatalf("incorrect chain height: expected %v, got %v",
|
2017-11-07 03:49:27 +03:00
|
|
|
112, chainHeight)
|
2017-11-07 03:06:11 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Now we'll re-start the ChannelRouter. It should recognize that it's
|
|
|
|
// behind the main chain and prune all the blocks that it missed while
|
|
|
|
// it was down.
|
|
|
|
ctx.RestartRouter()
|
|
|
|
|
|
|
|
// At this point, the channel that was pruned should no longer be known
|
|
|
|
// by the router.
|
|
|
|
_, _, hasChan, err = ctx.graph.HasChannelEdge(chanID1.ToUint64())
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("error looking for edge: %v", chanID1)
|
|
|
|
}
|
|
|
|
if hasChan {
|
|
|
|
t.Fatalf("channel was found in graph but shouldn't have been")
|
|
|
|
}
|
|
|
|
}
|
2018-02-13 03:27:30 +03:00
|
|
|
|
|
|
|
// TestFindPathFeeWeighting tests that the findPath method will properly prefer
|
|
|
|
// routes with lower fees over routes with lower time lock values. This is
|
|
|
|
// meant to exercise the fact that the internal findPath method ranks edges
|
|
|
|
// with the square of the total fee in order bias towards lower fees.
|
|
|
|
func TestFindPathFeeWeighting(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight, basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
var preImage [32]byte
|
|
|
|
copy(preImage[:], bytes.Repeat([]byte{9}, 32))
|
|
|
|
|
|
|
|
sourceNode, err := ctx.graph.SourceNode()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch source node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
ignoreVertex := make(map[Vertex]struct{})
|
|
|
|
ignoreEdge := make(map[uint64]struct{})
|
|
|
|
|
|
|
|
amt := lnwire.MilliSatoshi(100)
|
|
|
|
|
|
|
|
target := ctx.aliases["luoji"]
|
|
|
|
if target == nil {
|
|
|
|
t.Fatalf("unable to find target node")
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll now attempt a path finding attempt using this set up. Due to
|
|
|
|
// the edge weighting, we should select the direct path over the 2 hop
|
|
|
|
// path even though the direct path has a higher potential time lock.
|
|
|
|
path, err := findPath(
|
|
|
|
nil, ctx.graph, sourceNode, target, ignoreVertex, ignoreEdge,
|
|
|
|
amt,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to find path: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// The route that was chosen should be exactly one hop, and should be
|
|
|
|
// directly to luoji.
|
|
|
|
if len(path) != 1 {
|
|
|
|
t.Fatalf("expected path length of 1, instead was: %v", len(path))
|
|
|
|
}
|
|
|
|
if path[0].Node.Alias != "luoji" {
|
|
|
|
t.Fatalf("wrong node: %v", path[0].Node.Alias)
|
|
|
|
}
|
|
|
|
}
|
2018-02-25 06:34:03 +03:00
|
|
|
|
|
|
|
// TestIsStaleNode tests that the IsStaleNode method properly detects stale
|
|
|
|
// node announcements.
|
|
|
|
func TestIsStaleNode(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Before we can insert a node in to the database, we need to create a
|
|
|
|
// channel that it's linked to.
|
|
|
|
var (
|
|
|
|
pub1 [33]byte
|
|
|
|
pub2 [33]byte
|
|
|
|
)
|
|
|
|
copy(pub1[:], priv1.PubKey().SerializeCompressed())
|
|
|
|
copy(pub2[:], priv2.PubKey().SerializeCompressed())
|
|
|
|
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
10000, 500)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
fundingBlock := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{fundingTx},
|
|
|
|
}
|
|
|
|
ctx.chain.addBlock(fundingBlock, chanID.BlockHeight, chanID.BlockHeight)
|
|
|
|
|
|
|
|
edge := &channeldb.ChannelEdgeInfo{
|
|
|
|
ChannelID: chanID.ToUint64(),
|
|
|
|
NodeKey1Bytes: pub1,
|
|
|
|
NodeKey2Bytes: pub2,
|
|
|
|
BitcoinKey1Bytes: pub1,
|
|
|
|
BitcoinKey2Bytes: pub2,
|
|
|
|
AuthProof: nil,
|
|
|
|
}
|
|
|
|
if err := ctx.router.AddEdge(edge); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Before we add the node, if we query for staleness, we should get
|
|
|
|
// false, as we haven't added the full node.
|
|
|
|
updateTimeStamp := time.Unix(123, 0)
|
|
|
|
if ctx.router.IsStaleNode(pub1, updateTimeStamp) {
|
|
|
|
t.Fatalf("incorrectly detected node as stale")
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the node stub in the database, we'll add the fully node
|
|
|
|
// announcement to the database.
|
|
|
|
n1 := &channeldb.LightningNode{
|
|
|
|
HaveNodeAnnouncement: true,
|
|
|
|
LastUpdate: updateTimeStamp,
|
|
|
|
Addresses: testAddrs,
|
|
|
|
Color: color.RGBA{1, 2, 3, 0},
|
|
|
|
Alias: "node11",
|
|
|
|
AuthSigBytes: testSig.Serialize(),
|
|
|
|
Features: testFeatures,
|
|
|
|
}
|
|
|
|
copy(n1.PubKeyBytes[:], priv1.PubKey().SerializeCompressed())
|
|
|
|
if err := ctx.router.AddNode(n1); err != nil {
|
|
|
|
t.Fatalf("could not add node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we use the same timestamp and query for staleness, we should get
|
|
|
|
// true.
|
|
|
|
if !ctx.router.IsStaleNode(pub1, updateTimeStamp) {
|
|
|
|
t.Fatalf("failure to detect stale node update")
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we update the timestamp and once again query for staleness, it
|
|
|
|
// should report false.
|
|
|
|
newTimeStamp := time.Unix(1234, 0)
|
|
|
|
if ctx.router.IsStaleNode(pub1, newTimeStamp) {
|
|
|
|
t.Fatalf("incorrectly detected node as stale")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestIsKnownEdge tests that the IsKnownEdge method properly detects stale
|
|
|
|
// channel announcements.
|
|
|
|
func TestIsKnownEdge(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// First, we'll create a new channel edge (just the info) and insert it
|
|
|
|
// into the database.
|
|
|
|
var (
|
|
|
|
pub1 [33]byte
|
|
|
|
pub2 [33]byte
|
|
|
|
)
|
|
|
|
copy(pub1[:], priv1.PubKey().SerializeCompressed())
|
|
|
|
copy(pub2[:], priv2.PubKey().SerializeCompressed())
|
|
|
|
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
10000, 500)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
fundingBlock := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{fundingTx},
|
|
|
|
}
|
|
|
|
ctx.chain.addBlock(fundingBlock, chanID.BlockHeight, chanID.BlockHeight)
|
|
|
|
|
|
|
|
edge := &channeldb.ChannelEdgeInfo{
|
|
|
|
ChannelID: chanID.ToUint64(),
|
|
|
|
NodeKey1Bytes: pub1,
|
|
|
|
NodeKey2Bytes: pub2,
|
|
|
|
BitcoinKey1Bytes: pub1,
|
|
|
|
BitcoinKey2Bytes: pub2,
|
|
|
|
AuthProof: nil,
|
|
|
|
}
|
|
|
|
if err := ctx.router.AddEdge(edge); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now that the edge has been inserted, query is the router already
|
|
|
|
// knows of the edge should return true.
|
|
|
|
if !ctx.router.IsKnownEdge(*chanID) {
|
|
|
|
t.Fatalf("router should detect edge as known")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TestIsStaleEdgePolicy tests that the IsStaleEdgePolicy properly detects
|
|
|
|
// stale channel edge update announcements.
|
|
|
|
func TestIsStaleEdgePolicy(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
const startingBlockHeight = 101
|
|
|
|
ctx, cleanUp, err := createTestCtx(startingBlockHeight,
|
|
|
|
basicGraphFilePath)
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create router: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// First, we'll create a new channel edge (just the info) and insert it
|
|
|
|
// into the database.
|
|
|
|
var (
|
|
|
|
pub1 [33]byte
|
|
|
|
pub2 [33]byte
|
|
|
|
)
|
|
|
|
copy(pub1[:], priv1.PubKey().SerializeCompressed())
|
|
|
|
copy(pub2[:], priv2.PubKey().SerializeCompressed())
|
|
|
|
|
|
|
|
fundingTx, _, chanID, err := createChannelEdge(ctx,
|
|
|
|
bitcoinKey1.SerializeCompressed(),
|
|
|
|
bitcoinKey2.SerializeCompressed(),
|
|
|
|
10000, 500)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create channel edge: %v", err)
|
|
|
|
}
|
|
|
|
fundingBlock := &wire.MsgBlock{
|
|
|
|
Transactions: []*wire.MsgTx{fundingTx},
|
|
|
|
}
|
|
|
|
ctx.chain.addBlock(fundingBlock, chanID.BlockHeight, chanID.BlockHeight)
|
|
|
|
|
|
|
|
// If we query for staleness before adding the edge, we should get
|
|
|
|
// false.
|
|
|
|
updateTimeStamp := time.Unix(123, 0)
|
|
|
|
if ctx.router.IsStaleEdgePolicy(*chanID, updateTimeStamp, 0) {
|
|
|
|
t.Fatalf("router failed to detect fresh edge policy")
|
|
|
|
}
|
|
|
|
if ctx.router.IsStaleEdgePolicy(*chanID, updateTimeStamp, 1) {
|
|
|
|
t.Fatalf("router failed to detect fresh edge policy")
|
|
|
|
}
|
|
|
|
|
|
|
|
edge := &channeldb.ChannelEdgeInfo{
|
|
|
|
ChannelID: chanID.ToUint64(),
|
|
|
|
NodeKey1Bytes: pub1,
|
|
|
|
NodeKey2Bytes: pub2,
|
|
|
|
BitcoinKey1Bytes: pub1,
|
|
|
|
BitcoinKey2Bytes: pub2,
|
|
|
|
AuthProof: nil,
|
|
|
|
}
|
|
|
|
if err := ctx.router.AddEdge(edge); err != nil {
|
|
|
|
t.Fatalf("unable to add edge: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll also add two edge policies, one for each direction.
|
|
|
|
edgePolicy := &channeldb.ChannelEdgePolicy{
|
|
|
|
SigBytes: testSig.Serialize(),
|
|
|
|
ChannelID: edge.ChannelID,
|
|
|
|
LastUpdate: updateTimeStamp,
|
|
|
|
TimeLockDelta: 10,
|
|
|
|
MinHTLC: 1,
|
|
|
|
FeeBaseMSat: 10,
|
|
|
|
FeeProportionalMillionths: 10000,
|
|
|
|
}
|
|
|
|
edgePolicy.Flags = 0
|
|
|
|
if err := ctx.router.UpdateEdge(edgePolicy); err != nil {
|
|
|
|
t.Fatalf("unable to update edge policy: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
edgePolicy = &channeldb.ChannelEdgePolicy{
|
|
|
|
SigBytes: testSig.Serialize(),
|
|
|
|
ChannelID: edge.ChannelID,
|
|
|
|
LastUpdate: updateTimeStamp,
|
|
|
|
TimeLockDelta: 10,
|
|
|
|
MinHTLC: 1,
|
|
|
|
FeeBaseMSat: 10,
|
|
|
|
FeeProportionalMillionths: 10000,
|
|
|
|
}
|
|
|
|
edgePolicy.Flags = 1
|
|
|
|
if err := ctx.router.UpdateEdge(edgePolicy); err != nil {
|
|
|
|
t.Fatalf("unable to update edge policy: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now that the edges have been added, an identical (chanID, flag,
|
|
|
|
// timestamp) tuple for each edge should be detected as a stale edge.
|
|
|
|
if !ctx.router.IsStaleEdgePolicy(*chanID, updateTimeStamp, 0) {
|
|
|
|
t.Fatalf("router failed to detect stale edge policy")
|
|
|
|
}
|
|
|
|
if !ctx.router.IsStaleEdgePolicy(*chanID, updateTimeStamp, 1) {
|
|
|
|
t.Fatalf("router failed to detect stale edge policy")
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we now update the timestamp for both edges, the router should
|
|
|
|
// detect that this tuple represents a fresh edge.
|
|
|
|
updateTimeStamp = time.Unix(9999, 0)
|
|
|
|
if ctx.router.IsStaleEdgePolicy(*chanID, updateTimeStamp, 0) {
|
|
|
|
t.Fatalf("router failed to detect fresh edge policy")
|
|
|
|
}
|
|
|
|
if ctx.router.IsStaleEdgePolicy(*chanID, updateTimeStamp, 1) {
|
|
|
|
t.Fatalf("router failed to detect fresh edge policy")
|
|
|
|
}
|
|
|
|
}
|