2019-05-15 10:02:53 +03:00
|
|
|
package itest
|
2016-08-30 08:07:54 +03:00
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
2019-09-29 01:43:42 +03:00
|
|
|
"context"
|
2019-01-23 07:45:38 +03:00
|
|
|
"crypto/rand"
|
2020-11-04 13:03:29 +03:00
|
|
|
"flag"
|
2016-08-30 08:07:54 +03:00
|
|
|
"fmt"
|
2016-11-22 06:16:44 +03:00
|
|
|
"io"
|
|
|
|
"io/ioutil"
|
2019-02-12 00:02:00 +03:00
|
|
|
"math"
|
2016-11-22 06:16:44 +03:00
|
|
|
"os"
|
2017-01-08 07:25:45 +03:00
|
|
|
"strings"
|
2016-10-24 05:00:09 +03:00
|
|
|
"testing"
|
2016-09-14 05:00:47 +03:00
|
|
|
"time"
|
2016-10-24 05:00:09 +03:00
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
"github.com/btcsuite/btcd/btcjson"
|
2018-06-05 04:34:16 +03:00
|
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
|
|
"github.com/btcsuite/btcd/integration/rpctest"
|
|
|
|
"github.com/btcsuite/btcd/rpcclient"
|
|
|
|
"github.com/btcsuite/btcd/wire"
|
|
|
|
"github.com/btcsuite/btcutil"
|
2020-05-25 09:38:25 +03:00
|
|
|
"github.com/btcsuite/btcwallet/wallet"
|
2018-06-15 22:33:49 +03:00
|
|
|
"github.com/davecgh/go-spew/spew"
|
|
|
|
"github.com/go-errors/errors"
|
2020-10-06 21:56:10 +03:00
|
|
|
"github.com/lightningnetwork/lnd/chainreg"
|
2020-11-17 02:28:56 +03:00
|
|
|
"github.com/lightningnetwork/lnd/funding"
|
2019-12-11 07:02:49 +03:00
|
|
|
"github.com/lightningnetwork/lnd/input"
|
2020-07-29 10:27:22 +03:00
|
|
|
"github.com/lightningnetwork/lnd/labels"
|
2020-05-13 16:47:45 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lncfg"
|
2018-06-15 22:33:49 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnrpc"
|
2019-03-19 19:09:27 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnrpc/routerrpc"
|
2019-12-11 07:02:49 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnrpc/signrpc"
|
2020-05-05 22:46:26 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnrpc/walletrpc"
|
2018-06-15 22:33:49 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lntest"
|
2019-09-19 22:46:29 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lntest/wait"
|
2020-06-10 15:21:13 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwallet"
|
2020-03-27 09:55:13 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwallet/chainfee"
|
2018-06-15 22:33:49 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
2020-07-07 11:32:13 +03:00
|
|
|
"github.com/stretchr/testify/require"
|
2016-08-30 08:07:54 +03:00
|
|
|
)
|
|
|
|
|
2020-11-04 13:03:29 +03:00
|
|
|
const (
|
|
|
|
// defaultSplitTranches is the default number of tranches we split the
|
|
|
|
// test cases into.
|
|
|
|
defaultSplitTranches uint = 1
|
|
|
|
|
|
|
|
// defaultRunTranche is the default index of the test cases tranche that
|
|
|
|
// we run.
|
|
|
|
defaultRunTranche uint = 0
|
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
// testCasesSplitParts is the number of tranches the test cases should
|
|
|
|
// be split into. By default this is set to 1, so no splitting happens.
|
|
|
|
// If this value is increased, then the -runtranche flag must be
|
|
|
|
// specified as well to indicate which part should be run in the current
|
|
|
|
// invocation.
|
|
|
|
testCasesSplitTranches = flag.Uint(
|
|
|
|
"splittranches", defaultSplitTranches, "split the test cases "+
|
|
|
|
"in this many tranches and run the tranche at "+
|
|
|
|
"0-based index specified by the -runtranche flag",
|
|
|
|
)
|
|
|
|
|
|
|
|
// testCasesRunTranche is the 0-based index of the split test cases
|
|
|
|
// tranche to run in the current invocation.
|
|
|
|
testCasesRunTranche = flag.Uint(
|
|
|
|
"runtranche", defaultRunTranche, "run the tranche of the "+
|
|
|
|
"split test cases with the given (0-based) index",
|
|
|
|
)
|
2020-06-22 20:36:12 +03:00
|
|
|
|
2021-06-14 13:31:08 +03:00
|
|
|
// dbBackendFlag specifies the backend to use
|
|
|
|
dbBackendFlag = flag.String("dbbackend", "bbolt", "Database backend (bbolt, etcd)")
|
2020-11-04 13:03:29 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// getTestCaseSplitTranche returns the sub slice of the test cases that should
|
|
|
|
// be run as the current split tranche as well as the index and slice offset of
|
|
|
|
// the tranche.
|
|
|
|
func getTestCaseSplitTranche() ([]*testCase, uint, uint) {
|
|
|
|
numTranches := defaultSplitTranches
|
|
|
|
if testCasesSplitTranches != nil {
|
|
|
|
numTranches = *testCasesSplitTranches
|
|
|
|
}
|
|
|
|
runTranche := defaultRunTranche
|
|
|
|
if testCasesRunTranche != nil {
|
|
|
|
runTranche = *testCasesRunTranche
|
|
|
|
}
|
|
|
|
|
2020-11-07 15:23:30 +03:00
|
|
|
// There's a special flake-hunt mode where we run the same test multiple
|
|
|
|
// times in parallel. In that case the tranche index is equal to the
|
|
|
|
// thread ID, but we need to actually run all tests for the regex
|
|
|
|
// selection to work.
|
|
|
|
threadID := runTranche
|
|
|
|
if numTranches == 1 {
|
|
|
|
runTranche = 0
|
|
|
|
}
|
|
|
|
|
2020-11-04 13:03:29 +03:00
|
|
|
numCases := uint(len(allTestCases))
|
|
|
|
testsPerTranche := numCases / numTranches
|
|
|
|
trancheOffset := runTranche * testsPerTranche
|
|
|
|
trancheEnd := trancheOffset + testsPerTranche
|
|
|
|
if trancheEnd > numCases || runTranche == numTranches-1 {
|
|
|
|
trancheEnd = numCases
|
|
|
|
}
|
|
|
|
|
2020-11-07 15:23:30 +03:00
|
|
|
return allTestCases[trancheOffset:trancheEnd], threadID, trancheOffset
|
2020-11-04 13:03:29 +03:00
|
|
|
}
|
|
|
|
|
2019-03-12 02:24:45 +03:00
|
|
|
func rpcPointToWirePoint(t *harnessTest, chanPoint *lnrpc.ChannelPoint) wire.OutPoint {
|
2021-02-13 11:05:33 +03:00
|
|
|
txid, err := lnrpc.GetChanPointFundingTxid(chanPoint)
|
2019-03-12 02:24:45 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get txid: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return wire.OutPoint{
|
|
|
|
Hash: *txid,
|
|
|
|
Index: chanPoint.OutputIndex,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-10-26 01:32:00 +03:00
|
|
|
// completePaymentRequests sends payments from a lightning node to complete all
|
|
|
|
// payment requests. If the awaitResponse parameter is true, this function
|
|
|
|
// does not return until all payments successfully complete without errors.
|
|
|
|
func completePaymentRequests(ctx context.Context, client lnrpc.LightningClient,
|
2020-03-31 13:44:18 +03:00
|
|
|
routerClient routerrpc.RouterClient, paymentRequests []string,
|
|
|
|
awaitResponse bool) error {
|
2017-10-26 01:32:00 +03:00
|
|
|
|
2019-04-15 16:00:51 +03:00
|
|
|
// We start by getting the current state of the client's channels. This
|
|
|
|
// is needed to ensure the payments actually have been committed before
|
|
|
|
// we return.
|
|
|
|
ctxt, _ := context.WithTimeout(ctx, defaultTimeout)
|
|
|
|
req := &lnrpc.ListChannelsRequest{}
|
|
|
|
listResp, err := client.ListChannels(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2020-03-31 13:44:18 +03:00
|
|
|
// send sends a payment and returns an error if it doesn't succeeded.
|
|
|
|
send := func(payReq string) error {
|
|
|
|
ctxc, cancel := context.WithCancel(ctx)
|
|
|
|
defer cancel()
|
2017-10-26 01:32:00 +03:00
|
|
|
|
2020-03-31 13:44:18 +03:00
|
|
|
payStream, err := routerClient.SendPaymentV2(
|
|
|
|
ctxc,
|
|
|
|
&routerrpc.SendPaymentRequest{
|
|
|
|
PaymentRequest: payReq,
|
|
|
|
TimeoutSeconds: 60,
|
|
|
|
FeeLimitMsat: noFeeLimitMsat,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2019-03-12 02:24:45 +03:00
|
|
|
}
|
2020-03-31 13:44:18 +03:00
|
|
|
|
|
|
|
resp, err := getPaymentResult(payStream)
|
2017-10-26 01:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2020-03-31 13:44:18 +03:00
|
|
|
if resp.Status != lnrpc.Payment_SUCCEEDED {
|
|
|
|
return errors.New(resp.FailureReason)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Launch all payments simultaneously.
|
|
|
|
results := make(chan error)
|
|
|
|
for _, payReq := range paymentRequests {
|
|
|
|
payReqCopy := payReq
|
|
|
|
go func() {
|
|
|
|
err := send(payReqCopy)
|
|
|
|
if awaitResponse {
|
|
|
|
results <- err
|
|
|
|
}
|
|
|
|
}()
|
2017-10-26 01:32:00 +03:00
|
|
|
}
|
|
|
|
|
2020-03-31 13:44:18 +03:00
|
|
|
// If awaiting a response, verify that all payments succeeded.
|
2017-10-26 01:32:00 +03:00
|
|
|
if awaitResponse {
|
|
|
|
for range paymentRequests {
|
2020-03-31 13:44:18 +03:00
|
|
|
err := <-results
|
2017-10-26 01:32:00 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
2019-04-15 16:00:51 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// We are not waiting for feedback in the form of a response, but we
|
|
|
|
// should still wait long enough for the server to receive and handle
|
|
|
|
// the send before cancelling the request. We wait for the number of
|
|
|
|
// updates to one of our channels has increased before we return.
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
2019-04-15 16:00:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctx, defaultTimeout)
|
|
|
|
newListResp, err := client.ListChannels(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2020-10-16 22:11:12 +03:00
|
|
|
// If the number of open channels is now lower than before
|
|
|
|
// attempting the payments, it means one of the payments
|
|
|
|
// triggered a force closure (for example, due to an incorrect
|
|
|
|
// preimage). Return early since it's clear the payment was
|
|
|
|
// attempted.
|
|
|
|
if len(newListResp.Channels) < len(listResp.Channels) {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
2019-04-15 16:00:51 +03:00
|
|
|
for _, c1 := range listResp.Channels {
|
|
|
|
for _, c2 := range newListResp.Channels {
|
|
|
|
if c1.ChannelPoint != c2.ChannelPoint {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// If this channel has an increased numbr of
|
|
|
|
// updates, we assume the payments are
|
|
|
|
// committed, and we can return.
|
|
|
|
if c2.NumUpdates > c1.NumUpdates {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return false
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2019-04-15 16:00:51 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
2017-10-26 01:32:00 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-08-12 16:23:08 +03:00
|
|
|
// makeFakePayHash creates random pre image hash
|
|
|
|
func makeFakePayHash(t *harnessTest) []byte {
|
|
|
|
randBuf := make([]byte, 32)
|
|
|
|
|
|
|
|
if _, err := rand.Read(randBuf); err != nil {
|
|
|
|
t.Fatalf("internal error, cannot generate random string: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return randBuf
|
|
|
|
}
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// createPayReqs is a helper method that will create a slice of payment
|
|
|
|
// requests for the given node.
|
2019-03-28 05:00:27 +03:00
|
|
|
func createPayReqs(node *lntest.HarnessNode, paymentAmt btcutil.Amount,
|
|
|
|
numInvoices int) ([]string, [][]byte, []*lnrpc.Invoice, error) {
|
2018-11-20 17:09:44 +03:00
|
|
|
|
|
|
|
payReqs := make([]string, numInvoices)
|
|
|
|
rHashes := make([][]byte, numInvoices)
|
|
|
|
invoices := make([]*lnrpc.Invoice, numInvoices)
|
|
|
|
for i := 0; i < numInvoices; i++ {
|
|
|
|
preimage := make([]byte, 32)
|
|
|
|
_, err := rand.Read(preimage)
|
|
|
|
if err != nil {
|
|
|
|
return nil, nil, nil, fmt.Errorf("unable to generate "+
|
|
|
|
"preimage: %v", err)
|
|
|
|
}
|
|
|
|
invoice := &lnrpc.Invoice{
|
|
|
|
Memo: "testing",
|
|
|
|
RPreimage: preimage,
|
|
|
|
Value: int64(paymentAmt),
|
|
|
|
}
|
2019-03-28 05:00:27 +03:00
|
|
|
ctxt, _ := context.WithTimeout(
|
|
|
|
context.Background(), defaultTimeout,
|
|
|
|
)
|
|
|
|
resp, err := node.AddInvoice(ctxt, invoice)
|
2018-11-20 17:09:44 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, nil, nil, fmt.Errorf("unable to add "+
|
|
|
|
"invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-11-25 06:05:32 +03:00
|
|
|
// Set the payment address in the invoice so the caller can
|
|
|
|
// properly use it.
|
|
|
|
invoice.PaymentAddr = resp.PaymentAddr
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
payReqs[i] = resp.PaymentRequest
|
|
|
|
rHashes[i] = resp.RHash
|
|
|
|
invoices[i] = invoice
|
|
|
|
}
|
|
|
|
return payReqs, rHashes, invoices, nil
|
|
|
|
}
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// getChanInfo is a helper method for getting channel info for a node's sole
|
|
|
|
// channel.
|
|
|
|
func getChanInfo(ctx context.Context, node *lntest.HarnessNode) (
|
|
|
|
*lnrpc.Channel, error) {
|
|
|
|
|
|
|
|
req := &lnrpc.ListChannelsRequest{}
|
|
|
|
channelInfo, err := node.ListChannels(ctx, req)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if len(channelInfo.Channels) != 1 {
|
|
|
|
return nil, fmt.Errorf("node should only have a single "+
|
2019-05-24 15:17:49 +03:00
|
|
|
"channel, instead it has %v", len(channelInfo.Channels))
|
2018-11-20 17:09:44 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return channelInfo.Channels[0], nil
|
|
|
|
}
|
|
|
|
|
2020-06-09 09:47:17 +03:00
|
|
|
// testGetRecoveryInfo checks whether lnd gives the right information about
|
|
|
|
// the wallet recovery process.
|
|
|
|
func testGetRecoveryInfo(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
// First, create a new node with strong passphrase and grab the mnemonic
|
|
|
|
// used for key derivation. This will bring up Carol with an empty
|
|
|
|
// wallet, and such that she is synced up.
|
|
|
|
password := []byte("The Magic Words are Squeamish Ossifrage")
|
2020-10-06 18:23:40 +03:00
|
|
|
carol, mnemonic, _, err := net.NewNodeWithSeed(
|
|
|
|
"Carol", nil, password, false,
|
|
|
|
)
|
2020-06-09 09:47:17 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create node with seed; %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
shutdownAndAssert(net, t, carol)
|
|
|
|
|
|
|
|
checkInfo := func(expectedRecoveryMode, expectedRecoveryFinished bool,
|
|
|
|
expectedProgress float64, recoveryWindow int32) {
|
|
|
|
|
|
|
|
// Restore Carol, passing in the password, mnemonic, and
|
|
|
|
// desired recovery window.
|
|
|
|
node, err := net.RestoreNodeWithSeed(
|
|
|
|
"Carol", nil, password, mnemonic, recoveryWindow, nil,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to restore node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for Carol to sync to the chain.
|
2021-03-10 01:12:26 +03:00
|
|
|
_, minerHeight, err := net.Miner.Client.GetBestBlock()
|
2020-06-09 09:47:17 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get current blockheight %v", err)
|
|
|
|
}
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
err = waitForNodeBlockHeight(ctxt, node, minerHeight)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to sync to chain: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Query carol for her current wallet recovery progress.
|
|
|
|
var (
|
|
|
|
recoveryMode bool
|
|
|
|
recoveryFinished bool
|
|
|
|
progress float64
|
|
|
|
)
|
|
|
|
|
|
|
|
err = wait.Predicate(func() bool {
|
|
|
|
// Verify that recovery info gives the right response.
|
|
|
|
req := &lnrpc.GetRecoveryInfoRequest{}
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
resp, err := node.GetRecoveryInfo(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query recovery info: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
recoveryMode = resp.RecoveryMode
|
|
|
|
recoveryFinished = resp.RecoveryFinished
|
|
|
|
progress = resp.Progress
|
|
|
|
|
|
|
|
if recoveryMode != expectedRecoveryMode ||
|
|
|
|
recoveryFinished != expectedRecoveryFinished ||
|
|
|
|
progress != expectedProgress {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2020-06-09 09:47:17 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("expected recovery mode to be %v, got %v, "+
|
|
|
|
"expected recovery finished to be %v, got %v, "+
|
|
|
|
"expected progress %v, got %v",
|
|
|
|
expectedRecoveryMode, recoveryMode,
|
|
|
|
expectedRecoveryFinished, recoveryFinished,
|
|
|
|
expectedProgress, progress,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Lastly, shutdown this Carol so we can move on to the next
|
|
|
|
// restoration.
|
|
|
|
shutdownAndAssert(net, t, node)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restore Carol with a recovery window of 0. Since it's not in recovery
|
|
|
|
// mode, the recovery info will give a response with recoveryMode=false,
|
|
|
|
// recoveryFinished=false, and progress=0
|
|
|
|
checkInfo(false, false, 0, 0)
|
|
|
|
|
|
|
|
// Change the recovery windown to be 1 to turn on recovery mode. Since the
|
|
|
|
// current chain height is the same as the birthday height, it should
|
|
|
|
// indicate the recovery process is finished.
|
|
|
|
checkInfo(true, true, 1, 1)
|
|
|
|
|
|
|
|
// We now go ahead 5 blocks. Because the wallet's syncing process is
|
|
|
|
// controlled by a goroutine in the background, it will catch up quickly.
|
|
|
|
// This makes the recovery progress back to 1.
|
|
|
|
mineBlocks(t, net, 5, 0)
|
|
|
|
checkInfo(true, true, 1, 1)
|
|
|
|
}
|
|
|
|
|
2018-04-03 02:57:58 +03:00
|
|
|
// testOnchainFundRecovery checks lnd's ability to rescan for onchain outputs
|
|
|
|
// when providing a valid aezeed that owns outputs on the chain. This test
|
|
|
|
// performs multiple restorations using the same seed and various recovery
|
|
|
|
// windows to ensure we detect funds properly.
|
|
|
|
func testOnchainFundRecovery(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
// First, create a new node with strong passphrase and grab the mnemonic
|
|
|
|
// used for key derivation. This will bring up Carol with an empty
|
|
|
|
// wallet, and such that she is synced up.
|
|
|
|
password := []byte("The Magic Words are Squeamish Ossifrage")
|
2020-10-06 18:23:40 +03:00
|
|
|
carol, mnemonic, _, err := net.NewNodeWithSeed(
|
|
|
|
"Carol", nil, password, false,
|
|
|
|
)
|
2018-04-03 02:57:58 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create node with seed; %v", err)
|
|
|
|
}
|
2018-06-08 00:14:39 +03:00
|
|
|
shutdownAndAssert(net, t, carol)
|
2018-04-03 02:57:58 +03:00
|
|
|
|
|
|
|
// Create a closure for testing the recovery of Carol's wallet. This
|
|
|
|
// method takes the expected value of Carol's balance when using the
|
|
|
|
// given recovery window. Additionally, the caller can specify an action
|
|
|
|
// to perform on the restored node before the node is shutdown.
|
2020-03-31 19:36:37 +03:00
|
|
|
restoreCheckBalance := func(expAmount int64, expectedNumUTXOs uint32,
|
2019-02-12 00:02:00 +03:00
|
|
|
recoveryWindow int32, fn func(*lntest.HarnessNode)) {
|
2018-04-03 02:57:58 +03:00
|
|
|
|
|
|
|
// Restore Carol, passing in the password, mnemonic, and
|
|
|
|
// desired recovery window.
|
|
|
|
node, err := net.RestoreNodeWithSeed(
|
2019-03-12 02:26:26 +03:00
|
|
|
"Carol", nil, password, mnemonic, recoveryWindow, nil,
|
2018-04-03 02:57:58 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to restore node: %v", err)
|
|
|
|
}
|
|
|
|
|
2019-02-12 00:02:00 +03:00
|
|
|
// Query carol for her current wallet balance, and also that we
|
|
|
|
// gain the expected number of UTXOs.
|
|
|
|
var (
|
|
|
|
currBalance int64
|
|
|
|
currNumUTXOs uint32
|
|
|
|
)
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
2018-04-03 02:57:58 +03:00
|
|
|
req := &lnrpc.WalletBalanceRequest{}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2018-04-03 02:57:58 +03:00
|
|
|
resp, err := node.WalletBalance(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query wallet balance: %v",
|
|
|
|
err)
|
|
|
|
}
|
|
|
|
currBalance = resp.ConfirmedBalance
|
|
|
|
|
2019-02-12 00:02:00 +03:00
|
|
|
utxoReq := &lnrpc.ListUnspentRequest{
|
|
|
|
MaxConfs: math.MaxInt32,
|
|
|
|
}
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
utxoResp, err := node.ListUnspent(ctxt, utxoReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query utxos: %v", err)
|
|
|
|
}
|
2020-03-31 19:36:37 +03:00
|
|
|
currNumUTXOs = uint32(len(utxoResp.Utxos))
|
2019-02-12 00:02:00 +03:00
|
|
|
|
2020-03-31 19:36:37 +03:00
|
|
|
// Verify that Carol's balance and number of UTXOs
|
|
|
|
// matches what's expected.
|
|
|
|
if expAmount != currBalance {
|
|
|
|
return false
|
|
|
|
}
|
2019-02-12 00:02:00 +03:00
|
|
|
if currNumUTXOs != expectedNumUTXOs {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2018-04-03 02:57:58 +03:00
|
|
|
return true
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2018-04-03 02:57:58 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("expected restored node to have %d satoshis, "+
|
2019-02-12 00:02:00 +03:00
|
|
|
"instead has %d satoshis, expected %d utxos "+
|
|
|
|
"instead has %d", expAmount, currBalance,
|
|
|
|
expectedNumUTXOs, currNumUTXOs)
|
2018-04-03 02:57:58 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// If the user provided a callback, execute the commands against
|
|
|
|
// the restored Carol.
|
|
|
|
if fn != nil {
|
|
|
|
fn(node)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Lastly, shutdown this Carol so we can move on to the next
|
|
|
|
// restoration.
|
2018-06-08 00:14:39 +03:00
|
|
|
shutdownAndAssert(net, t, node)
|
2018-04-03 02:57:58 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Create a closure-factory for building closures that can generate and
|
|
|
|
// skip a configurable number of addresses, before finally sending coins
|
|
|
|
// to a next generated address. The returned closure will apply the same
|
|
|
|
// behavior to both default P2WKH and NP2WKH scopes.
|
|
|
|
skipAndSend := func(nskip int) func(*lntest.HarnessNode) {
|
|
|
|
return func(node *lntest.HarnessNode) {
|
|
|
|
newP2WKHAddrReq := &lnrpc.NewAddressRequest{
|
|
|
|
Type: AddrTypeWitnessPubkeyHash,
|
|
|
|
}
|
|
|
|
|
|
|
|
newNP2WKHAddrReq := &lnrpc.NewAddressRequest{
|
|
|
|
Type: AddrTypeNestedPubkeyHash,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Generate and skip the number of addresses requested.
|
|
|
|
for i := 0; i < nskip; i++ {
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2018-04-03 02:57:58 +03:00
|
|
|
_, err = node.NewAddress(ctxt, newP2WKHAddrReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to generate new "+
|
|
|
|
"p2wkh address: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-04-03 02:57:58 +03:00
|
|
|
_, err = node.NewAddress(ctxt, newNP2WKHAddrReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to generate new "+
|
|
|
|
"np2wkh address: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Send one BTC to the next P2WKH address.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(
|
|
|
|
ctxt, t.t, btcutil.SatoshiPerBitcoin, node,
|
2018-04-03 02:57:58 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// And another to the next NP2WKH address.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoinsNP2WKH(
|
|
|
|
ctxt, t.t, btcutil.SatoshiPerBitcoin, node,
|
2018-04-03 02:57:58 +03:00
|
|
|
)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restore Carol with a recovery window of 0. Since no coins have been
|
|
|
|
// sent, her balance should be zero.
|
|
|
|
//
|
|
|
|
// After, one BTC is sent to both her first external P2WKH and NP2WKH
|
|
|
|
// addresses.
|
2019-02-12 00:02:00 +03:00
|
|
|
restoreCheckBalance(0, 0, 0, skipAndSend(0))
|
2018-04-03 02:57:58 +03:00
|
|
|
|
|
|
|
// Check that restoring without a look-ahead results in having no funds
|
|
|
|
// in the wallet, even though they exist on-chain.
|
2019-02-12 00:02:00 +03:00
|
|
|
restoreCheckBalance(0, 0, 0, nil)
|
2018-04-03 02:57:58 +03:00
|
|
|
|
2019-02-12 00:02:00 +03:00
|
|
|
// Now, check that using a look-ahead of 1 recovers the balance from
|
|
|
|
// the two transactions above. We should also now have 2 UTXOs in the
|
|
|
|
// wallet at the end of the recovery attempt.
|
2018-04-03 02:57:58 +03:00
|
|
|
//
|
|
|
|
// After, we will generate and skip 9 P2WKH and NP2WKH addresses, and
|
|
|
|
// send another BTC to the subsequent 10th address in each derivation
|
|
|
|
// path.
|
2019-02-12 00:02:00 +03:00
|
|
|
restoreCheckBalance(2*btcutil.SatoshiPerBitcoin, 2, 1, skipAndSend(9))
|
2018-04-03 02:57:58 +03:00
|
|
|
|
|
|
|
// Check that using a recovery window of 9 does not find the two most
|
|
|
|
// recent txns.
|
2019-02-12 00:02:00 +03:00
|
|
|
restoreCheckBalance(2*btcutil.SatoshiPerBitcoin, 2, 9, nil)
|
2018-04-03 02:57:58 +03:00
|
|
|
|
|
|
|
// Extending our recovery window to 10 should find the most recent
|
2019-02-12 00:02:00 +03:00
|
|
|
// transactions, leaving the wallet with 4 BTC total. We should also
|
|
|
|
// learn of the two additional UTXOs created above.
|
2018-04-03 02:57:58 +03:00
|
|
|
//
|
|
|
|
// After, we will skip 19 more addrs, sending to the 20th address past
|
|
|
|
// our last found address, and repeat the same checks.
|
2019-02-12 00:02:00 +03:00
|
|
|
restoreCheckBalance(4*btcutil.SatoshiPerBitcoin, 4, 10, skipAndSend(19))
|
2018-04-03 02:57:58 +03:00
|
|
|
|
|
|
|
// Check that recovering with a recovery window of 19 fails to find the
|
|
|
|
// most recent transactions.
|
2019-02-12 00:02:00 +03:00
|
|
|
restoreCheckBalance(4*btcutil.SatoshiPerBitcoin, 4, 19, nil)
|
2018-04-03 02:57:58 +03:00
|
|
|
|
2019-02-12 00:02:00 +03:00
|
|
|
// Ensure that using a recovery window of 20 succeeds with all UTXOs
|
|
|
|
// found and the final balance reflected.
|
2020-03-31 19:36:37 +03:00
|
|
|
|
|
|
|
// After these checks are done, we'll want to make sure we can also
|
|
|
|
// recover change address outputs. This is mainly motivated by a now
|
|
|
|
// fixed bug in the wallet in which change addresses could at times be
|
|
|
|
// created outside of the default key scopes. Recovery only used to be
|
|
|
|
// performed on the default key scopes, so ideally this test case
|
|
|
|
// would've caught the bug earlier. Carol has received 6 BTC so far from
|
|
|
|
// the miner, we'll send 5 back to ensure all of her UTXOs get spent to
|
|
|
|
// avoid fee discrepancies and a change output is formed.
|
|
|
|
const minerAmt = 5 * btcutil.SatoshiPerBitcoin
|
|
|
|
const finalBalance = 6 * btcutil.SatoshiPerBitcoin
|
|
|
|
promptChangeAddr := func(node *lntest.HarnessNode) {
|
|
|
|
minerAddr, err := net.Miner.NewAddress()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create new miner address: %v", err)
|
|
|
|
}
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
resp, err := node.SendCoins(ctxt, &lnrpc.SendCoinsRequest{
|
|
|
|
Addr: minerAddr.String(),
|
|
|
|
Amount: minerAmt,
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send coins to miner: %v", err)
|
|
|
|
}
|
|
|
|
txid, err := waitForTxInMempool(
|
2021-03-10 01:12:26 +03:00
|
|
|
net.Miner.Client, minerMempoolTimeout,
|
2020-03-31 19:36:37 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("transaction not found in mempool: %v", err)
|
|
|
|
}
|
|
|
|
if resp.Txid != txid.String() {
|
|
|
|
t.Fatalf("txid mismatch: %v vs %v", resp.Txid,
|
|
|
|
txid.String())
|
|
|
|
}
|
|
|
|
block := mineBlocks(t, net, 1, 1)[0]
|
|
|
|
assertTxInBlock(t, block, txid)
|
|
|
|
}
|
|
|
|
restoreCheckBalance(finalBalance, 6, 20, promptChangeAddr)
|
|
|
|
|
|
|
|
// We should expect a static fee of 27750 satoshis for spending 6 inputs
|
|
|
|
// (3 P2WPKH, 3 NP2WPKH) to two P2WPKH outputs. Carol should therefore
|
|
|
|
// only have one UTXO present (the change output) of 6 - 5 - fee BTC.
|
|
|
|
const fee = 27750
|
|
|
|
restoreCheckBalance(finalBalance-minerAmt-fee, 1, 21, nil)
|
2018-04-03 02:57:58 +03:00
|
|
|
}
|
|
|
|
|
2020-03-06 18:11:48 +03:00
|
|
|
// commitType is a simple enum used to run though the basic funding flow with
|
|
|
|
// different commitment formats.
|
|
|
|
type commitType byte
|
|
|
|
|
|
|
|
const (
|
|
|
|
// commitTypeLegacy is the old school commitment type.
|
2020-03-14 04:52:33 +03:00
|
|
|
commitTypeLegacy commitType = iota
|
2020-03-06 18:11:48 +03:00
|
|
|
|
|
|
|
// commiTypeTweakless is the commitment type where the remote key is
|
|
|
|
// static (non-tweaked).
|
|
|
|
commitTypeTweakless
|
2020-03-14 04:52:33 +03:00
|
|
|
|
|
|
|
// commitTypeAnchors is the kind of commitment that has extra outputs
|
|
|
|
// used for anchoring down to commitment using CPFP.
|
|
|
|
commitTypeAnchors
|
2020-03-06 18:11:48 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// String returns that name of the commitment type.
|
|
|
|
func (c commitType) String() string {
|
|
|
|
switch c {
|
|
|
|
case commitTypeLegacy:
|
|
|
|
return "legacy"
|
|
|
|
case commitTypeTweakless:
|
|
|
|
return "tweakless"
|
2020-03-14 04:52:33 +03:00
|
|
|
case commitTypeAnchors:
|
|
|
|
return "anchors"
|
2020-03-06 18:11:48 +03:00
|
|
|
default:
|
|
|
|
return "invalid"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Args returns the command line flag to supply to enable this commitment type.
|
|
|
|
func (c commitType) Args() []string {
|
|
|
|
switch c {
|
|
|
|
case commitTypeLegacy:
|
2020-07-09 00:26:05 +03:00
|
|
|
return []string{"--protocol.legacy.committweak"}
|
2020-03-06 18:11:48 +03:00
|
|
|
case commitTypeTweakless:
|
|
|
|
return []string{}
|
2020-03-14 04:52:33 +03:00
|
|
|
case commitTypeAnchors:
|
|
|
|
return []string{"--protocol.anchors"}
|
2020-03-06 18:11:48 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-03-04 15:21:28 +03:00
|
|
|
// calcStaticFee calculates appropriate fees for commitment transactions. This
|
|
|
|
// function provides a simple way to allow test balance assertions to take fee
|
|
|
|
// calculations into account.
|
|
|
|
func (c commitType) calcStaticFee(numHTLCs int) btcutil.Amount {
|
|
|
|
const htlcWeight = input.HTLCWeight
|
|
|
|
var (
|
|
|
|
feePerKw = chainfee.SatPerKVByte(50000).FeePerKWeight()
|
|
|
|
commitWeight = input.CommitWeight
|
|
|
|
anchors = btcutil.Amount(0)
|
|
|
|
)
|
|
|
|
|
|
|
|
// The anchor commitment type is slightly heavier, and we must also add
|
|
|
|
// the value of the two anchors to the resulting fee the initiator
|
2020-12-15 13:22:14 +03:00
|
|
|
// pays. In addition the fee rate is capped at 10 sat/vbyte for anchor
|
|
|
|
// channels.
|
2020-03-04 15:21:28 +03:00
|
|
|
if c == commitTypeAnchors {
|
2020-12-15 13:22:14 +03:00
|
|
|
feePerKw = chainfee.SatPerKVByte(
|
|
|
|
lnwallet.DefaultAnchorsCommitMaxFeeRateSatPerVByte * 1000,
|
|
|
|
).FeePerKWeight()
|
2020-03-04 15:21:28 +03:00
|
|
|
commitWeight = input.AnchorCommitWeight
|
|
|
|
anchors = 2 * anchorSize
|
|
|
|
}
|
|
|
|
|
|
|
|
return feePerKw.FeeForWeight(int64(commitWeight+htlcWeight*numHTLCs)) +
|
|
|
|
anchors
|
|
|
|
}
|
|
|
|
|
|
|
|
// channelCommitType retrieves the active channel commitment type for the given
|
|
|
|
// chan point.
|
|
|
|
func channelCommitType(node *lntest.HarnessNode,
|
|
|
|
chanPoint *lnrpc.ChannelPoint) (commitType, error) {
|
|
|
|
|
|
|
|
ctxb := context.Background()
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
|
|
|
|
req := &lnrpc.ListChannelsRequest{}
|
|
|
|
channels, err := node.ListChannels(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("listchannels failed: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, c := range channels.Channels {
|
|
|
|
if c.ChannelPoint == txStr(chanPoint) {
|
|
|
|
switch c.CommitmentType {
|
|
|
|
|
|
|
|
// If the anchor output size is non-zero, we are
|
|
|
|
// dealing with the anchor type.
|
|
|
|
case lnrpc.CommitmentType_ANCHORS:
|
|
|
|
return commitTypeAnchors, nil
|
|
|
|
|
|
|
|
// StaticRemoteKey means it is tweakless,
|
|
|
|
case lnrpc.CommitmentType_STATIC_REMOTE_KEY:
|
|
|
|
return commitTypeTweakless, nil
|
|
|
|
|
|
|
|
// Otherwise legacy.
|
|
|
|
default:
|
|
|
|
return commitTypeLegacy, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0, fmt.Errorf("channel point %v not found", chanPoint)
|
|
|
|
}
|
|
|
|
|
2019-08-20 03:55:30 +03:00
|
|
|
// calculateMaxHtlc re-implements the RequiredRemoteChannelReserve of the
|
|
|
|
// funding manager's config, which corresponds to the maximum MaxHTLC value we
|
|
|
|
// allow users to set when updating a channel policy.
|
|
|
|
func calculateMaxHtlc(chanCap btcutil.Amount) uint64 {
|
|
|
|
reserve := lnwire.NewMSatFromSatoshis(chanCap / 100)
|
|
|
|
max := lnwire.NewMSatFromSatoshis(chanCap) - reserve
|
|
|
|
return uint64(max)
|
|
|
|
}
|
|
|
|
|
2017-12-19 19:10:00 +03:00
|
|
|
// testUpdateChannelPolicy tests that policy updates made to a channel
|
|
|
|
// gets propagated to other nodes in the network.
|
|
|
|
func testUpdateChannelPolicy(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
2018-08-22 10:32:43 +03:00
|
|
|
const (
|
|
|
|
defaultFeeBase = 1000
|
|
|
|
defaultFeeRate = 1
|
2020-10-06 21:56:10 +03:00
|
|
|
defaultTimeLockDelta = chainreg.DefaultBitcoinTimeLockDelta
|
2018-08-22 10:32:43 +03:00
|
|
|
defaultMinHtlc = 1000
|
|
|
|
)
|
2020-11-17 02:28:56 +03:00
|
|
|
defaultMaxHtlc := calculateMaxHtlc(funding.MaxBtcFundingAmount)
|
2018-08-22 10:32:43 +03:00
|
|
|
|
2017-12-19 19:10:00 +03:00
|
|
|
// Launch notification clients for all nodes, such that we can
|
2018-04-04 06:18:42 +03:00
|
|
|
// get notified when they discover new channels and updates in the
|
|
|
|
// graph.
|
2021-04-16 11:55:51 +03:00
|
|
|
aliceSub := subscribeGraphNotifications(ctxb, t, net.Alice)
|
2018-08-23 08:16:16 +03:00
|
|
|
defer close(aliceSub.quit)
|
2021-04-16 11:55:51 +03:00
|
|
|
bobSub := subscribeGraphNotifications(ctxb, t, net.Bob)
|
2018-08-23 08:16:16 +03:00
|
|
|
defer close(bobSub.quit)
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2020-11-17 02:28:56 +03:00
|
|
|
chanAmt := funding.MaxBtcFundingAmount
|
2018-08-22 10:32:43 +03:00
|
|
|
pushAmt := chanAmt / 2
|
2017-12-19 19:10:00 +03:00
|
|
|
|
|
|
|
// Create a channel Alice->Bob.
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoint := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, net.Bob,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
PushAmt: pushAmt,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-08-22 10:32:42 +03:00
|
|
|
// We add all the nodes' update channels to a slice, such that we can
|
|
|
|
// make sure they all receive the expected updates.
|
2018-08-23 08:16:16 +03:00
|
|
|
graphSubs := []graphSubscription{aliceSub, bobSub}
|
2018-08-22 10:32:42 +03:00
|
|
|
nodes := []*lntest.HarnessNode{net.Alice, net.Bob}
|
|
|
|
|
2018-08-22 10:32:43 +03:00
|
|
|
// Alice and Bob should see each other's ChannelUpdates, advertising the
|
|
|
|
// default routing policies.
|
|
|
|
expectedPolicy := &lnrpc.RoutingPolicy{
|
|
|
|
FeeBaseMsat: defaultFeeBase,
|
|
|
|
FeeRateMilliMsat: defaultFeeRate,
|
|
|
|
TimeLockDelta: defaultTimeLockDelta,
|
|
|
|
MinHtlc: defaultMinHtlc,
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: defaultMaxHtlc,
|
2018-08-22 10:32:43 +03:00
|
|
|
}
|
|
|
|
|
2018-08-23 08:16:16 +03:00
|
|
|
for _, graphSub := range graphSubs {
|
2018-08-22 10:32:43 +03:00
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, graphSub,
|
2018-08-22 10:32:43 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{net.Alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
{net.Bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
// They should now know about the default policies.
|
|
|
|
for _, node := range nodes {
|
|
|
|
assertChannelPolicy(
|
|
|
|
t, node, net.Alice.PubKeyStr, expectedPolicy, chanPoint,
|
|
|
|
)
|
|
|
|
assertChannelPolicy(
|
|
|
|
t, node, net.Bob.PubKeyStr, expectedPolicy, chanPoint,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-12-19 19:10:00 +03:00
|
|
|
err := net.Alice.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't report channel: %v", err)
|
|
|
|
}
|
|
|
|
err = net.Bob.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't report channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2021-03-04 02:15:02 +03:00
|
|
|
// Create Carol with options to rate limit channel updates up to 2 per
|
|
|
|
// day, and create a new channel Bob->Carol.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(
|
|
|
|
t.t, "Carol", []string{
|
2021-03-04 02:15:02 +03:00
|
|
|
"--gossip.max-channel-update-burst=2",
|
|
|
|
"--gossip.channel-update-interval=24h",
|
|
|
|
},
|
|
|
|
)
|
2018-06-08 00:14:39 +03:00
|
|
|
|
|
|
|
// Clean up carol's node when the test finishes.
|
|
|
|
defer shutdownAndAssert(net, t, carol)
|
|
|
|
|
2021-04-16 11:55:51 +03:00
|
|
|
carolSub := subscribeGraphNotifications(ctxb, t, carol)
|
2018-08-23 08:16:16 +03:00
|
|
|
defer close(carolSub.quit)
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-08-23 08:16:16 +03:00
|
|
|
graphSubs = append(graphSubs, carolSub)
|
2018-08-22 10:32:42 +03:00
|
|
|
nodes = append(nodes, carol)
|
|
|
|
|
2018-08-22 10:32:43 +03:00
|
|
|
// Send some coins to Carol that can be used for channel funding.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, carol)
|
2018-08-22 10:32:43 +03:00
|
|
|
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxb, t.t, carol, net.Bob)
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-08-22 10:32:43 +03:00
|
|
|
// Open the channel Carol->Bob with a custom min_htlc value set. Since
|
|
|
|
// Carol is opening the channel, she will require Bob to not forward
|
|
|
|
// HTLCs smaller than this value, and hence he should advertise it as
|
|
|
|
// part of his ChannelUpdate.
|
|
|
|
const customMinHtlc = 5000
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoint2 := openChannelAndAssert(
|
2018-08-22 10:32:43 +03:00
|
|
|
ctxt, t, net, carol, net.Bob,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
PushAmt: pushAmt,
|
2018-08-22 10:32:43 +03:00
|
|
|
MinHtlc: customMinHtlc,
|
2018-08-22 10:32:42 +03:00
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-08-22 10:32:43 +03:00
|
|
|
expectedPolicyBob := &lnrpc.RoutingPolicy{
|
|
|
|
FeeBaseMsat: defaultFeeBase,
|
|
|
|
FeeRateMilliMsat: defaultFeeRate,
|
|
|
|
TimeLockDelta: defaultTimeLockDelta,
|
|
|
|
MinHtlc: customMinHtlc,
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: defaultMaxHtlc,
|
2018-08-22 10:32:43 +03:00
|
|
|
}
|
|
|
|
expectedPolicyCarol := &lnrpc.RoutingPolicy{
|
|
|
|
FeeBaseMsat: defaultFeeBase,
|
|
|
|
FeeRateMilliMsat: defaultFeeRate,
|
|
|
|
TimeLockDelta: defaultTimeLockDelta,
|
|
|
|
MinHtlc: defaultMinHtlc,
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: defaultMaxHtlc,
|
2018-08-22 10:32:43 +03:00
|
|
|
}
|
|
|
|
|
2018-08-23 08:16:16 +03:00
|
|
|
for _, graphSub := range graphSubs {
|
2018-08-22 10:32:43 +03:00
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, graphSub,
|
2018-08-22 10:32:43 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{net.Bob.PubKeyStr, expectedPolicyBob, chanPoint2},
|
|
|
|
{carol.PubKeyStr, expectedPolicyCarol, chanPoint2},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check that all nodes now know about the updated policies.
|
|
|
|
for _, node := range nodes {
|
|
|
|
assertChannelPolicy(
|
|
|
|
t, node, net.Bob.PubKeyStr, expectedPolicyBob,
|
|
|
|
chanPoint2,
|
|
|
|
)
|
|
|
|
assertChannelPolicy(
|
|
|
|
t, node, carol.PubKeyStr, expectedPolicyCarol,
|
|
|
|
chanPoint2,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-04-14 01:35:54 +03:00
|
|
|
err = net.Alice.WaitForNetworkChannelOpen(ctxt, chanPoint2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't report channel: %v", err)
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-12-19 19:10:00 +03:00
|
|
|
err = net.Bob.WaitForNetworkChannelOpen(ctxt, chanPoint2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't report channel: %v", err)
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-12-19 19:10:00 +03:00
|
|
|
err = carol.WaitForNetworkChannelOpen(ctxt, chanPoint2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("carol didn't report channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-08-22 10:32:44 +03:00
|
|
|
// First we'll try to send a payment from Alice to Carol with an amount
|
|
|
|
// less than the min_htlc value required by Carol. This payment should
|
|
|
|
// fail, as the channel Bob->Carol cannot carry HTLCs this small.
|
|
|
|
payAmt := btcutil.Amount(4)
|
|
|
|
invoice := &lnrpc.Invoice{
|
|
|
|
Memo: "testing",
|
|
|
|
Value: int64(payAmt),
|
|
|
|
}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
resp, err := carol.AddInvoice(ctxt, invoice)
|
2018-08-22 10:32:44 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to add invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-08-22 10:32:44 +03:00
|
|
|
err = completePaymentRequests(
|
2020-03-31 13:44:18 +03:00
|
|
|
ctxt, net.Alice, net.Alice.RouterClient,
|
|
|
|
[]string{resp.PaymentRequest}, true,
|
2018-08-22 10:32:44 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// Alice knows about the channel policy of Carol and should therefore
|
|
|
|
// not be able to find a path during routing.
|
2020-03-31 13:44:18 +03:00
|
|
|
expErr := lnrpc.PaymentFailureReason_FAILURE_REASON_NO_ROUTE
|
|
|
|
if err.Error() != expErr.String() {
|
|
|
|
t.Fatalf("expected %v, instead got %v", expErr, err)
|
2018-08-22 10:32:44 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Now we try to send a payment over the channel with a value too low
|
|
|
|
// to be accepted. First we query for a route to route a payment of
|
|
|
|
// 5000 mSAT, as this is accepted.
|
|
|
|
payAmt = btcutil.Amount(5)
|
|
|
|
routesReq := &lnrpc.QueryRoutesRequest{
|
|
|
|
PubKey: carol.PubKeyStr,
|
|
|
|
Amt: int64(payAmt),
|
2019-03-12 04:43:37 +03:00
|
|
|
FinalCltvDelta: defaultTimeLockDelta,
|
2018-08-22 10:32:44 +03:00
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-08-22 10:32:44 +03:00
|
|
|
routes, err := net.Alice.QueryRoutes(ctxt, routesReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get route: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(routes.Routes) != 1 {
|
|
|
|
t.Fatalf("expected to find 1 route, got %v", len(routes.Routes))
|
|
|
|
}
|
|
|
|
|
|
|
|
// We change the route to carry a payment of 4000 mSAT instead of 5000
|
|
|
|
// mSAT.
|
|
|
|
payAmt = btcutil.Amount(4)
|
|
|
|
amtSat := int64(payAmt)
|
|
|
|
amtMSat := int64(lnwire.NewMSatFromSatoshis(payAmt))
|
|
|
|
routes.Routes[0].Hops[0].AmtToForward = amtSat
|
|
|
|
routes.Routes[0].Hops[0].AmtToForwardMsat = amtMSat
|
|
|
|
routes.Routes[0].Hops[1].AmtToForward = amtSat
|
|
|
|
routes.Routes[0].Hops[1].AmtToForwardMsat = amtMSat
|
|
|
|
|
|
|
|
// Send the payment with the modified value.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-08-22 10:32:44 +03:00
|
|
|
alicePayStream, err := net.Alice.SendToRoute(ctxt)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create payment stream for alice: %v", err)
|
|
|
|
}
|
|
|
|
sendReq := &lnrpc.SendToRouteRequest{
|
|
|
|
PaymentHash: resp.RHash,
|
2018-08-08 12:09:30 +03:00
|
|
|
Route: routes.Routes[0],
|
2018-08-22 10:32:44 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
err = alicePayStream.Send(sendReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We expect this payment to fail, and that the min_htlc value is
|
|
|
|
// communicated back to us, since the attempted HTLC value was too low.
|
|
|
|
sendResp, err := alicePayStream.Recv()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Expected as part of the error message.
|
|
|
|
substrs := []string{
|
|
|
|
"AmountBelowMinimum",
|
|
|
|
"HtlcMinimumMsat: (lnwire.MilliSatoshi) 5000 mSAT",
|
|
|
|
}
|
|
|
|
for _, s := range substrs {
|
|
|
|
if !strings.Contains(sendResp.PaymentError, s) {
|
|
|
|
t.Fatalf("expected error to contain \"%v\", instead "+
|
2018-08-30 04:00:54 +03:00
|
|
|
"got %v", s, sendResp.PaymentError)
|
2018-08-22 10:32:44 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure sending using the original value succeeds.
|
|
|
|
payAmt = btcutil.Amount(5)
|
|
|
|
amtSat = int64(payAmt)
|
|
|
|
amtMSat = int64(lnwire.NewMSatFromSatoshis(payAmt))
|
|
|
|
routes.Routes[0].Hops[0].AmtToForward = amtSat
|
|
|
|
routes.Routes[0].Hops[0].AmtToForwardMsat = amtMSat
|
|
|
|
routes.Routes[0].Hops[1].AmtToForward = amtSat
|
|
|
|
routes.Routes[0].Hops[1].AmtToForwardMsat = amtMSat
|
|
|
|
|
2020-11-25 07:31:11 +03:00
|
|
|
// Manually set the MPP payload a new for each payment since
|
|
|
|
// the payment addr will change with each invoice, although we
|
|
|
|
// can re-use the route itself.
|
|
|
|
route := routes.Routes[0]
|
|
|
|
route.Hops[len(route.Hops)-1].TlvPayload = true
|
|
|
|
route.Hops[len(route.Hops)-1].MppRecord = &lnrpc.MPPRecord{
|
|
|
|
PaymentAddr: resp.PaymentAddr,
|
|
|
|
TotalAmtMsat: amtMSat,
|
|
|
|
}
|
|
|
|
|
2018-08-22 10:32:44 +03:00
|
|
|
sendReq = &lnrpc.SendToRouteRequest{
|
|
|
|
PaymentHash: resp.RHash,
|
2020-11-25 07:31:11 +03:00
|
|
|
Route: route,
|
2018-08-22 10:32:44 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
err = alicePayStream.Send(sendReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
sendResp, err = alicePayStream.Recv()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if sendResp.PaymentError != "" {
|
|
|
|
t.Fatalf("expected payment to succeed, instead got %v",
|
|
|
|
sendResp.PaymentError)
|
|
|
|
}
|
|
|
|
|
2019-08-20 03:55:30 +03:00
|
|
|
// With our little cluster set up, we'll update the fees and the max htlc
|
|
|
|
// size for the Bob side of the Alice->Bob channel, and make sure
|
|
|
|
// all nodes learn about it.
|
2017-12-19 19:10:00 +03:00
|
|
|
baseFee := int64(1500)
|
|
|
|
feeRate := int64(12)
|
|
|
|
timeLockDelta := uint32(66)
|
2019-08-20 03:55:30 +03:00
|
|
|
maxHtlc := uint64(500000)
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-08-22 10:32:43 +03:00
|
|
|
expectedPolicy = &lnrpc.RoutingPolicy{
|
2018-04-30 23:55:31 +03:00
|
|
|
FeeBaseMsat: baseFee,
|
2018-06-16 15:53:44 +03:00
|
|
|
FeeRateMilliMsat: testFeeBase * feeRate,
|
2018-04-30 23:55:31 +03:00
|
|
|
TimeLockDelta: timeLockDelta,
|
2018-08-22 10:32:43 +03:00
|
|
|
MinHtlc: defaultMinHtlc,
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: maxHtlc,
|
2018-04-30 23:55:31 +03:00
|
|
|
}
|
|
|
|
|
2017-12-19 19:10:00 +03:00
|
|
|
req := &lnrpc.PolicyUpdateRequest{
|
|
|
|
BaseFeeMsat: baseFee,
|
|
|
|
FeeRate: float64(feeRate),
|
|
|
|
TimeLockDelta: timeLockDelta,
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: maxHtlc,
|
2018-04-30 23:55:31 +03:00
|
|
|
Scope: &lnrpc.PolicyUpdateRequest_ChanPoint{
|
|
|
|
ChanPoint: chanPoint,
|
|
|
|
},
|
2017-12-19 19:10:00 +03:00
|
|
|
}
|
|
|
|
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
if _, err := net.Bob.UpdateChannelPolicy(ctxt, req); err != nil {
|
2017-12-19 19:10:00 +03:00
|
|
|
t.Fatalf("unable to get alice's balance: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-04-04 06:18:42 +03:00
|
|
|
// Wait for all nodes to have seen the policy update done by Bob.
|
2018-08-23 08:16:16 +03:00
|
|
|
for _, graphSub := range graphSubs {
|
2018-08-22 10:32:42 +03:00
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, graphSub,
|
2018-08-22 10:32:42 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{net.Bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-04-04 06:18:42 +03:00
|
|
|
// Check that all nodes now know about Bob's updated policy.
|
2018-08-22 10:32:42 +03:00
|
|
|
for _, node := range nodes {
|
|
|
|
assertChannelPolicy(
|
|
|
|
t, node, net.Bob.PubKeyStr, expectedPolicy, chanPoint,
|
|
|
|
)
|
|
|
|
}
|
2018-04-04 06:18:42 +03:00
|
|
|
|
|
|
|
// Now that all nodes have received the new channel update, we'll try
|
|
|
|
// to send a payment from Alice to Carol to ensure that Alice has
|
|
|
|
// internalized this fee update. This shouldn't affect the route that
|
|
|
|
// Alice takes though: we updated the Alice -> Bob channel and she
|
|
|
|
// doesn't pay for transit over that channel as it's direct.
|
2018-08-22 10:32:43 +03:00
|
|
|
// Note that the payment amount is >= the min_htlc value for the
|
|
|
|
// channel Bob->Carol, so it should successfully be forwarded.
|
2018-08-22 10:32:44 +03:00
|
|
|
payAmt = btcutil.Amount(5)
|
|
|
|
invoice = &lnrpc.Invoice{
|
2018-04-04 06:18:42 +03:00
|
|
|
Memo: "testing",
|
|
|
|
Value: int64(payAmt),
|
|
|
|
}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
resp, err = carol.AddInvoice(ctxt, invoice)
|
2018-04-04 06:18:42 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to add invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-04-04 06:18:42 +03:00
|
|
|
err = completePaymentRequests(
|
2020-03-31 13:44:18 +03:00
|
|
|
ctxt, net.Alice, net.Alice.RouterClient,
|
|
|
|
[]string{resp.PaymentRequest}, true,
|
2018-04-04 06:18:42 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payment: %v", err)
|
|
|
|
}
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-04-04 06:18:42 +03:00
|
|
|
// We'll now open a channel from Alice directly to Carol.
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxb, t.t, net.Alice, carol)
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-04-04 06:18:42 +03:00
|
|
|
chanPoint3 := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, carol,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
PushAmt: pushAmt,
|
|
|
|
},
|
2018-04-04 06:18:42 +03:00
|
|
|
)
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-12-19 19:10:00 +03:00
|
|
|
err = net.Alice.WaitForNetworkChannelOpen(ctxt, chanPoint3)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't report channel: %v", err)
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-12-19 19:10:00 +03:00
|
|
|
err = carol.WaitForNetworkChannelOpen(ctxt, chanPoint3)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't report channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-04-04 06:18:42 +03:00
|
|
|
// Make a global update, and check that both channels' new policies get
|
|
|
|
// propagated.
|
2017-12-19 19:10:00 +03:00
|
|
|
baseFee = int64(800)
|
|
|
|
feeRate = int64(123)
|
|
|
|
timeLockDelta = uint32(22)
|
2020-01-10 17:27:50 +03:00
|
|
|
maxHtlc *= 2
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-04-30 23:55:31 +03:00
|
|
|
expectedPolicy.FeeBaseMsat = baseFee
|
2018-06-16 15:53:44 +03:00
|
|
|
expectedPolicy.FeeRateMilliMsat = testFeeBase * feeRate
|
2018-04-30 23:55:31 +03:00
|
|
|
expectedPolicy.TimeLockDelta = timeLockDelta
|
2019-08-20 03:55:30 +03:00
|
|
|
expectedPolicy.MaxHtlcMsat = maxHtlc
|
2018-04-30 23:55:31 +03:00
|
|
|
|
2017-12-19 19:10:00 +03:00
|
|
|
req = &lnrpc.PolicyUpdateRequest{
|
|
|
|
BaseFeeMsat: baseFee,
|
|
|
|
FeeRate: float64(feeRate),
|
|
|
|
TimeLockDelta: timeLockDelta,
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: maxHtlc,
|
2017-12-19 19:10:00 +03:00
|
|
|
}
|
|
|
|
req.Scope = &lnrpc.PolicyUpdateRequest_Global{}
|
|
|
|
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
_, err = net.Alice.UpdateChannelPolicy(ctxt, req)
|
2017-12-19 19:10:00 +03:00
|
|
|
if err != nil {
|
2019-08-20 03:55:30 +03:00
|
|
|
t.Fatalf("unable to update alice's channel policy: %v", err)
|
2017-12-19 19:10:00 +03:00
|
|
|
}
|
|
|
|
|
2018-04-04 06:18:42 +03:00
|
|
|
// Wait for all nodes to have seen the policy updates for both of
|
|
|
|
// Alice's channels.
|
2018-08-23 08:16:16 +03:00
|
|
|
for _, graphSub := range graphSubs {
|
2018-08-22 10:32:42 +03:00
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, graphSub,
|
2018-08-22 10:32:42 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{net.Alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
{net.Alice.PubKeyStr, expectedPolicy, chanPoint3},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2018-04-04 06:18:42 +03:00
|
|
|
// And finally check that all nodes remembers the policy update they
|
|
|
|
// received.
|
2018-08-22 10:32:42 +03:00
|
|
|
for _, node := range nodes {
|
|
|
|
assertChannelPolicy(
|
|
|
|
t, node, net.Alice.PubKeyStr, expectedPolicy,
|
|
|
|
chanPoint, chanPoint3,
|
|
|
|
)
|
|
|
|
}
|
2017-12-19 19:10:00 +03:00
|
|
|
|
2021-03-04 02:15:02 +03:00
|
|
|
// Now, to test that Carol is properly rate limiting incoming updates,
|
|
|
|
// we'll send two more update from Alice. Carol should accept the first,
|
|
|
|
// but not the second, as she only allows two updates per day and a day
|
|
|
|
// has yet to elapse from the previous update.
|
|
|
|
const numUpdatesTilRateLimit = 2
|
|
|
|
for i := 0; i < numUpdatesTilRateLimit; i++ {
|
|
|
|
prevAlicePolicy := *expectedPolicy
|
|
|
|
baseFee *= 2
|
|
|
|
expectedPolicy.FeeBaseMsat = baseFee
|
|
|
|
req.BaseFeeMsat = baseFee
|
|
|
|
|
|
|
|
ctxt, cancel := context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
defer cancel()
|
|
|
|
_, err = net.Alice.UpdateChannelPolicy(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to update alice's channel policy: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for all nodes to have seen the policy updates for both
|
|
|
|
// of Alice's channels. Carol will not see the last update as
|
|
|
|
// the limit has been reached.
|
|
|
|
for idx, graphSub := range graphSubs {
|
|
|
|
expUpdates := []expectedChanUpdate{
|
|
|
|
{net.Alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
{net.Alice.PubKeyStr, expectedPolicy, chanPoint3},
|
|
|
|
}
|
|
|
|
// Carol was added last, which is why we check the last
|
|
|
|
// index.
|
|
|
|
if i == numUpdatesTilRateLimit-1 && idx == len(graphSubs)-1 {
|
|
|
|
expUpdates = nil
|
|
|
|
}
|
|
|
|
waitForChannelUpdate(t, graphSub, expUpdates)
|
|
|
|
}
|
|
|
|
|
|
|
|
// And finally check that all nodes remembers the policy update
|
|
|
|
// they received. Since Carol didn't receive the last update,
|
|
|
|
// she still has Alice's old policy.
|
|
|
|
for idx, node := range nodes {
|
|
|
|
policy := expectedPolicy
|
|
|
|
// Carol was added last, which is why we check the last
|
|
|
|
// index.
|
|
|
|
if i == numUpdatesTilRateLimit-1 && idx == len(nodes)-1 {
|
|
|
|
policy = &prevAlicePolicy
|
|
|
|
}
|
|
|
|
assertChannelPolicy(
|
|
|
|
t, node, net.Alice.PubKeyStr, policy, chanPoint,
|
|
|
|
chanPoint3,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-12-19 19:10:00 +03:00
|
|
|
// Close the channels.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2017-12-19 19:10:00 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false)
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2017-12-19 19:10:00 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Bob, chanPoint2, false)
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-08-09 22:37:30 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint3, false)
|
2017-12-19 19:10:00 +03:00
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// waitForNodeBlockHeight queries the node for its current block height until
|
|
|
|
// it reaches the passed height.
|
|
|
|
func waitForNodeBlockHeight(ctx context.Context, node *lntest.HarnessNode,
|
|
|
|
height int32) error {
|
|
|
|
var predErr error
|
2019-09-19 22:46:29 +03:00
|
|
|
err := wait.Predicate(func() bool {
|
2020-12-08 18:27:01 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctx, defaultTimeout)
|
2018-12-14 11:24:00 +03:00
|
|
|
info, err := node.GetInfo(ctxt, &lnrpc.GetInfoRequest{})
|
|
|
|
if err != nil {
|
|
|
|
predErr = err
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
if int32(info.BlockHeight) != height {
|
|
|
|
predErr = fmt.Errorf("expected block height to "+
|
|
|
|
"be %v, was %v", height, info.BlockHeight)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return true
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2018-12-14 11:24:00 +03:00
|
|
|
if err != nil {
|
|
|
|
return predErr
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-10-02 18:56:03 +03:00
|
|
|
// testOpenChannelAfterReorg tests that in the case where we have an open
|
|
|
|
// channel where the funding tx gets reorged out, the channel will no
|
|
|
|
// longer be present in the node's routing table.
|
2017-11-03 21:52:02 +03:00
|
|
|
func testOpenChannelAfterReorg(net *lntest.NetworkHarness, t *harnessTest) {
|
2019-05-24 15:17:49 +03:00
|
|
|
// Skip test for neutrino, as we cannot disconnect the miner at will.
|
|
|
|
// TODO(halseth): remove when either can disconnect at will, or restart
|
|
|
|
// node with connection to new miner.
|
2021-02-20 04:42:20 +03:00
|
|
|
if net.BackendCfg.Name() == lntest.NeutrinoBackendName {
|
2019-05-24 15:17:49 +03:00
|
|
|
t.Skipf("skipping reorg test for neutrino backend")
|
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
var (
|
|
|
|
ctxb = context.Background()
|
|
|
|
temp = "temp"
|
|
|
|
)
|
2017-10-02 18:56:03 +03:00
|
|
|
|
|
|
|
// Set up a new miner that we can use to cause a reorg.
|
2020-11-04 13:03:30 +03:00
|
|
|
tempLogDir := fmt.Sprintf("%s/.tempminerlogs", lntest.GetLogDir())
|
2020-09-30 01:15:05 +03:00
|
|
|
logFilename := "output-open_channel_reorg-temp_miner.log"
|
|
|
|
tempMiner, tempMinerCleanUp, err := lntest.NewMiner(
|
2020-12-03 13:30:22 +03:00
|
|
|
tempLogDir, logFilename, harnessNetParams,
|
|
|
|
&rpcclient.NotificationHandlers{}, lntest.GetBtcdBinary(),
|
2020-07-31 11:32:30 +03:00
|
|
|
)
|
2020-09-30 01:15:05 +03:00
|
|
|
require.NoError(t.t, err, "failed to create temp miner")
|
2020-09-02 21:38:54 +03:00
|
|
|
defer func() {
|
|
|
|
require.NoError(
|
2020-09-30 01:15:05 +03:00
|
|
|
t.t, tempMinerCleanUp(),
|
|
|
|
"failed to clean up temp miner",
|
2020-09-02 21:38:54 +03:00
|
|
|
)
|
|
|
|
}()
|
2017-10-02 18:56:03 +03:00
|
|
|
|
2020-09-30 01:15:05 +03:00
|
|
|
// Setup the temp miner
|
|
|
|
require.NoError(
|
|
|
|
t.t, tempMiner.SetUp(false, 0), "unable to set up mining node",
|
|
|
|
)
|
|
|
|
|
2017-10-02 18:56:03 +03:00
|
|
|
// We start by connecting the new miner to our original miner,
|
|
|
|
// such that it will sync to our original chain.
|
2021-03-10 01:12:26 +03:00
|
|
|
err = net.Miner.Client.Node(
|
2018-12-14 11:24:00 +03:00
|
|
|
btcjson.NConnect, tempMiner.P2PAddress(), &temp,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to remove node: %v", err)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
2018-12-14 11:24:00 +03:00
|
|
|
nodeSlice := []*rpctest.Harness{net.Miner, tempMiner}
|
2017-10-02 18:56:03 +03:00
|
|
|
if err := rpctest.JoinNodes(nodeSlice, rpctest.Blocks); err != nil {
|
|
|
|
t.Fatalf("unable to join node on blocks: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// The two miners should be on the same blockheight.
|
|
|
|
assertMinerBlockHeightDelta(t, net.Miner, tempMiner, 0)
|
2017-10-02 18:56:03 +03:00
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// We disconnect the two miners, such that we can mine two different
|
|
|
|
// chains and can cause a reorg later.
|
2021-03-10 01:12:26 +03:00
|
|
|
err = net.Miner.Client.Node(
|
2018-12-14 11:24:00 +03:00
|
|
|
btcjson.NDisconnect, tempMiner.P2PAddress(), &temp,
|
|
|
|
)
|
2017-10-02 18:56:03 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to remove node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a new channel that requires 1 confs before it's considered
|
|
|
|
// open, then broadcast the funding transaction
|
2020-11-17 02:28:56 +03:00
|
|
|
chanAmt := funding.MaxBtcFundingAmount
|
2017-10-02 18:56:03 +03:00
|
|
|
pushAmt := btcutil.Amount(0)
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2017-10-02 18:56:03 +03:00
|
|
|
pendingUpdate, err := net.OpenPendingChannel(ctxt, net.Alice, net.Bob,
|
|
|
|
chanAmt, pushAmt)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to open channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// Wait for miner to have seen the funding tx. The temporary miner is
|
|
|
|
// disconnected, and won't see the transaction.
|
2021-03-10 01:12:26 +03:00
|
|
|
_, err = waitForTxInMempool(net.Miner.Client, minerMempoolTimeout)
|
2018-11-29 14:31:40 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("failed to find funding tx in mempool: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-10-02 18:56:03 +03:00
|
|
|
// At this point, the channel's funding transaction will have been
|
|
|
|
// broadcast, but not confirmed, and the channel should be pending.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-10-02 18:56:03 +03:00
|
|
|
assertNumOpenChannelsPending(ctxt, t, net.Alice, net.Bob, 1)
|
|
|
|
|
|
|
|
fundingTxID, err := chainhash.NewHash(pendingUpdate.Txid)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to convert funding txid into chainhash.Hash:"+
|
|
|
|
" %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We now cause a fork, by letting our original miner mine 10 blocks,
|
|
|
|
// and our new miner mine 15. This will also confirm our pending
|
2018-12-14 11:24:00 +03:00
|
|
|
// channel on the original miner's chain, which should be considered
|
|
|
|
// open.
|
2018-11-29 14:33:11 +03:00
|
|
|
block := mineBlocks(t, net, 10, 1)[0]
|
2017-10-02 18:56:03 +03:00
|
|
|
assertTxInBlock(t, block, fundingTxID)
|
2021-03-10 01:12:26 +03:00
|
|
|
if _, err := tempMiner.Client.Generate(15); err != nil {
|
2018-12-14 11:24:00 +03:00
|
|
|
t.Fatalf("unable to generate blocks: %v", err)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// Ensure the chain lengths are what we expect, with the temp miner
|
|
|
|
// being 5 blocks ahead.
|
|
|
|
assertMinerBlockHeightDelta(t, net.Miner, tempMiner, 5)
|
|
|
|
|
|
|
|
// Wait for Alice to sync to the original miner's chain.
|
2021-03-10 01:12:26 +03:00
|
|
|
_, minerHeight, err := net.Miner.Client.GetBestBlock()
|
2017-10-02 18:56:03 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get current blockheight %v", err)
|
|
|
|
}
|
2018-12-14 11:24:00 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
err = waitForNodeBlockHeight(ctxt, net.Alice, minerHeight)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to sync to chain: %v", err)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
chanPoint := &lnrpc.ChannelPoint{
|
2018-01-11 07:59:30 +03:00
|
|
|
FundingTxid: &lnrpc.ChannelPoint_FundingTxidBytes{
|
|
|
|
FundingTxidBytes: pendingUpdate.Txid,
|
|
|
|
},
|
2017-10-02 18:56:03 +03:00
|
|
|
OutputIndex: pendingUpdate.OutputIndex,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure channel is no longer pending.
|
|
|
|
assertNumOpenChannelsPending(ctxt, t, net.Alice, net.Bob, 0)
|
|
|
|
|
|
|
|
// Wait for Alice and Bob to recognize and advertise the new channel
|
|
|
|
// generated above.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-10-02 18:56:03 +03:00
|
|
|
err = net.Alice.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-10-02 18:56:03 +03:00
|
|
|
err = net.Bob.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Alice should now have 1 edge in her graph.
|
2018-09-26 18:30:09 +03:00
|
|
|
req := &lnrpc.ChannelGraphRequest{
|
|
|
|
IncludeUnannounced: true,
|
|
|
|
}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
chanGraph, err := net.Alice.DescribeGraph(ctxt, req)
|
2017-10-02 18:56:03 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query for alice's routing table: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
numEdges := len(chanGraph.Edges)
|
|
|
|
if numEdges != 1 {
|
|
|
|
t.Fatalf("expected to find one edge in the graph, found %d",
|
|
|
|
numEdges)
|
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// Now we disconnect Alice's chain backend from the original miner, and
|
|
|
|
// connect the two miners together. Since the temporary miner knows
|
|
|
|
// about a longer chain, both miners should sync to that chain.
|
2019-05-24 15:17:48 +03:00
|
|
|
err = net.BackendCfg.DisconnectMiner()
|
2018-12-14 11:24:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to remove node: %v", err)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// Connecting to the temporary miner should now cause our original
|
|
|
|
// chain to be re-orged out.
|
2021-03-10 01:12:26 +03:00
|
|
|
err = net.Miner.Client.Node(
|
2018-12-14 11:24:00 +03:00
|
|
|
btcjson.NConnect, tempMiner.P2PAddress(), &temp,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to remove node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
nodes := []*rpctest.Harness{tempMiner, net.Miner}
|
|
|
|
if err := rpctest.JoinNodes(nodes, rpctest.Blocks); err != nil {
|
2017-10-02 18:56:03 +03:00
|
|
|
t.Fatalf("unable to join node on blocks: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Once again they should be on the same chain.
|
2018-12-14 11:24:00 +03:00
|
|
|
assertMinerBlockHeightDelta(t, net.Miner, tempMiner, 0)
|
|
|
|
|
|
|
|
// Now we disconnect the two miners, and connect our original miner to
|
|
|
|
// our chain backend once again.
|
2021-03-10 01:12:26 +03:00
|
|
|
err = net.Miner.Client.Node(
|
2018-12-14 11:24:00 +03:00
|
|
|
btcjson.NDisconnect, tempMiner.P2PAddress(), &temp,
|
|
|
|
)
|
2017-10-02 18:56:03 +03:00
|
|
|
if err != nil {
|
2018-12-14 11:24:00 +03:00
|
|
|
t.Fatalf("unable to remove node: %v", err)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
|
|
|
|
2019-05-24 15:17:48 +03:00
|
|
|
err = net.BackendCfg.ConnectMiner()
|
2017-10-02 18:56:03 +03:00
|
|
|
if err != nil {
|
2018-12-14 11:24:00 +03:00
|
|
|
t.Fatalf("unable to remove node: %v", err)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// This should have caused a reorg, and Alice should sync to the longer
|
|
|
|
// chain, where the funding transaction is not confirmed.
|
2021-03-10 01:12:26 +03:00
|
|
|
_, tempMinerHeight, err := tempMiner.Client.GetBestBlock()
|
2018-12-14 11:24:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get current blockheight %v", err)
|
|
|
|
}
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
err = waitForNodeBlockHeight(ctxt, net.Alice, tempMinerHeight)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to sync to chain: %v", err)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
2017-11-03 07:30:19 +03:00
|
|
|
|
2017-10-02 18:56:03 +03:00
|
|
|
// Since the fundingtx was reorged out, Alice should now have no edges
|
|
|
|
// in her graph.
|
2018-09-26 18:30:09 +03:00
|
|
|
req = &lnrpc.ChannelGraphRequest{
|
|
|
|
IncludeUnannounced: true,
|
|
|
|
}
|
2017-10-02 18:56:03 +03:00
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
var predErr error
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
2018-12-14 11:24:00 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
chanGraph, err = net.Alice.DescribeGraph(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
predErr = fmt.Errorf("unable to query for alice's routing table: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
numEdges = len(chanGraph.Edges)
|
|
|
|
if numEdges != 0 {
|
|
|
|
predErr = fmt.Errorf("expected to find no edge in the graph, found %d",
|
|
|
|
numEdges)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return true
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2018-12-14 11:24:00 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf(predErr.Error())
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
|
|
|
|
2018-10-29 07:52:09 +03:00
|
|
|
// Cleanup by mining the funding tx again, then closing the channel.
|
2018-11-29 14:33:11 +03:00
|
|
|
block = mineBlocks(t, net, 1, 1)[0]
|
2018-10-29 07:52:09 +03:00
|
|
|
assertTxInBlock(t, block, fundingTxID)
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2019-03-13 03:34:42 +03:00
|
|
|
closeReorgedChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false)
|
2017-10-02 18:56:03 +03:00
|
|
|
}
|
|
|
|
|
2019-11-20 07:12:29 +03:00
|
|
|
// testDisconnectingTargetPeer performs a test which disconnects Alice-peer from
|
2019-11-20 07:13:04 +03:00
|
|
|
// Bob-peer and then re-connects them again. We expect Alice to be able to
|
|
|
|
// disconnect at any point.
|
2017-11-03 21:52:02 +03:00
|
|
|
func testDisconnectingTargetPeer(net *lntest.NetworkHarness, t *harnessTest) {
|
2017-05-05 10:00:39 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2019-11-20 07:13:04 +03:00
|
|
|
// We'll start both nodes with a high backoff so that they don't
|
|
|
|
// reconnect automatically during our test.
|
|
|
|
args := []string{
|
|
|
|
"--minbackoff=1m",
|
|
|
|
"--maxbackoff=1m",
|
|
|
|
}
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
alice := net.NewNode(t.t, "Alice", args)
|
2019-11-20 07:12:29 +03:00
|
|
|
defer shutdownAndAssert(net, t, alice)
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
bob := net.NewNode(t.t, "Bob", args)
|
2019-11-20 07:12:29 +03:00
|
|
|
defer shutdownAndAssert(net, t, bob)
|
|
|
|
|
|
|
|
// Start by connecting Alice and Bob with no channels.
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, alice, bob)
|
2019-11-20 07:12:29 +03:00
|
|
|
|
2017-05-05 10:00:39 +03:00
|
|
|
// Check existing connection.
|
2019-11-20 07:12:29 +03:00
|
|
|
assertNumConnections(t, alice, bob, 1)
|
|
|
|
|
|
|
|
// Give Alice some coins so she can fund a channel.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, alice)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
2020-11-17 02:28:56 +03:00
|
|
|
chanAmt := funding.MaxBtcFundingAmount
|
2017-05-05 10:00:39 +03:00
|
|
|
pushAmt := btcutil.Amount(0)
|
|
|
|
|
|
|
|
// Create a new channel that requires 1 confs before it's considered
|
|
|
|
// open, then broadcast the funding transaction
|
|
|
|
const numConfs = 1
|
2019-11-20 07:12:29 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
|
|
|
pendingUpdate, err := net.OpenPendingChannel(
|
|
|
|
ctxt, alice, bob, chanAmt, pushAmt,
|
|
|
|
)
|
2017-05-05 10:00:39 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to open channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2019-11-20 07:12:29 +03:00
|
|
|
// At this point, the channel's funding transaction will have been
|
|
|
|
// broadcast, but not confirmed. Alice and Bob's nodes should reflect
|
|
|
|
// this when queried via RPC.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2019-11-20 07:12:29 +03:00
|
|
|
assertNumOpenChannelsPending(ctxt, t, alice, bob, 1)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
2019-11-20 07:12:29 +03:00
|
|
|
// Disconnect Alice-peer from Bob-peer and get error causes by one
|
|
|
|
// pending channel with detach node is existing.
|
2019-11-20 07:13:04 +03:00
|
|
|
if err := net.DisconnectNodes(ctxt, alice, bob); err != nil {
|
2017-05-05 10:00:39 +03:00
|
|
|
t.Fatalf("Bob's peer was disconnected from Alice's"+
|
|
|
|
" while one pending channel is existing: err %v", err)
|
|
|
|
}
|
|
|
|
|
2017-07-06 03:48:44 +03:00
|
|
|
time.Sleep(time.Millisecond * 300)
|
|
|
|
|
2019-11-20 07:13:04 +03:00
|
|
|
// Assert that the connection was torn down.
|
|
|
|
assertNumConnections(t, alice, bob, 0)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
|
|
|
fundingTxID, err := chainhash.NewHash(pendingUpdate.Txid)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to convert funding txid into chainhash.Hash:"+
|
|
|
|
" %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Mine a block, then wait for Alice's node to notify us that the
|
|
|
|
// channel has been opened. The funding transaction should be found
|
|
|
|
// within the newly mined block.
|
2018-11-29 14:33:11 +03:00
|
|
|
block := mineBlocks(t, net, numConfs, 1)[0]
|
2017-05-05 10:00:39 +03:00
|
|
|
assertTxInBlock(t, block, fundingTxID)
|
|
|
|
|
2019-11-20 07:12:29 +03:00
|
|
|
// At this point, the channel should be fully opened and there should be
|
|
|
|
// no pending channels remaining for either node.
|
2017-07-12 02:27:11 +03:00
|
|
|
time.Sleep(time.Millisecond * 300)
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
2019-11-20 07:12:29 +03:00
|
|
|
assertNumOpenChannelsPending(ctxt, t, alice, bob, 0)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
2019-11-20 07:13:04 +03:00
|
|
|
// Reconnect the nodes so that the channel can become active.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, alice, bob)
|
2019-11-20 07:13:04 +03:00
|
|
|
|
2019-11-20 07:12:29 +03:00
|
|
|
// The channel should be listed in the peer information returned by both
|
|
|
|
// peers.
|
2017-05-05 10:00:39 +03:00
|
|
|
outPoint := wire.OutPoint{
|
|
|
|
Hash: *fundingTxID,
|
|
|
|
Index: pendingUpdate.OutputIndex,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check both nodes to ensure that the channel is ready for operation.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2019-11-20 07:12:29 +03:00
|
|
|
if err := net.AssertChannelExists(ctxt, alice, &outPoint); err != nil {
|
2017-05-05 10:00:39 +03:00
|
|
|
t.Fatalf("unable to assert channel existence: %v", err)
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2019-11-20 07:12:29 +03:00
|
|
|
if err := net.AssertChannelExists(ctxt, bob, &outPoint); err != nil {
|
2017-05-05 10:00:39 +03:00
|
|
|
t.Fatalf("unable to assert channel existence: %v", err)
|
|
|
|
}
|
|
|
|
|
2019-11-20 07:12:07 +03:00
|
|
|
// Disconnect Alice-peer from Bob-peer and get error causes by one
|
|
|
|
// active channel with detach node is existing.
|
2019-11-20 07:13:04 +03:00
|
|
|
if err := net.DisconnectNodes(ctxt, alice, bob); err != nil {
|
2019-11-20 07:12:07 +03:00
|
|
|
t.Fatalf("Bob's peer was disconnected from Alice's"+
|
|
|
|
" while one active channel is existing: err %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check existing connection.
|
2019-11-20 07:13:04 +03:00
|
|
|
assertNumConnections(t, alice, bob, 0)
|
|
|
|
|
|
|
|
// Reconnect both nodes before force closing the channel.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, alice, bob)
|
2019-11-20 07:12:07 +03:00
|
|
|
|
|
|
|
// Finally, immediately close the channel. This function will also block
|
|
|
|
// until the channel is closed and will additionally assert the relevant
|
|
|
|
// channel closing post conditions.
|
2017-05-05 10:00:39 +03:00
|
|
|
chanPoint := &lnrpc.ChannelPoint{
|
2018-01-11 07:59:30 +03:00
|
|
|
FundingTxid: &lnrpc.ChannelPoint_FundingTxidBytes{
|
|
|
|
FundingTxidBytes: pendingUpdate.Txid,
|
|
|
|
},
|
2017-05-05 10:00:39 +03:00
|
|
|
OutputIndex: pendingUpdate.OutputIndex,
|
|
|
|
}
|
|
|
|
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2019-11-20 07:12:29 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, alice, chanPoint, true)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
2019-11-20 07:12:29 +03:00
|
|
|
// Disconnect Alice-peer from Bob-peer without getting error about
|
|
|
|
// existing channels.
|
2019-11-20 07:13:04 +03:00
|
|
|
if err := net.DisconnectNodes(ctxt, alice, bob); err != nil {
|
2018-04-13 14:20:12 +03:00
|
|
|
t.Fatalf("unable to disconnect Bob's peer from Alice's: err %v",
|
2019-11-20 07:13:04 +03:00
|
|
|
err)
|
2017-05-05 10:00:39 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Check zero peer connections.
|
2019-11-20 07:12:29 +03:00
|
|
|
assertNumConnections(t, alice, bob, 0)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
|
|
|
// Finally, re-connect both nodes.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, alice, bob)
|
2017-05-05 10:00:39 +03:00
|
|
|
|
|
|
|
// Check existing connection.
|
2019-11-20 07:12:29 +03:00
|
|
|
assertNumConnections(t, alice, net.Bob, 1)
|
2017-10-07 02:58:13 +03:00
|
|
|
|
2018-11-06 13:03:01 +03:00
|
|
|
// Cleanup by mining the force close and sweep transaction.
|
2019-11-20 07:12:29 +03:00
|
|
|
cleanupForceClose(t, net, alice, chanPoint)
|
2017-05-05 10:00:39 +03:00
|
|
|
}
|
|
|
|
|
2017-01-31 07:21:52 +03:00
|
|
|
// testFundingPersistence is intended to ensure that the Funding Manager
|
|
|
|
// persists the state of new channels prior to broadcasting the channel's
|
|
|
|
// funding transaction. This ensures that the daemon maintains an up-to-date
|
|
|
|
// representation of channels if the system is restarted or disconnected.
|
|
|
|
// testFundingPersistence mirrors testBasicChannelFunding, but adds restarts
|
|
|
|
// and checks for the state of channels with unconfirmed funding transactions.
|
2017-11-03 21:52:02 +03:00
|
|
|
func testChannelFundingPersistence(net *lntest.NetworkHarness, t *harnessTest) {
|
2017-01-31 07:21:52 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2020-11-17 02:28:56 +03:00
|
|
|
chanAmt := funding.MaxBtcFundingAmount
|
2017-01-31 07:21:52 +03:00
|
|
|
pushAmt := btcutil.Amount(0)
|
|
|
|
|
2017-07-31 06:26:30 +03:00
|
|
|
// As we need to create a channel that requires more than 1
|
|
|
|
// confirmation before it's open, with the current set of defaults,
|
|
|
|
// we'll need to create a new node instance.
|
|
|
|
const numConfs = 5
|
2017-12-19 19:10:00 +03:00
|
|
|
carolArgs := []string{fmt.Sprintf("--bitcoin.defaultchanconfs=%v", numConfs)}
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", carolArgs)
|
2018-06-08 00:14:39 +03:00
|
|
|
|
|
|
|
// Clean up carol's node when the test finishes.
|
|
|
|
defer shutdownAndAssert(net, t, carol)
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, net.Alice, carol)
|
2017-07-31 06:26:30 +03:00
|
|
|
|
2017-02-25 03:30:58 +03:00
|
|
|
// Create a new channel that requires 5 confs before it's considered
|
|
|
|
// open, then broadcast the funding transaction
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2017-07-31 06:26:30 +03:00
|
|
|
pendingUpdate, err := net.OpenPendingChannel(ctxt, net.Alice, carol,
|
|
|
|
chanAmt, pushAmt)
|
2017-01-31 07:21:52 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to open channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-07-31 06:26:30 +03:00
|
|
|
// At this point, the channel's funding transaction will have been
|
|
|
|
// broadcast, but not confirmed. Alice and Bob's nodes should reflect
|
|
|
|
// this when queried via RPC.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-07-31 06:26:30 +03:00
|
|
|
assertNumOpenChannelsPending(ctxt, t, net.Alice, carol, 1)
|
2017-01-31 07:21:52 +03:00
|
|
|
|
|
|
|
// Restart both nodes to test that the appropriate state has been
|
|
|
|
// persisted and that both nodes recover gracefully.
|
|
|
|
if err := net.RestartNode(net.Alice, nil); err != nil {
|
|
|
|
t.Fatalf("Node restart failed: %v", err)
|
|
|
|
}
|
2017-07-31 06:26:30 +03:00
|
|
|
if err := net.RestartNode(carol, nil); err != nil {
|
2017-01-31 07:21:52 +03:00
|
|
|
t.Fatalf("Node restart failed: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
fundingTxID, err := chainhash.NewHash(pendingUpdate.Txid)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to convert funding txid into chainhash.Hash:"+
|
|
|
|
" %v", err)
|
|
|
|
}
|
2020-07-29 10:27:22 +03:00
|
|
|
fundingTxStr := fundingTxID.String()
|
2017-01-31 07:21:52 +03:00
|
|
|
|
|
|
|
// Mine a block, then wait for Alice's node to notify us that the
|
|
|
|
// channel has been opened. The funding transaction should be found
|
|
|
|
// within the newly mined block.
|
2018-11-29 14:33:11 +03:00
|
|
|
block := mineBlocks(t, net, 1, 1)[0]
|
2017-01-31 07:21:52 +03:00
|
|
|
assertTxInBlock(t, block, fundingTxID)
|
|
|
|
|
2020-07-29 10:27:22 +03:00
|
|
|
// Get the height that our transaction confirmed at.
|
2021-03-10 01:12:26 +03:00
|
|
|
_, height, err := net.Miner.Client.GetBestBlock()
|
2020-07-29 10:27:22 +03:00
|
|
|
require.NoError(t.t, err, "could not get best block")
|
|
|
|
|
2017-01-31 07:21:52 +03:00
|
|
|
// Restart both nodes to test that the appropriate state has been
|
|
|
|
// persisted and that both nodes recover gracefully.
|
|
|
|
if err := net.RestartNode(net.Alice, nil); err != nil {
|
|
|
|
t.Fatalf("Node restart failed: %v", err)
|
|
|
|
}
|
2017-07-31 06:26:30 +03:00
|
|
|
if err := net.RestartNode(carol, nil); err != nil {
|
2017-01-31 07:21:52 +03:00
|
|
|
t.Fatalf("Node restart failed: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// The following block ensures that after both nodes have restarted,
|
|
|
|
// they have reconnected before the execution of the next test.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, net.Alice, carol)
|
2017-01-31 07:21:52 +03:00
|
|
|
|
2017-02-25 03:30:58 +03:00
|
|
|
// Next, mine enough blocks s.t the channel will open with a single
|
|
|
|
// additional block mined.
|
2021-03-10 01:12:26 +03:00
|
|
|
if _, err := net.Miner.Client.Generate(3); err != nil {
|
2017-02-25 03:30:58 +03:00
|
|
|
t.Fatalf("unable to mine blocks: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-07-29 10:27:22 +03:00
|
|
|
// Assert that our wallet has our opening transaction with a label
|
|
|
|
// that does not have a channel ID set yet, because we have not
|
|
|
|
// reached our required confirmations.
|
2020-01-10 17:27:50 +03:00
|
|
|
tx := findTxAtHeight(ctxt, t, height, fundingTxStr, net.Alice)
|
2020-07-29 10:27:22 +03:00
|
|
|
|
|
|
|
// At this stage, we expect the transaction to be labelled, but not with
|
|
|
|
// our channel ID because our transaction has not yet confirmed.
|
|
|
|
label := labels.MakeLabel(labels.LabelTypeChannelOpen, nil)
|
|
|
|
require.Equal(t.t, label, tx.Label, "open channel label wrong")
|
|
|
|
|
2017-02-25 03:30:58 +03:00
|
|
|
// Both nodes should still show a single channel as pending.
|
2017-08-22 10:50:30 +03:00
|
|
|
time.Sleep(time.Second * 1)
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-07-31 06:26:30 +03:00
|
|
|
assertNumOpenChannelsPending(ctxt, t, net.Alice, carol, 1)
|
2017-02-25 03:30:58 +03:00
|
|
|
|
|
|
|
// Finally, mine the last block which should mark the channel as open.
|
2021-03-10 01:12:26 +03:00
|
|
|
if _, err := net.Miner.Client.Generate(1); err != nil {
|
2017-02-25 03:30:58 +03:00
|
|
|
t.Fatalf("unable to mine blocks: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-01-31 07:21:52 +03:00
|
|
|
// At this point, the channel should be fully opened and there should
|
|
|
|
// be no pending channels remaining for either node.
|
2017-08-22 10:50:30 +03:00
|
|
|
time.Sleep(time.Second * 1)
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-07-31 06:26:30 +03:00
|
|
|
assertNumOpenChannelsPending(ctxt, t, net.Alice, carol, 0)
|
2017-01-31 07:21:52 +03:00
|
|
|
|
|
|
|
// The channel should be listed in the peer information returned by
|
|
|
|
// both peers.
|
|
|
|
outPoint := wire.OutPoint{
|
|
|
|
Hash: *fundingTxID,
|
|
|
|
Index: pendingUpdate.OutputIndex,
|
|
|
|
}
|
|
|
|
|
2020-07-29 10:27:22 +03:00
|
|
|
// Re-lookup our transaction in the block that it confirmed in.
|
2020-01-10 17:27:50 +03:00
|
|
|
tx = findTxAtHeight(ctxt, t, height, fundingTxStr, net.Alice)
|
2020-07-29 10:27:22 +03:00
|
|
|
|
|
|
|
// Create an additional check for our channel assertion that will
|
|
|
|
// check that our label is as expected.
|
|
|
|
check := func(channel *lnrpc.Channel) {
|
|
|
|
shortChanID := lnwire.NewShortChanIDFromInt(
|
|
|
|
channel.ChanId,
|
|
|
|
)
|
|
|
|
|
|
|
|
label := labels.MakeLabel(
|
|
|
|
labels.LabelTypeChannelOpen, &shortChanID,
|
|
|
|
)
|
|
|
|
require.Equal(t.t, label, tx.Label,
|
|
|
|
"open channel label not updated")
|
|
|
|
}
|
|
|
|
|
2017-01-31 07:21:52 +03:00
|
|
|
// Check both nodes to ensure that the channel is ready for operation.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2020-07-29 10:27:22 +03:00
|
|
|
err = net.AssertChannelExists(ctxt, net.Alice, &outPoint, check)
|
|
|
|
if err != nil {
|
2017-01-31 07:21:52 +03:00
|
|
|
t.Fatalf("unable to assert channel existence: %v", err)
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-07-31 06:26:30 +03:00
|
|
|
if err := net.AssertChannelExists(ctxt, carol, &outPoint); err != nil {
|
2017-01-31 07:21:52 +03:00
|
|
|
t.Fatalf("unable to assert channel existence: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, immediately close the channel. This function will also
|
|
|
|
// block until the channel is closed and will additionally assert the
|
|
|
|
// relevant channel closing post conditions.
|
|
|
|
chanPoint := &lnrpc.ChannelPoint{
|
2018-01-11 07:59:30 +03:00
|
|
|
FundingTxid: &lnrpc.ChannelPoint_FundingTxidBytes{
|
|
|
|
FundingTxidBytes: pendingUpdate.Txid,
|
|
|
|
},
|
2017-01-31 07:21:52 +03:00
|
|
|
OutputIndex: pendingUpdate.OutputIndex,
|
|
|
|
}
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2017-05-05 04:06:12 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false)
|
2017-01-31 07:21:52 +03:00
|
|
|
}
|
|
|
|
|
2020-07-29 10:27:22 +03:00
|
|
|
// findTxAtHeight gets all of the transactions that a node's wallet has a record
|
|
|
|
// of at the target height, and finds and returns the tx with the target txid,
|
|
|
|
// failing if it is not found.
|
|
|
|
func findTxAtHeight(ctx context.Context, t *harnessTest, height int32,
|
2020-01-10 17:27:50 +03:00
|
|
|
target string, node *lntest.HarnessNode) *lnrpc.Transaction {
|
2020-07-29 10:27:22 +03:00
|
|
|
|
|
|
|
txns, err := node.LightningClient.GetTransactions(
|
|
|
|
ctx, &lnrpc.GetTransactionsRequest{
|
|
|
|
StartHeight: height,
|
|
|
|
EndHeight: height,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
require.NoError(t.t, err, "could not get transactions")
|
|
|
|
|
|
|
|
for _, tx := range txns.Transactions {
|
|
|
|
if tx.TxHash == target {
|
|
|
|
return tx
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-08-06 13:08:44 +03:00
|
|
|
// testChannelBalance creates a new channel between Alice and Bob, then checks
|
|
|
|
// channel balance to be equal amount specified while creation of channel.
|
2017-11-03 21:52:02 +03:00
|
|
|
func testChannelBalance(net *lntest.NetworkHarness, t *harnessTest) {
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxb := context.Background()
|
2016-10-15 16:47:09 +03:00
|
|
|
|
2017-08-08 02:30:29 +03:00
|
|
|
// Open a channel with 0.16 BTC between Alice and Bob, ensuring the
|
2016-10-15 16:47:09 +03:00
|
|
|
// channel has been opened properly.
|
2020-11-17 02:28:56 +03:00
|
|
|
amount := funding.MaxBtcFundingAmount
|
2016-08-31 05:36:33 +03:00
|
|
|
|
2016-09-15 22:24:52 +03:00
|
|
|
// Creates a helper closure to be used below which asserts the proper
|
|
|
|
// response to a channel balance RPC.
|
2020-08-06 13:08:44 +03:00
|
|
|
checkChannelBalance := func(node *lntest.HarnessNode,
|
|
|
|
local, remote btcutil.Amount) {
|
2016-08-31 05:36:33 +03:00
|
|
|
|
2020-08-06 13:08:44 +03:00
|
|
|
expectedResponse := &lnrpc.ChannelBalanceResponse{
|
|
|
|
LocalBalance: &lnrpc.Amount{
|
|
|
|
Sat: uint64(local),
|
|
|
|
Msat: uint64(lnwire.NewMSatFromSatoshis(local)),
|
|
|
|
},
|
|
|
|
RemoteBalance: &lnrpc.Amount{
|
|
|
|
Sat: uint64(remote),
|
|
|
|
Msat: uint64(lnwire.NewMSatFromSatoshis(
|
|
|
|
remote,
|
|
|
|
)),
|
|
|
|
},
|
|
|
|
UnsettledLocalBalance: &lnrpc.Amount{},
|
|
|
|
UnsettledRemoteBalance: &lnrpc.Amount{},
|
|
|
|
PendingOpenLocalBalance: &lnrpc.Amount{},
|
|
|
|
PendingOpenRemoteBalance: &lnrpc.Amount{},
|
|
|
|
// Deprecated fields.
|
|
|
|
Balance: int64(local),
|
2016-09-15 21:59:51 +03:00
|
|
|
}
|
2020-08-06 13:08:44 +03:00
|
|
|
assertChannelBalanceResp(t, node, expectedResponse)
|
2016-08-30 08:07:54 +03:00
|
|
|
}
|
2016-08-31 21:59:08 +03:00
|
|
|
|
2018-03-31 05:59:58 +03:00
|
|
|
// Before beginning, make sure alice and bob are connected.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, net.Alice, net.Bob)
|
2018-03-31 05:59:58 +03:00
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoint := openChannelAndAssert(
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, t, net, net.Alice, net.Bob,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: amount,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2016-09-15 21:59:51 +03:00
|
|
|
|
2017-03-15 06:06:33 +03:00
|
|
|
// Wait for both Alice and Bob to recognize this new channel.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-03-15 06:06:33 +03:00
|
|
|
err := net.Alice.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-03-15 06:06:33 +03:00
|
|
|
err = net.Bob.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-03-04 15:21:28 +03:00
|
|
|
cType, err := channelCommitType(net.Alice, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get channel type: %v", err)
|
|
|
|
}
|
|
|
|
|
2016-09-15 22:24:52 +03:00
|
|
|
// As this is a single funder channel, Alice's balance should be
|
|
|
|
// exactly 0.5 BTC since now state transitions have taken place yet.
|
2020-08-06 13:08:44 +03:00
|
|
|
checkChannelBalance(net.Alice, amount-cType.calcStaticFee(0), 0)
|
2016-09-15 21:59:51 +03:00
|
|
|
|
2016-09-15 22:24:52 +03:00
|
|
|
// Ensure Bob currently has no available balance within the channel.
|
2020-08-06 13:08:44 +03:00
|
|
|
checkChannelBalance(net.Bob, 0, amount-cType.calcStaticFee(0))
|
2016-09-15 21:59:51 +03:00
|
|
|
|
2016-09-15 22:24:52 +03:00
|
|
|
// Finally close the channel between Alice and Bob, asserting that the
|
|
|
|
// channel has been properly closed on-chain.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false)
|
2016-08-30 08:07:54 +03:00
|
|
|
}
|
|
|
|
|
2018-12-07 03:49:05 +03:00
|
|
|
// testChannelUnsettledBalance will test that the UnsettledBalance field
|
|
|
|
// is updated according to the number of Pending Htlcs.
|
|
|
|
// Alice will send Htlcs to Carol while she is in hodl mode. This will result
|
|
|
|
// in a build of pending Htlcs. We expect the channels unsettled balance to
|
|
|
|
// equal the sum of all the Pending Htlcs.
|
|
|
|
func testChannelUnsettledBalance(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
const chanAmt = btcutil.Amount(1000000)
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
2020-08-10 22:47:27 +03:00
|
|
|
// Creates a helper closure to be used below which asserts the proper
|
|
|
|
// response to a channel balance RPC.
|
|
|
|
checkChannelBalance := func(node *lntest.HarnessNode,
|
|
|
|
local, remote, unsettledLocal, unsettledRemote btcutil.Amount) {
|
|
|
|
|
|
|
|
expectedResponse := &lnrpc.ChannelBalanceResponse{
|
|
|
|
LocalBalance: &lnrpc.Amount{
|
|
|
|
Sat: uint64(local),
|
|
|
|
Msat: uint64(lnwire.NewMSatFromSatoshis(
|
|
|
|
local,
|
|
|
|
)),
|
|
|
|
},
|
|
|
|
RemoteBalance: &lnrpc.Amount{
|
|
|
|
Sat: uint64(remote),
|
|
|
|
Msat: uint64(lnwire.NewMSatFromSatoshis(
|
|
|
|
remote,
|
|
|
|
)),
|
|
|
|
},
|
|
|
|
UnsettledLocalBalance: &lnrpc.Amount{
|
|
|
|
Sat: uint64(unsettledLocal),
|
|
|
|
Msat: uint64(lnwire.NewMSatFromSatoshis(
|
|
|
|
unsettledLocal,
|
|
|
|
)),
|
|
|
|
},
|
|
|
|
UnsettledRemoteBalance: &lnrpc.Amount{
|
|
|
|
Sat: uint64(unsettledRemote),
|
|
|
|
Msat: uint64(lnwire.NewMSatFromSatoshis(
|
|
|
|
unsettledRemote,
|
|
|
|
)),
|
|
|
|
},
|
|
|
|
PendingOpenLocalBalance: &lnrpc.Amount{},
|
|
|
|
PendingOpenRemoteBalance: &lnrpc.Amount{},
|
|
|
|
// Deprecated fields.
|
|
|
|
Balance: int64(local),
|
|
|
|
}
|
|
|
|
assertChannelBalanceResp(t, node, expectedResponse)
|
|
|
|
}
|
|
|
|
|
2018-12-07 03:49:05 +03:00
|
|
|
// Create carol in hodl mode.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", []string{"--hodl.exit-settle"})
|
2018-12-07 03:49:05 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
|
|
|
|
|
|
|
// Connect Alice to Carol.
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxb, t.t, net.Alice, carol)
|
2018-12-07 03:49:05 +03:00
|
|
|
|
|
|
|
// Open a channel between Alice and Carol.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
|
|
|
chanPointAlice := openChannelAndAssert(
|
|
|
|
ctxt, t, net, net.Alice, carol,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Wait for Alice and Carol to receive the channel edge from the
|
|
|
|
// funding manager.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-07 23:05:12 +03:00
|
|
|
err := net.Alice.WaitForNetworkChannelOpen(ctxt, chanPointAlice)
|
2018-12-07 03:49:05 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't see the alice->carol channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
err = carol.WaitForNetworkChannelOpen(ctxt, chanPointAlice)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't see the alice->carol channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-08-10 22:47:27 +03:00
|
|
|
cType, err := channelCommitType(net.Alice, chanPointAlice)
|
|
|
|
require.NoError(t.t, err, "unable to get channel type")
|
|
|
|
|
|
|
|
// Check alice's channel balance, which should have zero remote and zero
|
|
|
|
// pending balance.
|
|
|
|
checkChannelBalance(net.Alice, chanAmt-cType.calcStaticFee(0), 0, 0, 0)
|
|
|
|
|
|
|
|
// Check carol's channel balance, which should have zero local and zero
|
|
|
|
// pending balance.
|
|
|
|
checkChannelBalance(carol, 0, chanAmt-cType.calcStaticFee(0), 0, 0)
|
|
|
|
|
2018-12-07 03:49:05 +03:00
|
|
|
// Channel should be ready for payments.
|
|
|
|
const (
|
|
|
|
payAmt = 100
|
|
|
|
numInvoices = 6
|
|
|
|
)
|
|
|
|
|
2020-03-31 13:44:18 +03:00
|
|
|
// Simulateneously send numInvoices payments from Alice to Carol.
|
2018-12-07 03:49:05 +03:00
|
|
|
carolPubKey := carol.PubKey[:]
|
2020-03-31 13:44:18 +03:00
|
|
|
errChan := make(chan error)
|
2018-12-07 03:49:05 +03:00
|
|
|
for i := 0; i < numInvoices; i++ {
|
2020-03-31 13:44:18 +03:00
|
|
|
go func() {
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
_, err := net.Alice.RouterClient.SendPaymentV2(ctxt,
|
|
|
|
&routerrpc.SendPaymentRequest{
|
|
|
|
Dest: carolPubKey,
|
|
|
|
Amt: int64(payAmt),
|
|
|
|
PaymentHash: makeFakePayHash(t),
|
2020-10-06 21:56:10 +03:00
|
|
|
FinalCltvDelta: chainreg.DefaultBitcoinTimeLockDelta,
|
2020-03-31 13:44:18 +03:00
|
|
|
TimeoutSeconds: 60,
|
|
|
|
FeeLimitMsat: noFeeLimitMsat,
|
|
|
|
})
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
errChan <- err
|
|
|
|
}
|
|
|
|
}()
|
2018-12-07 03:49:05 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Test that the UnsettledBalance for both Alice and Carol
|
|
|
|
// is equal to the amount of invoices * payAmt.
|
|
|
|
var unsettledErr error
|
|
|
|
nodes := []*lntest.HarnessNode{net.Alice, carol}
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
2018-12-07 03:49:05 +03:00
|
|
|
// There should be a number of PendingHtlcs equal
|
|
|
|
// to the amount of Invoices sent.
|
|
|
|
unsettledErr = assertNumActiveHtlcs(nodes, numInvoices)
|
|
|
|
if unsettledErr != nil {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set the amount expected for the Unsettled Balance for
|
|
|
|
// this channel.
|
|
|
|
expectedBalance := numInvoices * payAmt
|
|
|
|
|
|
|
|
// Check each nodes UnsettledBalance field.
|
|
|
|
for _, node := range nodes {
|
|
|
|
// Get channel info for the node.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
chanInfo, err := getChanInfo(ctxt, node)
|
|
|
|
if err != nil {
|
|
|
|
unsettledErr = err
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check that UnsettledBalance is what we expect.
|
|
|
|
if int(chanInfo.UnsettledBalance) != expectedBalance {
|
|
|
|
unsettledErr = fmt.Errorf("unsettled balance failed "+
|
|
|
|
"expected: %v, received: %v", expectedBalance,
|
|
|
|
chanInfo.UnsettledBalance)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}, defaultTimeout)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unsettled balace error: %v", unsettledErr)
|
|
|
|
}
|
|
|
|
|
2020-03-31 13:44:18 +03:00
|
|
|
// Check for payment errors.
|
|
|
|
select {
|
2021-06-28 21:10:10 +03:00
|
|
|
case err := <-errChan:
|
|
|
|
t.Fatalf("payment error: %v", err)
|
|
|
|
default:
|
2018-04-13 14:20:12 +03:00
|
|
|
}
|
2018-05-23 20:54:12 +03:00
|
|
|
|
2021-06-28 21:10:10 +03:00
|
|
|
// Check alice's channel balance, which should have a remote unsettled
|
|
|
|
// balance that equals to the amount of invoices * payAmt. The remote
|
|
|
|
// balance remains zero.
|
|
|
|
aliceLocal := chanAmt - cType.calcStaticFee(0) - numInvoices*payAmt
|
|
|
|
checkChannelBalance(net.Alice, aliceLocal, 0, 0, numInvoices*payAmt)
|
2018-04-13 14:20:12 +03:00
|
|
|
|
2021-06-28 21:10:10 +03:00
|
|
|
// Check carol's channel balance, which should have a local unsettled
|
|
|
|
// balance that equals to the amount of invoices * payAmt. The local
|
|
|
|
// balance remains zero.
|
|
|
|
checkChannelBalance(carol, 0, aliceLocal, numInvoices*payAmt, 0)
|
2017-10-07 02:58:13 +03:00
|
|
|
|
2021-06-28 21:10:10 +03:00
|
|
|
// Force and assert the channel closure.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, chanPointAlice, true)
|
2018-05-23 20:54:12 +03:00
|
|
|
|
2021-06-28 21:10:10 +03:00
|
|
|
// Cleanup by mining the force close and sweep transaction.
|
|
|
|
cleanupForceClose(t, net, net.Alice, chanPointAlice)
|
2017-10-07 02:58:13 +03:00
|
|
|
}
|
|
|
|
|
2018-02-23 01:29:06 +03:00
|
|
|
// testSphinxReplayPersistence verifies that replayed onion packets are rejected
|
|
|
|
// by a remote peer after a restart. We use a combination of unsafe
|
|
|
|
// configuration arguments to force Carol to replay the same sphinx packet after
|
|
|
|
// reconnecting to Dave, and compare the returned failure message with what we
|
|
|
|
// expect for replayed onion packets.
|
|
|
|
func testSphinxReplayPersistence(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
// Open a channel with 100k satoshis between Carol and Dave with Carol being
|
|
|
|
// the sole funder of the channel.
|
|
|
|
chanAmt := btcutil.Amount(100000)
|
|
|
|
|
|
|
|
// First, we'll create Dave, the receiver, and start him in hodl mode.
|
2021-06-07 23:05:12 +03:00
|
|
|
dave := net.NewNode(t.t, "Dave", []string{"--hodl.exit-settle"})
|
2018-02-23 01:29:06 +03:00
|
|
|
|
2018-06-08 00:14:39 +03:00
|
|
|
// We must remember to shutdown the nodes we created for the duration
|
|
|
|
// of the tests, only leaving the two seed nodes (Alice and Bob) within
|
|
|
|
// our test network.
|
|
|
|
defer shutdownAndAssert(net, t, dave)
|
|
|
|
|
2018-02-23 01:29:06 +03:00
|
|
|
// Next, we'll create Carol and establish a channel to from her to
|
2019-11-20 07:13:04 +03:00
|
|
|
// Dave. Carol is started in both unsafe-replay which will cause her to
|
|
|
|
// replay any pending Adds held in memory upon reconnection.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", []string{"--unsafe-replay"})
|
2018-06-08 00:14:39 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
|
|
|
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, carol, dave)
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, carol)
|
|
|
|
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoint := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, carol, dave,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2018-02-23 01:29:06 +03:00
|
|
|
|
2020-04-13 18:29:52 +03:00
|
|
|
// Next, we'll create Fred who is going to initiate the payment and
|
|
|
|
// establish a channel to from him to Carol. We can't perform this test
|
|
|
|
// by paying from Carol directly to Dave, because the '--unsafe-replay'
|
|
|
|
// setup doesn't apply to locally added htlcs. In that case, the
|
|
|
|
// mailbox, that is responsible for generating the replay, is bypassed.
|
2021-06-07 23:05:12 +03:00
|
|
|
fred := net.NewNode(t.t, "Fred", nil)
|
2020-04-13 18:29:52 +03:00
|
|
|
defer shutdownAndAssert(net, t, fred)
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, fred, carol)
|
2020-04-13 18:29:52 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, fred)
|
|
|
|
|
2020-04-13 18:29:52 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
|
|
|
chanPointFC := openChannelAndAssert(
|
|
|
|
ctxt, t, net, fred, carol,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
2018-02-23 01:29:06 +03:00
|
|
|
// Now that the channel is open, create an invoice for Dave which
|
|
|
|
// expects a payment of 1000 satoshis from Carol paid via a particular
|
|
|
|
// preimage.
|
|
|
|
const paymentAmt = 1000
|
|
|
|
preimage := bytes.Repeat([]byte("A"), 32)
|
|
|
|
invoice := &lnrpc.Invoice{
|
|
|
|
Memo: "testing",
|
|
|
|
RPreimage: preimage,
|
|
|
|
Value: paymentAmt,
|
|
|
|
}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
invoiceResp, err := dave.AddInvoice(ctxt, invoice)
|
2018-02-23 01:29:06 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to add invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-04-13 18:29:52 +03:00
|
|
|
// Wait for all channels to be recognized and advertized.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-02-23 01:29:06 +03:00
|
|
|
err = carol.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
err = dave.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
2020-04-13 18:29:52 +03:00
|
|
|
err = carol.WaitForNetworkChannelOpen(ctxt, chanPointFC)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
err = fred.WaitForNetworkChannelOpen(ctxt, chanPointFC)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't advertise channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
2018-02-23 01:29:06 +03:00
|
|
|
|
2020-04-13 18:29:52 +03:00
|
|
|
// With the invoice for Dave added, send a payment from Fred paying
|
2018-02-23 01:29:06 +03:00
|
|
|
// to the above generated invoice.
|
|
|
|
ctx, cancel := context.WithCancel(ctxb)
|
|
|
|
defer cancel()
|
|
|
|
|
2020-03-31 13:44:18 +03:00
|
|
|
payStream, err := fred.RouterClient.SendPaymentV2(
|
|
|
|
ctx,
|
|
|
|
&routerrpc.SendPaymentRequest{
|
|
|
|
PaymentRequest: invoiceResp.PaymentRequest,
|
|
|
|
TimeoutSeconds: 60,
|
|
|
|
FeeLimitMsat: noFeeLimitMsat,
|
|
|
|
},
|
|
|
|
)
|
2018-02-23 01:29:06 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to open payment stream: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
time.Sleep(200 * time.Millisecond)
|
|
|
|
|
|
|
|
// Dave's invoice should not be marked as settled.
|
|
|
|
payHash := &lnrpc.PaymentHash{
|
|
|
|
RHash: invoiceResp.RHash,
|
|
|
|
}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
dbInvoice, err := dave.LookupInvoice(ctxt, payHash)
|
2018-02-23 01:29:06 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to lookup invoice: %v", err)
|
|
|
|
}
|
|
|
|
if dbInvoice.Settled {
|
|
|
|
t.Fatalf("dave's invoice should not be marked as settled: %v",
|
|
|
|
spew.Sdump(dbInvoice))
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the payment sent but hedl, all balance related stats should not
|
|
|
|
// have changed.
|
2019-12-19 10:57:12 +03:00
|
|
|
err = wait.InvariantNoError(
|
|
|
|
assertAmountSent(0, carol, dave), 3*time.Second,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf(err.Error())
|
|
|
|
}
|
2018-02-23 01:29:06 +03:00
|
|
|
|
|
|
|
// With the first payment sent, restart dave to make sure he is
|
|
|
|
// persisting the information required to detect replayed sphinx
|
|
|
|
// packets.
|
|
|
|
if err := net.RestartNode(dave, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart dave: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Carol should retransmit the Add hedl in her mailbox on startup. Dave
|
|
|
|
// should not accept the replayed Add, and actually fail back the
|
|
|
|
// pending payment. Even though he still holds the original settle, if
|
|
|
|
// he does fail, it is almost certainly caused by the sphinx replay
|
|
|
|
// protection, as it is the only validation we do in hodl mode.
|
2020-03-31 13:44:18 +03:00
|
|
|
result, err := getPaymentResult(payStream)
|
2018-02-23 01:29:06 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to receive payment response: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-04-13 18:29:52 +03:00
|
|
|
// Assert that Fred receives the expected failure after Carol sent a
|
|
|
|
// duplicate packet that fails due to sphinx replay detection.
|
2020-03-31 13:44:18 +03:00
|
|
|
if result.Status == lnrpc.Payment_SUCCEEDED {
|
2020-04-01 01:13:25 +03:00
|
|
|
t.Fatalf("expected payment error")
|
2018-02-23 01:29:06 +03:00
|
|
|
}
|
2020-04-13 18:29:52 +03:00
|
|
|
assertLastHTLCError(t, fred, lnrpc.Failure_INVALID_ONION_KEY)
|
2018-02-23 01:29:06 +03:00
|
|
|
|
2018-03-31 00:53:05 +03:00
|
|
|
// Since the payment failed, the balance should still be left
|
|
|
|
// unaltered.
|
2019-12-19 10:57:12 +03:00
|
|
|
err = wait.InvariantNoError(
|
|
|
|
assertAmountSent(0, carol, dave), 3*time.Second,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf(err.Error())
|
|
|
|
}
|
2018-02-23 01:29:06 +03:00
|
|
|
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-03-31 00:53:05 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, carol, chanPoint, true)
|
2018-11-06 13:03:01 +03:00
|
|
|
|
|
|
|
// Cleanup by mining the force close and sweep transaction.
|
|
|
|
cleanupForceClose(t, net, carol, chanPoint)
|
2018-02-23 01:29:06 +03:00
|
|
|
}
|
|
|
|
|
2020-06-10 15:21:13 +03:00
|
|
|
// testListChannels checks that the response from ListChannels is correct. It
|
|
|
|
// tests the values in all ChannelConstraints are returned as expected. Once
|
|
|
|
// ListChannels becomes mature, a test against all fields in ListChannels should
|
|
|
|
// be performed.
|
|
|
|
func testListChannels(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
2020-08-13 01:45:57 +03:00
|
|
|
const aliceRemoteMaxHtlcs = 50
|
|
|
|
const bobRemoteMaxHtlcs = 100
|
|
|
|
|
2020-06-10 15:21:13 +03:00
|
|
|
// Create two fresh nodes and open a channel between them.
|
2021-06-07 23:05:12 +03:00
|
|
|
alice := net.NewNode(t.t, "Alice", nil)
|
2020-06-10 15:21:13 +03:00
|
|
|
defer shutdownAndAssert(net, t, alice)
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
bob := net.NewNode(
|
|
|
|
t.t, "Bob", []string{
|
|
|
|
fmt.Sprintf(
|
|
|
|
"--default-remote-max-htlcs=%v",
|
|
|
|
bobRemoteMaxHtlcs,
|
|
|
|
),
|
|
|
|
},
|
|
|
|
)
|
2020-06-10 15:21:13 +03:00
|
|
|
defer shutdownAndAssert(net, t, bob)
|
|
|
|
|
|
|
|
// Connect Alice to Bob.
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxb, t.t, alice, bob)
|
2020-06-10 15:21:13 +03:00
|
|
|
|
|
|
|
// Give Alice some coins so she can fund a channel.
|
2020-01-10 17:27:50 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, alice)
|
2020-06-10 15:21:13 +03:00
|
|
|
|
|
|
|
// Open a channel with 100k satoshis between Alice and Bob with Alice
|
|
|
|
// being the sole funder of the channel. The minial HTLC amount is set to
|
|
|
|
// 4200 msats.
|
|
|
|
const customizedMinHtlc = 4200
|
|
|
|
|
|
|
|
chanAmt := btcutil.Amount(100000)
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
|
|
|
chanPoint := openChannelAndAssert(
|
|
|
|
ctxt, t, net, alice, bob,
|
|
|
|
lntest.OpenChannelParams{
|
2020-08-13 01:46:17 +03:00
|
|
|
Amt: chanAmt,
|
|
|
|
MinHtlc: customizedMinHtlc,
|
|
|
|
RemoteMaxHtlcs: aliceRemoteMaxHtlcs,
|
2020-06-10 15:21:13 +03:00
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Wait for Alice and Bob to receive the channel edge from the
|
|
|
|
// funding manager.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
err := alice.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
2020-06-10 15:21:13 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't see the alice->bob channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
err = bob.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't see the bob->alice channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Alice should have one channel opened with Bob.
|
|
|
|
assertNodeNumChannels(t, alice, 1)
|
|
|
|
// Bob should have one channel opened with Alice.
|
|
|
|
assertNodeNumChannels(t, bob, 1)
|
|
|
|
|
|
|
|
// Get the ListChannel response from Alice.
|
|
|
|
listReq := &lnrpc.ListChannelsRequest{}
|
|
|
|
ctxb = context.Background()
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
resp, err := alice.ListChannels(ctxt, listReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query for %s's channel list: %v",
|
|
|
|
alice.Name(), err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check the returned response is correct.
|
|
|
|
aliceChannel := resp.Channels[0]
|
|
|
|
|
|
|
|
// defaultConstraints is a ChannelConstraints with default values. It is
|
|
|
|
// used to test against Alice's local channel constraints.
|
|
|
|
defaultConstraints := &lnrpc.ChannelConstraints{
|
|
|
|
CsvDelay: 4,
|
|
|
|
ChanReserveSat: 1000,
|
|
|
|
DustLimitSat: uint64(lnwallet.DefaultDustLimit()),
|
|
|
|
MaxPendingAmtMsat: 99000000,
|
|
|
|
MinHtlcMsat: 1,
|
2020-08-13 01:45:57 +03:00
|
|
|
MaxAcceptedHtlcs: bobRemoteMaxHtlcs,
|
2020-06-10 15:21:13 +03:00
|
|
|
}
|
|
|
|
assertChannelConstraintsEqual(
|
2020-08-13 01:45:29 +03:00
|
|
|
t, defaultConstraints, aliceChannel.LocalConstraints,
|
2020-06-10 15:21:13 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// customizedConstraints is a ChannelConstraints with customized values.
|
|
|
|
// Ideally, all these values can be passed in when creating the channel.
|
|
|
|
// Currently, only the MinHtlcMsat is customized. It is used to check
|
|
|
|
// against Alice's remote channel constratins.
|
|
|
|
customizedConstraints := &lnrpc.ChannelConstraints{
|
|
|
|
CsvDelay: 4,
|
|
|
|
ChanReserveSat: 1000,
|
|
|
|
DustLimitSat: uint64(lnwallet.DefaultDustLimit()),
|
|
|
|
MaxPendingAmtMsat: 99000000,
|
|
|
|
MinHtlcMsat: customizedMinHtlc,
|
2020-08-13 01:45:57 +03:00
|
|
|
MaxAcceptedHtlcs: aliceRemoteMaxHtlcs,
|
2020-06-10 15:21:13 +03:00
|
|
|
}
|
|
|
|
assertChannelConstraintsEqual(
|
2020-08-13 01:45:29 +03:00
|
|
|
t, customizedConstraints, aliceChannel.RemoteConstraints,
|
2020-06-10 15:21:13 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// Get the ListChannel response for Bob.
|
|
|
|
listReq = &lnrpc.ListChannelsRequest{}
|
|
|
|
ctxb = context.Background()
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
resp, err = bob.ListChannels(ctxt, listReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query for %s's channel "+
|
|
|
|
"list: %v", bob.Name(), err)
|
|
|
|
}
|
|
|
|
|
|
|
|
bobChannel := resp.Channels[0]
|
|
|
|
if bobChannel.ChannelPoint != aliceChannel.ChannelPoint {
|
|
|
|
t.Fatalf("Bob's channel point mismatched, want: %s, got: %s",
|
|
|
|
chanPoint.String(), bobChannel.ChannelPoint,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check channel constraints match. Alice's local channel constraint should
|
|
|
|
// be equal to Bob's remote channel constraint, and her remote one should
|
|
|
|
// be equal to Bob's local one.
|
|
|
|
assertChannelConstraintsEqual(
|
|
|
|
t, aliceChannel.LocalConstraints, bobChannel.RemoteConstraints,
|
|
|
|
)
|
|
|
|
assertChannelConstraintsEqual(
|
|
|
|
t, aliceChannel.RemoteConstraints, bobChannel.LocalConstraints,
|
|
|
|
)
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2021-02-17 01:06:54 +03:00
|
|
|
// testUpdateChanStatus checks that calls to the UpdateChanStatus RPC update
|
|
|
|
// the channel graph as expected, and that channel state is properly updated
|
|
|
|
// in the presence of interleaved node disconnects / reconnects.
|
|
|
|
func testUpdateChanStatus(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
// Create two fresh nodes and open a channel between them.
|
2021-06-07 23:05:12 +03:00
|
|
|
alice := net.NewNode(
|
|
|
|
t.t, "Alice", []string{
|
|
|
|
"--minbackoff=10s",
|
|
|
|
"--chan-enable-timeout=1.5s",
|
|
|
|
"--chan-disable-timeout=3s",
|
|
|
|
"--chan-status-sample-interval=.5s",
|
|
|
|
},
|
|
|
|
)
|
2021-02-17 01:06:54 +03:00
|
|
|
defer shutdownAndAssert(net, t, alice)
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
bob := net.NewNode(
|
|
|
|
t.t, "Bob", []string{
|
|
|
|
"--minbackoff=10s",
|
|
|
|
"--chan-enable-timeout=1.5s",
|
|
|
|
"--chan-disable-timeout=3s",
|
|
|
|
"--chan-status-sample-interval=.5s",
|
|
|
|
},
|
|
|
|
)
|
2021-02-17 01:06:54 +03:00
|
|
|
defer shutdownAndAssert(net, t, bob)
|
|
|
|
|
|
|
|
// Connect Alice to Bob.
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxb, t.t, alice, bob)
|
2021-02-17 01:06:54 +03:00
|
|
|
|
|
|
|
// Give Alice some coins so she can fund a channel.
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, alice)
|
2021-02-17 01:06:54 +03:00
|
|
|
|
|
|
|
// Open a channel with 100k satoshis between Alice and Bob with Alice
|
|
|
|
// being the sole funder of the channel.
|
|
|
|
chanAmt := btcutil.Amount(100000)
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
|
|
|
chanPoint := openChannelAndAssert(
|
|
|
|
ctxt, t, net, alice, bob,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Wait for Alice and Bob to receive the channel edge from the
|
|
|
|
// funding manager.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
err := alice.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
2021-02-17 01:06:54 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't see the alice->bob channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
err = bob.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't see the bob->alice channel before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Launch a node for Carol which will connect to Alice and Bob in
|
|
|
|
// order to receive graph updates. This will ensure that the
|
|
|
|
// channel updates are propagated throughout the network.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", nil)
|
2021-02-17 01:06:54 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, alice, carol)
|
2021-02-17 01:06:54 +03:00
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, bob, carol)
|
2021-02-17 01:06:54 +03:00
|
|
|
|
2021-04-16 11:55:51 +03:00
|
|
|
carolSub := subscribeGraphNotifications(ctxb, t, carol)
|
2021-02-17 01:06:54 +03:00
|
|
|
defer close(carolSub.quit)
|
|
|
|
|
|
|
|
// sendReq sends an UpdateChanStatus request to the given node.
|
|
|
|
sendReq := func(node *lntest.HarnessNode, chanPoint *lnrpc.ChannelPoint,
|
|
|
|
action routerrpc.ChanStatusAction) {
|
|
|
|
|
|
|
|
req := &routerrpc.UpdateChanStatusRequest{
|
|
|
|
ChanPoint: chanPoint,
|
|
|
|
Action: action,
|
|
|
|
}
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
_, err = node.RouterClient.UpdateChanStatus(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to call UpdateChanStatus for %s's node: %v",
|
|
|
|
node.Name(), err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// assertEdgeDisabled ensures that a given node has the correct
|
|
|
|
// Disabled state for a channel.
|
|
|
|
assertEdgeDisabled := func(node *lntest.HarnessNode,
|
|
|
|
chanPoint *lnrpc.ChannelPoint, disabled bool) {
|
|
|
|
|
|
|
|
var predErr error
|
|
|
|
err = wait.Predicate(func() bool {
|
|
|
|
req := &lnrpc.ChannelGraphRequest{
|
|
|
|
IncludeUnannounced: true,
|
|
|
|
}
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
chanGraph, err := node.DescribeGraph(ctxt, req)
|
|
|
|
if err != nil {
|
|
|
|
predErr = fmt.Errorf("unable to query node %v's graph: %v", node, err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
numEdges := len(chanGraph.Edges)
|
|
|
|
if numEdges != 1 {
|
|
|
|
predErr = fmt.Errorf("expected to find 1 edge in the graph, found %d", numEdges)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
edge := chanGraph.Edges[0]
|
|
|
|
if edge.ChanPoint != chanPoint.GetFundingTxidStr() {
|
|
|
|
predErr = fmt.Errorf("expected chan_point %v, got %v",
|
|
|
|
chanPoint.GetFundingTxidStr(), edge.ChanPoint)
|
|
|
|
}
|
|
|
|
var policy *lnrpc.RoutingPolicy
|
|
|
|
if node.PubKeyStr == edge.Node1Pub {
|
|
|
|
policy = edge.Node1Policy
|
|
|
|
} else {
|
|
|
|
policy = edge.Node2Policy
|
|
|
|
}
|
|
|
|
if disabled != policy.Disabled {
|
|
|
|
predErr = fmt.Errorf("expected policy.Disabled to be %v, "+
|
|
|
|
"but policy was %v", disabled, policy)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return true
|
|
|
|
}, defaultTimeout)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("%v", predErr)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// When updating the state of the channel between Alice and Bob, we
|
|
|
|
// should expect to see channel updates with the default routing
|
|
|
|
// policy. The value of "Disabled" will depend on the specific
|
|
|
|
// scenario being tested.
|
|
|
|
expectedPolicy := &lnrpc.RoutingPolicy{
|
|
|
|
FeeBaseMsat: int64(chainreg.DefaultBitcoinBaseFeeMSat),
|
|
|
|
FeeRateMilliMsat: int64(chainreg.DefaultBitcoinFeeRate),
|
|
|
|
TimeLockDelta: chainreg.DefaultBitcoinTimeLockDelta,
|
|
|
|
MinHtlc: 1000, // default value
|
|
|
|
MaxHtlcMsat: calculateMaxHtlc(chanAmt),
|
|
|
|
}
|
|
|
|
|
|
|
|
// Initially, the channel between Alice and Bob should not be
|
|
|
|
// disabled.
|
|
|
|
assertEdgeDisabled(alice, chanPoint, false)
|
|
|
|
|
|
|
|
// Manually disable the channel and ensure that a "Disabled = true"
|
|
|
|
// update is propagated.
|
|
|
|
sendReq(alice, chanPoint, routerrpc.ChanStatusAction_DISABLE)
|
|
|
|
expectedPolicy.Disabled = true
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Re-enable the channel and ensure that a "Disabled = false" update
|
|
|
|
// is propagated.
|
|
|
|
sendReq(alice, chanPoint, routerrpc.ChanStatusAction_ENABLE)
|
|
|
|
expectedPolicy.Disabled = false
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Manually enabling a channel should NOT prevent subsequent
|
|
|
|
// disconnections from automatically disabling the channel again
|
|
|
|
// (we don't want to clutter the network with channels that are
|
|
|
|
// falsely advertised as enabled when they don't work).
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
if err := net.DisconnectNodes(ctxt, alice, bob); err != nil {
|
|
|
|
t.Fatalf("unable to disconnect Alice from Bob: %v", err)
|
|
|
|
}
|
|
|
|
expectedPolicy.Disabled = true
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
{bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Reconnecting the nodes should propagate a "Disabled = false" update.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, alice, bob)
|
2021-02-17 01:06:54 +03:00
|
|
|
expectedPolicy.Disabled = false
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
{bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Manually disabling the channel should prevent a subsequent
|
|
|
|
// disconnect / reconnect from re-enabling the channel on
|
|
|
|
// Alice's end. Note the asymmetry between manual enable and
|
|
|
|
// manual disable!
|
|
|
|
sendReq(alice, chanPoint, routerrpc.ChanStatusAction_DISABLE)
|
|
|
|
|
|
|
|
// Alice sends out the "Disabled = true" update in response to
|
|
|
|
// the ChanStatusAction_DISABLE request.
|
|
|
|
expectedPolicy.Disabled = true
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
if err := net.DisconnectNodes(ctxt, alice, bob); err != nil {
|
|
|
|
t.Fatalf("unable to disconnect Alice from Bob: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Bob sends a "Disabled = true" update upon detecting the
|
|
|
|
// disconnect.
|
|
|
|
expectedPolicy.Disabled = true
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Bob sends a "Disabled = false" update upon detecting the
|
|
|
|
// reconnect.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, alice, bob)
|
2021-02-17 01:06:54 +03:00
|
|
|
expectedPolicy.Disabled = false
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// However, since we manually disabled the channel on Alice's end,
|
|
|
|
// the policy on Alice's end should still be "Disabled = true". Again,
|
|
|
|
// note the asymmetry between manual enable and manual disable!
|
|
|
|
assertEdgeDisabled(alice, chanPoint, true)
|
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
if err := net.DisconnectNodes(ctxt, alice, bob); err != nil {
|
|
|
|
t.Fatalf("unable to disconnect Alice from Bob: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Bob sends a "Disabled = true" update upon detecting the
|
|
|
|
// disconnect.
|
|
|
|
expectedPolicy.Disabled = true
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// After restoring automatic channel state management on Alice's end,
|
|
|
|
// BOTH Alice and Bob should set the channel state back to "enabled"
|
|
|
|
// on reconnect.
|
|
|
|
sendReq(alice, chanPoint, routerrpc.ChanStatusAction_AUTO)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, alice, bob)
|
2021-02-17 01:06:54 +03:00
|
|
|
expectedPolicy.Disabled = false
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, carolSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{alice.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
{bob.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
assertEdgeDisabled(alice, chanPoint, false)
|
|
|
|
}
|
|
|
|
|
2018-06-16 15:53:44 +03:00
|
|
|
// updateChannelPolicy updates the channel policy of node to the
|
|
|
|
// given fees and timelock delta. This function blocks until
|
|
|
|
// listenerNode has received the policy update.
|
|
|
|
func updateChannelPolicy(t *harnessTest, node *lntest.HarnessNode,
|
|
|
|
chanPoint *lnrpc.ChannelPoint, baseFee int64, feeRate int64,
|
2019-08-20 03:55:30 +03:00
|
|
|
timeLockDelta uint32, maxHtlc uint64, listenerNode *lntest.HarnessNode) {
|
|
|
|
|
2018-06-16 15:53:44 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
expectedPolicy := &lnrpc.RoutingPolicy{
|
|
|
|
FeeBaseMsat: baseFee,
|
|
|
|
FeeRateMilliMsat: feeRate,
|
|
|
|
TimeLockDelta: timeLockDelta,
|
2018-08-22 10:32:43 +03:00
|
|
|
MinHtlc: 1000, // default value
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: maxHtlc,
|
2018-06-16 15:53:44 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
updateFeeReq := &lnrpc.PolicyUpdateRequest{
|
|
|
|
BaseFeeMsat: baseFee,
|
|
|
|
FeeRate: float64(feeRate) / testFeeBase,
|
|
|
|
TimeLockDelta: timeLockDelta,
|
|
|
|
Scope: &lnrpc.PolicyUpdateRequest_ChanPoint{
|
|
|
|
ChanPoint: chanPoint,
|
|
|
|
},
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: maxHtlc,
|
2018-06-16 15:53:44 +03:00
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2018-06-16 15:53:44 +03:00
|
|
|
if _, err := node.UpdateChannelPolicy(ctxt, updateFeeReq); err != nil {
|
|
|
|
t.Fatalf("unable to update chan policy: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for listener node to receive the channel update from node.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-04-16 11:55:51 +03:00
|
|
|
graphSub := subscribeGraphNotifications(ctxt, t, listenerNode)
|
2018-08-23 08:16:16 +03:00
|
|
|
defer close(graphSub.quit)
|
2018-06-16 15:53:44 +03:00
|
|
|
|
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, graphSub,
|
2018-08-22 10:32:42 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{node.PubKeyStr, expectedPolicy, chanPoint},
|
|
|
|
},
|
2018-06-16 15:53:44 +03:00
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// testUnannouncedChannels checks unannounced channels are not returned by
|
|
|
|
// describeGraph RPC request unless explicitly asked for.
|
|
|
|
func testUnannouncedChannels(net *lntest.NetworkHarness, t *harnessTest) {
|
2019-11-05 02:11:55 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
amount := funding.MaxBtcFundingAmount
|
2019-11-05 02:11:55 +03:00
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// Open a channel between Alice and Bob, ensuring the
|
|
|
|
// channel has been opened properly.
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
|
|
|
chanOpenUpdate := openChannelStream(
|
|
|
|
ctxt, t, net, net.Alice, net.Bob,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
2021-04-16 11:41:51 +03:00
|
|
|
Amt: amount,
|
2018-08-22 10:32:42 +03:00
|
|
|
},
|
|
|
|
)
|
2018-02-03 13:21:16 +03:00
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// Mine 2 blocks, and check that the channel is opened but not yet
|
|
|
|
// announced to the network.
|
|
|
|
mineBlocks(t, net, 2, 1)
|
2018-02-03 13:21:16 +03:00
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// One block is enough to make the channel ready for use, since the
|
|
|
|
// nodes have defaultNumConfs=1 set.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
fundingChanPoint, err := net.WaitForChannelOpen(ctxt, chanOpenUpdate)
|
2019-11-05 02:11:55 +03:00
|
|
|
if err != nil {
|
2021-04-16 11:41:51 +03:00
|
|
|
t.Fatalf("error while waiting for channel open: %v", err)
|
2019-11-05 02:11:55 +03:00
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// Alice should have 1 edge in her graph.
|
|
|
|
req := &lnrpc.ChannelGraphRequest{
|
|
|
|
IncludeUnannounced: true,
|
2019-09-03 13:23:39 +03:00
|
|
|
}
|
2021-04-16 11:41:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
chanGraph, err := net.Alice.DescribeGraph(ctxt, req)
|
2020-10-26 22:25:25 +03:00
|
|
|
if err != nil {
|
2021-04-16 11:41:51 +03:00
|
|
|
t.Fatalf("unable to query alice's graph: %v", err)
|
2020-10-26 22:25:25 +03:00
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
numEdges := len(chanGraph.Edges)
|
|
|
|
if numEdges != 1 {
|
|
|
|
t.Fatalf("expected to find 1 edge in the graph, found %d", numEdges)
|
2018-02-03 13:21:16 +03:00
|
|
|
}
|
2021-04-16 11:41:51 +03:00
|
|
|
|
|
|
|
// Channels should not be announced yet, hence Alice should have no
|
|
|
|
// announced edges in her graph.
|
|
|
|
req.IncludeUnannounced = false
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-04-16 11:41:51 +03:00
|
|
|
chanGraph, err = net.Alice.DescribeGraph(ctxt, req)
|
2018-02-03 13:21:16 +03:00
|
|
|
if err != nil {
|
2021-04-16 11:41:51 +03:00
|
|
|
t.Fatalf("unable to query alice's graph: %v", err)
|
2018-02-03 13:21:16 +03:00
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
numEdges = len(chanGraph.Edges)
|
|
|
|
if numEdges != 0 {
|
|
|
|
t.Fatalf("expected to find 0 announced edges in the graph, found %d",
|
|
|
|
numEdges)
|
2019-11-05 02:11:55 +03:00
|
|
|
}
|
2018-02-03 13:21:16 +03:00
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// Mine 4 more blocks, and check that the channel is now announced.
|
|
|
|
mineBlocks(t, net, 4, 0)
|
2019-11-05 02:11:55 +03:00
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// Give the network a chance to learn that auth proof is confirmed.
|
|
|
|
var predErr error
|
|
|
|
err = wait.Predicate(func() bool {
|
|
|
|
// The channel should now be announced. Check that Alice has 1
|
|
|
|
// announced edge.
|
|
|
|
req.IncludeUnannounced = false
|
2019-11-05 02:11:55 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-04-16 11:41:51 +03:00
|
|
|
chanGraph, err = net.Alice.DescribeGraph(ctxt, req)
|
2018-02-03 13:21:16 +03:00
|
|
|
if err != nil {
|
2021-04-16 11:41:51 +03:00
|
|
|
predErr = fmt.Errorf("unable to query alice's graph: %v", err)
|
|
|
|
return false
|
2018-02-03 13:21:16 +03:00
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
numEdges = len(chanGraph.Edges)
|
|
|
|
if numEdges != 1 {
|
|
|
|
predErr = fmt.Errorf("expected to find 1 announced edge in "+
|
|
|
|
"the graph, found %d", numEdges)
|
|
|
|
return false
|
2019-11-05 02:11:55 +03:00
|
|
|
}
|
2021-04-16 11:41:51 +03:00
|
|
|
return true
|
|
|
|
}, defaultTimeout)
|
2019-05-30 02:31:49 +03:00
|
|
|
if err != nil {
|
2021-04-16 11:41:51 +03:00
|
|
|
t.Fatalf("%v", predErr)
|
2019-11-05 02:11:55 +03:00
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// The channel should now be announced. Check that Alice has 1 announced
|
|
|
|
// edge.
|
|
|
|
req.IncludeUnannounced = false
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
chanGraph, err = net.Alice.DescribeGraph(ctxt, req)
|
2019-11-05 02:11:55 +03:00
|
|
|
if err != nil {
|
2021-04-16 11:41:51 +03:00
|
|
|
t.Fatalf("unable to query alice's graph: %v", err)
|
2019-11-05 02:11:55 +03:00
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
numEdges = len(chanGraph.Edges)
|
|
|
|
if numEdges != 1 {
|
|
|
|
t.Fatalf("expected to find 1 announced edge in the graph, found %d",
|
|
|
|
numEdges)
|
2019-05-30 02:31:49 +03:00
|
|
|
}
|
|
|
|
|
2021-04-16 11:41:51 +03:00
|
|
|
// Close the channel used during the test.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2021-04-16 11:41:51 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, fundingChanPoint, false)
|
2018-02-03 13:21:16 +03:00
|
|
|
}
|
|
|
|
|
2018-11-02 06:20:21 +03:00
|
|
|
// channelSubscription houses the proxied update and error chans for a node's
|
|
|
|
// channel subscriptions.
|
|
|
|
type channelSubscription struct {
|
|
|
|
updateChan chan *lnrpc.ChannelEventUpdate
|
|
|
|
errChan chan error
|
|
|
|
quit chan struct{}
|
|
|
|
}
|
|
|
|
|
|
|
|
// subscribeChannelNotifications subscribes to channel updates and launches a
|
|
|
|
// goroutine that forwards these to the returned channel.
|
|
|
|
func subscribeChannelNotifications(ctxb context.Context, t *harnessTest,
|
|
|
|
node *lntest.HarnessNode) channelSubscription {
|
|
|
|
|
|
|
|
// We'll first start by establishing a notification client which will
|
|
|
|
// send us notifications upon channels becoming active, inactive or
|
|
|
|
// closed.
|
|
|
|
req := &lnrpc.ChannelEventSubscription{}
|
|
|
|
ctx, cancelFunc := context.WithCancel(ctxb)
|
|
|
|
|
|
|
|
chanUpdateClient, err := node.SubscribeChannelEvents(ctx, req)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create channel update client: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll launch a goroutine that will be responsible for proxying all
|
|
|
|
// notifications recv'd from the client into the channel below.
|
|
|
|
errChan := make(chan error, 1)
|
|
|
|
quit := make(chan struct{})
|
|
|
|
chanUpdates := make(chan *lnrpc.ChannelEventUpdate, 20)
|
|
|
|
go func() {
|
|
|
|
defer cancelFunc()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-quit:
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
chanUpdate, err := chanUpdateClient.Recv()
|
|
|
|
select {
|
|
|
|
case <-quit:
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
|
|
|
|
if err == io.EOF {
|
|
|
|
return
|
|
|
|
} else if err != nil {
|
|
|
|
select {
|
|
|
|
case errChan <- err:
|
|
|
|
case <-quit:
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
select {
|
|
|
|
case chanUpdates <- chanUpdate:
|
|
|
|
case <-quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
return channelSubscription{
|
|
|
|
updateChan: chanUpdates,
|
|
|
|
errChan: errChan,
|
|
|
|
quit: quit,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// testBasicChannelCreationAndUpdates tests multiple channel opening and closing,
|
|
|
|
// and ensures that if a node is subscribed to channel updates they will be
|
|
|
|
// received correctly for both cooperative and force closed channels.
|
|
|
|
func testBasicChannelCreationAndUpdates(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
2016-10-24 05:00:09 +03:00
|
|
|
const (
|
|
|
|
numChannels = 2
|
2020-11-17 02:28:56 +03:00
|
|
|
amount = funding.MaxBtcFundingAmount
|
2016-10-24 05:00:09 +03:00
|
|
|
)
|
2016-10-17 07:36:43 +03:00
|
|
|
|
2019-02-12 04:23:47 +03:00
|
|
|
// Subscribe Bob and Alice to channel event notifications.
|
2018-11-02 06:20:21 +03:00
|
|
|
bobChanSub := subscribeChannelNotifications(ctxb, t, net.Bob)
|
|
|
|
defer close(bobChanSub.quit)
|
|
|
|
|
2019-02-12 04:23:47 +03:00
|
|
|
aliceChanSub := subscribeChannelNotifications(ctxb, t, net.Alice)
|
|
|
|
defer close(aliceChanSub.quit)
|
|
|
|
|
2016-10-17 07:36:43 +03:00
|
|
|
// Open the channel between Alice and Bob, asserting that the
|
|
|
|
// channel has been properly open on-chain.
|
2016-10-24 05:00:09 +03:00
|
|
|
chanPoints := make([]*lnrpc.ChannelPoint, numChannels)
|
|
|
|
for i := 0; i < numChannels; i++ {
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoints[i] = openChannelAndAssert(
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, t, net, net.Alice, net.Bob,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: amount,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2016-10-17 07:36:43 +03:00
|
|
|
}
|
|
|
|
|
2019-02-12 04:23:47 +03:00
|
|
|
// Since each of the channels just became open, Bob and Alice should
|
|
|
|
// each receive an open and an active notification for each channel.
|
2018-11-02 06:20:21 +03:00
|
|
|
var numChannelUpds int
|
2019-09-29 13:13:01 +03:00
|
|
|
const totalNtfns = 3 * numChannels
|
2019-02-12 04:23:47 +03:00
|
|
|
verifyOpenUpdatesReceived := func(sub channelSubscription) error {
|
|
|
|
numChannelUpds = 0
|
|
|
|
for numChannelUpds < totalNtfns {
|
|
|
|
select {
|
|
|
|
case update := <-sub.updateChan:
|
|
|
|
switch update.Type {
|
2019-09-29 13:13:01 +03:00
|
|
|
case lnrpc.ChannelEventUpdate_PENDING_OPEN_CHANNEL:
|
|
|
|
if numChannelUpds%3 != 0 {
|
|
|
|
return fmt.Errorf("expected " +
|
|
|
|
"open or active" +
|
|
|
|
"channel ntfn, got pending open " +
|
2019-02-12 04:23:47 +03:00
|
|
|
"channel ntfn instead")
|
|
|
|
}
|
|
|
|
case lnrpc.ChannelEventUpdate_OPEN_CHANNEL:
|
2019-09-29 13:13:01 +03:00
|
|
|
if numChannelUpds%3 != 1 {
|
|
|
|
return fmt.Errorf("expected " +
|
|
|
|
"pending open or active" +
|
2019-02-12 04:23:47 +03:00
|
|
|
"channel ntfn, got open" +
|
|
|
|
"channel ntfn instead")
|
|
|
|
}
|
2019-09-29 13:13:01 +03:00
|
|
|
case lnrpc.ChannelEventUpdate_ACTIVE_CHANNEL:
|
|
|
|
if numChannelUpds%3 != 2 {
|
|
|
|
return fmt.Errorf("expected " +
|
|
|
|
"pending open or open" +
|
|
|
|
"channel ntfn, got active " +
|
|
|
|
"channel ntfn instead")
|
|
|
|
}
|
2019-02-12 04:23:47 +03:00
|
|
|
default:
|
|
|
|
return fmt.Errorf("update type mismatch: "+
|
|
|
|
"expected open or active channel "+
|
|
|
|
"notification, got: %v",
|
|
|
|
update.Type)
|
|
|
|
}
|
|
|
|
numChannelUpds++
|
|
|
|
case <-time.After(time.Second * 10):
|
|
|
|
return fmt.Errorf("timeout waiting for channel "+
|
|
|
|
"notifications, only received %d/%d "+
|
|
|
|
"chanupds", numChannelUpds,
|
|
|
|
totalNtfns)
|
2018-11-02 06:20:21 +03:00
|
|
|
}
|
|
|
|
}
|
2019-02-12 04:23:47 +03:00
|
|
|
|
|
|
|
return nil
|
2018-11-02 06:20:21 +03:00
|
|
|
}
|
|
|
|
|
2019-02-12 04:23:47 +03:00
|
|
|
if err := verifyOpenUpdatesReceived(bobChanSub); err != nil {
|
|
|
|
t.Fatalf("error verifying open updates: %v", err)
|
|
|
|
}
|
|
|
|
if err := verifyOpenUpdatesReceived(aliceChanSub); err != nil {
|
|
|
|
t.Fatalf("error verifying open updates: %v", err)
|
|
|
|
}
|
2018-11-02 06:20:21 +03:00
|
|
|
|
|
|
|
// Close the channel between Alice and Bob, asserting that the channel
|
|
|
|
// has been properly closed on-chain.
|
|
|
|
for i, chanPoint := range chanPoints {
|
|
|
|
ctx, _ := context.WithTimeout(context.Background(), defaultTimeout)
|
|
|
|
|
|
|
|
// Force close half of the channels.
|
|
|
|
force := i%2 == 0
|
|
|
|
closeChannelAndAssert(ctx, t, net, net.Alice, chanPoint, force)
|
|
|
|
if force {
|
|
|
|
cleanupForceClose(t, net, net.Alice, chanPoint)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// verifyCloseUpdatesReceived is used to verify that Alice and Bob
|
|
|
|
// receive the correct channel updates in order.
|
|
|
|
verifyCloseUpdatesReceived := func(sub channelSubscription,
|
2020-02-21 14:24:24 +03:00
|
|
|
forceType lnrpc.ChannelCloseSummary_ClosureType,
|
2020-04-03 10:09:39 +03:00
|
|
|
closeInitiator lnrpc.Initiator) error {
|
2018-11-02 06:20:21 +03:00
|
|
|
|
|
|
|
// Ensure one inactive and one closed notification is received for each
|
|
|
|
// closed channel.
|
|
|
|
numChannelUpds := 0
|
|
|
|
for numChannelUpds < 2*numChannels {
|
2020-02-21 14:24:24 +03:00
|
|
|
expectedCloseType := lnrpc.ChannelCloseSummary_COOPERATIVE_CLOSE
|
|
|
|
|
|
|
|
// Every other channel should be force closed. If this
|
|
|
|
// channel was force closed, set the expected close type
|
|
|
|
// the the type passed in.
|
2018-11-02 06:20:21 +03:00
|
|
|
force := (numChannelUpds/2)%2 == 0
|
2020-02-21 14:24:24 +03:00
|
|
|
if force {
|
|
|
|
expectedCloseType = forceType
|
|
|
|
}
|
2018-11-02 06:20:21 +03:00
|
|
|
|
|
|
|
select {
|
|
|
|
case chanUpdate := <-sub.updateChan:
|
2020-02-21 14:24:24 +03:00
|
|
|
err := verifyCloseUpdate(
|
|
|
|
chanUpdate, expectedCloseType,
|
|
|
|
closeInitiator,
|
|
|
|
)
|
2018-11-02 06:20:21 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
numChannelUpds++
|
|
|
|
case err := <-sub.errChan:
|
|
|
|
return err
|
|
|
|
case <-time.After(time.Second * 10):
|
2020-02-21 14:24:24 +03:00
|
|
|
return fmt.Errorf("timeout waiting "+
|
|
|
|
"for channel notifications, only "+
|
|
|
|
"received %d/%d chanupds",
|
|
|
|
numChannelUpds, 2*numChannels)
|
2018-11-02 06:20:21 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Verify Bob receives all closed channel notifications. He should
|
|
|
|
// receive a remote force close notification for force closed channels.
|
2020-02-21 14:24:24 +03:00
|
|
|
// All channels (cooperatively and force closed) should have a remote
|
|
|
|
// close initiator because Alice closed the channels.
|
2018-11-02 06:20:21 +03:00
|
|
|
if err := verifyCloseUpdatesReceived(bobChanSub,
|
2020-02-21 14:24:24 +03:00
|
|
|
lnrpc.ChannelCloseSummary_REMOTE_FORCE_CLOSE,
|
2020-04-03 10:09:39 +03:00
|
|
|
lnrpc.Initiator_INITIATOR_REMOTE); err != nil {
|
2018-11-02 06:20:21 +03:00
|
|
|
t.Fatalf("errored verifying close updates: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Verify Alice receives all closed channel notifications. She should
|
|
|
|
// receive a remote force close notification for force closed channels.
|
2020-02-21 14:24:24 +03:00
|
|
|
// All channels (cooperatively and force closed) should have a local
|
|
|
|
// close initiator because Alice closed the channels.
|
2018-11-02 06:20:21 +03:00
|
|
|
if err := verifyCloseUpdatesReceived(aliceChanSub,
|
2020-02-21 14:24:24 +03:00
|
|
|
lnrpc.ChannelCloseSummary_LOCAL_FORCE_CLOSE,
|
2020-04-03 10:09:39 +03:00
|
|
|
lnrpc.Initiator_INITIATOR_LOCAL); err != nil {
|
2018-11-02 06:20:21 +03:00
|
|
|
t.Fatalf("errored verifying close updates: %v", err)
|
2016-10-17 07:36:43 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-10-15 16:24:56 +03:00
|
|
|
// testMaxPendingChannels checks that error is returned from remote peer if
|
|
|
|
// max pending channel number was exceeded and that '--maxpendingchannels' flag
|
|
|
|
// exists and works properly.
|
2017-11-03 21:52:02 +03:00
|
|
|
func testMaxPendingChannels(net *lntest.NetworkHarness, t *harnessTest) {
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2020-05-13 16:47:45 +03:00
|
|
|
maxPendingChannels := lncfg.DefaultMaxPendingChannels + 1
|
2020-11-17 02:28:56 +03:00
|
|
|
amount := funding.MaxBtcFundingAmount
|
2016-10-15 16:24:56 +03:00
|
|
|
|
|
|
|
// Create a new node (Carol) with greater number of max pending
|
|
|
|
// channels.
|
|
|
|
args := []string{
|
|
|
|
fmt.Sprintf("--maxpendingchannels=%v", maxPendingChannels),
|
|
|
|
}
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", args)
|
2018-06-08 00:14:39 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
2016-10-24 05:00:09 +03:00
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, net.Alice, carol)
|
2016-10-15 16:24:56 +03:00
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2016-10-15 16:24:56 +03:00
|
|
|
carolBalance := btcutil.Amount(maxPendingChannels) * amount
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, carolBalance, carol)
|
2016-10-15 16:24:56 +03:00
|
|
|
|
|
|
|
// Send open channel requests without generating new blocks thereby
|
2017-03-15 06:06:33 +03:00
|
|
|
// increasing pool of pending channels. Then check that we can't open
|
|
|
|
// the channel if the number of pending channels exceed max value.
|
2016-10-15 16:24:56 +03:00
|
|
|
openStreams := make([]lnrpc.Lightning_OpenChannelClient, maxPendingChannels)
|
|
|
|
for i := 0; i < maxPendingChannels; i++ {
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2020-04-03 00:29:50 +03:00
|
|
|
stream := openChannelStream(
|
|
|
|
ctxt, t, net, net.Alice, carol,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: amount,
|
|
|
|
},
|
2018-08-10 07:17:06 +03:00
|
|
|
)
|
2016-10-15 16:24:56 +03:00
|
|
|
openStreams[i] = stream
|
|
|
|
}
|
|
|
|
|
|
|
|
// Carol exhausted available amount of pending channels, next open
|
|
|
|
// channel request should cause ErrorGeneric to be sent back to Alice.
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2021-06-07 23:05:12 +03:00
|
|
|
_, err := net.OpenChannel(
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, net.Alice, carol,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: amount,
|
|
|
|
},
|
2018-08-10 07:17:06 +03:00
|
|
|
)
|
2018-08-22 10:32:42 +03:00
|
|
|
|
2016-10-15 16:24:56 +03:00
|
|
|
if err == nil {
|
2016-10-24 05:00:09 +03:00
|
|
|
t.Fatalf("error wasn't received")
|
2019-09-20 11:55:20 +03:00
|
|
|
} else if !strings.Contains(
|
|
|
|
err.Error(), lnwire.ErrMaxPendingChannels.Error(),
|
|
|
|
) {
|
2016-10-26 15:25:42 +03:00
|
|
|
t.Fatalf("not expected error was received: %v", err)
|
2016-10-15 16:24:56 +03:00
|
|
|
}
|
|
|
|
|
2017-03-15 06:06:33 +03:00
|
|
|
// For now our channels are in pending state, in order to not interfere
|
|
|
|
// with other tests we should clean up - complete opening of the
|
|
|
|
// channel and then close it.
|
2016-10-15 16:24:56 +03:00
|
|
|
|
2017-11-14 04:37:20 +03:00
|
|
|
// Mine 6 blocks, then wait for node's to notify us that the channel has
|
2017-03-15 06:06:33 +03:00
|
|
|
// been opened. The funding transactions should be found within the
|
2017-11-14 04:37:20 +03:00
|
|
|
// first newly mined block. 6 blocks make sure the funding transaction
|
2018-02-07 06:11:11 +03:00
|
|
|
// has enough confirmations to be announced publicly.
|
2018-11-29 14:33:11 +03:00
|
|
|
block := mineBlocks(t, net, 6, maxPendingChannels)[0]
|
2016-10-15 16:24:56 +03:00
|
|
|
|
|
|
|
chanPoints := make([]*lnrpc.ChannelPoint, maxPendingChannels)
|
|
|
|
for i, stream := range openStreams {
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2016-12-08 09:38:46 +03:00
|
|
|
fundingChanPoint, err := net.WaitForChannelOpen(ctxt, stream)
|
2016-10-15 16:24:56 +03:00
|
|
|
if err != nil {
|
2016-10-24 05:00:09 +03:00
|
|
|
t.Fatalf("error while waiting for channel open: %v", err)
|
2016-10-15 16:24:56 +03:00
|
|
|
}
|
|
|
|
|
2021-02-13 11:05:33 +03:00
|
|
|
fundingTxID, err := lnrpc.GetChanPointFundingTxid(fundingChanPoint)
|
2018-01-11 07:59:30 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get txid: %v", err)
|
|
|
|
}
|
2016-10-15 16:24:56 +03:00
|
|
|
|
2017-03-15 06:06:33 +03:00
|
|
|
// Ensure that the funding transaction enters a block, and is
|
|
|
|
// properly advertised by Alice.
|
2016-10-24 05:00:09 +03:00
|
|
|
assertTxInBlock(t, block, fundingTxID)
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2017-03-15 06:06:33 +03:00
|
|
|
err = net.Alice.WaitForNetworkChannelOpen(ctxt, fundingChanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("channel not seen on network before "+
|
|
|
|
"timeout: %v", err)
|
|
|
|
}
|
2016-10-15 16:24:56 +03:00
|
|
|
|
|
|
|
// The channel should be listed in the peer information
|
|
|
|
// returned by both peers.
|
|
|
|
chanPoint := wire.OutPoint{
|
|
|
|
Hash: *fundingTxID,
|
|
|
|
Index: fundingChanPoint.OutputIndex,
|
|
|
|
}
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-26 18:09:39 +03:00
|
|
|
if err := net.AssertChannelExists(ctxt, net.Alice, &chanPoint); err != nil {
|
2016-10-24 05:00:09 +03:00
|
|
|
t.Fatalf("unable to assert channel existence: %v", err)
|
2016-10-15 16:24:56 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
chanPoints[i] = fundingChanPoint
|
|
|
|
}
|
|
|
|
|
2017-01-07 23:48:15 +03:00
|
|
|
// Next, close the channel between Alice and Carol, asserting that the
|
|
|
|
// channel has been properly closed on-chain.
|
2016-10-15 16:24:56 +03:00
|
|
|
for _, chanPoint := range chanPoints {
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2017-02-23 01:49:04 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false)
|
2016-10-15 16:24:56 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-03-19 13:01:42 +03:00
|
|
|
// getNTxsFromMempool polls until finding the desired number of transactions in
|
|
|
|
// the provided miner's mempool and returns the full transactions to the caller.
|
|
|
|
func getNTxsFromMempool(miner *rpcclient.Client, n int,
|
|
|
|
timeout time.Duration) ([]*wire.MsgTx, error) {
|
|
|
|
|
|
|
|
txids, err := waitForNTxsInMempool(miner, n, timeout)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
var txes []*wire.MsgTx
|
|
|
|
for _, txid := range txids {
|
|
|
|
tx, err := miner.GetRawTransaction(txid)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
txes = append(txes, tx.MsgTx())
|
|
|
|
}
|
|
|
|
return txes, nil
|
|
|
|
}
|
|
|
|
|
2020-09-04 12:19:27 +03:00
|
|
|
// getTxFee retrieves parent transactions and reconstructs the fee paid.
|
|
|
|
func getTxFee(miner *rpcclient.Client, tx *wire.MsgTx) (btcutil.Amount, error) {
|
|
|
|
var balance btcutil.Amount
|
|
|
|
for _, in := range tx.TxIn {
|
|
|
|
parentHash := in.PreviousOutPoint.Hash
|
|
|
|
rawTx, err := miner.GetRawTransaction(&parentHash)
|
|
|
|
if err != nil {
|
|
|
|
return 0, err
|
|
|
|
}
|
|
|
|
parent := rawTx.MsgTx()
|
|
|
|
balance += btcutil.Amount(
|
|
|
|
parent.TxOut[in.PreviousOutPoint.Index].Value,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, out := range tx.TxOut {
|
|
|
|
balance -= btcutil.Amount(out.Value)
|
|
|
|
}
|
|
|
|
|
|
|
|
return balance, nil
|
|
|
|
}
|
|
|
|
|
2018-06-15 22:33:49 +03:00
|
|
|
// testGarbageCollectLinkNodes tests that we properly garbase collect link nodes
|
|
|
|
// from the database and the set of persistent connections within the server.
|
|
|
|
func testGarbageCollectLinkNodes(net *lntest.NetworkHarness, t *harnessTest) {
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2018-06-15 22:33:49 +03:00
|
|
|
const (
|
|
|
|
chanAmt = 1000000
|
|
|
|
)
|
|
|
|
|
|
|
|
// Open a channel between Alice and Bob which will later be
|
|
|
|
// cooperatively closed.
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-06-15 22:33:49 +03:00
|
|
|
coopChanPoint := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, net.Bob,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-06-15 22:33:49 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// Create Carol's node and connect Alice to her.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", nil)
|
2018-06-15 22:33:49 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, net.Alice, carol)
|
2018-06-15 22:33:49 +03:00
|
|
|
|
|
|
|
// Open a channel between Alice and Carol which will later be force
|
|
|
|
// closed.
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-06-15 22:33:49 +03:00
|
|
|
forceCloseChanPoint := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, carol,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-06-15 22:33:49 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// Now, create Dave's a node and also open a channel between Alice and
|
|
|
|
// him. This link will serve as the only persistent link throughout
|
|
|
|
// restarts in this test.
|
2021-06-07 23:05:12 +03:00
|
|
|
dave := net.NewNode(t.t, "Dave", nil)
|
2018-06-15 22:33:49 +03:00
|
|
|
defer shutdownAndAssert(net, t, dave)
|
2021-06-07 23:05:12 +03:00
|
|
|
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, net.Alice, dave)
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-06-15 22:33:49 +03:00
|
|
|
persistentChanPoint := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, dave,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-06-15 22:33:49 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// isConnected is a helper closure that checks if a peer is connected to
|
|
|
|
// Alice.
|
|
|
|
isConnected := func(pubKey string) bool {
|
|
|
|
req := &lnrpc.ListPeersRequest{}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
resp, err := net.Alice.ListPeers(ctxt, req)
|
2018-06-15 22:33:49 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to retrieve alice's peers: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, peer := range resp.Peers {
|
|
|
|
if peer.PubKey == pubKey {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restart both Bob and Carol to ensure Alice is able to reconnect to
|
|
|
|
// them.
|
|
|
|
if err := net.RestartNode(net.Bob, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart bob's node: %v", err)
|
|
|
|
}
|
|
|
|
if err := net.RestartNode(carol, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart carol's node: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-12-08 20:54:18 +03:00
|
|
|
require.Eventually(t.t, func() bool {
|
2018-06-15 22:33:49 +03:00
|
|
|
return isConnected(net.Bob.PubKeyStr)
|
2020-12-08 20:54:18 +03:00
|
|
|
}, defaultTimeout, 20*time.Millisecond)
|
|
|
|
require.Eventually(t.t, func() bool {
|
2018-06-15 22:33:49 +03:00
|
|
|
return isConnected(carol.PubKeyStr)
|
2020-12-08 20:54:18 +03:00
|
|
|
}, defaultTimeout, 20*time.Millisecond)
|
2018-06-15 22:33:49 +03:00
|
|
|
|
|
|
|
// We'll also restart Alice to ensure she can reconnect to her peers
|
|
|
|
// with open channels.
|
|
|
|
if err := net.RestartNode(net.Alice, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart alice's node: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-12-08 20:54:18 +03:00
|
|
|
require.Eventually(t.t, func() bool {
|
2018-06-15 22:33:49 +03:00
|
|
|
return isConnected(net.Bob.PubKeyStr)
|
2020-12-08 20:54:18 +03:00
|
|
|
}, defaultTimeout, 20*time.Millisecond)
|
|
|
|
require.Eventually(t.t, func() bool {
|
2018-06-15 22:33:49 +03:00
|
|
|
return isConnected(carol.PubKeyStr)
|
2020-12-08 20:54:18 +03:00
|
|
|
}, defaultTimeout, 20*time.Millisecond)
|
|
|
|
require.Eventually(t.t, func() bool {
|
|
|
|
return isConnected(dave.PubKeyStr)
|
|
|
|
}, defaultTimeout, 20*time.Millisecond)
|
2021-06-07 23:05:12 +03:00
|
|
|
err := wait.Predicate(func() bool {
|
2018-09-11 11:25:48 +03:00
|
|
|
return isConnected(dave.PubKeyStr)
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2018-06-15 22:33:49 +03:00
|
|
|
|
|
|
|
// testReconnection is a helper closure that restarts the nodes at both
|
|
|
|
// ends of a channel to ensure they do not reconnect after restarting.
|
|
|
|
// When restarting Alice, we'll first need to ensure she has
|
|
|
|
// reestablished her connection with Dave, as they still have an open
|
|
|
|
// channel together.
|
|
|
|
testReconnection := func(node *lntest.HarnessNode) {
|
2018-09-11 11:25:48 +03:00
|
|
|
// Restart both nodes, to trigger the pruning logic.
|
|
|
|
if err := net.RestartNode(node, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart %v's node: %v",
|
|
|
|
node.Name(), err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := net.RestartNode(net.Alice, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart alice's node: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now restart both nodes and make sure they don't reconnect.
|
2018-06-15 22:33:49 +03:00
|
|
|
if err := net.RestartNode(node, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart %v's node: %v", node.Name(),
|
|
|
|
err)
|
|
|
|
}
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Invariant(func() bool {
|
2018-06-15 22:33:49 +03:00
|
|
|
return !isConnected(node.PubKeyStr)
|
2018-09-11 11:25:48 +03:00
|
|
|
}, 5*time.Second)
|
2018-06-15 22:33:49 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice reconnected to %v", node.Name())
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := net.RestartNode(net.Alice, nil); err != nil {
|
|
|
|
t.Fatalf("unable to restart alice's node: %v", err)
|
|
|
|
}
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
2018-09-11 11:25:48 +03:00
|
|
|
return isConnected(dave.PubKeyStr)
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2018-09-11 11:25:48 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't reconnect to Dave")
|
|
|
|
}
|
|
|
|
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Invariant(func() bool {
|
2018-09-11 11:25:48 +03:00
|
|
|
return !isConnected(node.PubKeyStr)
|
|
|
|
}, 5*time.Second)
|
2018-06-15 22:33:49 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice reconnected to %v", node.Name())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now, we'll close the channel between Alice and Bob and ensure there
|
|
|
|
// is no reconnection logic between the both once the channel is fully
|
|
|
|
// closed.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-06-15 22:33:49 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, coopChanPoint, false)
|
|
|
|
|
|
|
|
testReconnection(net.Bob)
|
|
|
|
|
|
|
|
// We'll do the same with Alice and Carol, but this time we'll force
|
|
|
|
// close the channel instead.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-06-15 22:33:49 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, forceCloseChanPoint, true)
|
|
|
|
|
2018-11-06 13:03:01 +03:00
|
|
|
// Cleanup by mining the force close and sweep transaction.
|
|
|
|
cleanupForceClose(t, net, net.Alice, forceCloseChanPoint)
|
|
|
|
|
2018-06-15 22:33:49 +03:00
|
|
|
// We'll need to mine some blocks in order to mark the channel fully
|
|
|
|
// closed.
|
2021-03-10 01:12:26 +03:00
|
|
|
_, err = net.Miner.Client.Generate(chainreg.DefaultBitcoinTimeLockDelta - defaultCSV)
|
2018-06-15 22:33:49 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to generate blocks: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-09-07 02:57:28 +03:00
|
|
|
// Before we test reconnection, we'll ensure that the channel has been
|
|
|
|
// fully cleaned up for both Carol and Alice.
|
|
|
|
var predErr error
|
|
|
|
pendingChansRequest := &lnrpc.PendingChannelsRequest{}
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-09-07 02:57:28 +03:00
|
|
|
pendingChanResp, err := net.Alice.PendingChannels(
|
|
|
|
ctxt, pendingChansRequest,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
predErr = fmt.Errorf("unable to query for pending "+
|
|
|
|
"channels: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
predErr = checkNumForceClosedChannels(pendingChanResp, 0)
|
|
|
|
if predErr != nil {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-09-07 02:57:28 +03:00
|
|
|
pendingChanResp, err = carol.PendingChannels(
|
|
|
|
ctxt, pendingChansRequest,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
predErr = fmt.Errorf("unable to query for pending "+
|
|
|
|
"channels: %v", err)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
predErr = checkNumForceClosedChannels(pendingChanResp, 0)
|
|
|
|
|
2021-04-16 11:55:51 +03:00
|
|
|
return predErr == nil
|
|
|
|
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2018-09-07 02:57:28 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("channels not marked as fully resolved: %v", predErr)
|
|
|
|
}
|
|
|
|
|
2018-06-15 22:33:49 +03:00
|
|
|
testReconnection(carol)
|
|
|
|
|
|
|
|
// Finally, we'll ensure that Bob and Carol no longer show in Alice's
|
|
|
|
// channel graph.
|
2018-09-26 18:30:09 +03:00
|
|
|
describeGraphReq := &lnrpc.ChannelGraphRequest{
|
|
|
|
IncludeUnannounced: true,
|
|
|
|
}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
channelGraph, err := net.Alice.DescribeGraph(ctxt, describeGraphReq)
|
2018-06-15 22:33:49 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query for alice's channel graph: %v", err)
|
|
|
|
}
|
|
|
|
for _, node := range channelGraph.Nodes {
|
|
|
|
if node.PubKey == net.Bob.PubKeyStr {
|
|
|
|
t.Fatalf("did not expect to find bob in the channel " +
|
|
|
|
"graph, but did")
|
|
|
|
}
|
|
|
|
if node.PubKey == carol.PubKeyStr {
|
|
|
|
t.Fatalf("did not expect to find carol in the channel " +
|
|
|
|
"graph, but did")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now that the test is done, we can also close the persistent link.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-06-15 22:33:49 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, persistentChanPoint, false)
|
|
|
|
}
|
|
|
|
|
2018-07-12 12:02:54 +03:00
|
|
|
// testDataLossProtection tests that if one of the nodes in a channel
|
|
|
|
// relationship lost state, they will detect this during channel sync, and the
|
|
|
|
// up-to-date party will force close the channel, giving the outdated party the
|
2019-03-12 02:26:26 +03:00
|
|
|
// opportunity to sweep its output.
|
2018-07-12 12:02:54 +03:00
|
|
|
func testDataLossProtection(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
const (
|
2020-11-17 02:28:56 +03:00
|
|
|
chanAmt = funding.MaxBtcFundingAmount
|
2018-07-12 12:02:54 +03:00
|
|
|
paymentAmt = 10000
|
|
|
|
numInvoices = 6
|
|
|
|
)
|
|
|
|
|
|
|
|
// Carol will be the up-to-date party. We set --nolisten to ensure Dave
|
|
|
|
// won't be able to connect to her and trigger the channel data
|
2020-01-18 01:19:56 +03:00
|
|
|
// protection logic automatically. We also can't have Carol
|
|
|
|
// automatically re-connect too early, otherwise DLP would be initiated
|
|
|
|
// at the wrong moment.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(
|
|
|
|
t.t, "Carol", []string{"--nolisten", "--minbackoff=1h"},
|
2020-01-18 01:19:56 +03:00
|
|
|
)
|
2018-07-12 12:02:54 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
|
|
|
|
|
|
|
// Dave will be the party losing his state.
|
2021-06-07 23:05:12 +03:00
|
|
|
dave := net.NewNode(t.t, "Dave", nil)
|
2018-07-12 12:02:54 +03:00
|
|
|
defer shutdownAndAssert(net, t, dave)
|
|
|
|
|
|
|
|
// Before we make a channel, we'll load up Carol with some coins sent
|
|
|
|
// directly from the miner.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, carol)
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// timeTravel is a method that will make Carol open a channel to the
|
|
|
|
// passed node, settle a series of payments, then reset the node back
|
|
|
|
// to the state before the payments happened. When this method returns
|
|
|
|
// the node will be unaware of the new state updates. The returned
|
|
|
|
// function can be used to restart the node in this state.
|
|
|
|
timeTravel := func(node *lntest.HarnessNode) (func() error,
|
|
|
|
*lnrpc.ChannelPoint, int64, error) {
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// We must let the node communicate with Carol before they are
|
|
|
|
// able to open channel, so we connect them.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, carol, node)
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// We'll first open up a channel between them with a 0.5 BTC
|
|
|
|
// value.
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-11-20 17:09:44 +03:00
|
|
|
chanPoint := openChannelAndAssert(
|
|
|
|
ctxt, t, net, carol, node,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
|
|
|
)
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// With the channel open, we'll create a few invoices for the
|
|
|
|
// node that Carol will pay to in order to advance the state of
|
|
|
|
// the channel.
|
|
|
|
// TODO(halseth): have dangling HTLCs on the commitment, able to
|
2021-04-16 11:55:51 +03:00
|
|
|
// retrieve funds?
|
2018-11-20 17:09:44 +03:00
|
|
|
payReqs, _, _, err := createPayReqs(
|
2019-03-28 05:00:27 +03:00
|
|
|
node, paymentAmt, numInvoices,
|
2018-11-20 17:09:44 +03:00
|
|
|
)
|
2018-07-12 12:02:54 +03:00
|
|
|
if err != nil {
|
2018-11-20 17:09:44 +03:00
|
|
|
t.Fatalf("unable to create pay reqs: %v", err)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Wait for Carol to receive the channel edge from the funding
|
|
|
|
// manager.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-20 17:09:44 +03:00
|
|
|
err = carol.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("carol didn't see the carol->%s channel "+
|
|
|
|
"before timeout: %v", node.Name(), err)
|
|
|
|
}
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Send payments from Carol using 3 of the payment hashes
|
|
|
|
// generated above.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2020-03-31 13:44:18 +03:00
|
|
|
err = completePaymentRequests(
|
|
|
|
ctxt, carol, carol.RouterClient,
|
|
|
|
payReqs[:numInvoices/2], true,
|
|
|
|
)
|
2018-07-12 12:02:54 +03:00
|
|
|
if err != nil {
|
2018-11-20 17:09:44 +03:00
|
|
|
t.Fatalf("unable to send payments: %v", err)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
2018-11-20 17:09:44 +03:00
|
|
|
|
|
|
|
// Next query for the node's channel state, as we sent 3
|
|
|
|
// payments of 10k satoshis each, it should now see his balance
|
|
|
|
// as being 30k satoshis.
|
|
|
|
var nodeChan *lnrpc.Channel
|
|
|
|
var predErr error
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-20 17:09:44 +03:00
|
|
|
bChan, err := getChanInfo(ctxt, node)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get channel info: %v", err)
|
|
|
|
}
|
|
|
|
if bChan.LocalBalance != 30000 {
|
|
|
|
predErr = fmt.Errorf("balance is incorrect, "+
|
|
|
|
"got %v, expected %v",
|
|
|
|
bChan.LocalBalance, 30000)
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeChan = bChan
|
|
|
|
return true
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2018-11-20 17:09:44 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("%v", predErr)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Grab the current commitment height (update number), we'll
|
|
|
|
// later revert him to this state after additional updates to
|
|
|
|
// revoke this state.
|
|
|
|
stateNumPreCopy := nodeChan.NumUpdates
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// With the temporary file created, copy the current state into
|
|
|
|
// the temporary file we created above. Later after more
|
|
|
|
// updates, we'll restore this state.
|
2021-06-16 08:57:07 +03:00
|
|
|
if err := net.BackupDb(node); err != nil {
|
2018-11-20 17:09:44 +03:00
|
|
|
t.Fatalf("unable to copy database files: %v", err)
|
|
|
|
}
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Finally, send more payments from , using the remaining
|
|
|
|
// payment hashes.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2020-03-31 13:44:18 +03:00
|
|
|
err = completePaymentRequests(
|
|
|
|
ctxt, carol, carol.RouterClient,
|
|
|
|
payReqs[numInvoices/2:], true,
|
|
|
|
)
|
2018-07-12 12:02:54 +03:00
|
|
|
if err != nil {
|
2018-11-20 17:09:44 +03:00
|
|
|
t.Fatalf("unable to send payments: %v", err)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
2018-11-20 17:09:44 +03:00
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-20 17:09:44 +03:00
|
|
|
nodeChan, err = getChanInfo(ctxt, node)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get dave chan info: %v", err)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Now we shutdown the node, copying over the its temporary
|
|
|
|
// database state which has the *prior* channel state over his
|
|
|
|
// current most up to date state. With this, we essentially
|
|
|
|
// force the node to travel back in time within the channel's
|
|
|
|
// history.
|
|
|
|
if err = net.RestartNode(node, func() error {
|
2021-06-16 08:57:07 +03:00
|
|
|
return net.RestoreDb(node)
|
2018-11-20 17:09:44 +03:00
|
|
|
}); err != nil {
|
|
|
|
t.Fatalf("unable to restart node: %v", err)
|
|
|
|
}
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2019-02-06 23:47:41 +03:00
|
|
|
// Make sure the channel is still there from the PoV of the
|
|
|
|
// node.
|
|
|
|
assertNodeNumChannels(t, node, 1)
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Now query for the channel state, it should show that it's at
|
|
|
|
// a state number in the past, not the *latest* state.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-20 17:09:44 +03:00
|
|
|
nodeChan, err = getChanInfo(ctxt, node)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get dave chan info: %v", err)
|
|
|
|
}
|
|
|
|
if nodeChan.NumUpdates != stateNumPreCopy {
|
|
|
|
t.Fatalf("db copy failed: %v", nodeChan.NumUpdates)
|
|
|
|
}
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
balReq := &lnrpc.WalletBalanceRequest{}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
balResp, err := node.WalletBalance(ctxt, balReq)
|
2018-11-20 17:09:44 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get dave's balance: %v", err)
|
|
|
|
}
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
restart, err := net.SuspendNode(node)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to suspend node: %v", err)
|
|
|
|
}
|
2020-03-14 04:52:33 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
return restart, chanPoint, balResp.ConfirmedBalance, nil
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Reset Dave to a state where he has an outdated channel state.
|
|
|
|
restartDave, _, daveStartingBalance, err := timeTravel(dave)
|
2018-07-12 12:02:54 +03:00
|
|
|
if err != nil {
|
2018-11-20 17:09:44 +03:00
|
|
|
t.Fatalf("unable to time travel dave: %v", err)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
|
|
|
|
2019-03-12 02:26:26 +03:00
|
|
|
// We make a note of the nodes' current on-chain balances, to make sure
|
|
|
|
// they are able to retrieve the channel funds eventually,
|
2018-11-20 17:09:44 +03:00
|
|
|
balReq := &lnrpc.WalletBalanceRequest{}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
carolBalResp, err := carol.WalletBalance(ctxt, balReq)
|
2018-07-12 12:02:54 +03:00
|
|
|
if err != nil {
|
2018-11-20 17:09:44 +03:00
|
|
|
t.Fatalf("unable to get carol's balance: %v", err)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
2018-11-20 17:09:44 +03:00
|
|
|
carolStartingBalance := carolBalResp.ConfirmedBalance
|
2018-07-12 12:02:54 +03:00
|
|
|
|
2018-11-20 17:09:44 +03:00
|
|
|
// Restart Dave to trigger a channel resync.
|
|
|
|
if err := restartDave(); err != nil {
|
|
|
|
t.Fatalf("unable to restart dave: %v", err)
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
|
|
|
|
2019-03-12 02:28:47 +03:00
|
|
|
// Assert that once Dave comes up, they reconnect, Carol force closes
|
|
|
|
// on chain, and both of them properly carry out the DLP protocol.
|
|
|
|
assertDLPExecuted(
|
|
|
|
net, t, carol, carolStartingBalance, dave, daveStartingBalance,
|
2020-03-10 16:01:39 +03:00
|
|
|
false,
|
2019-03-12 02:28:47 +03:00
|
|
|
)
|
2018-11-20 17:09:46 +03:00
|
|
|
|
2019-03-12 02:28:47 +03:00
|
|
|
// As a second part of this test, we will test the scenario where a
|
2018-11-20 17:09:46 +03:00
|
|
|
// channel is closed while Dave is offline, loses his state and comes
|
|
|
|
// back online. In this case the node should attempt to resync the
|
|
|
|
// channel, and the peer should resend a channel sync message for the
|
|
|
|
// closed channel, such that Dave can retrieve his funds.
|
|
|
|
//
|
|
|
|
// We start by letting Dave time travel back to an outdated state.
|
|
|
|
restartDave, chanPoint2, daveStartingBalance, err := timeTravel(dave)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to time travel eve: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
carolBalResp, err = carol.WalletBalance(ctxt, balReq)
|
2018-11-20 17:09:46 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get carol's balance: %v", err)
|
|
|
|
}
|
|
|
|
carolStartingBalance = carolBalResp.ConfirmedBalance
|
|
|
|
|
|
|
|
// Now let Carol force close the channel while Dave is offline.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-11-20 17:09:46 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, carol, chanPoint2, true)
|
|
|
|
|
|
|
|
// Wait for the channel to be marked pending force close.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-20 17:09:46 +03:00
|
|
|
err = waitForChannelPendingForceClose(ctxt, carol, chanPoint2)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("channel not pending force close: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Mine enough blocks for Carol to sweep her funds.
|
2020-03-09 20:43:47 +03:00
|
|
|
mineBlocks(t, net, defaultCSV-1, 0)
|
2018-11-20 17:09:46 +03:00
|
|
|
|
2021-03-10 01:12:26 +03:00
|
|
|
carolSweep, err := waitForTxInMempool(net.Miner.Client, minerMempoolTimeout)
|
2018-11-20 17:09:46 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to find Carol's sweep tx in mempool: %v", err)
|
|
|
|
}
|
2019-03-12 02:28:47 +03:00
|
|
|
block := mineBlocks(t, net, 1, 1)[0]
|
2018-11-20 17:09:46 +03:00
|
|
|
assertTxInBlock(t, block, carolSweep)
|
|
|
|
|
|
|
|
// Now the channel should be fully closed also from Carol's POV.
|
|
|
|
assertNumPendingChannels(t, carol, 0, 0)
|
|
|
|
|
|
|
|
// Make sure Carol got her balance back.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
carolBalResp, err = carol.WalletBalance(ctxt, balReq)
|
2018-11-20 17:09:46 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get carol's balance: %v", err)
|
|
|
|
}
|
2019-03-12 02:28:47 +03:00
|
|
|
carolBalance := carolBalResp.ConfirmedBalance
|
2018-11-20 17:09:46 +03:00
|
|
|
if carolBalance <= carolStartingBalance {
|
|
|
|
t.Fatalf("expected carol to have balance above %d, "+
|
|
|
|
"instead had %v", carolStartingBalance,
|
|
|
|
carolBalance)
|
|
|
|
}
|
|
|
|
|
2018-11-28 11:25:27 +03:00
|
|
|
assertNodeNumChannels(t, carol, 0)
|
2018-11-20 17:09:46 +03:00
|
|
|
|
|
|
|
// When Dave comes online, he will reconnect to Carol, try to resync
|
|
|
|
// the channel, but it will already be closed. Carol should resend the
|
|
|
|
// information Dave needs to sweep his funds.
|
|
|
|
if err := restartDave(); err != nil {
|
2019-05-05 01:35:37 +03:00
|
|
|
t.Fatalf("unable to restart Eve: %v", err)
|
2018-11-20 17:09:46 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Dave should sweep his funds.
|
2021-03-10 01:12:26 +03:00
|
|
|
_, err = waitForTxInMempool(net.Miner.Client, minerMempoolTimeout)
|
2018-11-20 17:09:46 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to find Dave's sweep tx in mempool: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Mine a block to confirm the sweep, and make sure Dave got his
|
|
|
|
// balance back.
|
2018-11-29 14:33:11 +03:00
|
|
|
mineBlocks(t, net, 1, 1)
|
2018-11-28 11:25:27 +03:00
|
|
|
assertNodeNumChannels(t, dave, 0)
|
2018-11-20 17:09:46 +03:00
|
|
|
|
2019-09-19 22:46:29 +03:00
|
|
|
err = wait.NoError(func() error {
|
2019-04-23 14:07:43 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
daveBalResp, err := dave.WalletBalance(ctxt, balReq)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("unable to get dave's balance: %v",
|
|
|
|
err)
|
|
|
|
}
|
2018-11-20 17:09:46 +03:00
|
|
|
|
2019-04-23 14:07:43 +03:00
|
|
|
daveBalance := daveBalResp.ConfirmedBalance
|
|
|
|
if daveBalance <= daveStartingBalance {
|
|
|
|
return fmt.Errorf("expected dave to have balance "+
|
|
|
|
"above %d, intead had %v", daveStartingBalance,
|
|
|
|
daveBalance)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
2020-12-08 18:27:01 +03:00
|
|
|
}, defaultTimeout)
|
2019-04-23 14:07:43 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("%v", err)
|
2018-11-20 17:09:46 +03:00
|
|
|
}
|
2018-07-12 12:02:54 +03:00
|
|
|
}
|
|
|
|
|
2018-11-19 20:37:50 +03:00
|
|
|
// testRejectHTLC tests that a node can be created with the flag --rejecthtlc.
|
|
|
|
// This means that the node will reject all forwarded HTLCs but can still
|
|
|
|
// accept direct HTLCs as well as send HTLCs.
|
|
|
|
func testRejectHTLC(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
// RejectHTLC
|
|
|
|
// Alice ------> Carol ------> Bob
|
|
|
|
//
|
|
|
|
const chanAmt = btcutil.Amount(1000000)
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
// Create Carol with reject htlc flag.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", []string{"--rejecthtlc"})
|
2018-11-19 20:37:50 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
|
|
|
|
|
|
|
// Connect Alice to Carol.
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxb, t.t, net.Alice, carol)
|
2018-11-19 20:37:50 +03:00
|
|
|
|
|
|
|
// Connect Carol to Bob.
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxb, t.t, carol, net.Bob)
|
2018-11-19 20:37:50 +03:00
|
|
|
|
|
|
|
// Send coins to Carol.
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxb, t.t, btcutil.SatoshiPerBitcoin, carol)
|
2018-11-19 20:37:50 +03:00
|
|
|
|
|
|
|
// Send coins to Alice.
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxb, t.t, btcutil.SatoshiPerBitcent, net.Alice)
|
2018-11-19 20:37:50 +03:00
|
|
|
|
|
|
|
// Open a channel between Alice and Carol.
|
2019-09-02 13:05:51 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-11-19 20:37:50 +03:00
|
|
|
chanPointAlice := openChannelAndAssert(
|
|
|
|
ctxt, t, net, net.Alice, carol,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Open a channel between Carol and Bob.
|
2019-09-02 13:05:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-11-19 20:37:50 +03:00
|
|
|
chanPointCarol := openChannelAndAssert(
|
|
|
|
ctxt, t, net, carol, net.Bob,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Channel should be ready for payments.
|
|
|
|
const payAmt = 100
|
|
|
|
|
|
|
|
// Helper closure to generate a random pre image.
|
|
|
|
genPreImage := func() []byte {
|
|
|
|
preimage := make([]byte, 32)
|
|
|
|
|
2021-06-09 20:29:22 +03:00
|
|
|
_, err := rand.Read(preimage)
|
2018-11-19 20:37:50 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to generate preimage: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return preimage
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create an invoice from Carol of 100 satoshis.
|
|
|
|
// We expect Alice to be able to pay this invoice.
|
|
|
|
preimage := genPreImage()
|
|
|
|
|
|
|
|
carolInvoice := &lnrpc.Invoice{
|
|
|
|
Memo: "testing - alice should pay carol",
|
|
|
|
RPreimage: preimage,
|
|
|
|
Value: payAmt,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Carol adds the invoice to her database.
|
|
|
|
resp, err := carol.AddInvoice(ctxb, carolInvoice)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to add invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Alice pays Carols invoice.
|
2019-09-02 13:05:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-19 20:37:50 +03:00
|
|
|
err = completePaymentRequests(
|
2020-03-31 13:44:18 +03:00
|
|
|
ctxt, net.Alice, net.Alice.RouterClient,
|
|
|
|
[]string{resp.PaymentRequest}, true,
|
2018-11-19 20:37:50 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payments from alice to carol: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create an invoice from Bob of 100 satoshis.
|
|
|
|
// We expect Carol to be able to pay this invoice.
|
|
|
|
preimage = genPreImage()
|
|
|
|
|
|
|
|
bobInvoice := &lnrpc.Invoice{
|
|
|
|
Memo: "testing - carol should pay bob",
|
|
|
|
RPreimage: preimage,
|
|
|
|
Value: payAmt,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Bob adds the invoice to his database.
|
|
|
|
resp, err = net.Bob.AddInvoice(ctxb, bobInvoice)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to add invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Carol pays Bobs invoice.
|
2019-09-02 13:05:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-19 20:37:50 +03:00
|
|
|
err = completePaymentRequests(
|
2020-03-31 13:44:18 +03:00
|
|
|
ctxt, carol, carol.RouterClient,
|
|
|
|
[]string{resp.PaymentRequest}, true,
|
2018-11-19 20:37:50 +03:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to send payments from carol to bob: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create an invoice from Bob of 100 satoshis.
|
|
|
|
// Alice attempts to pay Bob but this should fail, since we are
|
|
|
|
// using Carol as a hop and her node will reject onward HTLCs.
|
|
|
|
preimage = genPreImage()
|
|
|
|
|
|
|
|
bobInvoice = &lnrpc.Invoice{
|
|
|
|
Memo: "testing - alice tries to pay bob",
|
|
|
|
RPreimage: preimage,
|
|
|
|
Value: payAmt,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Bob adds the invoice to his database.
|
|
|
|
resp, err = net.Bob.AddInvoice(ctxb, bobInvoice)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to add invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Alice attempts to pay Bobs invoice. This payment should be rejected since
|
|
|
|
// we are using Carol as an intermediary hop, Carol is running lnd with
|
|
|
|
// --rejecthtlc.
|
2019-09-02 13:05:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-11-19 20:37:50 +03:00
|
|
|
err = completePaymentRequests(
|
2020-03-31 13:44:18 +03:00
|
|
|
ctxt, net.Alice, net.Alice.RouterClient,
|
|
|
|
[]string{resp.PaymentRequest}, true,
|
2018-11-19 20:37:50 +03:00
|
|
|
)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf(
|
|
|
|
"should have been rejected, carol will not accept forwarded htlcs",
|
|
|
|
)
|
|
|
|
}
|
2020-04-01 01:13:25 +03:00
|
|
|
|
|
|
|
assertLastHTLCError(t, net.Alice, lnrpc.Failure_CHANNEL_DISABLED)
|
2018-11-19 20:37:50 +03:00
|
|
|
|
|
|
|
// Close all channels.
|
2019-09-02 13:05:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-11-19 20:37:50 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, chanPointAlice, false)
|
2019-09-02 13:05:51 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-11-19 20:37:50 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, carol, chanPointCarol, false)
|
|
|
|
}
|
|
|
|
|
2017-12-22 15:49:49 +03:00
|
|
|
func testGraphTopologyNotifications(net *lntest.NetworkHarness, t *harnessTest) {
|
2021-01-29 11:14:38 +03:00
|
|
|
t.t.Run("pinned", func(t *testing.T) {
|
|
|
|
ht := newHarnessTest(t, net)
|
|
|
|
testGraphTopologyNtfns(net, ht, true)
|
|
|
|
})
|
|
|
|
t.t.Run("unpinned", func(t *testing.T) {
|
|
|
|
ht := newHarnessTest(t, net)
|
|
|
|
testGraphTopologyNtfns(net, ht, false)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func testGraphTopologyNtfns(net *lntest.NetworkHarness, t *harnessTest, pinned bool) {
|
2017-12-22 15:49:49 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2020-11-17 02:28:56 +03:00
|
|
|
const chanAmt = funding.MaxBtcFundingAmount
|
2018-11-28 11:25:46 +03:00
|
|
|
|
2021-01-29 11:14:38 +03:00
|
|
|
// Spin up Bob first, since we will need to grab his pubkey when
|
|
|
|
// starting Alice to test pinned syncing.
|
2021-06-07 23:05:12 +03:00
|
|
|
bob := net.NewNode(t.t, "bob", nil)
|
2021-01-29 11:14:29 +03:00
|
|
|
defer shutdownAndAssert(net, t, bob)
|
|
|
|
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-01-29 11:14:38 +03:00
|
|
|
bobInfo, err := bob.GetInfo(ctxt, &lnrpc.GetInfoRequest{})
|
|
|
|
require.NoError(t.t, err)
|
|
|
|
bobPubkey := bobInfo.IdentityPubkey
|
|
|
|
|
|
|
|
// For unpinned syncing, start Alice as usual. Otherwise grab Bob's
|
|
|
|
// pubkey to include in his pinned syncer set.
|
|
|
|
var aliceArgs []string
|
|
|
|
if pinned {
|
|
|
|
aliceArgs = []string{
|
2021-01-23 00:01:29 +03:00
|
|
|
"--numgraphsyncpeers=0",
|
2021-01-29 11:14:38 +03:00
|
|
|
fmt.Sprintf("--gossip.pinned-syncers=%s", bobPubkey),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
alice := net.NewNode(t.t, "alice", aliceArgs)
|
2021-01-29 11:14:38 +03:00
|
|
|
defer shutdownAndAssert(net, t, alice)
|
|
|
|
|
|
|
|
// Connect Alice and Bob.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, alice, bob)
|
2021-01-29 11:14:29 +03:00
|
|
|
|
|
|
|
// Alice stimmy.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, alice)
|
2021-01-29 11:14:29 +03:00
|
|
|
|
|
|
|
// Bob stimmy.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, bob)
|
2021-01-29 11:14:29 +03:00
|
|
|
|
2021-01-29 11:14:38 +03:00
|
|
|
// Assert that Bob has the correct sync type before proceeeding.
|
|
|
|
if pinned {
|
|
|
|
assertSyncType(t, alice, bobPubkey, lnrpc.Peer_PINNED_SYNC)
|
|
|
|
} else {
|
|
|
|
assertSyncType(t, alice, bobPubkey, lnrpc.Peer_ACTIVE_SYNC)
|
|
|
|
}
|
|
|
|
|
2021-01-29 11:14:50 +03:00
|
|
|
// Regardless of syncer type, ensure that both peers report having
|
|
|
|
// completed their initial sync before continuing to make a channel.
|
|
|
|
waitForGraphSync(t, alice)
|
|
|
|
|
2017-12-22 15:49:49 +03:00
|
|
|
// Let Alice subscribe to graph notifications.
|
2021-04-16 11:55:51 +03:00
|
|
|
graphSub := subscribeGraphNotifications(ctxb, t, alice)
|
2018-08-23 08:16:16 +03:00
|
|
|
defer close(graphSub.quit)
|
2017-12-22 15:49:49 +03:00
|
|
|
|
|
|
|
// Open a new channel between Alice and Bob.
|
2021-01-29 11:14:29 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoint := openChannelAndAssert(
|
2021-01-29 11:14:29 +03:00
|
|
|
ctxt, t, net, alice, bob,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2017-03-14 08:23:19 +03:00
|
|
|
|
2018-02-07 06:11:11 +03:00
|
|
|
// The channel opening above should have triggered a few notifications
|
2017-07-14 22:47:52 +03:00
|
|
|
// sent to the notification client. We'll expect two channel updates,
|
|
|
|
// and two node announcements.
|
2018-10-02 02:45:11 +03:00
|
|
|
var numChannelUpds int
|
|
|
|
var numNodeAnns int
|
|
|
|
for numChannelUpds < 2 && numNodeAnns < 2 {
|
2017-03-14 08:23:19 +03:00
|
|
|
select {
|
|
|
|
// Ensure that a new update for both created edges is properly
|
|
|
|
// dispatched to our registered client.
|
2018-08-23 08:16:16 +03:00
|
|
|
case graphUpdate := <-graphSub.updateChan:
|
2018-10-02 02:45:11 +03:00
|
|
|
// Process all channel updates prsented in this update
|
|
|
|
// message.
|
|
|
|
for _, chanUpdate := range graphUpdate.ChannelUpdates {
|
2017-07-14 22:47:52 +03:00
|
|
|
switch chanUpdate.AdvertisingNode {
|
2021-01-29 11:14:29 +03:00
|
|
|
case alice.PubKeyStr:
|
|
|
|
case bob.PubKeyStr:
|
2017-07-14 22:47:52 +03:00
|
|
|
default:
|
|
|
|
t.Fatalf("unknown advertising node: %v",
|
|
|
|
chanUpdate.AdvertisingNode)
|
|
|
|
}
|
|
|
|
switch chanUpdate.ConnectingNode {
|
2021-01-29 11:14:29 +03:00
|
|
|
case alice.PubKeyStr:
|
|
|
|
case bob.PubKeyStr:
|
2017-07-14 22:47:52 +03:00
|
|
|
default:
|
|
|
|
t.Fatalf("unknown connecting node: %v",
|
|
|
|
chanUpdate.ConnectingNode)
|
|
|
|
}
|
2018-10-02 02:45:11 +03:00
|
|
|
|
|
|
|
if chanUpdate.Capacity != int64(chanAmt) {
|
|
|
|
t.Fatalf("channel capacities mismatch:"+
|
|
|
|
" expected %v, got %v", chanAmt,
|
|
|
|
btcutil.Amount(chanUpdate.Capacity))
|
|
|
|
}
|
|
|
|
numChannelUpds++
|
2017-03-14 08:23:19 +03:00
|
|
|
}
|
|
|
|
|
2018-10-02 02:45:11 +03:00
|
|
|
for _, nodeUpdate := range graphUpdate.NodeUpdates {
|
2017-07-14 22:47:52 +03:00
|
|
|
switch nodeUpdate.IdentityKey {
|
2021-01-29 11:14:29 +03:00
|
|
|
case alice.PubKeyStr:
|
|
|
|
case bob.PubKeyStr:
|
2017-07-14 22:47:52 +03:00
|
|
|
default:
|
|
|
|
t.Fatalf("unknown node: %v",
|
|
|
|
nodeUpdate.IdentityKey)
|
|
|
|
}
|
2018-10-02 02:45:11 +03:00
|
|
|
numNodeAnns++
|
2017-03-14 08:23:19 +03:00
|
|
|
}
|
2018-08-23 08:16:16 +03:00
|
|
|
case err := <-graphSub.errChan:
|
|
|
|
t.Fatalf("unable to recv graph update: %v", err)
|
2017-03-15 06:06:33 +03:00
|
|
|
case <-time.After(time.Second * 10):
|
2018-10-02 02:45:11 +03:00
|
|
|
t.Fatalf("timeout waiting for graph notifications, "+
|
|
|
|
"only received %d/2 chanupds and %d/2 nodeanns",
|
|
|
|
numChannelUpds, numNodeAnns)
|
2017-03-14 08:23:19 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-03-10 01:12:26 +03:00
|
|
|
_, blockHeight, err := net.Miner.Client.GetBestBlock()
|
2017-03-14 08:23:19 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get current blockheight %v", err)
|
|
|
|
}
|
|
|
|
|
2017-10-26 01:32:00 +03:00
|
|
|
// Now we'll test that updates are properly sent after channels are closed
|
|
|
|
// within the network.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2021-01-29 11:14:29 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, alice, chanPoint, false)
|
2017-03-14 08:23:19 +03:00
|
|
|
|
2018-06-15 04:16:28 +03:00
|
|
|
// Now that the channel has been closed, we should receive a
|
|
|
|
// notification indicating so.
|
|
|
|
out:
|
|
|
|
for {
|
|
|
|
select {
|
2018-08-23 08:16:16 +03:00
|
|
|
case graphUpdate := <-graphSub.updateChan:
|
2018-06-15 04:16:28 +03:00
|
|
|
if len(graphUpdate.ClosedChans) != 1 {
|
|
|
|
continue
|
|
|
|
}
|
2017-03-14 08:23:19 +03:00
|
|
|
|
2018-06-15 04:16:28 +03:00
|
|
|
closedChan := graphUpdate.ClosedChans[0]
|
|
|
|
if closedChan.ClosedHeight != uint32(blockHeight+1) {
|
|
|
|
t.Fatalf("close heights of channel mismatch: "+
|
|
|
|
"expected %v, got %v", blockHeight+1,
|
|
|
|
closedChan.ClosedHeight)
|
|
|
|
}
|
2021-02-13 11:05:33 +03:00
|
|
|
chanPointTxid, err := lnrpc.GetChanPointFundingTxid(chanPoint)
|
2018-06-15 04:16:28 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get txid: %v", err)
|
|
|
|
}
|
2021-02-13 11:05:33 +03:00
|
|
|
closedChanTxid, err := lnrpc.GetChanPointFundingTxid(
|
2018-06-15 04:16:28 +03:00
|
|
|
closedChan.ChanPoint,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get txid: %v", err)
|
|
|
|
}
|
2019-03-12 02:28:47 +03:00
|
|
|
if !bytes.Equal(closedChanTxid[:], chanPointTxid[:]) {
|
2018-06-15 04:16:28 +03:00
|
|
|
t.Fatalf("channel point hash mismatch: "+
|
|
|
|
"expected %v, got %v", chanPointTxid,
|
|
|
|
closedChanTxid)
|
|
|
|
}
|
|
|
|
if closedChan.ChanPoint.OutputIndex != chanPoint.OutputIndex {
|
|
|
|
t.Fatalf("output index mismatch: expected %v, "+
|
|
|
|
"got %v", chanPoint.OutputIndex,
|
|
|
|
closedChan.ChanPoint)
|
|
|
|
}
|
|
|
|
|
|
|
|
break out
|
2018-08-23 08:16:16 +03:00
|
|
|
|
|
|
|
case err := <-graphSub.errChan:
|
|
|
|
t.Fatalf("unable to recv graph update: %v", err)
|
2018-06-15 04:16:28 +03:00
|
|
|
case <-time.After(time.Second * 10):
|
|
|
|
t.Fatalf("notification for channel closure not " +
|
|
|
|
"sent")
|
2017-03-14 08:23:19 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// For the final portion of the test, we'll ensure that once a new node
|
2017-07-14 22:47:52 +03:00
|
|
|
// appears in the network, the proper notification is dispatched. Note
|
|
|
|
// that a node that does not have any channels open is ignored, so first
|
|
|
|
// we disconnect Alice and Bob, open a channel between Bob and Carol,
|
|
|
|
// and finally connect Alice to Bob again.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-01-29 11:14:29 +03:00
|
|
|
if err := net.DisconnectNodes(ctxt, alice, bob); err != nil {
|
2017-07-14 22:47:52 +03:00
|
|
|
t.Fatalf("unable to disconnect alice and bob: %v", err)
|
|
|
|
}
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", nil)
|
2018-06-08 00:14:39 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
2017-03-14 08:23:19 +03:00
|
|
|
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, bob, carol)
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoint = openChannelAndAssert(
|
2021-01-29 11:14:29 +03:00
|
|
|
ctxt, t, net, bob, carol,
|
2018-08-22 10:32:42 +03:00
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2017-03-14 08:23:19 +03:00
|
|
|
|
2017-07-14 22:47:52 +03:00
|
|
|
// Reconnect Alice and Bob. This should result in the nodes syncing up
|
|
|
|
// their respective graph state, with the new addition being the
|
|
|
|
// existence of Carol in the graph, and also the channel between Bob
|
|
|
|
// and Carol. Note that we will also receive a node announcement from
|
|
|
|
// Bob, since a node will update its node announcement after a new
|
|
|
|
// channel is opened.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, alice, bob)
|
2017-03-14 08:23:19 +03:00
|
|
|
|
2017-07-14 22:47:52 +03:00
|
|
|
// We should receive an update advertising the newly connected node,
|
|
|
|
// Bob's new node announcement, and the channel between Bob and Carol.
|
2018-10-02 02:45:11 +03:00
|
|
|
numNodeAnns = 0
|
|
|
|
numChannelUpds = 0
|
|
|
|
for numChannelUpds < 2 && numNodeAnns < 1 {
|
2017-07-14 22:47:52 +03:00
|
|
|
select {
|
2018-08-23 08:16:16 +03:00
|
|
|
case graphUpdate := <-graphSub.updateChan:
|
2018-10-02 02:45:11 +03:00
|
|
|
for _, nodeUpdate := range graphUpdate.NodeUpdates {
|
2017-07-14 22:47:52 +03:00
|
|
|
switch nodeUpdate.IdentityKey {
|
|
|
|
case carol.PubKeyStr:
|
2021-01-29 11:14:29 +03:00
|
|
|
case bob.PubKeyStr:
|
2017-07-14 22:47:52 +03:00
|
|
|
default:
|
|
|
|
t.Fatalf("unknown node update pubey: %v",
|
|
|
|
nodeUpdate.IdentityKey)
|
|
|
|
}
|
2018-10-02 02:45:11 +03:00
|
|
|
numNodeAnns++
|
2017-07-14 22:47:52 +03:00
|
|
|
}
|
|
|
|
|
2018-10-02 02:45:11 +03:00
|
|
|
for _, chanUpdate := range graphUpdate.ChannelUpdates {
|
2017-07-14 22:47:52 +03:00
|
|
|
switch chanUpdate.AdvertisingNode {
|
|
|
|
case carol.PubKeyStr:
|
2021-01-29 11:14:29 +03:00
|
|
|
case bob.PubKeyStr:
|
2017-07-14 22:47:52 +03:00
|
|
|
default:
|
|
|
|
t.Fatalf("unknown advertising node: %v",
|
|
|
|
chanUpdate.AdvertisingNode)
|
|
|
|
}
|
|
|
|
switch chanUpdate.ConnectingNode {
|
|
|
|
case carol.PubKeyStr:
|
2021-01-29 11:14:29 +03:00
|
|
|
case bob.PubKeyStr:
|
2017-07-14 22:47:52 +03:00
|
|
|
default:
|
|
|
|
t.Fatalf("unknown connecting node: %v",
|
|
|
|
chanUpdate.ConnectingNode)
|
|
|
|
}
|
2018-10-02 02:45:11 +03:00
|
|
|
|
|
|
|
if chanUpdate.Capacity != int64(chanAmt) {
|
|
|
|
t.Fatalf("channel capacities mismatch:"+
|
|
|
|
" expected %v, got %v", chanAmt,
|
|
|
|
btcutil.Amount(chanUpdate.Capacity))
|
|
|
|
}
|
|
|
|
numChannelUpds++
|
2017-07-14 22:47:52 +03:00
|
|
|
}
|
2018-08-23 08:16:16 +03:00
|
|
|
case err := <-graphSub.errChan:
|
|
|
|
t.Fatalf("unable to recv graph update: %v", err)
|
2017-07-14 22:47:52 +03:00
|
|
|
case <-time.After(time.Second * 10):
|
2018-10-02 02:45:11 +03:00
|
|
|
t.Fatalf("timeout waiting for graph notifications, "+
|
|
|
|
"only received %d/2 chanupds and %d/2 nodeanns",
|
|
|
|
numChannelUpds, numNodeAnns)
|
2017-03-14 08:23:19 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-07-14 22:47:52 +03:00
|
|
|
// Close the channel between Bob and Carol.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2021-01-29 11:14:29 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, bob, chanPoint, false)
|
2017-03-14 08:23:19 +03:00
|
|
|
}
|
|
|
|
|
2017-02-23 03:24:22 +03:00
|
|
|
// testNodeAnnouncement ensures that when a node is started with one or more
|
|
|
|
// external IP addresses specified on the command line, that those addresses
|
|
|
|
// announced to the network and reported in the network graph.
|
2017-11-03 21:52:02 +03:00
|
|
|
func testNodeAnnouncement(net *lntest.NetworkHarness, t *harnessTest) {
|
2017-02-23 03:24:22 +03:00
|
|
|
ctxb := context.Background()
|
2018-11-28 11:25:46 +03:00
|
|
|
|
2021-04-16 11:55:51 +03:00
|
|
|
aliceSub := subscribeGraphNotifications(ctxb, t, net.Alice)
|
2018-08-23 08:16:16 +03:00
|
|
|
defer close(aliceSub.quit)
|
2017-02-23 03:24:22 +03:00
|
|
|
|
2018-05-24 09:54:12 +03:00
|
|
|
advertisedAddrs := []string{
|
|
|
|
"192.168.1.1:8333",
|
|
|
|
"[2001:db8:85a3:8d3:1319:8a2e:370:7348]:8337",
|
2018-09-21 05:23:34 +03:00
|
|
|
"bkb6azqggsaiskzi.onion:9735",
|
|
|
|
"fomvuglh6h6vcag73xo5t5gv56ombih3zr2xvplkpbfd7wrog4swjwid.onion:1234",
|
2017-02-23 03:24:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
var lndArgs []string
|
2018-05-24 09:54:12 +03:00
|
|
|
for _, addr := range advertisedAddrs {
|
|
|
|
lndArgs = append(lndArgs, "--externalip="+addr)
|
2017-02-23 03:24:22 +03:00
|
|
|
}
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
dave := net.NewNode(t.t, "Dave", lndArgs)
|
2018-06-08 00:14:39 +03:00
|
|
|
defer shutdownAndAssert(net, t, dave)
|
2017-02-23 03:24:22 +03:00
|
|
|
|
2017-07-14 22:47:52 +03:00
|
|
|
// We must let Dave have an open channel before he can send a node
|
|
|
|
// announcement, so we open a channel with Bob,
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, net.Bob, dave)
|
2017-07-14 22:47:52 +03:00
|
|
|
|
2019-04-25 02:06:14 +03:00
|
|
|
// Alice shouldn't receive any new updates yet since the channel has yet
|
|
|
|
// to be opened.
|
|
|
|
select {
|
|
|
|
case <-aliceSub.updateChan:
|
|
|
|
t.Fatalf("received unexpected update from dave")
|
|
|
|
case <-time.After(time.Second):
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll then go ahead and open a channel between Bob and Dave. This
|
|
|
|
// ensures that Alice receives the node announcement from Bob as part of
|
|
|
|
// the announcement broadcast.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
chanPoint := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Bob, dave,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: 1000000,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2017-07-14 22:47:52 +03:00
|
|
|
|
2018-05-24 09:54:12 +03:00
|
|
|
assertAddrs := func(addrsFound []string, targetAddrs ...string) {
|
|
|
|
addrs := make(map[string]struct{}, len(addrsFound))
|
|
|
|
for _, addr := range addrsFound {
|
|
|
|
addrs[addr] = struct{}{}
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, addr := range targetAddrs {
|
|
|
|
if _, ok := addrs[addr]; !ok {
|
|
|
|
t.Fatalf("address %v not found in node "+
|
|
|
|
"announcement", addr)
|
|
|
|
}
|
|
|
|
}
|
2017-02-23 03:24:22 +03:00
|
|
|
}
|
|
|
|
|
2018-08-23 08:16:16 +03:00
|
|
|
waitForAddrsInUpdate := func(graphSub graphSubscription,
|
2018-05-24 09:54:12 +03:00
|
|
|
nodePubKey string, targetAddrs ...string) {
|
2017-02-23 03:24:22 +03:00
|
|
|
|
2018-05-24 09:54:12 +03:00
|
|
|
for {
|
|
|
|
select {
|
2018-08-23 08:16:16 +03:00
|
|
|
case graphUpdate := <-graphSub.updateChan:
|
2018-05-24 09:54:12 +03:00
|
|
|
for _, update := range graphUpdate.NodeUpdates {
|
|
|
|
if update.IdentityKey == nodePubKey {
|
|
|
|
assertAddrs(
|
2021-05-05 13:16:05 +03:00
|
|
|
update.Addresses, // nolint:staticcheck
|
2018-05-24 09:54:12 +03:00
|
|
|
targetAddrs...,
|
|
|
|
)
|
|
|
|
return
|
2017-03-25 11:40:33 +03:00
|
|
|
}
|
2017-02-23 03:24:22 +03:00
|
|
|
}
|
2018-08-23 08:16:16 +03:00
|
|
|
case err := <-graphSub.errChan:
|
|
|
|
t.Fatalf("unable to recv graph update: %v", err)
|
2020-12-08 18:27:01 +03:00
|
|
|
case <-time.After(defaultTimeout):
|
2018-05-24 09:54:12 +03:00
|
|
|
t.Fatalf("did not receive node ann update")
|
2017-02-23 03:24:22 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2018-05-24 09:54:12 +03:00
|
|
|
|
2019-04-25 02:06:14 +03:00
|
|
|
// We'll then wait for Alice to receive Dave's node announcement
|
|
|
|
// including the expected advertised addresses from Bob since they
|
|
|
|
// should already be connected.
|
2018-08-23 08:16:16 +03:00
|
|
|
waitForAddrsInUpdate(
|
|
|
|
aliceSub, dave.PubKeyStr, advertisedAddrs...,
|
|
|
|
)
|
2017-04-17 01:27:28 +03:00
|
|
|
|
2017-07-14 22:47:52 +03:00
|
|
|
// Close the channel between Bob and Dave.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2017-07-14 22:47:52 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Bob, chanPoint, false)
|
2017-02-23 03:24:22 +03:00
|
|
|
}
|
|
|
|
|
2017-11-03 21:52:02 +03:00
|
|
|
func testNodeSignVerify(net *lntest.NetworkHarness, t *harnessTest) {
|
2017-04-20 05:28:10 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2020-11-17 02:28:56 +03:00
|
|
|
chanAmt := funding.MaxBtcFundingAmount
|
2017-04-29 14:44:29 +03:00
|
|
|
pushAmt := btcutil.Amount(100000)
|
2017-04-20 05:28:10 +03:00
|
|
|
|
2017-04-29 14:44:29 +03:00
|
|
|
// Create a channel between alice and bob.
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-03-15 20:02:40 +03:00
|
|
|
aliceBobCh := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, net.Bob,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
PushAmt: pushAmt,
|
|
|
|
},
|
2018-03-15 20:02:40 +03:00
|
|
|
)
|
2017-04-20 05:28:10 +03:00
|
|
|
|
2017-04-29 14:44:29 +03:00
|
|
|
aliceMsg := []byte("alice msg")
|
2017-04-20 05:28:10 +03:00
|
|
|
|
2017-04-29 14:44:29 +03:00
|
|
|
// alice signs "alice msg" and sends her signature to bob.
|
2017-04-20 05:28:10 +03:00
|
|
|
sigReq := &lnrpc.SignMessageRequest{Msg: aliceMsg}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
sigResp, err := net.Alice.SignMessage(ctxt, sigReq)
|
2017-04-20 05:28:10 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("SignMessage rpc call failed: %v", err)
|
|
|
|
}
|
|
|
|
aliceSig := sigResp.Signature
|
|
|
|
|
2017-04-29 14:44:29 +03:00
|
|
|
// bob verifying alice's signature should succeed since alice and bob are
|
|
|
|
// connected.
|
|
|
|
verifyReq := &lnrpc.VerifyMessageRequest{Msg: aliceMsg, Signature: aliceSig}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
verifyResp, err := net.Bob.VerifyMessage(ctxt, verifyReq)
|
2017-04-20 05:28:10 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("VerifyMessage failed: %v", err)
|
|
|
|
}
|
|
|
|
if !verifyResp.Valid {
|
|
|
|
t.Fatalf("alice's signature didn't validate")
|
|
|
|
}
|
2017-04-29 14:44:29 +03:00
|
|
|
if verifyResp.Pubkey != net.Alice.PubKeyStr {
|
|
|
|
t.Fatalf("alice's signature doesn't contain alice's pubkey.")
|
|
|
|
}
|
2017-04-20 05:28:10 +03:00
|
|
|
|
2017-04-29 14:44:29 +03:00
|
|
|
// carol is a new node that is unconnected to alice or bob.
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(t.t, "Carol", nil)
|
2018-06-08 00:14:39 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
2017-04-20 05:28:10 +03:00
|
|
|
|
|
|
|
carolMsg := []byte("carol msg")
|
|
|
|
|
2017-04-29 14:44:29 +03:00
|
|
|
// carol signs "carol msg" and sends her signature to bob.
|
2017-04-20 05:28:10 +03:00
|
|
|
sigReq = &lnrpc.SignMessageRequest{Msg: carolMsg}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
sigResp, err = carol.SignMessage(ctxt, sigReq)
|
2017-04-20 05:28:10 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("SignMessage rpc call failed: %v", err)
|
|
|
|
}
|
|
|
|
carolSig := sigResp.Signature
|
|
|
|
|
2017-04-29 14:44:29 +03:00
|
|
|
// bob verifying carol's signature should fail since they are not connected.
|
|
|
|
verifyReq = &lnrpc.VerifyMessageRequest{Msg: carolMsg, Signature: carolSig}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
verifyResp, err = net.Bob.VerifyMessage(ctxt, verifyReq)
|
2017-04-20 05:28:10 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("VerifyMessage failed: %v", err)
|
|
|
|
}
|
|
|
|
if verifyResp.Valid {
|
|
|
|
t.Fatalf("carol's signature should not be valid")
|
|
|
|
}
|
2017-04-29 14:44:29 +03:00
|
|
|
if verifyResp.Pubkey != carol.PubKeyStr {
|
|
|
|
t.Fatalf("carol's signature doesn't contain her pubkey")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Close the channel between alice and bob.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2017-04-29 14:44:29 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Alice, aliceBobCh, false)
|
2017-04-20 05:28:10 +03:00
|
|
|
}
|
|
|
|
|
2018-06-14 05:36:08 +03:00
|
|
|
// testSendUpdateDisableChannel ensures that a channel update with the disable
|
|
|
|
// flag set is sent once a channel has been either unilaterally or cooperatively
|
|
|
|
// closed.
|
|
|
|
func testSendUpdateDisableChannel(net *lntest.NetworkHarness, t *harnessTest) {
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxb := context.Background()
|
|
|
|
|
2018-06-14 05:36:08 +03:00
|
|
|
const (
|
|
|
|
chanAmt = 100000
|
|
|
|
)
|
|
|
|
|
|
|
|
// Open a channel between Alice and Bob and Alice and Carol. These will
|
|
|
|
// be closed later on in order to trigger channel update messages
|
|
|
|
// marking the channels as disabled.
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-06-14 05:36:08 +03:00
|
|
|
chanPointAliceBob := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, net.Bob,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-06-14 05:36:08 +03:00
|
|
|
)
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
carol := net.NewNode(
|
|
|
|
t.t, "Carol", []string{
|
|
|
|
"--minbackoff=10s",
|
|
|
|
"--chan-enable-timeout=1.5s",
|
|
|
|
"--chan-disable-timeout=3s",
|
|
|
|
"--chan-status-sample-interval=.5s",
|
|
|
|
})
|
2018-06-14 05:36:08 +03:00
|
|
|
defer shutdownAndAssert(net, t, carol)
|
2018-08-09 11:03:15 +03:00
|
|
|
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, net.Alice, carol)
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-06-14 05:36:08 +03:00
|
|
|
chanPointAliceCarol := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, net.Alice, carol,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-06-14 05:36:08 +03:00
|
|
|
)
|
|
|
|
|
2018-08-09 11:03:15 +03:00
|
|
|
// We create a new node Eve that has an inactive channel timeout of
|
|
|
|
// just 2 seconds (down from the default 20m). It will be used to test
|
|
|
|
// channel updates for channels going inactive.
|
2021-06-07 23:05:12 +03:00
|
|
|
eve := net.NewNode(
|
|
|
|
t.t, "Eve", []string{
|
|
|
|
"--minbackoff=10s",
|
|
|
|
"--chan-enable-timeout=1.5s",
|
|
|
|
"--chan-disable-timeout=3s",
|
|
|
|
"--chan-status-sample-interval=.5s",
|
|
|
|
})
|
2018-08-09 11:03:15 +03:00
|
|
|
defer shutdownAndAssert(net, t, eve)
|
|
|
|
|
|
|
|
// Give Eve some coins.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, eve)
|
2018-08-09 11:03:15 +03:00
|
|
|
|
|
|
|
// Connect Eve to Carol and Bob, and open a channel to carol.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, eve, carol)
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, eve, net.Bob)
|
2018-08-09 11:03:15 +03:00
|
|
|
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-08-09 11:03:15 +03:00
|
|
|
chanPointEveCarol := openChannelAndAssert(
|
2018-08-22 10:32:42 +03:00
|
|
|
ctxt, t, net, eve, carol,
|
|
|
|
lntest.OpenChannelParams{
|
|
|
|
Amt: chanAmt,
|
|
|
|
},
|
2018-08-09 11:03:15 +03:00
|
|
|
)
|
|
|
|
|
2018-06-14 05:36:08 +03:00
|
|
|
// Launch a node for Dave which will connect to Bob in order to receive
|
|
|
|
// graph updates from. This will ensure that the channel updates are
|
|
|
|
// propagated throughout the network.
|
2021-06-07 23:05:12 +03:00
|
|
|
dave := net.NewNode(t.t, "Dave", nil)
|
2018-06-14 05:36:08 +03:00
|
|
|
defer shutdownAndAssert(net, t, dave)
|
2018-11-28 11:25:46 +03:00
|
|
|
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.ConnectNodes(ctxt, t.t, net.Bob, dave)
|
2018-06-14 05:36:08 +03:00
|
|
|
|
2021-04-16 11:55:51 +03:00
|
|
|
daveSub := subscribeGraphNotifications(ctxb, t, dave)
|
2018-08-23 08:16:16 +03:00
|
|
|
defer close(daveSub.quit)
|
2018-06-14 05:36:08 +03:00
|
|
|
|
|
|
|
// We should expect to see a channel update with the default routing
|
|
|
|
// policy, except that it should indicate the channel is disabled.
|
|
|
|
expectedPolicy := &lnrpc.RoutingPolicy{
|
2020-10-06 21:56:10 +03:00
|
|
|
FeeBaseMsat: int64(chainreg.DefaultBitcoinBaseFeeMSat),
|
|
|
|
FeeRateMilliMsat: int64(chainreg.DefaultBitcoinFeeRate),
|
|
|
|
TimeLockDelta: chainreg.DefaultBitcoinTimeLockDelta,
|
2018-08-22 10:32:43 +03:00
|
|
|
MinHtlc: 1000, // default value
|
2019-08-20 03:55:30 +03:00
|
|
|
MaxHtlcMsat: calculateMaxHtlc(chanAmt),
|
2018-06-14 05:36:08 +03:00
|
|
|
Disabled: true,
|
|
|
|
}
|
|
|
|
|
2018-08-09 11:03:15 +03:00
|
|
|
// Let Carol go offline. Since Eve has an inactive timeout of 2s, we
|
|
|
|
// expect her to send an update disabling the channel.
|
|
|
|
restartCarol, err := net.SuspendNode(carol)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to suspend carol: %v", err)
|
|
|
|
}
|
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, daveSub,
|
2018-08-22 10:32:42 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{eve.PubKeyStr, expectedPolicy, chanPointEveCarol},
|
|
|
|
},
|
2018-08-09 11:03:15 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
// We restart Carol. Since the channel now becomes active again, Eve
|
|
|
|
// should send a ChannelUpdate setting the channel no longer disabled.
|
|
|
|
if err := restartCarol(); err != nil {
|
|
|
|
t.Fatalf("unable to restart carol: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
expectedPolicy.Disabled = false
|
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, daveSub,
|
2018-08-22 10:32:42 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{eve.PubKeyStr, expectedPolicy, chanPointEveCarol},
|
|
|
|
},
|
2018-08-09 11:03:15 +03:00
|
|
|
)
|
|
|
|
|
2019-02-15 04:13:20 +03:00
|
|
|
// Now we'll test a long disconnection. Disconnect Carol and Eve and
|
|
|
|
// ensure they both detect each other as disabled. Their min backoffs
|
|
|
|
// are high enough to not interfere with disabling logic.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
if err := net.DisconnectNodes(ctxt, carol, eve); err != nil {
|
|
|
|
t.Fatalf("unable to disconnect Carol from Eve: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for a disable from both Carol and Eve to come through.
|
|
|
|
expectedPolicy.Disabled = true
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, daveSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{eve.PubKeyStr, expectedPolicy, chanPointEveCarol},
|
|
|
|
{carol.PubKeyStr, expectedPolicy, chanPointEveCarol},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Reconnect Carol and Eve, this should cause them to reenable the
|
|
|
|
// channel from both ends after a short delay.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, carol, eve)
|
2019-02-15 04:13:20 +03:00
|
|
|
|
|
|
|
expectedPolicy.Disabled = false
|
|
|
|
waitForChannelUpdate(
|
|
|
|
t, daveSub,
|
|
|
|
[]expectedChanUpdate{
|
|
|
|
{eve.PubKeyStr, expectedPolicy, chanPointEveCarol},
|
|
|
|
{carol.PubKeyStr, expectedPolicy, chanPointEveCarol},
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
// Now we'll test a short disconnection. Disconnect Carol and Eve, then
|
|
|
|
// reconnect them after one second so that their scheduled disables are
|
|
|
|
// aborted. One second is twice the status sample interval, so this
|
|
|
|
// should allow for the disconnect to be detected, but still leave time
|
|
|
|
// to cancel the announcement before the 3 second inactive timeout is
|
|
|
|
// hit.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
if err := net.DisconnectNodes(ctxt, carol, eve); err != nil {
|
|
|
|
t.Fatalf("unable to disconnect Carol from Eve: %v", err)
|
|
|
|
}
|
|
|
|
time.Sleep(time.Second)
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-11 07:35:30 +03:00
|
|
|
net.EnsureConnected(ctxt, t.t, eve, carol)
|
2019-02-15 04:13:20 +03:00
|
|
|
|
|
|
|
// Since the disable should have been canceled by both Carol and Eve, we
|
|
|
|
// expect no channel updates to appear on the network.
|
|
|
|
assertNoChannelUpdates(t, daveSub, 4*time.Second)
|
|
|
|
|
2018-06-14 05:36:08 +03:00
|
|
|
// Close Alice's channels with Bob and Carol cooperatively and
|
|
|
|
// unilaterally respectively.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-08-09 11:03:13 +03:00
|
|
|
_, _, err = net.CloseChannel(ctxt, net.Alice, chanPointAliceBob, false)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to close channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-08-09 11:03:13 +03:00
|
|
|
_, _, err = net.CloseChannel(ctxt, net.Alice, chanPointAliceCarol, true)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to close channel: %v", err)
|
|
|
|
}
|
2018-06-14 05:36:08 +03:00
|
|
|
|
2018-08-09 11:03:13 +03:00
|
|
|
// Now that the channel close processes have been started, we should
|
|
|
|
// receive an update marking each as disabled.
|
2018-08-09 11:03:15 +03:00
|
|
|
expectedPolicy.Disabled = true
|
2018-06-14 05:36:08 +03:00
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, daveSub,
|
2018-08-22 10:32:42 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{net.Alice.PubKeyStr, expectedPolicy, chanPointAliceBob},
|
|
|
|
{net.Alice.PubKeyStr, expectedPolicy, chanPointAliceCarol},
|
|
|
|
},
|
2018-06-14 05:36:08 +03:00
|
|
|
)
|
2018-08-09 11:03:13 +03:00
|
|
|
|
|
|
|
// Finally, close the channels by mining the closing transactions.
|
2018-11-29 14:33:11 +03:00
|
|
|
mineBlocks(t, net, 1, 2)
|
2018-08-09 11:03:15 +03:00
|
|
|
|
|
|
|
// Also do this check for Eve's channel with Carol.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-08-09 11:03:15 +03:00
|
|
|
_, _, err = net.CloseChannel(ctxt, eve, chanPointEveCarol, false)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to close channel: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
waitForChannelUpdate(
|
2018-08-23 08:16:16 +03:00
|
|
|
t, daveSub,
|
2018-08-22 10:32:42 +03:00
|
|
|
[]expectedChanUpdate{
|
|
|
|
{eve.PubKeyStr, expectedPolicy, chanPointEveCarol},
|
|
|
|
},
|
2018-08-09 11:03:15 +03:00
|
|
|
)
|
2018-11-29 14:33:11 +03:00
|
|
|
mineBlocks(t, net, 1, 1)
|
2019-08-07 14:12:32 +03:00
|
|
|
|
|
|
|
// And finally, clean up the force closed channel by mining the
|
|
|
|
// sweeping transaction.
|
|
|
|
cleanupForceClose(t, net, net.Alice, chanPointAliceCarol)
|
2018-06-14 05:36:08 +03:00
|
|
|
}
|
|
|
|
|
2018-06-20 10:09:02 +03:00
|
|
|
// testAbandonChannel abandones a channel and asserts that it is no
|
|
|
|
// longer open and not in one of the pending closure states. It also
|
|
|
|
// verifies that the abandoned channel is reported as closed with close
|
|
|
|
// type 'abandoned'.
|
|
|
|
func testAbandonChannel(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
// First establish a channel between Alice and Bob.
|
|
|
|
channelParam := lntest.OpenChannelParams{
|
2020-11-17 02:28:56 +03:00
|
|
|
Amt: funding.MaxBtcFundingAmount,
|
2018-06-20 10:09:02 +03:00
|
|
|
PushAmt: btcutil.Amount(100000),
|
|
|
|
}
|
|
|
|
|
2018-11-26 18:09:39 +03:00
|
|
|
ctxt, _ := context.WithTimeout(ctxb, channelOpenTimeout)
|
2018-06-20 10:09:02 +03:00
|
|
|
chanPoint := openChannelAndAssert(
|
2019-11-14 03:09:20 +03:00
|
|
|
ctxt, t, net, net.Alice, net.Bob, channelParam,
|
|
|
|
)
|
2021-02-13 11:05:33 +03:00
|
|
|
txid, err := lnrpc.GetChanPointFundingTxid(chanPoint)
|
2019-11-14 03:09:20 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get txid: %v", err)
|
|
|
|
}
|
|
|
|
chanPointStr := fmt.Sprintf("%v:%v", txid, chanPoint.OutputIndex)
|
2018-06-20 10:09:02 +03:00
|
|
|
|
|
|
|
// Wait for channel to be confirmed open.
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2019-11-14 03:09:20 +03:00
|
|
|
err = net.Alice.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
2018-06-20 10:09:02 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("alice didn't report channel: %v", err)
|
|
|
|
}
|
|
|
|
err = net.Bob.WaitForNetworkChannelOpen(ctxt, chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("bob didn't report channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2019-11-14 03:09:20 +03:00
|
|
|
// Now that the channel is open, we'll obtain its channel ID real quick
|
|
|
|
// so we can use it to query the graph below.
|
|
|
|
listReq := &lnrpc.ListChannelsRequest{}
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
aliceChannelList, err := net.Alice.ListChannels(ctxt, listReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch alice's channels: %v", err)
|
|
|
|
}
|
|
|
|
var chanID uint64
|
|
|
|
for _, channel := range aliceChannelList.Channels {
|
|
|
|
if channel.ChannelPoint == chanPointStr {
|
|
|
|
chanID = channel.ChanId
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if chanID == 0 {
|
|
|
|
t.Fatalf("unable to find channel")
|
|
|
|
}
|
|
|
|
|
2020-03-04 12:57:06 +03:00
|
|
|
// To make sure the channel is removed from the backup file as well when
|
|
|
|
// being abandoned, grab a backup snapshot so we can compare it with the
|
|
|
|
// later state.
|
|
|
|
bkupBefore, err := ioutil.ReadFile(net.Alice.ChanBackupPath())
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("could not get channel backup before abandoning "+
|
|
|
|
"channel: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-06-20 10:09:02 +03:00
|
|
|
// Send request to abandon channel.
|
|
|
|
abandonChannelRequest := &lnrpc.AbandonChannelRequest{
|
|
|
|
ChannelPoint: chanPoint,
|
|
|
|
}
|
|
|
|
|
2018-11-27 11:18:35 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2018-06-20 10:09:02 +03:00
|
|
|
_, err = net.Alice.AbandonChannel(ctxt, abandonChannelRequest)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to abandon channel: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Assert that channel in no longer open.
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2019-11-14 03:09:20 +03:00
|
|
|
aliceChannelList, err = net.Alice.ListChannels(ctxt, listReq)
|
2018-06-20 10:09:02 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to list channels: %v", err)
|
|
|
|
}
|
|
|
|
if len(aliceChannelList.Channels) != 0 {
|
|
|
|
t.Fatalf("alice should only have no channels open, "+
|
|
|
|
"instead she has %v",
|
|
|
|
len(aliceChannelList.Channels))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Assert that channel is not pending closure.
|
|
|
|
pendingReq := &lnrpc.PendingChannelsRequest{}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
alicePendingList, err := net.Alice.PendingChannels(ctxt, pendingReq)
|
2018-06-20 10:09:02 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to list pending channels: %v", err)
|
|
|
|
}
|
2020-01-10 17:27:50 +03:00
|
|
|
if len(alicePendingList.PendingClosingChannels) != 0 { //nolint:staticcheck
|
2018-06-20 10:09:02 +03:00
|
|
|
t.Fatalf("alice should only have no pending closing channels, "+
|
|
|
|
"instead she has %v",
|
2020-01-10 17:27:50 +03:00
|
|
|
len(alicePendingList.PendingClosingChannels)) //nolint:staticcheck
|
2018-06-20 10:09:02 +03:00
|
|
|
}
|
|
|
|
if len(alicePendingList.PendingForceClosingChannels) != 0 {
|
|
|
|
t.Fatalf("alice should only have no pending force closing "+
|
|
|
|
"channels instead she has %v",
|
|
|
|
len(alicePendingList.PendingForceClosingChannels))
|
|
|
|
}
|
|
|
|
if len(alicePendingList.WaitingCloseChannels) != 0 {
|
|
|
|
t.Fatalf("alice should only have no waiting close "+
|
|
|
|
"channels instead she has %v",
|
|
|
|
len(alicePendingList.WaitingCloseChannels))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Assert that channel is listed as abandoned.
|
|
|
|
closedReq := &lnrpc.ClosedChannelsRequest{
|
|
|
|
Abandoned: true,
|
|
|
|
}
|
2018-11-28 11:25:46 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
aliceClosedList, err := net.Alice.ClosedChannels(ctxt, closedReq)
|
2018-06-20 10:09:02 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to list closed channels: %v", err)
|
|
|
|
}
|
|
|
|
if len(aliceClosedList.Channels) != 1 {
|
|
|
|
t.Fatalf("alice should only have a single abandoned channel, "+
|
|
|
|
"instead she has %v",
|
|
|
|
len(aliceClosedList.Channels))
|
|
|
|
}
|
|
|
|
|
2019-11-14 03:09:20 +03:00
|
|
|
// Ensure that the channel can no longer be found in the channel graph.
|
|
|
|
_, err = net.Alice.GetChanInfo(ctxb, &lnrpc.ChanInfoRequest{
|
|
|
|
ChanId: chanID,
|
|
|
|
})
|
|
|
|
if !strings.Contains(err.Error(), "marked as zombie") {
|
|
|
|
t.Fatalf("channel shouldn't be found in the channel " +
|
|
|
|
"graph!")
|
|
|
|
}
|
|
|
|
|
2020-03-04 12:57:06 +03:00
|
|
|
// Make sure the channel is no longer in the channel backup list.
|
2020-03-24 14:18:42 +03:00
|
|
|
err = wait.Predicate(func() bool {
|
|
|
|
bkupAfter, err := ioutil.ReadFile(net.Alice.ChanBackupPath())
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("could not get channel backup before "+
|
|
|
|
"abandoning channel: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
return len(bkupAfter) < len(bkupBefore)
|
|
|
|
}, defaultTimeout)
|
2020-03-04 12:57:06 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("channel wasn't removed from channel backup file")
|
|
|
|
}
|
|
|
|
|
2019-11-14 03:09:20 +03:00
|
|
|
// Calling AbandonChannel again, should result in no new errors, as the
|
|
|
|
// channel has already been removed.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
_, err = net.Alice.AbandonChannel(ctxt, abandonChannelRequest)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to abandon channel a second time: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now that we're done with the test, the channel can be closed. This
|
|
|
|
// is necessary to avoid unexpected outcomes of other tests that use
|
|
|
|
// Bob's lnd instance.
|
2018-11-27 10:57:32 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, channelCloseTimeout)
|
2018-06-20 10:09:02 +03:00
|
|
|
closeChannelAndAssert(ctxt, t, net, net.Bob, chanPoint, true)
|
2018-11-06 13:03:01 +03:00
|
|
|
|
|
|
|
// Cleanup by mining the force close and sweep transaction.
|
|
|
|
cleanupForceClose(t, net, net.Bob, chanPoint)
|
2018-06-20 10:09:02 +03:00
|
|
|
}
|
|
|
|
|
2018-12-25 01:05:48 +03:00
|
|
|
// testSweepAllCoins tests that we're able to properly sweep all coins from the
|
|
|
|
// wallet into a single target address at the specified fee rate.
|
|
|
|
func testSweepAllCoins(net *lntest.NetworkHarness, t *harnessTest) {
|
|
|
|
ctxb := context.Background()
|
|
|
|
|
|
|
|
// First, we'll make a new node, ainz who'll we'll use to test wallet
|
|
|
|
// sweeping.
|
2021-06-07 23:05:12 +03:00
|
|
|
ainz := net.NewNode(t.t, "Ainz", nil)
|
2018-12-25 01:05:48 +03:00
|
|
|
defer shutdownAndAssert(net, t, ainz)
|
|
|
|
|
|
|
|
// Next, we'll give Ainz exactly 2 utxos of 1 BTC each, with one of
|
|
|
|
// them being p2wkh and the other being a n2wpkh address.
|
|
|
|
ctxt, _ := context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoins(ctxt, t.t, btcutil.SatoshiPerBitcoin, ainz)
|
|
|
|
|
2018-12-25 01:05:48 +03:00
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2021-06-09 20:29:22 +03:00
|
|
|
net.SendCoinsNP2WKH(ctxt, t.t, btcutil.SatoshiPerBitcoin, ainz)
|
2018-12-25 01:05:48 +03:00
|
|
|
|
2019-04-10 05:45:55 +03:00
|
|
|
// Ensure that we can't send coins to our own Pubkey.
|
|
|
|
info, err := ainz.GetInfo(ctxt, &lnrpc.GetInfoRequest{})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get node info: %v", err)
|
|
|
|
}
|
|
|
|
|
2020-05-18 15:13:24 +03:00
|
|
|
// Create a label that we will used to label the transaction with.
|
|
|
|
sendCoinsLabel := "send all coins"
|
|
|
|
|
2019-04-10 05:45:55 +03:00
|
|
|
sweepReq := &lnrpc.SendCoinsRequest{
|
|
|
|
Addr: info.IdentityPubkey,
|
|
|
|
SendAll: true,
|
2020-05-18 15:13:24 +03:00
|
|
|
Label: sendCoinsLabel,
|
2019-04-10 05:45:55 +03:00
|
|
|
}
|
|
|
|
_, err = ainz.SendCoins(ctxt, sweepReq)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("expected SendCoins to users own pubkey to fail")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Ensure that we can't send coins to another users Pubkey.
|
|
|
|
info, err = net.Alice.GetInfo(ctxt, &lnrpc.GetInfoRequest{})
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get node info: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
sweepReq = &lnrpc.SendCoinsRequest{
|
|
|
|
Addr: info.IdentityPubkey,
|
|
|
|
SendAll: true,
|
2020-05-18 15:13:24 +03:00
|
|
|
Label: sendCoinsLabel,
|
2019-04-10 05:45:55 +03:00
|
|
|
}
|
|
|
|
_, err = ainz.SendCoins(ctxt, sweepReq)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("expected SendCoins to Alices pubkey to fail")
|
|
|
|
}
|
|
|
|
|
2018-12-25 01:05:48 +03:00
|
|
|
// With the two coins above mined, we'll now instruct ainz to sweep all
|
|
|
|
// the coins to an external address not under its control.
|
2019-02-27 22:16:34 +03:00
|
|
|
// We will first attempt to send the coins to addresses that are not
|
|
|
|
// compatible with the current network. This is to test that the wallet
|
|
|
|
// will prevent any onchain transactions to addresses that are not on the
|
|
|
|
// same network as the user.
|
|
|
|
|
|
|
|
// Send coins to a testnet3 address.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
2019-04-10 05:45:55 +03:00
|
|
|
sweepReq = &lnrpc.SendCoinsRequest{
|
2019-02-27 22:16:34 +03:00
|
|
|
Addr: "tb1qfc8fusa98jx8uvnhzavxccqlzvg749tvjw82tg",
|
|
|
|
SendAll: true,
|
2020-05-18 15:13:24 +03:00
|
|
|
Label: sendCoinsLabel,
|
2019-02-27 22:16:34 +03:00
|
|
|
}
|
|
|
|
_, err = ainz.SendCoins(ctxt, sweepReq)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("expected SendCoins to different network to fail")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Send coins to a mainnet address.
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
sweepReq = &lnrpc.SendCoinsRequest{
|
|
|
|
Addr: "1MPaXKp5HhsLNjVSqaL7fChE3TVyrTMRT3",
|
|
|
|
SendAll: true,
|
2020-05-18 15:13:24 +03:00
|
|
|
Label: sendCoinsLabel,
|
2019-02-27 22:16:34 +03:00
|
|
|
}
|
|
|
|
_, err = ainz.SendCoins(ctxt, sweepReq)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("expected SendCoins to different network to fail")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Send coins to a compatible address.
|
2018-12-25 01:05:48 +03:00
|
|
|
minerAddr, err := net.Miner.NewAddress()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create new miner addr: %v", err)
|
|
|
|
}
|
|
|
|
|
2019-02-27 22:16:34 +03:00
|
|
|
sweepReq = &lnrpc.SendCoinsRequest{
|
2018-12-25 01:05:48 +03:00
|
|
|
Addr: minerAddr.String(),
|
|
|
|
SendAll: true,
|
2020-05-18 15:13:24 +03:00
|
|
|
Label: sendCoinsLabel,
|
2018-12-25 01:05:48 +03:00
|
|
|
}
|
|
|
|
ctxt, _ = context.WithTimeout(ctxb, defaultTimeout)
|
|
|
|
_, err = ainz.SendCoins(ctxt, sweepReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to sweep coins: %v", err)
|
|
|
|
}
|
|
|
|
|
2019-02-27 22:16:34 +03:00
|
|
|
// We'll mine a block which should include the sweep transaction we
|
2018-12-25 01:05:48 +03:00
|
|
|
// generated above.
|
|
|
|
block := mineBlocks(t, net, 1, 1)[0]
|
|
|
|
|
|
|
|
// The sweep transaction should have exactly two inputs as we only had
|
|
|
|
// two UTXOs in the wallet.
|
|
|
|
sweepTx := block.Transactions[1]
|
|
|
|
if len(sweepTx.TxIn) != 2 {
|
|
|
|
t.Fatalf("expected 2 inputs instead have %v", len(sweepTx.TxIn))
|
|
|
|
}
|
|
|
|
|
2020-05-25 09:38:25 +03:00
|
|
|
sweepTxStr := sweepTx.TxHash().String()
|
|
|
|
assertTxLabel(ctxb, t, ainz, sweepTxStr, sendCoinsLabel)
|
|
|
|
|
|
|
|
// While we are looking at labels, we test our label transaction command
|
|
|
|
// to make sure it is behaving as expected. First, we try to label our
|
|
|
|
// transaction with an empty label, and check that we fail as expected.
|
|
|
|
sweepHash := sweepTx.TxHash()
|
|
|
|
_, err = ainz.WalletKitClient.LabelTransaction(
|
|
|
|
ctxt, &walletrpc.LabelTransactionRequest{
|
|
|
|
Txid: sweepHash[:],
|
|
|
|
Label: "",
|
|
|
|
Overwrite: false,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("expected error for zero transaction label")
|
2020-05-18 15:13:24 +03:00
|
|
|
}
|
|
|
|
|
2020-05-25 09:38:25 +03:00
|
|
|
// Our error will be wrapped in a rpc error, so we check that it
|
|
|
|
// contains the error we expect.
|
2020-09-04 14:47:56 +03:00
|
|
|
errZeroLabel := "cannot label transaction with empty label"
|
|
|
|
if !strings.Contains(err.Error(), errZeroLabel) {
|
2020-05-25 09:38:25 +03:00
|
|
|
t.Fatalf("expected: zero label error, got: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next, we try to relabel our transaction without setting the overwrite
|
|
|
|
// boolean. We expect this to fail, because the wallet requires setting
|
|
|
|
// of this param to prevent accidental overwrite of labels.
|
|
|
|
_, err = ainz.WalletKitClient.LabelTransaction(
|
|
|
|
ctxt, &walletrpc.LabelTransactionRequest{
|
|
|
|
Txid: sweepHash[:],
|
|
|
|
Label: "label that will not work",
|
|
|
|
Overwrite: false,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("expected error for tx already labelled")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Our error will be wrapped in a rpc error, so we check that it
|
|
|
|
// contains the error we expect.
|
|
|
|
if !strings.Contains(err.Error(), wallet.ErrTxLabelExists.Error()) {
|
|
|
|
t.Fatalf("expected: label exists, got: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, we overwrite our label with a new label, which should not
|
|
|
|
// fail.
|
|
|
|
newLabel := "new sweep tx label"
|
|
|
|
_, err = ainz.WalletKitClient.LabelTransaction(
|
|
|
|
ctxt, &walletrpc.LabelTransactionRequest{
|
|
|
|
Txid: sweepHash[:],
|
|
|
|
Label: newLabel,
|
|
|
|
Overwrite: true,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("could not label tx: %v", err)
|
2020-05-18 15:13:24 +03:00
|
|
|
}
|
|
|
|
|
2020-05-25 09:38:25 +03:00
|
|
|
assertTxLabel(ctxb, t, ainz, sweepTxStr, newLabel)
|
|
|
|
|
2018-12-25 01:05:48 +03:00
|
|
|
// Finally, Ainz should now have no coins at all within his wallet.
|
|
|
|
balReq := &lnrpc.WalletBalanceRequest{}
|
|
|
|
resp, err := ainz.WalletBalance(ctxt, balReq)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to get ainz's balance: %v", err)
|
|
|
|
}
|
|
|
|
switch {
|
|
|
|
case resp.ConfirmedBalance != 0:
|
|
|
|
t.Fatalf("expected no confirmed balance, instead have %v",
|
|
|
|
resp.ConfirmedBalance)
|
|
|
|
|
|
|
|
case resp.UnconfirmedBalance != 0:
|
|
|
|
t.Fatalf("expected no unconfirmed balance, instead have %v",
|
|
|
|
resp.UnconfirmedBalance)
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we try again, but this time specifying an amount, then the call
|
|
|
|
// should fail.
|
|
|
|
sweepReq.Amount = 10000
|
|
|
|
_, err = ainz.SendCoins(ctxt, sweepReq)
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("sweep attempt should fail")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-08-21 14:44:25 +03:00
|
|
|
// deriveFundingShim creates a channel funding shim by deriving the necessary
|
|
|
|
// keys on both sides.
|
|
|
|
func deriveFundingShim(net *lntest.NetworkHarness, t *harnessTest,
|
2020-08-21 15:12:30 +03:00
|
|
|
carol, dave *lntest.HarnessNode, chanSize btcutil.Amount,
|
|
|
|
thawHeight uint32, keyIndex int32, publish bool) (*lnrpc.FundingShim,
|
|
|
|
*lnrpc.ChannelPoint, *chainhash.Hash) {
|
2020-08-21 14:44:25 +03:00
|
|
|
|
|
|
|
ctxb := context.Background()
|
2019-12-11 07:02:49 +03:00
|
|
|
keyLoc := &signrpc.KeyLocator{
|
|
|
|
KeyFamily: 9999,
|
2020-08-21 14:44:25 +03:00
|
|
|
KeyIndex: keyIndex,
|
2019-12-11 07:02:49 +03:00
|
|
|
}
|
|
|
|
carolFundingKey, err := carol.WalletKitClient.DeriveKey(ctxb, keyLoc)
|
2020-08-21 14:47:18 +03:00
|
|
|
require.NoError(t.t, err)
|
2019-12-11 07:02:49 +03:00
|
|
|
daveFundingKey, err := dave.WalletKitClient.DeriveKey(ctxb, keyLoc)
|
2020-08-21 14:47:18 +03:00
|
|
|
require.NoError(t.t, err)
|
2019-12-11 07:02:49 +03:00
|
|
|
|
|
|
|
// Now that we have the multi-sig keys for each party, we can manually
|
|
|
|
// construct the funding transaction. We'll instruct the backend to
|
|
|
|
// immediately create and broadcast a transaction paying out an exact
|
|
|
|
// amount. Normally this would reside in the mempool, but we just
|
|
|
|
// confirm it now for simplicity.
|
|
|
|
_, fundingOutput, err := input.GenFundingPkScript(
|
|
|
|
carolFundingKey.RawKeyBytes, daveFundingKey.RawKeyBytes,
|
|
|
|
int64(chanSize),
|
|
|
|
)
|
2020-08-21 14:47:18 +03:00
|
|
|
require.NoError(t.t, err)
|
2020-08-21 15:12:30 +03:00
|
|
|
|
|
|
|
var txid *chainhash.Hash
|
|
|
|
targetOutputs := []*wire.TxOut{fundingOutput}
|
|
|
|
if publish {
|
|
|
|
txid, err = net.Miner.SendOutputsWithoutChange(
|
|
|
|
targetOutputs, 5,
|
|
|
|
)
|
|
|
|
require.NoError(t.t, err)
|
|
|
|
} else {
|
|
|
|
tx, err := net.Miner.CreateTransaction(targetOutputs, 5, false)
|
|
|
|
require.NoError(t.t, err)
|
|
|
|
|
|
|
|
txHash := tx.TxHash()
|
|
|
|
txid = &txHash
|
|
|
|
}
|
2019-12-11 07:02:49 +03:00
|
|
|
|
|
|
|
// At this point, we can being our external channel funding workflow.
|
|
|
|
// We'll start by generating a pending channel ID externally that will
|
|
|
|
// be used to track this new funding type.
|
|
|
|
var pendingChanID [32]byte
|
2020-08-21 14:47:18 +03:00
|
|
|
_, err = rand.Read(pendingChanID[:])
|
|
|
|
require.NoError(t.t, err)
|
2019-12-11 07:02:49 +03:00
|
|
|
|
|
|
|
// Now that we have the pending channel ID, Dave (our responder) will
|
|
|
|
// register the intent to receive a new channel funding workflow using
|
|
|
|
// the pending channel ID.
|
2020-03-14 03:14:30 +03:00
|
|
|
chanPoint := &lnrpc.ChannelPoint{
|
|
|
|
FundingTxid: &lnrpc.ChannelPoint_FundingTxidBytes{
|
|
|
|
FundingTxidBytes: txid[:],
|
2019-12-11 07:02:49 +03:00
|
|
|
},
|
2020-03-14 03:14:30 +03:00
|
|
|
}
|
|
|
|
chanPointShim := &lnrpc.ChanPointShim{
|
|
|
|
Amt: int64(chanSize),
|
|
|
|
ChanPoint: chanPoint,
|
2019-12-11 07:02:49 +03:00
|
|
|
LocalKey: &lnrpc.KeyDescriptor{
|
|
|
|
RawKeyBytes: daveFundingKey.RawKeyBytes,
|
|
|
|
KeyLoc: &lnrpc.KeyLocator{
|
|
|
|
KeyFamily: daveFundingKey.KeyLoc.KeyFamily,
|
|
|
|
KeyIndex: daveFundingKey.KeyLoc.KeyIndex,
|
|
|
|
},
|
|
|
|
},
|
|
|
|
RemoteKey: carolFundingKey.RawKeyBytes,
|
|
|
|
PendingChanId: pendingChanID[:],
|
2020-03-14 03:14:30 +03:00
|
|
|
ThawHeight: thawHeight,
|
2019-12-11 07:02:49 +03:00
|
|
|
}
|
|
|
|
fundingShim := &lnrpc.FundingShim{
|
|
|
|
Shim: &lnrpc.FundingShim_ChanPointShim{
|
|
|
|
ChanPointShim: chanPointShim,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
_, err = dave.FundingStateStep(ctxb, &lnrpc.FundingTransitionMsg{
|
|
|
|
Trigger: &lnrpc.FundingTransitionMsg_ShimRegister{
|
|
|
|
ShimRegister: fundingShim,
|
|
|
|
},
|
|
|
|
})
|
2020-08-21 14:47:18 +03:00
|
|
|
require.NoError(t.t, err)
|
2019-12-11 07:02:49 +03:00
|
|
|
|
|
|
|
// If we attempt to register the same shim (has the same pending chan
|
|
|
|
// ID), then we should get an error.
|
|
|
|
_, err = dave.FundingStateStep(ctxb, &lnrpc.FundingTransitionMsg{
|
|
|
|
Trigger: &lnrpc.FundingTransitionMsg_ShimRegister{
|
|
|
|
ShimRegister: fundingShim,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
if err == nil {
|
|
|
|
t.Fatalf("duplicate pending channel ID funding shim " +
|
|
|
|
"registration should trigger an error")
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll take the chan point shim we just registered for Dave (the
|
|
|
|
// responder), and swap the local/remote keys before we feed it in as
|
|
|
|
// Carol's funding shim as the initiator.
|
|
|
|
fundingShim.GetChanPointShim().LocalKey = &lnrpc.KeyDescriptor{
|
|
|
|
RawKeyBytes: carolFundingKey.RawKeyBytes,
|
|
|
|
KeyLoc: &lnrpc.KeyLocator{
|
|
|
|
KeyFamily: carolFundingKey.KeyLoc.KeyFamily,
|
|
|
|
KeyIndex: carolFundingKey.KeyLoc.KeyIndex,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
fundingShim.GetChanPointShim().RemoteKey = daveFundingKey.RawKeyBytes
|
|
|
|
|
2020-08-21 15:12:30 +03:00
|
|
|
return fundingShim, chanPoint, txid
|
2019-12-11 07:02:49 +03:00
|
|
|
}
|
|
|
|
|
2016-08-30 08:07:54 +03:00
|
|
|
// TestLightningNetworkDaemon performs a series of integration tests amongst a
|
2016-09-15 21:59:51 +03:00
|
|
|
// programmatically driven network of lnd nodes.
|
2016-08-30 08:07:54 +03:00
|
|
|
func TestLightningNetworkDaemon(t *testing.T) {
|
2020-10-26 16:21:05 +03:00
|
|
|
// If no tests are registered, then we can exit early.
|
2020-11-04 13:03:29 +03:00
|
|
|
if len(allTestCases) == 0 {
|
2020-01-10 17:27:48 +03:00
|
|
|
t.Skip("integration tests not selected with flag 'rpctest'")
|
|
|
|
}
|
|
|
|
|
2020-11-04 13:03:29 +03:00
|
|
|
// Parse testing flags that influence our test execution.
|
2020-11-04 13:03:30 +03:00
|
|
|
logDir := lntest.GetLogDir()
|
|
|
|
require.NoError(t, os.MkdirAll(logDir, 0700))
|
2020-11-04 13:03:29 +03:00
|
|
|
testCases, trancheIndex, trancheOffset := getTestCaseSplitTranche()
|
|
|
|
lntest.ApplyPortOffset(uint32(trancheIndex) * 1000)
|
|
|
|
|
2020-12-03 13:30:23 +03:00
|
|
|
// Before we start any node, we need to make sure that any btcd node
|
|
|
|
// that is started through the RPC harness uses a unique port as well to
|
|
|
|
// avoid any port collisions.
|
|
|
|
rpctest.ListenAddressGenerator = lntest.GenerateBtcdListenerAddresses
|
2016-08-30 08:07:54 +03:00
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// Declare the network harness here to gain access to its
|
|
|
|
// 'OnTxAccepted' call back.
|
2017-11-17 04:31:41 +03:00
|
|
|
var lndHarness *lntest.NetworkHarness
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// Create an instance of the btcd's rpctest.Harness that will act as
|
|
|
|
// the miner for all tests. This will be used to fund the wallets of
|
|
|
|
// the nodes within the test network and to drive blockchain related
|
|
|
|
// events within the network. Revert the default setting of accepting
|
|
|
|
// non-standard transactions on simnet to reject them. Transactions on
|
|
|
|
// the lightning network should always be standard to get better
|
|
|
|
// guarantees of getting included in to blocks.
|
|
|
|
//
|
|
|
|
// We will also connect it to our chain backend.
|
2020-11-04 13:03:30 +03:00
|
|
|
minerLogDir := fmt.Sprintf("%s/.minerlogs", logDir)
|
2020-09-30 01:15:05 +03:00
|
|
|
miner, minerCleanUp, err := lntest.NewMiner(
|
2020-12-03 13:30:22 +03:00
|
|
|
minerLogDir, "output_btcd_miner.log", harnessNetParams,
|
|
|
|
&rpcclient.NotificationHandlers{}, lntest.GetBtcdBinary(),
|
2020-09-30 01:15:05 +03:00
|
|
|
)
|
|
|
|
require.NoError(t, err, "failed to create new miner")
|
2018-10-10 14:04:13 +03:00
|
|
|
defer func() {
|
2020-09-30 01:15:05 +03:00
|
|
|
require.NoError(t, minerCleanUp(), "failed to clean up miner")
|
2018-10-10 14:04:13 +03:00
|
|
|
}()
|
2017-11-17 04:31:41 +03:00
|
|
|
|
2019-05-24 15:17:48 +03:00
|
|
|
// Start a chain backend.
|
2019-07-11 14:51:22 +03:00
|
|
|
chainBackend, cleanUp, err := lntest.NewBackend(
|
|
|
|
miner.P2PAddress(), harnessNetParams,
|
|
|
|
)
|
2020-12-03 13:30:23 +03:00
|
|
|
require.NoError(t, err, "new backend")
|
2020-09-02 22:15:04 +03:00
|
|
|
defer func() {
|
2020-12-03 13:30:23 +03:00
|
|
|
require.NoError(t, cleanUp(), "cleanup")
|
2020-09-02 22:15:04 +03:00
|
|
|
}()
|
2019-05-24 15:17:48 +03:00
|
|
|
|
2020-12-03 13:30:23 +03:00
|
|
|
// Before we start anything, we want to overwrite some of the connection
|
|
|
|
// settings to make the tests more robust. We might need to restart the
|
|
|
|
// miner while there are already blocks present, which will take a bit
|
|
|
|
// longer than the 1 second the default settings amount to. Doubling
|
|
|
|
// both values will give us retries up to 4 seconds.
|
|
|
|
miner.MaxConnRetries = rpctest.DefaultMaxConnectionRetries * 2
|
|
|
|
miner.ConnectionRetryTimeout = rpctest.DefaultConnectionRetryTimeout * 2
|
2018-12-14 11:24:00 +03:00
|
|
|
|
2020-12-03 13:30:23 +03:00
|
|
|
// Set up miner and connect chain backend to it.
|
|
|
|
require.NoError(t, miner.SetUp(true, 50))
|
2021-03-10 01:12:26 +03:00
|
|
|
require.NoError(t, miner.Client.NotifyNewTransactions(false))
|
2020-12-03 13:30:23 +03:00
|
|
|
require.NoError(t, chainBackend.ConnectMiner(), "connect miner")
|
2020-09-03 13:59:08 +03:00
|
|
|
|
2021-06-14 13:31:08 +03:00
|
|
|
// Parse database backend
|
|
|
|
var dbBackend lntest.DatabaseBackend
|
|
|
|
switch *dbBackendFlag {
|
|
|
|
case "bbolt":
|
|
|
|
dbBackend = lntest.BackendBbolt
|
|
|
|
|
|
|
|
case "etcd":
|
|
|
|
dbBackend = lntest.BackendEtcd
|
|
|
|
|
|
|
|
default:
|
|
|
|
require.Fail(t, "unknown db backend")
|
|
|
|
}
|
|
|
|
|
2018-12-14 11:24:00 +03:00
|
|
|
// Now we can set up our test harness (LND instance), with the chain
|
|
|
|
// backend we just created.
|
2020-12-03 13:30:23 +03:00
|
|
|
ht := newHarnessTest(t, nil)
|
2020-11-04 13:03:31 +03:00
|
|
|
binary := ht.getLndBinary()
|
2020-06-22 20:36:12 +03:00
|
|
|
lndHarness, err = lntest.NewNetworkHarness(
|
2021-06-14 13:31:08 +03:00
|
|
|
miner, chainBackend, binary, dbBackend,
|
2020-06-22 20:36:12 +03:00
|
|
|
)
|
2016-08-31 05:34:13 +03:00
|
|
|
if err != nil {
|
2016-10-26 15:28:05 +03:00
|
|
|
ht.Fatalf("unable to create lightning network harness: %v", err)
|
2016-08-31 05:34:13 +03:00
|
|
|
}
|
2020-10-29 17:37:17 +03:00
|
|
|
defer lndHarness.Stop()
|
2016-08-31 05:34:13 +03:00
|
|
|
|
2016-11-24 11:49:18 +03:00
|
|
|
// Spawn a new goroutine to watch for any fatal errors that any of the
|
|
|
|
// running lnd processes encounter. If an error occurs, then the test
|
2017-10-25 05:35:54 +03:00
|
|
|
// case should naturally as a result and we log the server error here to
|
|
|
|
// help debug.
|
2016-11-24 11:49:18 +03:00
|
|
|
go func() {
|
2017-10-25 05:35:54 +03:00
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case err, more := <-lndHarness.ProcessErrors():
|
|
|
|
if !more {
|
|
|
|
return
|
|
|
|
}
|
2020-11-04 13:03:29 +03:00
|
|
|
ht.Logf("lnd finished with error (stderr):\n%v",
|
|
|
|
err)
|
2017-10-25 05:35:54 +03:00
|
|
|
}
|
2016-11-24 11:49:18 +03:00
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
2017-01-06 00:17:49 +03:00
|
|
|
// Next mine enough blocks in order for segwit and the CSV package
|
|
|
|
// soft-fork to activate on SimNet.
|
2019-07-11 14:51:22 +03:00
|
|
|
numBlocks := harnessNetParams.MinerConfirmationWindow * 2
|
2021-03-10 01:12:26 +03:00
|
|
|
if _, err := miner.Client.Generate(numBlocks); err != nil {
|
2017-01-06 00:17:49 +03:00
|
|
|
ht.Fatalf("unable to generate blocks: %v", err)
|
|
|
|
}
|
|
|
|
|
2016-08-31 05:34:13 +03:00
|
|
|
// With the btcd harness created, we can now complete the
|
2016-09-15 22:24:52 +03:00
|
|
|
// initialization of the network. args - list of lnd arguments,
|
|
|
|
// example: "--debuglevel=debug"
|
2016-09-26 20:40:58 +03:00
|
|
|
// TODO(roasbeef): create master balanced channel with all the monies?
|
2020-08-13 01:45:57 +03:00
|
|
|
aliceBobArgs := []string{
|
|
|
|
"--default-remote-max-htlcs=483",
|
|
|
|
}
|
2016-08-30 08:07:54 +03:00
|
|
|
|
2020-11-04 13:03:29 +03:00
|
|
|
// Run the subset of the test cases selected in this tranche.
|
|
|
|
for idx, testCase := range testCases {
|
|
|
|
testCase := testCase
|
2020-10-29 17:37:17 +03:00
|
|
|
name := fmt.Sprintf("%02d-of-%d/%s/%s",
|
|
|
|
trancheOffset+uint(idx)+1, len(allTestCases),
|
|
|
|
chainBackend.Name(), testCase.name)
|
2018-06-15 22:35:12 +03:00
|
|
|
|
2020-10-29 17:37:17 +03:00
|
|
|
success := t.Run(name, func(t1 *testing.T) {
|
|
|
|
cleanTestCaseName := strings.ReplaceAll(
|
|
|
|
testCase.name, " ", "_",
|
|
|
|
)
|
|
|
|
|
2021-06-07 23:05:12 +03:00
|
|
|
err = lndHarness.SetUp(
|
|
|
|
t1, cleanTestCaseName, aliceBobArgs,
|
|
|
|
)
|
2020-10-29 17:37:17 +03:00
|
|
|
require.NoError(t1,
|
|
|
|
err, "unable to set up test lightning network",
|
|
|
|
)
|
|
|
|
defer func() {
|
|
|
|
require.NoError(t1, lndHarness.TearDown())
|
|
|
|
}()
|
2018-06-15 22:35:12 +03:00
|
|
|
|
2021-06-11 07:35:30 +03:00
|
|
|
lndHarness.EnsureConnected(
|
|
|
|
context.Background(), t1,
|
|
|
|
lndHarness.Alice, lndHarness.Bob,
|
2020-10-29 17:37:17 +03:00
|
|
|
)
|
2020-10-29 17:35:02 +03:00
|
|
|
|
2020-10-29 17:37:17 +03:00
|
|
|
logLine := fmt.Sprintf(
|
|
|
|
"STARTING ============ %v ============\n",
|
|
|
|
testCase.name,
|
|
|
|
)
|
2018-01-09 14:59:32 +03:00
|
|
|
|
2020-11-21 14:52:02 +03:00
|
|
|
AddToNodeLog(t, lndHarness.Alice, logLine)
|
|
|
|
AddToNodeLog(t, lndHarness.Bob, logLine)
|
2020-10-29 17:37:17 +03:00
|
|
|
|
|
|
|
// Start every test with the default static fee estimate.
|
|
|
|
lndHarness.SetFeeEstimate(12500)
|
|
|
|
|
|
|
|
// Create a separate harness test for the testcase to
|
|
|
|
// avoid overwriting the external harness test that is
|
|
|
|
// tied to the parent test.
|
2019-04-15 17:19:47 +03:00
|
|
|
ht := newHarnessTest(t1, lndHarness)
|
|
|
|
ht.RunTestCase(testCase)
|
2017-06-19 16:53:52 +03:00
|
|
|
})
|
2017-06-19 17:39:07 +03:00
|
|
|
|
2017-06-19 16:53:52 +03:00
|
|
|
// Stop at the first failure. Mimic behavior of original test
|
2017-06-19 17:39:07 +03:00
|
|
|
// framework.
|
2017-06-19 16:53:52 +03:00
|
|
|
if !success {
|
2020-04-06 22:32:26 +03:00
|
|
|
// Log failure time to help relate the lnd logs to the
|
|
|
|
// failure.
|
2020-11-04 13:03:29 +03:00
|
|
|
t.Logf("Failure time: %v", time.Now().Format(
|
|
|
|
"2006-01-02 15:04:05.000",
|
|
|
|
))
|
2017-06-19 16:53:52 +03:00
|
|
|
break
|
|
|
|
}
|
2016-08-30 08:07:54 +03:00
|
|
|
}
|
|
|
|
}
|