e4188ba9c2
This fixes a long-standing force close bug. When we receive a revocation, store the updates that the remote should sign next under a new database key. Previously, these were not persisted which would lead to force closure.
741 lines
21 KiB
Go
741 lines
21 KiB
Go
package channeldb
|
|
|
|
import (
|
|
"io/ioutil"
|
|
"math"
|
|
"math/rand"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"reflect"
|
|
"testing"
|
|
|
|
"github.com/btcsuite/btcd/btcec"
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
"github.com/btcsuite/btcd/wire"
|
|
"github.com/btcsuite/btcutil"
|
|
"github.com/davecgh/go-spew/spew"
|
|
"github.com/lightningnetwork/lnd/channeldb/kvdb"
|
|
"github.com/lightningnetwork/lnd/keychain"
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
"github.com/lightningnetwork/lnd/shachain"
|
|
)
|
|
|
|
func TestOpenWithCreate(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, create a temporary directory to be used for the duration of
|
|
// this test.
|
|
tempDirName, err := ioutil.TempDir("", "channeldb")
|
|
if err != nil {
|
|
t.Fatalf("unable to create temp dir: %v", err)
|
|
}
|
|
defer os.RemoveAll(tempDirName)
|
|
|
|
// Next, open thereby creating channeldb for the first time.
|
|
dbPath := filepath.Join(tempDirName, "cdb")
|
|
backend, cleanup, err := kvdb.GetTestBackend(dbPath, "cdb")
|
|
if err != nil {
|
|
t.Fatalf("unable to get test db backend: %v", err)
|
|
}
|
|
defer cleanup()
|
|
|
|
cdb, err := CreateWithBackend(backend)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channeldb: %v", err)
|
|
}
|
|
if err := cdb.Close(); err != nil {
|
|
t.Fatalf("unable to close channeldb: %v", err)
|
|
}
|
|
|
|
// The path should have been successfully created.
|
|
if !fileExists(dbPath) {
|
|
t.Fatalf("channeldb failed to create data directory")
|
|
}
|
|
|
|
// Now, reopen the same db in dry run migration mode. Since we have not
|
|
// applied any migrations, this should ignore the flag and not fail.
|
|
cdb, err = Open(dbPath, OptionDryRunMigration(true))
|
|
if err != nil {
|
|
t.Fatalf("unable to create channeldb: %v", err)
|
|
}
|
|
if err := cdb.Close(); err != nil {
|
|
t.Fatalf("unable to close channeldb: %v", err)
|
|
}
|
|
}
|
|
|
|
// TestWipe tests that the database wipe operation completes successfully
|
|
// and that the buckets are deleted. It also checks that attempts to fetch
|
|
// information while the buckets are not set return the correct errors.
|
|
func TestWipe(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, create a temporary directory to be used for the duration of
|
|
// this test.
|
|
tempDirName, err := ioutil.TempDir("", "channeldb")
|
|
if err != nil {
|
|
t.Fatalf("unable to create temp dir: %v", err)
|
|
}
|
|
defer os.RemoveAll(tempDirName)
|
|
|
|
// Next, open thereby creating channeldb for the first time.
|
|
dbPath := filepath.Join(tempDirName, "cdb")
|
|
backend, cleanup, err := kvdb.GetTestBackend(dbPath, "cdb")
|
|
if err != nil {
|
|
t.Fatalf("unable to get test db backend: %v", err)
|
|
}
|
|
defer cleanup()
|
|
|
|
cdb, err := CreateWithBackend(backend)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channeldb: %v", err)
|
|
}
|
|
defer cdb.Close()
|
|
|
|
if err := cdb.Wipe(); err != nil {
|
|
t.Fatalf("unable to wipe channeldb: %v", err)
|
|
}
|
|
// Check correct errors are returned
|
|
_, err = cdb.FetchAllOpenChannels()
|
|
if err != ErrNoActiveChannels {
|
|
t.Fatalf("fetching open channels: expected '%v' instead got '%v'",
|
|
ErrNoActiveChannels, err)
|
|
}
|
|
_, err = cdb.FetchClosedChannels(false)
|
|
if err != ErrNoClosedChannels {
|
|
t.Fatalf("fetching closed channels: expected '%v' instead got '%v'",
|
|
ErrNoClosedChannels, err)
|
|
}
|
|
}
|
|
|
|
// TestFetchClosedChannelForID tests that we are able to properly retrieve a
|
|
// ChannelCloseSummary from the DB given a ChannelID.
|
|
func TestFetchClosedChannelForID(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
const numChans = 101
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create the test channel state, that we will mutate the index of the
|
|
// funding point.
|
|
state := createTestChannelState(t, cdb)
|
|
|
|
// Now run through the number of channels, and modify the outpoint index
|
|
// to create new channel IDs.
|
|
for i := uint32(0); i < numChans; i++ {
|
|
// Save the open channel to disk.
|
|
state.FundingOutpoint.Index = i
|
|
|
|
// Write the channel to disk in a pending state.
|
|
createTestChannel(
|
|
t, cdb,
|
|
fundingPointOption(state.FundingOutpoint),
|
|
openChannelOption(),
|
|
)
|
|
|
|
// Close the channel. To make sure we retrieve the correct
|
|
// summary later, we make them differ in the SettledBalance.
|
|
closeSummary := &ChannelCloseSummary{
|
|
ChanPoint: state.FundingOutpoint,
|
|
RemotePub: state.IdentityPub,
|
|
SettledBalance: btcutil.Amount(500 + i),
|
|
}
|
|
if err := state.CloseChannel(closeSummary); err != nil {
|
|
t.Fatalf("unable to close channel: %v", err)
|
|
}
|
|
}
|
|
|
|
// Now run though them all again and make sure we are able to retrieve
|
|
// summaries from the DB.
|
|
for i := uint32(0); i < numChans; i++ {
|
|
state.FundingOutpoint.Index = i
|
|
|
|
// We calculate the ChannelID and use it to fetch the summary.
|
|
cid := lnwire.NewChanIDFromOutPoint(&state.FundingOutpoint)
|
|
fetchedSummary, err := cdb.FetchClosedChannelForID(cid)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch close summary: %v", err)
|
|
}
|
|
|
|
// Make sure we retrieved the correct one by checking the
|
|
// SettledBalance.
|
|
if fetchedSummary.SettledBalance != btcutil.Amount(500+i) {
|
|
t.Fatalf("summaries don't match: expected %v got %v",
|
|
btcutil.Amount(500+i),
|
|
fetchedSummary.SettledBalance)
|
|
}
|
|
}
|
|
|
|
// As a final test we make sure that we get ErrClosedChannelNotFound
|
|
// for a ChannelID we didn't add to the DB.
|
|
state.FundingOutpoint.Index++
|
|
cid := lnwire.NewChanIDFromOutPoint(&state.FundingOutpoint)
|
|
_, err = cdb.FetchClosedChannelForID(cid)
|
|
if err != ErrClosedChannelNotFound {
|
|
t.Fatalf("expected ErrClosedChannelNotFound, instead got: %v", err)
|
|
}
|
|
}
|
|
|
|
// TestAddrsForNode tests the we're able to properly obtain all the addresses
|
|
// for a target node.
|
|
func TestAddrsForNode(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
graph := cdb.ChannelGraph()
|
|
|
|
// We'll make a test vertex to insert into the database, as the source
|
|
// node, but this node will only have half the number of addresses it
|
|
// usually does.
|
|
testNode, err := createTestVertex(cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
}
|
|
testNode.Addresses = []net.Addr{testAddr}
|
|
if err := graph.SetSourceNode(testNode); err != nil {
|
|
t.Fatalf("unable to set source node: %v", err)
|
|
}
|
|
|
|
// Next, we'll make a link node with the same pubkey, but with an
|
|
// additional address.
|
|
nodePub, err := testNode.PubKey()
|
|
if err != nil {
|
|
t.Fatalf("unable to recv node pub: %v", err)
|
|
}
|
|
linkNode := cdb.NewLinkNode(
|
|
wire.MainNet, nodePub, anotherAddr,
|
|
)
|
|
if err := linkNode.Sync(); err != nil {
|
|
t.Fatalf("unable to sync link node: %v", err)
|
|
}
|
|
|
|
// Now that we've created a link node, as well as a vertex for the
|
|
// node, we'll query for all its addresses.
|
|
nodeAddrs, err := cdb.AddrsForNode(nodePub)
|
|
if err != nil {
|
|
t.Fatalf("unable to obtain node addrs: %v", err)
|
|
}
|
|
|
|
expectedAddrs := make(map[string]struct{})
|
|
expectedAddrs[testAddr.String()] = struct{}{}
|
|
expectedAddrs[anotherAddr.String()] = struct{}{}
|
|
|
|
// Finally, ensure that all the expected addresses are found.
|
|
if len(nodeAddrs) != len(expectedAddrs) {
|
|
t.Fatalf("expected %v addrs, got %v",
|
|
len(expectedAddrs), len(nodeAddrs))
|
|
}
|
|
for _, addr := range nodeAddrs {
|
|
if _, ok := expectedAddrs[addr.String()]; !ok {
|
|
t.Fatalf("unexpected addr: %v", addr)
|
|
}
|
|
}
|
|
}
|
|
|
|
// TestFetchChannel tests that we're able to fetch an arbitrary channel from
|
|
// disk.
|
|
func TestFetchChannel(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create an open channel.
|
|
channelState := createTestChannel(t, cdb, openChannelOption())
|
|
|
|
// Next, attempt to fetch the channel by its chan point.
|
|
dbChannel, err := cdb.FetchChannel(channelState.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch channel: %v", err)
|
|
}
|
|
|
|
// The decoded channel state should be identical to what we stored
|
|
// above.
|
|
if !reflect.DeepEqual(channelState, dbChannel) {
|
|
t.Fatalf("channel state doesn't match:: %v vs %v",
|
|
spew.Sdump(channelState), spew.Sdump(dbChannel))
|
|
}
|
|
|
|
// If we attempt to query for a non-exist ante channel, then we should
|
|
// get an error.
|
|
channelState2 := createTestChannelState(t, cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channel state: %v", err)
|
|
}
|
|
channelState2.FundingOutpoint.Index ^= 1
|
|
|
|
_, err = cdb.FetchChannel(channelState2.FundingOutpoint)
|
|
if err == nil {
|
|
t.Fatalf("expected query to fail")
|
|
}
|
|
}
|
|
|
|
func genRandomChannelShell() (*ChannelShell, error) {
|
|
var testPriv [32]byte
|
|
if _, err := rand.Read(testPriv[:]); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
_, pub := btcec.PrivKeyFromBytes(btcec.S256(), testPriv[:])
|
|
|
|
var chanPoint wire.OutPoint
|
|
if _, err := rand.Read(chanPoint.Hash[:]); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
pub.Curve = nil
|
|
|
|
chanPoint.Index = uint32(rand.Intn(math.MaxUint16))
|
|
|
|
chanStatus := ChanStatusDefault | ChanStatusRestored
|
|
|
|
var shaChainPriv [32]byte
|
|
if _, err := rand.Read(testPriv[:]); err != nil {
|
|
return nil, err
|
|
}
|
|
revRoot, err := chainhash.NewHash(shaChainPriv[:])
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
shaChainProducer := shachain.NewRevocationProducer(*revRoot)
|
|
|
|
return &ChannelShell{
|
|
NodeAddrs: []net.Addr{&net.TCPAddr{
|
|
IP: net.ParseIP("127.0.0.1"),
|
|
Port: 18555,
|
|
}},
|
|
Chan: &OpenChannel{
|
|
chanStatus: chanStatus,
|
|
ChainHash: rev,
|
|
FundingOutpoint: chanPoint,
|
|
ShortChannelID: lnwire.NewShortChanIDFromInt(
|
|
uint64(rand.Int63()),
|
|
),
|
|
IdentityPub: pub,
|
|
LocalChanCfg: ChannelConfig{
|
|
ChannelConstraints: ChannelConstraints{
|
|
CsvDelay: uint16(rand.Int63()),
|
|
},
|
|
PaymentBasePoint: keychain.KeyDescriptor{
|
|
KeyLocator: keychain.KeyLocator{
|
|
Family: keychain.KeyFamily(rand.Int63()),
|
|
Index: uint32(rand.Int63()),
|
|
},
|
|
},
|
|
},
|
|
RemoteCurrentRevocation: pub,
|
|
IsPending: false,
|
|
RevocationStore: shachain.NewRevocationStore(),
|
|
RevocationProducer: shaChainProducer,
|
|
},
|
|
}, nil
|
|
}
|
|
|
|
// TestRestoreChannelShells tests that we're able to insert a partially channel
|
|
// populated to disk. This is useful for channel recovery purposes. We should
|
|
// find the new channel shell on disk, and also the db should be populated with
|
|
// an edge for that channel.
|
|
func TestRestoreChannelShells(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// First, we'll make our channel shell, it will only have the minimal
|
|
// amount of information required for us to initiate the data loss
|
|
// protection feature.
|
|
channelShell, err := genRandomChannelShell()
|
|
if err != nil {
|
|
t.Fatalf("unable to gen channel shell: %v", err)
|
|
}
|
|
|
|
// With the channel shell constructed, we'll now insert it into the
|
|
// database with the restoration method.
|
|
if err := cdb.RestoreChannelShells(channelShell); err != nil {
|
|
t.Fatalf("unable to restore channel shell: %v", err)
|
|
}
|
|
|
|
// Now that the channel has been inserted, we'll attempt to query for
|
|
// it to ensure we can properly locate it via various means.
|
|
//
|
|
// First, we'll attempt to query for all channels that we have with the
|
|
// node public key that was restored.
|
|
nodeChans, err := cdb.FetchOpenChannels(channelShell.Chan.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable find channel: %v", err)
|
|
}
|
|
|
|
// We should now find a single channel from the database.
|
|
if len(nodeChans) != 1 {
|
|
t.Fatalf("unable to find restored channel by node "+
|
|
"pubkey: %v", err)
|
|
}
|
|
|
|
// Ensure that it isn't possible to modify the commitment state machine
|
|
// of this restored channel.
|
|
channel := nodeChans[0]
|
|
err = channel.UpdateCommitment(nil, nil)
|
|
if err != ErrNoRestoredChannelMutation {
|
|
t.Fatalf("able to mutate restored channel")
|
|
}
|
|
err = channel.AppendRemoteCommitChain(nil)
|
|
if err != ErrNoRestoredChannelMutation {
|
|
t.Fatalf("able to mutate restored channel")
|
|
}
|
|
err = channel.AdvanceCommitChainTail(nil, nil)
|
|
if err != ErrNoRestoredChannelMutation {
|
|
t.Fatalf("able to mutate restored channel")
|
|
}
|
|
|
|
// That single channel should have the proper channel point, and also
|
|
// the expected set of flags to indicate that it was a restored
|
|
// channel.
|
|
if nodeChans[0].FundingOutpoint != channelShell.Chan.FundingOutpoint {
|
|
t.Fatalf("wrong funding outpoint: expected %v, got %v",
|
|
nodeChans[0].FundingOutpoint,
|
|
channelShell.Chan.FundingOutpoint)
|
|
}
|
|
if !nodeChans[0].HasChanStatus(ChanStatusRestored) {
|
|
t.Fatalf("node has wrong status flags: %v",
|
|
nodeChans[0].chanStatus)
|
|
}
|
|
|
|
// We should also be able to find the channel if we query for it
|
|
// directly.
|
|
_, err = cdb.FetchChannel(channelShell.Chan.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch channel: %v", err)
|
|
}
|
|
|
|
// We should also be able to find the link node that was inserted by
|
|
// its public key.
|
|
linkNode, err := cdb.FetchLinkNode(channelShell.Chan.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch link node: %v", err)
|
|
}
|
|
|
|
// The node should have the same address, as specified in the channel
|
|
// shell.
|
|
if reflect.DeepEqual(linkNode.Addresses, channelShell.NodeAddrs) {
|
|
t.Fatalf("addr mismach: expected %v, got %v",
|
|
linkNode.Addresses, channelShell.NodeAddrs)
|
|
}
|
|
}
|
|
|
|
// TestAbandonChannel tests that the AbandonChannel method is able to properly
|
|
// remove a channel from the database and add a close channel summary. If
|
|
// called after a channel has already been removed, the method shouldn't return
|
|
// an error.
|
|
func TestAbandonChannel(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// If we attempt to abandon the state of a channel that doesn't exist
|
|
// in the open or closed channel bucket, then we should receive an
|
|
// error.
|
|
err = cdb.AbandonChannel(&wire.OutPoint{}, 0)
|
|
if err == nil {
|
|
t.Fatalf("removing non-existent channel should have failed")
|
|
}
|
|
|
|
// We'll now create a new channel in a pending state to abandon
|
|
// shortly.
|
|
chanState := createTestChannel(t, cdb)
|
|
|
|
// We should now be able to abandon the channel without any errors.
|
|
closeHeight := uint32(11)
|
|
err = cdb.AbandonChannel(&chanState.FundingOutpoint, closeHeight)
|
|
if err != nil {
|
|
t.Fatalf("unable to abandon channel: %v", err)
|
|
}
|
|
|
|
// At this point, the channel should no longer be found in the set of
|
|
// open channels.
|
|
_, err = cdb.FetchChannel(chanState.FundingOutpoint)
|
|
if err != ErrChannelNotFound {
|
|
t.Fatalf("channel should not have been found: %v", err)
|
|
}
|
|
|
|
// However we should be able to retrieve a close channel summary for
|
|
// the channel.
|
|
_, err = cdb.FetchClosedChannel(&chanState.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch closed channel: %v", err)
|
|
}
|
|
|
|
// Finally, if we attempt to abandon the channel again, we should get a
|
|
// nil error as the channel has already been abandoned.
|
|
err = cdb.AbandonChannel(&chanState.FundingOutpoint, closeHeight)
|
|
if err != nil {
|
|
t.Fatalf("unable to abandon channel: %v", err)
|
|
}
|
|
}
|
|
|
|
// TestFetchChannels tests the filtering of open channels in fetchChannels.
|
|
// It tests the case where no filters are provided (which is equivalent to
|
|
// FetchAllOpenChannels) and every combination of pending and waiting close.
|
|
func TestFetchChannels(t *testing.T) {
|
|
// Create static channel IDs for each kind of channel retrieved by
|
|
// fetchChannels so that the expected channel IDs can be set in tests.
|
|
var (
|
|
// Pending is a channel that is pending open, and has not had
|
|
// a close initiated.
|
|
pendingChan = lnwire.NewShortChanIDFromInt(1)
|
|
|
|
// pendingWaitingClose is a channel that is pending open and
|
|
// has has its closing transaction broadcast.
|
|
pendingWaitingChan = lnwire.NewShortChanIDFromInt(2)
|
|
|
|
// openChan is a channel that has confirmed on chain.
|
|
openChan = lnwire.NewShortChanIDFromInt(3)
|
|
|
|
// openWaitingChan is a channel that has confirmed on chain,
|
|
// and it waiting for its close transaction to confirm.
|
|
openWaitingChan = lnwire.NewShortChanIDFromInt(4)
|
|
)
|
|
|
|
tests := []struct {
|
|
name string
|
|
filters []fetchChannelsFilter
|
|
expectedChannels map[lnwire.ShortChannelID]bool
|
|
}{
|
|
{
|
|
name: "get all channels",
|
|
filters: []fetchChannelsFilter{},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
pendingWaitingChan: true,
|
|
openChan: true,
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "pending channels",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
pendingWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "open channels",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
openChan: true,
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "waiting close channels",
|
|
filters: []fetchChannelsFilter{
|
|
waitingCloseFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingWaitingChan: true,
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "not waiting close channels",
|
|
filters: []fetchChannelsFilter{
|
|
waitingCloseFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
openChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "pending waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(true),
|
|
waitingCloseFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "pending, not waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(true),
|
|
waitingCloseFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "open waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(false),
|
|
waitingCloseFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "open, not waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(false),
|
|
waitingCloseFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
openChan: true,
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
test := test
|
|
|
|
t.Run(test.name, func(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test "+
|
|
"database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create a pending channel that is not awaiting close.
|
|
createTestChannel(
|
|
t, cdb, channelIDOption(pendingChan),
|
|
)
|
|
|
|
// Create a pending channel which has has been marked as
|
|
// broadcast, indicating that its closing transaction is
|
|
// waiting to confirm.
|
|
pendingClosing := createTestChannel(
|
|
t, cdb,
|
|
channelIDOption(pendingWaitingChan),
|
|
)
|
|
|
|
err = pendingClosing.MarkCoopBroadcasted(nil, true)
|
|
if err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
// Create a open channel that is not awaiting close.
|
|
createTestChannel(
|
|
t, cdb,
|
|
channelIDOption(openChan),
|
|
openChannelOption(),
|
|
)
|
|
|
|
// Create a open channel which has has been marked as
|
|
// broadcast, indicating that its closing transaction is
|
|
// waiting to confirm.
|
|
openClosing := createTestChannel(
|
|
t, cdb,
|
|
channelIDOption(openWaitingChan),
|
|
openChannelOption(),
|
|
)
|
|
err = openClosing.MarkCoopBroadcasted(nil, true)
|
|
if err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
channels, err := fetchChannels(cdb, test.filters...)
|
|
if err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
if len(channels) != len(test.expectedChannels) {
|
|
t.Fatalf("expected: %v channels, "+
|
|
"got: %v", len(test.expectedChannels),
|
|
len(channels))
|
|
}
|
|
|
|
for _, ch := range channels {
|
|
_, ok := test.expectedChannels[ch.ShortChannelID]
|
|
if !ok {
|
|
t.Fatalf("fetch channels unexpected "+
|
|
"channel: %v", ch.ShortChannelID)
|
|
}
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
// TestFetchHistoricalChannel tests lookup of historical channels.
|
|
func TestFetchHistoricalChannel(t *testing.T) {
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create a an open channel in the database.
|
|
channel := createTestChannel(t, cdb, openChannelOption())
|
|
|
|
// First, try to lookup a channel when the bucket does not
|
|
// exist.
|
|
_, err = cdb.FetchHistoricalChannel(&channel.FundingOutpoint)
|
|
if err != ErrNoHistoricalBucket {
|
|
t.Fatalf("expected no bucket, got: %v", err)
|
|
}
|
|
|
|
// Close the channel so that it will be written to the historical
|
|
// bucket. The values provided in the channel close summary are the
|
|
// minimum required for this call to run without panicking.
|
|
if err := channel.CloseChannel(&ChannelCloseSummary{
|
|
ChanPoint: channel.FundingOutpoint,
|
|
RemotePub: channel.IdentityPub,
|
|
SettledBalance: btcutil.Amount(500),
|
|
}); err != nil {
|
|
t.Fatalf("unexpected error closing channel: %v", err)
|
|
}
|
|
|
|
histChannel, err := cdb.FetchHistoricalChannel(&channel.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unexepected error getting channel: %v", err)
|
|
}
|
|
|
|
// Set the db on our channel to nil so that we can check that all other
|
|
// fields on the channel equal those on the historical channel.
|
|
channel.Db = nil
|
|
|
|
if !reflect.DeepEqual(histChannel, channel) {
|
|
t.Fatalf("expected: %v, got: %v", channel, histChannel)
|
|
}
|
|
|
|
// Create an outpoint that will not be in the db and look it up.
|
|
badOutpoint := &wire.OutPoint{
|
|
Hash: channel.FundingOutpoint.Hash,
|
|
Index: channel.FundingOutpoint.Index + 1,
|
|
}
|
|
_, err = cdb.FetchHistoricalChannel(badOutpoint)
|
|
if err != ErrChannelNotFound {
|
|
t.Fatalf("expected chan not found, got: %v", err)
|
|
}
|
|
|
|
}
|