d23be59f53
This was initially done as there were a few assertions throughout the codebase requiring a channel's policy to be known. Now that these have been addressed, we no longer need to store restored channels in the graph, as their policies where incomplete anyway.
741 lines
21 KiB
Go
741 lines
21 KiB
Go
package channeldb
|
|
|
|
import (
|
|
"io/ioutil"
|
|
"math"
|
|
"math/rand"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"reflect"
|
|
"testing"
|
|
|
|
"github.com/btcsuite/btcd/btcec"
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
|
"github.com/btcsuite/btcd/wire"
|
|
"github.com/btcsuite/btcutil"
|
|
"github.com/davecgh/go-spew/spew"
|
|
"github.com/lightningnetwork/lnd/channeldb/kvdb"
|
|
"github.com/lightningnetwork/lnd/keychain"
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
"github.com/lightningnetwork/lnd/shachain"
|
|
)
|
|
|
|
func TestOpenWithCreate(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, create a temporary directory to be used for the duration of
|
|
// this test.
|
|
tempDirName, err := ioutil.TempDir("", "channeldb")
|
|
if err != nil {
|
|
t.Fatalf("unable to create temp dir: %v", err)
|
|
}
|
|
defer os.RemoveAll(tempDirName)
|
|
|
|
// Next, open thereby creating channeldb for the first time.
|
|
dbPath := filepath.Join(tempDirName, "cdb")
|
|
backend, cleanup, err := kvdb.GetTestBackend(dbPath, "cdb")
|
|
if err != nil {
|
|
t.Fatalf("unable to get test db backend: %v", err)
|
|
}
|
|
defer cleanup()
|
|
|
|
cdb, err := CreateWithBackend(backend)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channeldb: %v", err)
|
|
}
|
|
if err := cdb.Close(); err != nil {
|
|
t.Fatalf("unable to close channeldb: %v", err)
|
|
}
|
|
|
|
// The path should have been successfully created.
|
|
if !fileExists(dbPath) {
|
|
t.Fatalf("channeldb failed to create data directory")
|
|
}
|
|
|
|
// Now, reopen the same db in dry run migration mode. Since we have not
|
|
// applied any migrations, this should ignore the flag and not fail.
|
|
cdb, err = Open(dbPath, OptionDryRunMigration(true))
|
|
if err != nil {
|
|
t.Fatalf("unable to create channeldb: %v", err)
|
|
}
|
|
if err := cdb.Close(); err != nil {
|
|
t.Fatalf("unable to close channeldb: %v", err)
|
|
}
|
|
}
|
|
|
|
// TestWipe tests that the database wipe operation completes successfully
|
|
// and that the buckets are deleted. It also checks that attempts to fetch
|
|
// information while the buckets are not set return the correct errors.
|
|
func TestWipe(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
// First, create a temporary directory to be used for the duration of
|
|
// this test.
|
|
tempDirName, err := ioutil.TempDir("", "channeldb")
|
|
if err != nil {
|
|
t.Fatalf("unable to create temp dir: %v", err)
|
|
}
|
|
defer os.RemoveAll(tempDirName)
|
|
|
|
// Next, open thereby creating channeldb for the first time.
|
|
dbPath := filepath.Join(tempDirName, "cdb")
|
|
backend, cleanup, err := kvdb.GetTestBackend(dbPath, "cdb")
|
|
if err != nil {
|
|
t.Fatalf("unable to get test db backend: %v", err)
|
|
}
|
|
defer cleanup()
|
|
|
|
cdb, err := CreateWithBackend(backend)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channeldb: %v", err)
|
|
}
|
|
defer cdb.Close()
|
|
|
|
if err := cdb.Wipe(); err != nil {
|
|
t.Fatalf("unable to wipe channeldb: %v", err)
|
|
}
|
|
// Check correct errors are returned
|
|
_, err = cdb.FetchAllOpenChannels()
|
|
if err != ErrNoActiveChannels {
|
|
t.Fatalf("fetching open channels: expected '%v' instead got '%v'",
|
|
ErrNoActiveChannels, err)
|
|
}
|
|
_, err = cdb.FetchClosedChannels(false)
|
|
if err != ErrNoClosedChannels {
|
|
t.Fatalf("fetching closed channels: expected '%v' instead got '%v'",
|
|
ErrNoClosedChannels, err)
|
|
}
|
|
}
|
|
|
|
// TestFetchClosedChannelForID tests that we are able to properly retrieve a
|
|
// ChannelCloseSummary from the DB given a ChannelID.
|
|
func TestFetchClosedChannelForID(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
const numChans = 101
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create the test channel state, that we will mutate the index of the
|
|
// funding point.
|
|
state := createTestChannelState(t, cdb)
|
|
|
|
// Now run through the number of channels, and modify the outpoint index
|
|
// to create new channel IDs.
|
|
for i := uint32(0); i < numChans; i++ {
|
|
// Save the open channel to disk.
|
|
state.FundingOutpoint.Index = i
|
|
|
|
// Write the channel to disk in a pending state.
|
|
createTestChannel(
|
|
t, cdb,
|
|
fundingPointOption(state.FundingOutpoint),
|
|
openChannelOption(),
|
|
)
|
|
|
|
// Close the channel. To make sure we retrieve the correct
|
|
// summary later, we make them differ in the SettledBalance.
|
|
closeSummary := &ChannelCloseSummary{
|
|
ChanPoint: state.FundingOutpoint,
|
|
RemotePub: state.IdentityPub,
|
|
SettledBalance: btcutil.Amount(500 + i),
|
|
}
|
|
if err := state.CloseChannel(closeSummary); err != nil {
|
|
t.Fatalf("unable to close channel: %v", err)
|
|
}
|
|
}
|
|
|
|
// Now run though them all again and make sure we are able to retrieve
|
|
// summaries from the DB.
|
|
for i := uint32(0); i < numChans; i++ {
|
|
state.FundingOutpoint.Index = i
|
|
|
|
// We calculate the ChannelID and use it to fetch the summary.
|
|
cid := lnwire.NewChanIDFromOutPoint(&state.FundingOutpoint)
|
|
fetchedSummary, err := cdb.FetchClosedChannelForID(cid)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch close summary: %v", err)
|
|
}
|
|
|
|
// Make sure we retrieved the correct one by checking the
|
|
// SettledBalance.
|
|
if fetchedSummary.SettledBalance != btcutil.Amount(500+i) {
|
|
t.Fatalf("summaries don't match: expected %v got %v",
|
|
btcutil.Amount(500+i),
|
|
fetchedSummary.SettledBalance)
|
|
}
|
|
}
|
|
|
|
// As a final test we make sure that we get ErrClosedChannelNotFound
|
|
// for a ChannelID we didn't add to the DB.
|
|
state.FundingOutpoint.Index++
|
|
cid := lnwire.NewChanIDFromOutPoint(&state.FundingOutpoint)
|
|
_, err = cdb.FetchClosedChannelForID(cid)
|
|
if err != ErrClosedChannelNotFound {
|
|
t.Fatalf("expected ErrClosedChannelNotFound, instead got: %v", err)
|
|
}
|
|
}
|
|
|
|
// TestAddrsForNode tests the we're able to properly obtain all the addresses
|
|
// for a target node.
|
|
func TestAddrsForNode(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
graph := cdb.ChannelGraph()
|
|
|
|
// We'll make a test vertex to insert into the database, as the source
|
|
// node, but this node will only have half the number of addresses it
|
|
// usually does.
|
|
testNode, err := createTestVertex(cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create test node: %v", err)
|
|
}
|
|
testNode.Addresses = []net.Addr{testAddr}
|
|
if err := graph.SetSourceNode(testNode); err != nil {
|
|
t.Fatalf("unable to set source node: %v", err)
|
|
}
|
|
|
|
// Next, we'll make a link node with the same pubkey, but with an
|
|
// additional address.
|
|
nodePub, err := testNode.PubKey()
|
|
if err != nil {
|
|
t.Fatalf("unable to recv node pub: %v", err)
|
|
}
|
|
linkNode := cdb.NewLinkNode(
|
|
wire.MainNet, nodePub, anotherAddr,
|
|
)
|
|
if err := linkNode.Sync(); err != nil {
|
|
t.Fatalf("unable to sync link node: %v", err)
|
|
}
|
|
|
|
// Now that we've created a link node, as well as a vertex for the
|
|
// node, we'll query for all its addresses.
|
|
nodeAddrs, err := cdb.AddrsForNode(nodePub)
|
|
if err != nil {
|
|
t.Fatalf("unable to obtain node addrs: %v", err)
|
|
}
|
|
|
|
expectedAddrs := make(map[string]struct{})
|
|
expectedAddrs[testAddr.String()] = struct{}{}
|
|
expectedAddrs[anotherAddr.String()] = struct{}{}
|
|
|
|
// Finally, ensure that all the expected addresses are found.
|
|
if len(nodeAddrs) != len(expectedAddrs) {
|
|
t.Fatalf("expected %v addrs, got %v",
|
|
len(expectedAddrs), len(nodeAddrs))
|
|
}
|
|
for _, addr := range nodeAddrs {
|
|
if _, ok := expectedAddrs[addr.String()]; !ok {
|
|
t.Fatalf("unexpected addr: %v", addr)
|
|
}
|
|
}
|
|
}
|
|
|
|
// TestFetchChannel tests that we're able to fetch an arbitrary channel from
|
|
// disk.
|
|
func TestFetchChannel(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create an open channel.
|
|
channelState := createTestChannel(t, cdb, openChannelOption())
|
|
|
|
// Next, attempt to fetch the channel by its chan point.
|
|
dbChannel, err := cdb.FetchChannel(channelState.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch channel: %v", err)
|
|
}
|
|
|
|
// The decoded channel state should be identical to what we stored
|
|
// above.
|
|
if !reflect.DeepEqual(channelState, dbChannel) {
|
|
t.Fatalf("channel state doesn't match:: %v vs %v",
|
|
spew.Sdump(channelState), spew.Sdump(dbChannel))
|
|
}
|
|
|
|
// If we attempt to query for a non-exist ante channel, then we should
|
|
// get an error.
|
|
channelState2 := createTestChannelState(t, cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channel state: %v", err)
|
|
}
|
|
channelState2.FundingOutpoint.Index ^= 1
|
|
|
|
_, err = cdb.FetchChannel(channelState2.FundingOutpoint)
|
|
if err == nil {
|
|
t.Fatalf("expected query to fail")
|
|
}
|
|
}
|
|
|
|
func genRandomChannelShell() (*ChannelShell, error) {
|
|
var testPriv [32]byte
|
|
if _, err := rand.Read(testPriv[:]); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
_, pub := btcec.PrivKeyFromBytes(btcec.S256(), testPriv[:])
|
|
|
|
var chanPoint wire.OutPoint
|
|
if _, err := rand.Read(chanPoint.Hash[:]); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
pub.Curve = nil
|
|
|
|
chanPoint.Index = uint32(rand.Intn(math.MaxUint16))
|
|
|
|
chanStatus := ChanStatusDefault | ChanStatusRestored
|
|
|
|
var shaChainPriv [32]byte
|
|
if _, err := rand.Read(testPriv[:]); err != nil {
|
|
return nil, err
|
|
}
|
|
revRoot, err := chainhash.NewHash(shaChainPriv[:])
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
shaChainProducer := shachain.NewRevocationProducer(*revRoot)
|
|
|
|
return &ChannelShell{
|
|
NodeAddrs: []net.Addr{&net.TCPAddr{
|
|
IP: net.ParseIP("127.0.0.1"),
|
|
Port: 18555,
|
|
}},
|
|
Chan: &OpenChannel{
|
|
chanStatus: chanStatus,
|
|
ChainHash: rev,
|
|
FundingOutpoint: chanPoint,
|
|
ShortChannelID: lnwire.NewShortChanIDFromInt(
|
|
uint64(rand.Int63()),
|
|
),
|
|
IdentityPub: pub,
|
|
LocalChanCfg: ChannelConfig{
|
|
ChannelConstraints: ChannelConstraints{
|
|
CsvDelay: uint16(rand.Int63()),
|
|
},
|
|
PaymentBasePoint: keychain.KeyDescriptor{
|
|
KeyLocator: keychain.KeyLocator{
|
|
Family: keychain.KeyFamily(rand.Int63()),
|
|
Index: uint32(rand.Int63()),
|
|
},
|
|
},
|
|
},
|
|
RemoteCurrentRevocation: pub,
|
|
IsPending: false,
|
|
RevocationStore: shachain.NewRevocationStore(),
|
|
RevocationProducer: shaChainProducer,
|
|
},
|
|
}, nil
|
|
}
|
|
|
|
// TestRestoreChannelShells tests that we're able to insert a partially channel
|
|
// populated to disk. This is useful for channel recovery purposes. We should
|
|
// find the new channel shell on disk, and also the db should be populated with
|
|
// an edge for that channel.
|
|
func TestRestoreChannelShells(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// First, we'll make our channel shell, it will only have the minimal
|
|
// amount of information required for us to initiate the data loss
|
|
// protection feature.
|
|
channelShell, err := genRandomChannelShell()
|
|
if err != nil {
|
|
t.Fatalf("unable to gen channel shell: %v", err)
|
|
}
|
|
|
|
// With the channel shell constructed, we'll now insert it into the
|
|
// database with the restoration method.
|
|
if err := cdb.RestoreChannelShells(channelShell); err != nil {
|
|
t.Fatalf("unable to restore channel shell: %v", err)
|
|
}
|
|
|
|
// Now that the channel has been inserted, we'll attempt to query for
|
|
// it to ensure we can properly locate it via various means.
|
|
//
|
|
// First, we'll attempt to query for all channels that we have with the
|
|
// node public key that was restored.
|
|
nodeChans, err := cdb.FetchOpenChannels(channelShell.Chan.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable find channel: %v", err)
|
|
}
|
|
|
|
// We should now find a single channel from the database.
|
|
if len(nodeChans) != 1 {
|
|
t.Fatalf("unable to find restored channel by node "+
|
|
"pubkey: %v", err)
|
|
}
|
|
|
|
// Ensure that it isn't possible to modify the commitment state machine
|
|
// of this restored channel.
|
|
channel := nodeChans[0]
|
|
err = channel.UpdateCommitment(nil, nil)
|
|
if err != ErrNoRestoredChannelMutation {
|
|
t.Fatalf("able to mutate restored channel")
|
|
}
|
|
err = channel.AppendRemoteCommitChain(nil)
|
|
if err != ErrNoRestoredChannelMutation {
|
|
t.Fatalf("able to mutate restored channel")
|
|
}
|
|
err = channel.AdvanceCommitChainTail(nil)
|
|
if err != ErrNoRestoredChannelMutation {
|
|
t.Fatalf("able to mutate restored channel")
|
|
}
|
|
|
|
// That single channel should have the proper channel point, and also
|
|
// the expected set of flags to indicate that it was a restored
|
|
// channel.
|
|
if nodeChans[0].FundingOutpoint != channelShell.Chan.FundingOutpoint {
|
|
t.Fatalf("wrong funding outpoint: expected %v, got %v",
|
|
nodeChans[0].FundingOutpoint,
|
|
channelShell.Chan.FundingOutpoint)
|
|
}
|
|
if !nodeChans[0].HasChanStatus(ChanStatusRestored) {
|
|
t.Fatalf("node has wrong status flags: %v",
|
|
nodeChans[0].chanStatus)
|
|
}
|
|
|
|
// We should also be able to find the channel if we query for it
|
|
// directly.
|
|
_, err = cdb.FetchChannel(channelShell.Chan.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch channel: %v", err)
|
|
}
|
|
|
|
// We should also be able to find the link node that was inserted by
|
|
// its public key.
|
|
linkNode, err := cdb.FetchLinkNode(channelShell.Chan.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch link node: %v", err)
|
|
}
|
|
|
|
// The node should have the same address, as specified in the channel
|
|
// shell.
|
|
if reflect.DeepEqual(linkNode.Addresses, channelShell.NodeAddrs) {
|
|
t.Fatalf("addr mismach: expected %v, got %v",
|
|
linkNode.Addresses, channelShell.NodeAddrs)
|
|
}
|
|
}
|
|
|
|
// TestAbandonChannel tests that the AbandonChannel method is able to properly
|
|
// remove a channel from the database and add a close channel summary. If
|
|
// called after a channel has already been removed, the method shouldn't return
|
|
// an error.
|
|
func TestAbandonChannel(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// If we attempt to abandon the state of a channel that doesn't exist
|
|
// in the open or closed channel bucket, then we should receive an
|
|
// error.
|
|
err = cdb.AbandonChannel(&wire.OutPoint{}, 0)
|
|
if err == nil {
|
|
t.Fatalf("removing non-existent channel should have failed")
|
|
}
|
|
|
|
// We'll now create a new channel in a pending state to abandon
|
|
// shortly.
|
|
chanState := createTestChannel(t, cdb)
|
|
|
|
// We should now be able to abandon the channel without any errors.
|
|
closeHeight := uint32(11)
|
|
err = cdb.AbandonChannel(&chanState.FundingOutpoint, closeHeight)
|
|
if err != nil {
|
|
t.Fatalf("unable to abandon channel: %v", err)
|
|
}
|
|
|
|
// At this point, the channel should no longer be found in the set of
|
|
// open channels.
|
|
_, err = cdb.FetchChannel(chanState.FundingOutpoint)
|
|
if err != ErrChannelNotFound {
|
|
t.Fatalf("channel should not have been found: %v", err)
|
|
}
|
|
|
|
// However we should be able to retrieve a close channel summary for
|
|
// the channel.
|
|
_, err = cdb.FetchClosedChannel(&chanState.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch closed channel: %v", err)
|
|
}
|
|
|
|
// Finally, if we attempt to abandon the channel again, we should get a
|
|
// nil error as the channel has already been abandoned.
|
|
err = cdb.AbandonChannel(&chanState.FundingOutpoint, closeHeight)
|
|
if err != nil {
|
|
t.Fatalf("unable to abandon channel: %v", err)
|
|
}
|
|
}
|
|
|
|
// TestFetchChannels tests the filtering of open channels in fetchChannels.
|
|
// It tests the case where no filters are provided (which is equivalent to
|
|
// FetchAllOpenChannels) and every combination of pending and waiting close.
|
|
func TestFetchChannels(t *testing.T) {
|
|
// Create static channel IDs for each kind of channel retrieved by
|
|
// fetchChannels so that the expected channel IDs can be set in tests.
|
|
var (
|
|
// Pending is a channel that is pending open, and has not had
|
|
// a close initiated.
|
|
pendingChan = lnwire.NewShortChanIDFromInt(1)
|
|
|
|
// pendingWaitingClose is a channel that is pending open and
|
|
// has has its closing transaction broadcast.
|
|
pendingWaitingChan = lnwire.NewShortChanIDFromInt(2)
|
|
|
|
// openChan is a channel that has confirmed on chain.
|
|
openChan = lnwire.NewShortChanIDFromInt(3)
|
|
|
|
// openWaitingChan is a channel that has confirmed on chain,
|
|
// and it waiting for its close transaction to confirm.
|
|
openWaitingChan = lnwire.NewShortChanIDFromInt(4)
|
|
)
|
|
|
|
tests := []struct {
|
|
name string
|
|
filters []fetchChannelsFilter
|
|
expectedChannels map[lnwire.ShortChannelID]bool
|
|
}{
|
|
{
|
|
name: "get all channels",
|
|
filters: []fetchChannelsFilter{},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
pendingWaitingChan: true,
|
|
openChan: true,
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "pending channels",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
pendingWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "open channels",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
openChan: true,
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "waiting close channels",
|
|
filters: []fetchChannelsFilter{
|
|
waitingCloseFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingWaitingChan: true,
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "not waiting close channels",
|
|
filters: []fetchChannelsFilter{
|
|
waitingCloseFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
openChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "pending waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(true),
|
|
waitingCloseFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "pending, not waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(true),
|
|
waitingCloseFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
pendingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "open waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(false),
|
|
waitingCloseFilter(true),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
openWaitingChan: true,
|
|
},
|
|
},
|
|
{
|
|
name: "open, not waiting",
|
|
filters: []fetchChannelsFilter{
|
|
pendingChannelFilter(false),
|
|
waitingCloseFilter(false),
|
|
},
|
|
expectedChannels: map[lnwire.ShortChannelID]bool{
|
|
openChan: true,
|
|
},
|
|
},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
test := test
|
|
|
|
t.Run(test.name, func(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test "+
|
|
"database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create a pending channel that is not awaiting close.
|
|
createTestChannel(
|
|
t, cdb, channelIDOption(pendingChan),
|
|
)
|
|
|
|
// Create a pending channel which has has been marked as
|
|
// broadcast, indicating that its closing transaction is
|
|
// waiting to confirm.
|
|
pendingClosing := createTestChannel(
|
|
t, cdb,
|
|
channelIDOption(pendingWaitingChan),
|
|
)
|
|
|
|
err = pendingClosing.MarkCoopBroadcasted(nil, true)
|
|
if err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
// Create a open channel that is not awaiting close.
|
|
createTestChannel(
|
|
t, cdb,
|
|
channelIDOption(openChan),
|
|
openChannelOption(),
|
|
)
|
|
|
|
// Create a open channel which has has been marked as
|
|
// broadcast, indicating that its closing transaction is
|
|
// waiting to confirm.
|
|
openClosing := createTestChannel(
|
|
t, cdb,
|
|
channelIDOption(openWaitingChan),
|
|
openChannelOption(),
|
|
)
|
|
err = openClosing.MarkCoopBroadcasted(nil, true)
|
|
if err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
channels, err := fetchChannels(cdb, test.filters...)
|
|
if err != nil {
|
|
t.Fatalf("unexpected error: %v", err)
|
|
}
|
|
|
|
if len(channels) != len(test.expectedChannels) {
|
|
t.Fatalf("expected: %v channels, "+
|
|
"got: %v", len(test.expectedChannels),
|
|
len(channels))
|
|
}
|
|
|
|
for _, ch := range channels {
|
|
_, ok := test.expectedChannels[ch.ShortChannelID]
|
|
if !ok {
|
|
t.Fatalf("fetch channels unexpected "+
|
|
"channel: %v", ch.ShortChannelID)
|
|
}
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
// TestFetchHistoricalChannel tests lookup of historical channels.
|
|
func TestFetchHistoricalChannel(t *testing.T) {
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create a an open channel in the database.
|
|
channel := createTestChannel(t, cdb, openChannelOption())
|
|
|
|
// First, try to lookup a channel when the bucket does not
|
|
// exist.
|
|
_, err = cdb.FetchHistoricalChannel(&channel.FundingOutpoint)
|
|
if err != ErrNoHistoricalBucket {
|
|
t.Fatalf("expected no bucket, got: %v", err)
|
|
}
|
|
|
|
// Close the channel so that it will be written to the historical
|
|
// bucket. The values provided in the channel close summary are the
|
|
// minimum required for this call to run without panicking.
|
|
if err := channel.CloseChannel(&ChannelCloseSummary{
|
|
ChanPoint: channel.FundingOutpoint,
|
|
RemotePub: channel.IdentityPub,
|
|
SettledBalance: btcutil.Amount(500),
|
|
}); err != nil {
|
|
t.Fatalf("unexpected error closing channel: %v", err)
|
|
}
|
|
|
|
histChannel, err := cdb.FetchHistoricalChannel(&channel.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("unexepected error getting channel: %v", err)
|
|
}
|
|
|
|
// Set the db on our channel to nil so that we can check that all other
|
|
// fields on the channel equal those on the historical channel.
|
|
channel.Db = nil
|
|
|
|
if !reflect.DeepEqual(histChannel, channel) {
|
|
t.Fatalf("expected: %v, got: %v", channel, histChannel)
|
|
}
|
|
|
|
// Create an outpoint that will not be in the db and look it up.
|
|
badOutpoint := &wire.OutPoint{
|
|
Hash: channel.FundingOutpoint.Hash,
|
|
Index: channel.FundingOutpoint.Index + 1,
|
|
}
|
|
_, err = cdb.FetchHistoricalChannel(badOutpoint)
|
|
if err != ErrChannelNotFound {
|
|
t.Fatalf("expected chan not found, got: %v", err)
|
|
}
|
|
|
|
}
|