7aa7cb1caf
In this commit htlc channeldb representation have been augmented with onion blob field, and (de)serialisaion functions have been changed to make the onion blob persistant.
747 lines
23 KiB
Go
747 lines
23 KiB
Go
package channeldb
|
|
|
|
import (
|
|
"bytes"
|
|
"io/ioutil"
|
|
"math/rand"
|
|
"net"
|
|
"os"
|
|
"reflect"
|
|
"testing"
|
|
|
|
"github.com/davecgh/go-spew/spew"
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
"github.com/lightningnetwork/lnd/shachain"
|
|
"github.com/roasbeef/btcd/btcec"
|
|
"github.com/roasbeef/btcd/chaincfg"
|
|
"github.com/roasbeef/btcd/chaincfg/chainhash"
|
|
"github.com/roasbeef/btcd/wire"
|
|
"github.com/roasbeef/btcutil"
|
|
_ "github.com/roasbeef/btcwallet/walletdb/bdb"
|
|
)
|
|
|
|
var (
|
|
netParams = &chaincfg.TestNet3Params
|
|
|
|
key = [chainhash.HashSize]byte{
|
|
0x81, 0xb6, 0x37, 0xd8, 0xfc, 0xd2, 0xc6, 0xda,
|
|
0x68, 0x59, 0xe6, 0x96, 0x31, 0x13, 0xa1, 0x17,
|
|
0xd, 0xe7, 0x93, 0xe4, 0xb7, 0x25, 0xb8, 0x4d,
|
|
0x1e, 0xb, 0x4c, 0xf9, 0x9e, 0xc5, 0x8c, 0xe9,
|
|
}
|
|
id = &wire.OutPoint{
|
|
Hash: [chainhash.HashSize]byte{
|
|
0x51, 0xb6, 0x37, 0xd8, 0xfc, 0xd2, 0xc6, 0xda,
|
|
0x48, 0x59, 0xe6, 0x96, 0x31, 0x13, 0xa1, 0x17,
|
|
0x2d, 0xe7, 0x93, 0xe4, 0xb7, 0x25, 0xb8, 0x4d,
|
|
0x1f, 0xb, 0x4c, 0xf9, 0x9e, 0xc5, 0x8c, 0xe9,
|
|
},
|
|
Index: 9,
|
|
}
|
|
rev = [chainhash.HashSize]byte{
|
|
0x51, 0xb6, 0x37, 0xd8, 0xfc, 0xd2, 0xc6, 0xda,
|
|
0x48, 0x59, 0xe6, 0x96, 0x31, 0x13, 0xa1, 0x17,
|
|
0x2d, 0xe7, 0x93, 0xe4,
|
|
}
|
|
testTx = &wire.MsgTx{
|
|
Version: 1,
|
|
TxIn: []*wire.TxIn{
|
|
{
|
|
PreviousOutPoint: wire.OutPoint{
|
|
Hash: chainhash.Hash{},
|
|
Index: 0xffffffff,
|
|
},
|
|
SignatureScript: []byte{0x04, 0x31, 0xdc, 0x00, 0x1b, 0x01, 0x62},
|
|
Sequence: 0xffffffff,
|
|
},
|
|
},
|
|
TxOut: []*wire.TxOut{
|
|
{
|
|
Value: 5000000000,
|
|
PkScript: []byte{
|
|
0x41, // OP_DATA_65
|
|
0x04, 0xd6, 0x4b, 0xdf, 0xd0, 0x9e, 0xb1, 0xc5,
|
|
0xfe, 0x29, 0x5a, 0xbd, 0xeb, 0x1d, 0xca, 0x42,
|
|
0x81, 0xbe, 0x98, 0x8e, 0x2d, 0xa0, 0xb6, 0xc1,
|
|
0xc6, 0xa5, 0x9d, 0xc2, 0x26, 0xc2, 0x86, 0x24,
|
|
0xe1, 0x81, 0x75, 0xe8, 0x51, 0xc9, 0x6b, 0x97,
|
|
0x3d, 0x81, 0xb0, 0x1c, 0xc3, 0x1f, 0x04, 0x78,
|
|
0x34, 0xbc, 0x06, 0xd6, 0xd6, 0xed, 0xf6, 0x20,
|
|
0xd1, 0x84, 0x24, 0x1a, 0x6a, 0xed, 0x8b, 0x63,
|
|
0xa6, // 65-byte signature
|
|
0xac, // OP_CHECKSIG
|
|
},
|
|
},
|
|
},
|
|
LockTime: 5,
|
|
}
|
|
testOutpoint = &wire.OutPoint{
|
|
Hash: key,
|
|
Index: 0,
|
|
}
|
|
privKey, pubKey = btcec.PrivKeyFromBytes(btcec.S256(), key[:])
|
|
)
|
|
|
|
// makeTestDB creates a new instance of the ChannelDB for testing purposes. A
|
|
// callback which cleans up the created temporary directories is also returned
|
|
// and intended to be executed after the test completes.
|
|
func makeTestDB() (*DB, func(), error) {
|
|
// First, create a temporary directory to be used for the duration of
|
|
// this test.
|
|
tempDirName, err := ioutil.TempDir("", "channeldb")
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
// Next, create channeldb for the first time.
|
|
cdb, err := Open(tempDirName)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
cleanUp := func() {
|
|
cdb.Close()
|
|
os.RemoveAll(tempDirName)
|
|
}
|
|
|
|
return cdb, cleanUp, nil
|
|
}
|
|
|
|
func createTestChannelState(cdb *DB) (*OpenChannel, error) {
|
|
// Simulate 1000 channel updates.
|
|
producer, err := shachain.NewRevocationProducerFromBytes(key[:])
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
store := shachain.NewRevocationStore()
|
|
for i := 0; i < 1; i++ {
|
|
preImage, err := producer.AtIndex(uint64(i))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if store.AddNextEntry(preImage); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
localCfg := ChannelConfig{
|
|
ChannelConstraints: ChannelConstraints{
|
|
DustLimit: btcutil.Amount(rand.Int63()),
|
|
MaxPendingAmount: lnwire.MilliSatoshi(rand.Int63()),
|
|
ChanReserve: btcutil.Amount(rand.Int63()),
|
|
MinHTLC: lnwire.MilliSatoshi(rand.Int63()),
|
|
MaxAcceptedHtlcs: uint16(rand.Int31()),
|
|
},
|
|
CsvDelay: uint16(rand.Int31()),
|
|
MultiSigKey: privKey.PubKey(),
|
|
RevocationBasePoint: privKey.PubKey(),
|
|
PaymentBasePoint: privKey.PubKey(),
|
|
DelayBasePoint: privKey.PubKey(),
|
|
}
|
|
remoteCfg := ChannelConfig{
|
|
ChannelConstraints: ChannelConstraints{
|
|
DustLimit: btcutil.Amount(rand.Int63()),
|
|
MaxPendingAmount: lnwire.MilliSatoshi(rand.Int63()),
|
|
ChanReserve: btcutil.Amount(rand.Int63()),
|
|
MinHTLC: lnwire.MilliSatoshi(rand.Int63()),
|
|
MaxAcceptedHtlcs: uint16(rand.Int31()),
|
|
},
|
|
CsvDelay: uint16(rand.Int31()),
|
|
MultiSigKey: privKey.PubKey(),
|
|
RevocationBasePoint: privKey.PubKey(),
|
|
PaymentBasePoint: privKey.PubKey(),
|
|
DelayBasePoint: privKey.PubKey(),
|
|
}
|
|
|
|
chanID := lnwire.NewShortChanIDFromInt(uint64(rand.Int63()))
|
|
|
|
return &OpenChannel{
|
|
ChanType: SingleFunder,
|
|
ChainHash: key,
|
|
FundingOutpoint: *testOutpoint,
|
|
ShortChanID: chanID,
|
|
IsInitiator: true,
|
|
IsPending: true,
|
|
IdentityPub: pubKey,
|
|
LocalChanCfg: localCfg,
|
|
RemoteChanCfg: remoteCfg,
|
|
CommitFee: btcutil.Amount(rand.Int63()),
|
|
FeePerKw: btcutil.Amount(5000),
|
|
Capacity: btcutil.Amount(10000),
|
|
LocalBalance: lnwire.MilliSatoshi(3000),
|
|
RemoteBalance: lnwire.MilliSatoshi(9000),
|
|
CommitTx: *testTx,
|
|
CommitSig: bytes.Repeat([]byte{1}, 71),
|
|
NumConfsRequired: 4,
|
|
RemoteCurrentRevocation: privKey.PubKey(),
|
|
RemoteNextRevocation: privKey.PubKey(),
|
|
RevocationProducer: producer,
|
|
RevocationStore: store,
|
|
NumUpdates: 0,
|
|
TotalMSatSent: 8,
|
|
TotalMSatReceived: 2,
|
|
Db: cdb,
|
|
}, nil
|
|
}
|
|
|
|
func TestOpenChannelPutGetDelete(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
defer cleanUp()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
|
|
// Create the test channel state, then add an additional fake HTLC
|
|
// before syncing to disk.
|
|
state, err := createTestChannelState(cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channel state: %v", err)
|
|
}
|
|
state.Htlcs = []*HTLC{
|
|
{
|
|
Signature: testSig.Serialize(),
|
|
Incoming: true,
|
|
Amt: 10,
|
|
RHash: key,
|
|
RefundTimeout: 1,
|
|
OnionBlob: []byte("onionblob"),
|
|
},
|
|
}
|
|
if err := state.FullSync(); err != nil {
|
|
t.Fatalf("unable to save and serialize channel state: %v", err)
|
|
}
|
|
|
|
openChannels, err := cdb.FetchOpenChannels(state.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch open channel: %v", err)
|
|
}
|
|
|
|
newState := openChannels[0]
|
|
|
|
// The decoded channel state should be identical to what we stored
|
|
// above.
|
|
if !reflect.DeepEqual(state, newState) {
|
|
state.LocalChanCfg.MultiSigKey.Curve = nil
|
|
state.LocalChanCfg.RevocationBasePoint.Curve = nil
|
|
state.LocalChanCfg.PaymentBasePoint.Curve = nil
|
|
state.LocalChanCfg.DelayBasePoint.Curve = nil
|
|
|
|
state.RemoteChanCfg.MultiSigKey.Curve = nil
|
|
state.RemoteChanCfg.RevocationBasePoint.Curve = nil
|
|
state.RemoteChanCfg.PaymentBasePoint.Curve = nil
|
|
state.RemoteChanCfg.DelayBasePoint.Curve = nil
|
|
|
|
state.IdentityPub.Curve = nil
|
|
state.RemoteNextRevocation.Curve = nil
|
|
state.RemoteCurrentRevocation.Curve = nil
|
|
|
|
newState.LocalChanCfg.MultiSigKey.Curve = nil
|
|
newState.LocalChanCfg.RevocationBasePoint.Curve = nil
|
|
newState.LocalChanCfg.PaymentBasePoint.Curve = nil
|
|
newState.LocalChanCfg.DelayBasePoint.Curve = nil
|
|
|
|
newState.RemoteChanCfg.MultiSigKey.Curve = nil
|
|
newState.RemoteChanCfg.RevocationBasePoint.Curve = nil
|
|
newState.RemoteChanCfg.PaymentBasePoint.Curve = nil
|
|
newState.RemoteChanCfg.DelayBasePoint.Curve = nil
|
|
|
|
newState.IdentityPub.Curve = nil
|
|
newState.RemoteCurrentRevocation.Curve = nil
|
|
newState.RemoteNextRevocation.Curve = nil
|
|
t.Fatalf("channel state doesn't match:: %v vs %v",
|
|
spew.Sdump(state), spew.Sdump(newState))
|
|
}
|
|
|
|
// We'll also test that the channel is properly able to hot swap the
|
|
// next revocation for the state machine. This tests the initial
|
|
// post-funding revocation exchange.
|
|
nextRevKey, err := btcec.NewPrivateKey(btcec.S256())
|
|
if err != nil {
|
|
t.Fatalf("unable to create new private key: %v", err)
|
|
}
|
|
if err := state.InsertNextRevocation(nextRevKey.PubKey()); err != nil {
|
|
t.Fatalf("unable to update revocation: %v", err)
|
|
}
|
|
|
|
openChannels, err = cdb.FetchOpenChannels(state.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch open channel: %v", err)
|
|
}
|
|
updatedChan := openChannels[0]
|
|
|
|
// Ensure that the revocation was set properly.
|
|
if !nextRevKey.PubKey().IsEqual(updatedChan.RemoteNextRevocation) {
|
|
t.Fatalf("next revocation wasn't updated")
|
|
}
|
|
|
|
// Finally to wrap up the test, delete the state of the channel within
|
|
// the database. This involves "closing" the channel which removes all
|
|
// written state, and creates a small "summary" elsewhere within the
|
|
// database.
|
|
closeSummary := &ChannelCloseSummary{
|
|
ChanPoint: state.FundingOutpoint,
|
|
RemotePub: state.IdentityPub,
|
|
SettledBalance: btcutil.Amount(500),
|
|
TimeLockedBalance: btcutil.Amount(10000),
|
|
IsPending: false,
|
|
CloseType: CooperativeClose,
|
|
}
|
|
if err := state.CloseChannel(closeSummary); err != nil {
|
|
t.Fatalf("unable to close channel: %v", err)
|
|
}
|
|
|
|
// As the channel is now closed, attempting to fetch all open channels
|
|
// for our fake node ID should return an empty slice.
|
|
openChans, err := cdb.FetchOpenChannels(state.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch open channels: %v", err)
|
|
}
|
|
if len(openChans) != 0 {
|
|
t.Fatalf("all channels not deleted, found %v", len(openChans))
|
|
}
|
|
|
|
// Additionally, attempting to fetch all the open channels globally
|
|
// should yield no results.
|
|
openChans, err = cdb.FetchAllChannels()
|
|
if err != nil {
|
|
t.Fatal("unable to fetch all open chans")
|
|
}
|
|
if len(openChans) != 0 {
|
|
t.Fatalf("all channels not deleted, found %v", len(openChans))
|
|
}
|
|
}
|
|
|
|
func TestChannelStateTransition(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
defer cleanUp()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
|
|
// First create a minimal channel, then perform a full sync in order to
|
|
// persist the data.
|
|
channel, err := createTestChannelState(cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channel state: %v", err)
|
|
}
|
|
if err := channel.FullSync(); err != nil {
|
|
t.Fatalf("unable to save and serialize channel state: %v", err)
|
|
}
|
|
|
|
// Add some HTLCs which were added during this new state transition.
|
|
// Half of the HTLCs are incoming, while the other half are outgoing.
|
|
var (
|
|
htlcs []*HTLC
|
|
htlcAmt lnwire.MilliSatoshi
|
|
)
|
|
for i := uint32(0); i < 10; i++ {
|
|
var incoming bool
|
|
if i > 5 {
|
|
incoming = true
|
|
}
|
|
htlc := &HTLC{
|
|
Signature: testSig.Serialize(),
|
|
Incoming: incoming,
|
|
Amt: 10,
|
|
RHash: key,
|
|
RefundTimeout: i,
|
|
OutputIndex: int32(i * 3),
|
|
}
|
|
htlcs = append(htlcs, htlc)
|
|
htlcAmt += htlc.Amt
|
|
}
|
|
|
|
// TODO(roasbeef): ensure that expiry matches
|
|
|
|
// Create a new channel delta which includes the above HTLCs, some
|
|
// balance updates, and an increment of the current commitment height.
|
|
// Additionally, modify the signature and commitment transaction.
|
|
newSequence := uint32(129498)
|
|
newSig := bytes.Repeat([]byte{3}, 71)
|
|
newTx := channel.CommitTx.Copy()
|
|
newTx.TxIn[0].Sequence = newSequence
|
|
delta := &ChannelDelta{
|
|
LocalBalance: lnwire.MilliSatoshi(1e8),
|
|
RemoteBalance: lnwire.MilliSatoshi(1e8),
|
|
Htlcs: htlcs,
|
|
UpdateNum: 1,
|
|
}
|
|
|
|
// First update the local node's broadcastable state.
|
|
if err := channel.UpdateCommitment(newTx, newSig, delta); err != nil {
|
|
t.Fatalf("unable to update commitment: %v", err)
|
|
}
|
|
|
|
// The balances, new update, the HTLCs and the changes to the fake
|
|
// commitment transaction along with the modified signature should all
|
|
// have been updated.
|
|
updatedChannel, err := cdb.FetchOpenChannels(channel.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch updated channel: %v", err)
|
|
}
|
|
if !bytes.Equal(updatedChannel[0].CommitSig, newSig) {
|
|
t.Fatalf("sigs don't match %x vs %x",
|
|
updatedChannel[0].CommitSig, newSig)
|
|
}
|
|
if updatedChannel[0].CommitTx.TxIn[0].Sequence != newSequence {
|
|
t.Fatalf("sequence numbers don't match: %v vs %v",
|
|
updatedChannel[0].CommitTx.TxIn[0].Sequence, newSequence)
|
|
}
|
|
if updatedChannel[0].LocalBalance != delta.LocalBalance {
|
|
t.Fatalf("local balances don't match: %v vs %v",
|
|
updatedChannel[0].LocalBalance, delta.LocalBalance)
|
|
}
|
|
if updatedChannel[0].RemoteBalance != delta.RemoteBalance {
|
|
t.Fatalf("remote balances don't match: %v vs %v",
|
|
updatedChannel[0].RemoteBalance, delta.RemoteBalance)
|
|
}
|
|
if updatedChannel[0].NumUpdates != uint64(delta.UpdateNum) {
|
|
t.Fatalf("update # doesn't match: %v vs %v",
|
|
updatedChannel[0].NumUpdates, delta.UpdateNum)
|
|
}
|
|
numDiskUpdates, err := updatedChannel[0].CommitmentHeight()
|
|
if err != nil {
|
|
t.Fatalf("unable to read commitment height from disk: %v", err)
|
|
}
|
|
if numDiskUpdates != uint64(delta.UpdateNum) {
|
|
t.Fatalf("num disk updates doesn't match: %v vs %v",
|
|
numDiskUpdates, delta.UpdateNum)
|
|
}
|
|
for i := 0; i < len(updatedChannel[0].Htlcs); i++ {
|
|
originalHTLC := updatedChannel[0].Htlcs[i]
|
|
diskHTLC := channel.Htlcs[i]
|
|
if !reflect.DeepEqual(originalHTLC, diskHTLC) {
|
|
t.Fatalf("htlc's dont match: %v vs %v",
|
|
spew.Sdump(originalHTLC),
|
|
spew.Sdump(diskHTLC))
|
|
}
|
|
}
|
|
|
|
// Next, write to the log which tracks the necessary revocation state
|
|
// needed to rectify any fishy behavior by the remote party. Modify the
|
|
// current uncollapsed revocation state to simulate a state transition
|
|
// by the remote party.
|
|
channel.RemoteCurrentRevocation = channel.RemoteNextRevocation
|
|
newPriv, err := btcec.NewPrivateKey(btcec.S256())
|
|
if err != nil {
|
|
t.Fatalf("unable to generate key: %v", err)
|
|
}
|
|
channel.RemoteNextRevocation = newPriv.PubKey()
|
|
if err := channel.AppendToRevocationLog(delta); err != nil {
|
|
t.Fatalf("unable to append to revocation log: %v", err)
|
|
}
|
|
|
|
// We should be able to fetch the channel delta created above by it's
|
|
// update number with all the state properly reconstructed.
|
|
diskDelta, err := channel.FindPreviousState(uint64(delta.UpdateNum))
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch past delta: %v", err)
|
|
}
|
|
|
|
// The two deltas (the original vs the on-disk version) should
|
|
// identical, and all HTLC data should properly be retained.
|
|
if delta.LocalBalance != diskDelta.LocalBalance {
|
|
t.Fatal("local balances don't match")
|
|
}
|
|
if delta.RemoteBalance != diskDelta.RemoteBalance {
|
|
t.Fatal("remote balances don't match")
|
|
}
|
|
if delta.UpdateNum != diskDelta.UpdateNum {
|
|
t.Fatal("update number doesn't match")
|
|
}
|
|
for i := 0; i < len(delta.Htlcs); i++ {
|
|
originalHTLC := delta.Htlcs[i]
|
|
diskHTLC := diskDelta.Htlcs[i]
|
|
if !reflect.DeepEqual(originalHTLC, diskHTLC) {
|
|
t.Fatalf("htlc's dont match: %v vs %v",
|
|
spew.Sdump(originalHTLC),
|
|
spew.Sdump(diskHTLC))
|
|
}
|
|
}
|
|
|
|
// The state number recovered from the tail of the revocation log
|
|
// should be identical to this current state.
|
|
logTail, err := channel.RevocationLogTail()
|
|
if err != nil {
|
|
t.Fatalf("unable to retrieve log: %v", err)
|
|
}
|
|
if logTail.UpdateNum != delta.UpdateNum {
|
|
t.Fatal("update number doesn't match")
|
|
}
|
|
|
|
// Next modify the delta slightly, then create a new entry within the
|
|
// revocation log.
|
|
delta.UpdateNum = 2
|
|
delta.LocalBalance -= htlcAmt
|
|
delta.RemoteBalance += htlcAmt
|
|
delta.Htlcs = nil
|
|
if err := channel.AppendToRevocationLog(delta); err != nil {
|
|
t.Fatalf("unable to append to revocation log: %v", err)
|
|
}
|
|
|
|
// Once again, fetch the state and ensure it has been properly updated.
|
|
diskDelta, err = channel.FindPreviousState(uint64(delta.UpdateNum))
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch past delta: %v", err)
|
|
}
|
|
if len(diskDelta.Htlcs) != 0 {
|
|
t.Fatalf("expected %v htlcs, got %v", 0, len(diskDelta.Htlcs))
|
|
}
|
|
if delta.LocalBalance != 1e8-htlcAmt {
|
|
t.Fatalf("mismatched balances, expected %v got %v", 1e8-htlcAmt,
|
|
delta.LocalBalance)
|
|
}
|
|
if delta.RemoteBalance != 1e8+htlcAmt {
|
|
t.Fatalf("mismatched balances, expected %v got %v", 1e8+htlcAmt,
|
|
delta.RemoteBalance)
|
|
}
|
|
|
|
// Once again, state number recovered from the tail of the revocation
|
|
// log should be identical to this current state.
|
|
logTail, err = channel.RevocationLogTail()
|
|
if err != nil {
|
|
t.Fatalf("unable to retrieve log: %v", err)
|
|
}
|
|
if logTail.UpdateNum != delta.UpdateNum {
|
|
t.Fatal("update number doesn't match")
|
|
}
|
|
|
|
// The revocation state stored on-disk should now also be identical.
|
|
updatedChannel, err = cdb.FetchOpenChannels(channel.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch updated channel: %v", err)
|
|
}
|
|
if !channel.RemoteCurrentRevocation.IsEqual(updatedChannel[0].RemoteCurrentRevocation) {
|
|
t.Fatalf("revocation state was not synced")
|
|
}
|
|
if !channel.RemoteNextRevocation.IsEqual(updatedChannel[0].RemoteNextRevocation) {
|
|
t.Fatalf("revocation state was not synced")
|
|
}
|
|
|
|
// Now attempt to delete the channel from the database.
|
|
closeSummary := &ChannelCloseSummary{
|
|
ChanPoint: channel.FundingOutpoint,
|
|
RemotePub: channel.IdentityPub,
|
|
SettledBalance: btcutil.Amount(500),
|
|
TimeLockedBalance: btcutil.Amount(10000),
|
|
IsPending: false,
|
|
CloseType: ForceClose,
|
|
}
|
|
if err := updatedChannel[0].CloseChannel(closeSummary); err != nil {
|
|
t.Fatalf("unable to delete updated channel: %v", err)
|
|
}
|
|
|
|
// If we attempt to fetch the target channel again, it shouldn't be
|
|
// found.
|
|
channels, err := cdb.FetchOpenChannels(channel.IdentityPub)
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch updated channels: %v", err)
|
|
}
|
|
if len(channels) != 0 {
|
|
t.Fatalf("%v channels, found, but none should be",
|
|
len(channels))
|
|
}
|
|
|
|
// Attempting to find previous states on the channel should fail as the
|
|
// revocation log has been deleted.
|
|
_, err = updatedChannel[0].FindPreviousState(uint64(delta.UpdateNum))
|
|
if err == nil {
|
|
t.Fatal("revocation log search should've failed")
|
|
}
|
|
}
|
|
|
|
func TestFetchPendingChannels(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("uanble to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// Create first test channel state
|
|
state, err := createTestChannelState(cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channel state: %v", err)
|
|
}
|
|
|
|
addr := &net.TCPAddr{
|
|
IP: net.ParseIP("127.0.0.1"),
|
|
Port: 18555,
|
|
}
|
|
|
|
const broadcastHeight = 99
|
|
if err := state.SyncPending(addr, broadcastHeight); err != nil {
|
|
t.Fatalf("unable to save and serialize channel state: %v", err)
|
|
}
|
|
|
|
pendingChannels, err := cdb.FetchPendingChannels()
|
|
if err != nil {
|
|
t.Fatalf("unable to list pending channels: %v", err)
|
|
}
|
|
|
|
if len(pendingChannels) != 1 {
|
|
t.Fatalf("incorrect number of pending channels: expecting %v,"+
|
|
"got %v", 1, len(pendingChannels))
|
|
}
|
|
|
|
// The broadcast height of the pending channel should've been set
|
|
// properly.
|
|
if pendingChannels[0].FundingBroadcastHeight != broadcastHeight {
|
|
t.Fatalf("broadcast height mismatch: expected %v, got %v",
|
|
pendingChannels[0].FundingBroadcastHeight,
|
|
broadcastHeight)
|
|
}
|
|
|
|
chanOpenLoc := lnwire.ShortChannelID{
|
|
BlockHeight: 5,
|
|
TxIndex: 10,
|
|
TxPosition: 15,
|
|
}
|
|
err = cdb.MarkChannelAsOpen(&pendingChannels[0].FundingOutpoint, chanOpenLoc)
|
|
if err != nil {
|
|
t.Fatalf("unable to mark channel as open: %v", err)
|
|
}
|
|
|
|
// Next, we'll re-fetch the channel to ensure that the open height was
|
|
// properly set.
|
|
openChans, err := cdb.FetchAllChannels()
|
|
if err != nil {
|
|
t.Fatalf("unable to fetch channels: %v", err)
|
|
}
|
|
if openChans[0].ShortChanID != chanOpenLoc {
|
|
t.Fatalf("channel opening heights don't match: expected %v, "+
|
|
"got %v", spew.Sdump(openChans[0].ShortChanID),
|
|
chanOpenLoc)
|
|
}
|
|
if openChans[0].FundingBroadcastHeight != broadcastHeight {
|
|
t.Fatalf("broadcast height mismatch: expected %v, got %v",
|
|
openChans[0].FundingBroadcastHeight,
|
|
broadcastHeight)
|
|
}
|
|
|
|
pendingChannels, err = cdb.FetchPendingChannels()
|
|
if err != nil {
|
|
t.Fatalf("unable to list pending channels: %v", err)
|
|
}
|
|
|
|
if len(pendingChannels) != 0 {
|
|
t.Fatalf("incorrect number of pending channels: expecting %v,"+
|
|
"got %v", 0, len(pendingChannels))
|
|
}
|
|
}
|
|
|
|
func TestFetchClosedChannels(t *testing.T) {
|
|
t.Parallel()
|
|
|
|
cdb, cleanUp, err := makeTestDB()
|
|
if err != nil {
|
|
t.Fatalf("unable to make test database: %v", err)
|
|
}
|
|
defer cleanUp()
|
|
|
|
// First create a test channel, that we'll be closing within this pull
|
|
// request.
|
|
state, err := createTestChannelState(cdb)
|
|
if err != nil {
|
|
t.Fatalf("unable to create channel state: %v", err)
|
|
}
|
|
|
|
// Next sync the channel to disk, marking it as being in a pending open
|
|
// state.
|
|
addr := &net.TCPAddr{
|
|
IP: net.ParseIP("127.0.0.1"),
|
|
Port: 18555,
|
|
}
|
|
const broadcastHeight = 99
|
|
if err := state.SyncPending(addr, broadcastHeight); err != nil {
|
|
t.Fatalf("unable to save and serialize channel state: %v", err)
|
|
}
|
|
|
|
// Next, simulate the confirmation of the channel by marking it as
|
|
// pending within the database.
|
|
chanOpenLoc := lnwire.ShortChannelID{
|
|
BlockHeight: 5,
|
|
TxIndex: 10,
|
|
TxPosition: 15,
|
|
}
|
|
err = cdb.MarkChannelAsOpen(&state.FundingOutpoint, chanOpenLoc)
|
|
if err != nil {
|
|
t.Fatalf("unable to mark channel as open: %v", err)
|
|
}
|
|
|
|
// Next, close the channel by including a close channel summary in the
|
|
// database.
|
|
summary := &ChannelCloseSummary{
|
|
ChanPoint: state.FundingOutpoint,
|
|
ClosingTXID: rev,
|
|
RemotePub: state.IdentityPub,
|
|
Capacity: state.Capacity,
|
|
SettledBalance: state.LocalBalance.ToSatoshis(),
|
|
TimeLockedBalance: state.LocalBalance.ToSatoshis() + 10000,
|
|
CloseType: ForceClose,
|
|
IsPending: true,
|
|
}
|
|
if err := state.CloseChannel(summary); err != nil {
|
|
t.Fatalf("unable to close channel: %v", err)
|
|
}
|
|
|
|
// Query the database to ensure that the channel has now been properly
|
|
// closed. We should get the same result whether querying for pending
|
|
// channels only, or not.
|
|
pendingClosed, err := cdb.FetchClosedChannels(true)
|
|
if err != nil {
|
|
t.Fatalf("failed fetcing closed channels: %v", err)
|
|
}
|
|
if len(pendingClosed) != 1 {
|
|
t.Fatalf("incorrect number of pending closed channels: expecting %v,"+
|
|
"got %v", 1, len(pendingClosed))
|
|
}
|
|
if !reflect.DeepEqual(summary, pendingClosed[0]) {
|
|
t.Fatalf("database summaries don't match: expected %v got %v",
|
|
spew.Sdump(summary), spew.Sdump(pendingClosed[0]))
|
|
}
|
|
closed, err := cdb.FetchClosedChannels(false)
|
|
if err != nil {
|
|
t.Fatalf("failed fetching all closed channels: %v", err)
|
|
}
|
|
if len(closed) != 1 {
|
|
t.Fatalf("incorrect number of closed channels: expecting %v, "+
|
|
"got %v", 1, len(closed))
|
|
}
|
|
if !reflect.DeepEqual(summary, closed[0]) {
|
|
t.Fatalf("database summaries don't match: expected %v got %v",
|
|
spew.Sdump(summary), spew.Sdump(closed[0]))
|
|
}
|
|
|
|
// Mark the channel as fully closed
|
|
err = cdb.MarkChanFullyClosed(&state.FundingOutpoint)
|
|
if err != nil {
|
|
t.Fatalf("failed fully closing channel: %v", err)
|
|
}
|
|
|
|
// The channel should no longer be considered pending, but should still
|
|
// be retrieved when fetching all the closed channels.
|
|
closed, err = cdb.FetchClosedChannels(false)
|
|
if err != nil {
|
|
t.Fatalf("failed fetcing closed channels: %v", err)
|
|
}
|
|
if len(closed) != 1 {
|
|
t.Fatalf("incorrect number of closed channels: expecting %v, "+
|
|
"got %v", 1, len(closed))
|
|
}
|
|
pendingClose, err := cdb.FetchClosedChannels(true)
|
|
if err != nil {
|
|
t.Fatalf("failed fetching channels pending close: %v", err)
|
|
}
|
|
if len(pendingClose) != 0 {
|
|
t.Fatalf("incorrect number of closed channels: expecting %v, "+
|
|
"got %v", 0, len(closed))
|
|
}
|
|
}
|