package wtdb_test import ( "bytes" crand "crypto/rand" "io" "io/ioutil" "net" "os" "reflect" "testing" "github.com/btcsuite/btcd/btcec" "github.com/lightningnetwork/lnd/kvdb" "github.com/lightningnetwork/lnd/lnwire" "github.com/lightningnetwork/lnd/watchtower/blob" "github.com/lightningnetwork/lnd/watchtower/wtclient" "github.com/lightningnetwork/lnd/watchtower/wtdb" "github.com/lightningnetwork/lnd/watchtower/wtmock" "github.com/lightningnetwork/lnd/watchtower/wtpolicy" ) // clientDBInit is a closure used to initialize a wtclient.DB instance its // cleanup function. type clientDBInit func(t *testing.T) (wtclient.DB, func()) type clientDBHarness struct { t *testing.T db wtclient.DB } func newClientDBHarness(t *testing.T, init clientDBInit) (*clientDBHarness, func()) { db, cleanup := init(t) h := &clientDBHarness{ t: t, db: db, } return h, cleanup } func (h *clientDBHarness) insertSession(session *wtdb.ClientSession, expErr error) { h.t.Helper() err := h.db.CreateClientSession(session) if err != expErr { h.t.Fatalf("expected create client session error: %v, got: %v", expErr, err) } } func (h *clientDBHarness) listSessions(id *wtdb.TowerID) map[wtdb.SessionID]*wtdb.ClientSession { h.t.Helper() sessions, err := h.db.ListClientSessions(id) if err != nil { h.t.Fatalf("unable to list client sessions: %v", err) } return sessions } func (h *clientDBHarness) nextKeyIndex(id wtdb.TowerID, blobType blob.Type) uint32 { h.t.Helper() index, err := h.db.NextSessionKeyIndex(id, blobType) if err != nil { h.t.Fatalf("unable to create next session key index: %v", err) } if index == 0 { h.t.Fatalf("next key index should never be 0") } return index } func (h *clientDBHarness) createTower(lnAddr *lnwire.NetAddress, expErr error) *wtdb.Tower { h.t.Helper() tower, err := h.db.CreateTower(lnAddr) if err != expErr { h.t.Fatalf("expected create tower error: %v, got: %v", expErr, err) } if tower.ID == 0 { h.t.Fatalf("tower id should never be 0") } for _, session := range h.listSessions(&tower.ID) { if session.Status != wtdb.CSessionActive { h.t.Fatalf("expected status for session %v to be %v, "+ "got %v", session.ID, wtdb.CSessionActive, session.Status) } } return tower } func (h *clientDBHarness) removeTower(pubKey *btcec.PublicKey, addr net.Addr, hasSessions bool, expErr error) { h.t.Helper() if err := h.db.RemoveTower(pubKey, addr); err != expErr { h.t.Fatalf("expected remove tower error: %v, got %v", expErr, err) } if expErr != nil { return } if addr != nil { tower, err := h.db.LoadTower(pubKey) if err != nil { h.t.Fatalf("expected tower %x to still exist", pubKey.SerializeCompressed()) } removedAddr := addr.String() for _, towerAddr := range tower.Addresses { if towerAddr.String() == removedAddr { h.t.Fatalf("address %v not removed for tower %x", removedAddr, pubKey.SerializeCompressed()) } } } else { tower, err := h.db.LoadTower(pubKey) if hasSessions && err != nil { h.t.Fatalf("expected tower %x with sessions to still "+ "exist", pubKey.SerializeCompressed()) } if !hasSessions && err == nil { h.t.Fatalf("expected tower %x with no sessions to not "+ "exist", pubKey.SerializeCompressed()) } if !hasSessions { return } for _, session := range h.listSessions(&tower.ID) { if session.Status != wtdb.CSessionInactive { h.t.Fatalf("expected status for session %v to "+ "be %v, got %v", session.ID, wtdb.CSessionInactive, session.Status) } } } } func (h *clientDBHarness) loadTower(pubKey *btcec.PublicKey, expErr error) *wtdb.Tower { h.t.Helper() tower, err := h.db.LoadTower(pubKey) if err != expErr { h.t.Fatalf("expected load tower error: %v, got: %v", expErr, err) } return tower } func (h *clientDBHarness) loadTowerByID(id wtdb.TowerID, expErr error) *wtdb.Tower { h.t.Helper() tower, err := h.db.LoadTowerByID(id) if err != expErr { h.t.Fatalf("expected load tower error: %v, got: %v", expErr, err) } return tower } func (h *clientDBHarness) fetchChanSummaries() map[lnwire.ChannelID]wtdb.ClientChanSummary { h.t.Helper() summaries, err := h.db.FetchChanSummaries() if err != nil { h.t.Fatalf("unable to fetch chan summaries: %v", err) } return summaries } func (h *clientDBHarness) registerChan(chanID lnwire.ChannelID, sweepPkScript []byte, expErr error) { h.t.Helper() err := h.db.RegisterChannel(chanID, sweepPkScript) if err != expErr { h.t.Fatalf("expected register channel error: %v, got: %v", expErr, err) } } func (h *clientDBHarness) commitUpdate(id *wtdb.SessionID, update *wtdb.CommittedUpdate, expErr error) uint16 { h.t.Helper() lastApplied, err := h.db.CommitUpdate(id, update) if err != expErr { h.t.Fatalf("expected commit update error: %v, got: %v", expErr, err) } return lastApplied } func (h *clientDBHarness) ackUpdate(id *wtdb.SessionID, seqNum uint16, lastApplied uint16, expErr error) { h.t.Helper() err := h.db.AckUpdate(id, seqNum, lastApplied) if err != expErr { h.t.Fatalf("expected commit update error: %v, got: %v", expErr, err) } } // testCreateClientSession asserts various conditions regarding the creation of // a new ClientSession. The test asserts: // - client sessions can only be created if a session key index is reserved. // - client sessions cannot be created with an incorrect session key index . // - inserting duplicate sessions fails. func testCreateClientSession(h *clientDBHarness) { const blobType = blob.TypeAltruistAnchorCommit // Create a test client session to insert. session := &wtdb.ClientSession{ ClientSessionBody: wtdb.ClientSessionBody{ TowerID: wtdb.TowerID(3), Policy: wtpolicy.Policy{ TxPolicy: wtpolicy.TxPolicy{ BlobType: blobType, }, MaxUpdates: 100, }, RewardPkScript: []byte{0x01, 0x02, 0x03}, }, ID: wtdb.SessionID([33]byte{0x01}), } // First, assert that this session is not already present in the // database. if _, ok := h.listSessions(nil)[session.ID]; ok { h.t.Fatalf("session for id %x should not exist yet", session.ID) } // Attempting to insert the client session without reserving a session // key index should fail. h.insertSession(session, wtdb.ErrNoReservedKeyIndex) // Now, reserve a session key for this tower. keyIndex := h.nextKeyIndex(session.TowerID, blobType) // The client session hasn't been updated with the reserved key index // (since it's still zero). Inserting should fail due to the mismatch. h.insertSession(session, wtdb.ErrIncorrectKeyIndex) // Reserve another key for the same index. Since no session has been // successfully created, it should return the same index to maintain // idempotency across restarts. keyIndex2 := h.nextKeyIndex(session.TowerID, blobType) if keyIndex != keyIndex2 { h.t.Fatalf("next key index should be idempotent: want: %v, "+ "got %v", keyIndex, keyIndex2) } // Now, set the client session's key index so that it is proper and // insert it. This should succeed. session.KeyIndex = keyIndex h.insertSession(session, nil) // Verify that the session now exists in the database. if _, ok := h.listSessions(nil)[session.ID]; !ok { h.t.Fatalf("session for id %x should exist now", session.ID) } // Attempt to insert the session again, which should fail due to the // session already existing. h.insertSession(session, wtdb.ErrClientSessionAlreadyExists) // Finally, assert that reserving another key index succeeds with a // different key index, now that the first one has been finalized. keyIndex3 := h.nextKeyIndex(session.TowerID, blobType) if keyIndex == keyIndex3 { h.t.Fatalf("key index still reserved after creating session") } } // testFilterClientSessions asserts that we can correctly filter client sessions // for a specific tower. func testFilterClientSessions(h *clientDBHarness) { // We'll create three client sessions, the first two belonging to one // tower, and the last belonging to another one. const numSessions = 3 const blobType = blob.TypeAltruistCommit towerSessions := make(map[wtdb.TowerID][]wtdb.SessionID) for i := 0; i < numSessions; i++ { towerID := wtdb.TowerID(1) if i == numSessions-1 { towerID = wtdb.TowerID(2) } keyIndex := h.nextKeyIndex(towerID, blobType) sessionID := wtdb.SessionID([33]byte{byte(i)}) h.insertSession(&wtdb.ClientSession{ ClientSessionBody: wtdb.ClientSessionBody{ TowerID: towerID, Policy: wtpolicy.Policy{ TxPolicy: wtpolicy.TxPolicy{ BlobType: blobType, }, MaxUpdates: 100, }, RewardPkScript: []byte{0x01, 0x02, 0x03}, KeyIndex: keyIndex, }, ID: sessionID, }, nil) towerSessions[towerID] = append(towerSessions[towerID], sessionID) } // We should see the expected sessions for each tower when filtering // them. for towerID, expectedSessions := range towerSessions { sessions := h.listSessions(&towerID) if len(sessions) != len(expectedSessions) { h.t.Fatalf("expected %v sessions for tower %v, got %v", len(expectedSessions), towerID, len(sessions)) } for _, expectedSession := range expectedSessions { if _, ok := sessions[expectedSession]; !ok { h.t.Fatalf("expected session %v for tower %v", expectedSession, towerID) } } } } // testCreateTower asserts the behavior of creating new Tower objects within the // database, and that the latest address is always prepended to the list of // known addresses for the tower. func testCreateTower(h *clientDBHarness) { // Test that loading a tower with an arbitrary tower id fails. h.loadTowerByID(20, wtdb.ErrTowerNotFound) pk, err := randPubKey() if err != nil { h.t.Fatalf("unable to generate pubkey: %v", err) } addr1 := &net.TCPAddr{IP: []byte{0x01, 0x00, 0x00, 0x00}, Port: 9911} lnAddr := &lnwire.NetAddress{ IdentityKey: pk, Address: addr1, } // Insert a random tower into the database. tower := h.createTower(lnAddr, nil) // Load the tower from the database and assert that it matches the tower // we created. tower2 := h.loadTowerByID(tower.ID, nil) if !reflect.DeepEqual(tower, tower2) { h.t.Fatalf("loaded tower mismatch, want: %v, got: %v", tower, tower2) } tower2 = h.loadTower(pk, err) if !reflect.DeepEqual(tower, tower2) { h.t.Fatalf("loaded tower mismatch, want: %v, got: %v", tower, tower2) } // Insert the address again into the database. Since the address is the // same, this should result in an unmodified tower record. towerDupAddr := h.createTower(lnAddr, nil) if len(towerDupAddr.Addresses) != 1 { h.t.Fatalf("duplicate address should be deduped") } if !reflect.DeepEqual(tower, towerDupAddr) { h.t.Fatalf("mismatch towers, want: %v, got: %v", tower, towerDupAddr) } // Generate a new address for this tower. addr2 := &net.TCPAddr{IP: []byte{0x02, 0x00, 0x00, 0x00}, Port: 9911} lnAddr2 := &lnwire.NetAddress{ IdentityKey: pk, Address: addr2, } // Insert the updated address, which should produce a tower with a new // address. towerNewAddr := h.createTower(lnAddr2, nil) // Load the tower from the database, and assert that it matches the // tower returned from creation. towerNewAddr2 := h.loadTowerByID(tower.ID, nil) if !reflect.DeepEqual(towerNewAddr, towerNewAddr2) { h.t.Fatalf("loaded tower mismatch, want: %v, got: %v", towerNewAddr, towerNewAddr2) } towerNewAddr2 = h.loadTower(pk, nil) if !reflect.DeepEqual(towerNewAddr, towerNewAddr2) { h.t.Fatalf("loaded tower mismatch, want: %v, got: %v", towerNewAddr, towerNewAddr2) } // Assert that there are now two addresses on the tower object. if len(towerNewAddr.Addresses) != 2 { h.t.Fatalf("new address should be added") } // Finally, assert that the new address was prepended since it is deemed // fresher. if !reflect.DeepEqual(tower.Addresses, towerNewAddr.Addresses[1:]) { h.t.Fatalf("new address should be prepended") } } // testRemoveTower asserts the behavior of removing Tower objects as a whole and // removing addresses from Tower objects within the database. func testRemoveTower(h *clientDBHarness) { // Generate a random public key we'll use for our tower. pk, err := randPubKey() if err != nil { h.t.Fatalf("unable to generate pubkey: %v", err) } // Removing a tower that does not exist within the database should // result in a NOP. h.removeTower(pk, nil, false, nil) // We'll create a tower with two addresses. addr1 := &net.TCPAddr{IP: []byte{0x01, 0x00, 0x00, 0x00}, Port: 9911} addr2 := &net.TCPAddr{IP: []byte{0x02, 0x00, 0x00, 0x00}, Port: 9911} h.createTower(&lnwire.NetAddress{ IdentityKey: pk, Address: addr1, }, nil) h.createTower(&lnwire.NetAddress{ IdentityKey: pk, Address: addr2, }, nil) // We'll then remove the second address. We should now only see the // first. h.removeTower(pk, addr2, false, nil) // We'll then remove the first address. We should now see that the tower // has no addresses left. h.removeTower(pk, addr1, false, wtdb.ErrLastTowerAddr) // Removing the tower as a whole from the database should succeed since // there aren't any active sessions for it. h.removeTower(pk, nil, false, nil) // We'll then recreate the tower, but this time we'll create a session // for it. tower := h.createTower(&lnwire.NetAddress{ IdentityKey: pk, Address: addr1, }, nil) const blobType = blob.TypeAltruistCommit session := &wtdb.ClientSession{ ClientSessionBody: wtdb.ClientSessionBody{ TowerID: tower.ID, Policy: wtpolicy.Policy{ TxPolicy: wtpolicy.TxPolicy{ BlobType: blobType, }, MaxUpdates: 100, }, RewardPkScript: []byte{0x01, 0x02, 0x03}, KeyIndex: h.nextKeyIndex(tower.ID, blobType), }, ID: wtdb.SessionID([33]byte{0x01}), } h.insertSession(session, nil) update := randCommittedUpdate(h.t, 1) h.commitUpdate(&session.ID, update, nil) // We should not be able to fully remove it from the database since // there's a session and it has unacked updates. h.removeTower(pk, nil, true, wtdb.ErrTowerUnackedUpdates) // Removing the tower after all sessions no longer have unacked updates // should result in the sessions becoming inactive. h.ackUpdate(&session.ID, 1, 1, nil) h.removeTower(pk, nil, true, nil) // Creating the tower again should mark all of the sessions active once // again. h.createTower(&lnwire.NetAddress{ IdentityKey: pk, Address: addr1, }, nil) } // testChanSummaries tests the process of a registering a channel and its // associated sweep pkscript. func testChanSummaries(h *clientDBHarness) { // First, assert that this channel is not already registered. var chanID lnwire.ChannelID if _, ok := h.fetchChanSummaries()[chanID]; ok { h.t.Fatalf("pkscript for channel %x should not exist yet", chanID) } // Generate a random sweep pkscript and register it for this channel. expPkScript := make([]byte, 22) if _, err := io.ReadFull(crand.Reader, expPkScript); err != nil { h.t.Fatalf("unable to generate pkscript: %v", err) } h.registerChan(chanID, expPkScript, nil) // Assert that the channel exists and that its sweep pkscript matches // the one we registered. summary, ok := h.fetchChanSummaries()[chanID] if !ok { h.t.Fatalf("pkscript for channel %x should not exist yet", chanID) } else if bytes.Compare(expPkScript, summary.SweepPkScript) != 0 { h.t.Fatalf("pkscript mismatch, want: %x, got: %x", expPkScript, summary.SweepPkScript) } // Finally, assert that re-registering the same channel produces a // failure. h.registerChan(chanID, expPkScript, wtdb.ErrChannelAlreadyRegistered) } // testCommitUpdate tests the behavior of CommitUpdate, ensuring that they can func testCommitUpdate(h *clientDBHarness) { const blobType = blob.TypeAltruistCommit session := &wtdb.ClientSession{ ClientSessionBody: wtdb.ClientSessionBody{ TowerID: wtdb.TowerID(3), Policy: wtpolicy.Policy{ TxPolicy: wtpolicy.TxPolicy{ BlobType: blobType, }, MaxUpdates: 100, }, RewardPkScript: []byte{0x01, 0x02, 0x03}, }, ID: wtdb.SessionID([33]byte{0x02}), } // Generate a random update and try to commit before inserting the // session, which should fail. update1 := randCommittedUpdate(h.t, 1) h.commitUpdate(&session.ID, update1, wtdb.ErrClientSessionNotFound) // Reserve a session key index and insert the session. session.KeyIndex = h.nextKeyIndex(session.TowerID, blobType) h.insertSession(session, nil) // Now, try to commit the update that failed initially which should // succeed. The lastApplied value should be 0 since we have not received // an ack from the tower. lastApplied := h.commitUpdate(&session.ID, update1, nil) if lastApplied != 0 { h.t.Fatalf("last applied mismatch, want: 0, got: %v", lastApplied) } // Assert that the committed update appears in the client session's // CommittedUpdates map when loaded from disk and that there are no // AckedUpdates. dbSession := h.listSessions(nil)[session.ID] checkCommittedUpdates(h.t, dbSession, []wtdb.CommittedUpdate{ *update1, }) checkAckedUpdates(h.t, dbSession, nil) // Try to commit the same update, which should succeed due to // idempotency (which is preserved when the breach hint is identical to // the on-disk update's hint). The lastApplied value should remain // unchanged. lastApplied2 := h.commitUpdate(&session.ID, update1, nil) if lastApplied2 != lastApplied { h.t.Fatalf("last applied should not have changed, got %v", lastApplied2) } // Assert that the loaded ClientSession is the same as before. dbSession = h.listSessions(nil)[session.ID] checkCommittedUpdates(h.t, dbSession, []wtdb.CommittedUpdate{ *update1, }) checkAckedUpdates(h.t, dbSession, nil) // Generate another random update and try to commit it at the identical // sequence number. Since the breach hint has changed, this should fail. update2 := randCommittedUpdate(h.t, 1) h.commitUpdate(&session.ID, update2, wtdb.ErrUpdateAlreadyCommitted) // Next, insert the new update at the next unallocated sequence number // which should succeed. update2.SeqNum = 2 lastApplied3 := h.commitUpdate(&session.ID, update2, nil) if lastApplied3 != lastApplied { h.t.Fatalf("last applied should not have changed, got %v", lastApplied3) } // Check that both updates now appear as committed on the ClientSession // loaded from disk. dbSession = h.listSessions(nil)[session.ID] checkCommittedUpdates(h.t, dbSession, []wtdb.CommittedUpdate{ *update1, *update2, }) checkAckedUpdates(h.t, dbSession, nil) // Finally, create one more random update and try to commit it at index // 4, which should be rejected since 3 is the next slot the database // expects. update4 := randCommittedUpdate(h.t, 4) h.commitUpdate(&session.ID, update4, wtdb.ErrCommitUnorderedUpdate) // Assert that the ClientSession loaded from disk remains unchanged. dbSession = h.listSessions(nil)[session.ID] checkCommittedUpdates(h.t, dbSession, []wtdb.CommittedUpdate{ *update1, *update2, }) checkAckedUpdates(h.t, dbSession, nil) } // testAckUpdate asserts the behavior of AckUpdate. func testAckUpdate(h *clientDBHarness) { const blobType = blob.TypeAltruistCommit // Create a new session that the updates in this will be tied to. session := &wtdb.ClientSession{ ClientSessionBody: wtdb.ClientSessionBody{ TowerID: wtdb.TowerID(3), Policy: wtpolicy.Policy{ TxPolicy: wtpolicy.TxPolicy{ BlobType: blobType, }, MaxUpdates: 100, }, RewardPkScript: []byte{0x01, 0x02, 0x03}, }, ID: wtdb.SessionID([33]byte{0x03}), } // Try to ack an update before inserting the client session, which // should fail. h.ackUpdate(&session.ID, 1, 0, wtdb.ErrClientSessionNotFound) // Reserve a session key and insert the client session. session.KeyIndex = h.nextKeyIndex(session.TowerID, blobType) h.insertSession(session, nil) // Now, try to ack update 1. This should fail since update 1 was never // committed. h.ackUpdate(&session.ID, 1, 0, wtdb.ErrCommittedUpdateNotFound) // Commit to a random update at seqnum 1. update1 := randCommittedUpdate(h.t, 1) lastApplied := h.commitUpdate(&session.ID, update1, nil) if lastApplied != 0 { h.t.Fatalf("last applied mismatch, want: 0, got: %v", lastApplied) } // Acking seqnum 1 should succeed. h.ackUpdate(&session.ID, 1, 1, nil) // Acking seqnum 1 again should fail. h.ackUpdate(&session.ID, 1, 1, wtdb.ErrCommittedUpdateNotFound) // Acking a valid seqnum with a reverted last applied value should fail. h.ackUpdate(&session.ID, 1, 0, wtdb.ErrLastAppliedReversion) // Acking with a last applied greater than any allocated seqnum should // fail. h.ackUpdate(&session.ID, 4, 3, wtdb.ErrUnallocatedLastApplied) // Assert that the ClientSession loaded from disk has one update in it's // AckedUpdates map, and that the committed update has been removed. dbSession := h.listSessions(nil)[session.ID] checkCommittedUpdates(h.t, dbSession, nil) checkAckedUpdates(h.t, dbSession, map[uint16]wtdb.BackupID{ 1: update1.BackupID, }) // Commit to another random update, and assert that the last applied // value is 1, since this was what was provided in the last successful // ack. update2 := randCommittedUpdate(h.t, 2) lastApplied = h.commitUpdate(&session.ID, update2, nil) if lastApplied != 1 { h.t.Fatalf("last applied mismatch, want: 1, got: %v", lastApplied) } // Ack seqnum 2. h.ackUpdate(&session.ID, 2, 2, nil) // Assert that both updates exist as AckedUpdates when loaded from disk. dbSession = h.listSessions(nil)[session.ID] checkCommittedUpdates(h.t, dbSession, nil) checkAckedUpdates(h.t, dbSession, map[uint16]wtdb.BackupID{ 1: update1.BackupID, 2: update2.BackupID, }) // Acking again with a lower last applied should fail. h.ackUpdate(&session.ID, 2, 1, wtdb.ErrLastAppliedReversion) // Acking an unallocated seqnum should fail. h.ackUpdate(&session.ID, 4, 2, wtdb.ErrCommittedUpdateNotFound) // Acking with a last applied greater than any allocated seqnum should // fail. h.ackUpdate(&session.ID, 4, 3, wtdb.ErrUnallocatedLastApplied) } // checkCommittedUpdates asserts that the CommittedUpdates on session match the // expUpdates provided. func checkCommittedUpdates(t *testing.T, session *wtdb.ClientSession, expUpdates []wtdb.CommittedUpdate) { t.Helper() // We promote nil expUpdates to an initialized slice since the database // should never return a nil slice. This promotion is done purely out of // convenience for the testing framework. if expUpdates == nil { expUpdates = make([]wtdb.CommittedUpdate, 0) } if !reflect.DeepEqual(session.CommittedUpdates, expUpdates) { t.Fatalf("committed updates mismatch, want: %v, got: %v", expUpdates, session.CommittedUpdates) } } // checkAckedUpdates asserts that the AckedUpdates on a sessio match the // expUpdates provided. func checkAckedUpdates(t *testing.T, session *wtdb.ClientSession, expUpdates map[uint16]wtdb.BackupID) { // We promote nil expUpdates to an initialized map since the database // should never return a nil map. This promotion is done purely out of // convenience for the testing framework. if expUpdates == nil { expUpdates = make(map[uint16]wtdb.BackupID) } if !reflect.DeepEqual(session.AckedUpdates, expUpdates) { t.Fatalf("acked updates mismatch, want: %v, got: %v", expUpdates, session.AckedUpdates) } } // TestClientDB asserts the behavior of a fresh client db, a reopened client db, // and the mock implementation. This ensures that all databases function // identically, especially in the negative paths. func TestClientDB(t *testing.T) { dbs := []struct { name string init clientDBInit }{ { name: "fresh clientdb", init: func(t *testing.T) (wtclient.DB, func()) { path, err := ioutil.TempDir("", "clientdb") if err != nil { t.Fatalf("unable to make temp dir: %v", err) } db, err := wtdb.OpenClientDB( path, kvdb.DefaultDBTimeout, ) if err != nil { os.RemoveAll(path) t.Fatalf("unable to open db: %v", err) } cleanup := func() { db.Close() os.RemoveAll(path) } return db, cleanup }, }, { name: "reopened clientdb", init: func(t *testing.T) (wtclient.DB, func()) { path, err := ioutil.TempDir("", "clientdb") if err != nil { t.Fatalf("unable to make temp dir: %v", err) } db, err := wtdb.OpenClientDB( path, kvdb.DefaultDBTimeout, ) if err != nil { os.RemoveAll(path) t.Fatalf("unable to open db: %v", err) } db.Close() db, err = wtdb.OpenClientDB( path, kvdb.DefaultDBTimeout, ) if err != nil { os.RemoveAll(path) t.Fatalf("unable to reopen db: %v", err) } cleanup := func() { db.Close() os.RemoveAll(path) } return db, cleanup }, }, { name: "mock", init: func(t *testing.T) (wtclient.DB, func()) { return wtmock.NewClientDB(), func() {} }, }, } tests := []struct { name string run func(*clientDBHarness) }{ { name: "create client session", run: testCreateClientSession, }, { name: "filter client sessions", run: testFilterClientSessions, }, { name: "create tower", run: testCreateTower, }, { name: "remove tower", run: testRemoveTower, }, { name: "chan summaries", run: testChanSummaries, }, { name: "commit update", run: testCommitUpdate, }, { name: "ack update", run: testAckUpdate, }, } for _, database := range dbs { db := database t.Run(db.name, func(t *testing.T) { t.Parallel() for _, test := range tests { t.Run(test.name, func(t *testing.T) { h, cleanup := newClientDBHarness( t, db.init, ) defer cleanup() test.run(h) }) } }) } } // randCommittedUpdate generates a random committed update. func randCommittedUpdate(t *testing.T, seqNum uint16) *wtdb.CommittedUpdate { var chanID lnwire.ChannelID if _, err := io.ReadFull(crand.Reader, chanID[:]); err != nil { t.Fatalf("unable to generate chan id: %v", err) } var hint blob.BreachHint if _, err := io.ReadFull(crand.Reader, hint[:]); err != nil { t.Fatalf("unable to generate breach hint: %v", err) } encBlob := make([]byte, blob.Size(blob.FlagCommitOutputs.Type())) if _, err := io.ReadFull(crand.Reader, encBlob); err != nil { t.Fatalf("unable to generate encrypted blob: %v", err) } return &wtdb.CommittedUpdate{ SeqNum: seqNum, CommittedUpdateBody: wtdb.CommittedUpdateBody{ BackupID: wtdb.BackupID{ ChanID: chanID, CommitHeight: 666, }, Hint: hint, EncryptedBlob: encBlob, }, } }