2016-09-17 03:23:37 +03:00
|
|
|
package channeldb
|
|
|
|
|
|
|
|
import (
|
2016-09-19 21:51:37 +03:00
|
|
|
"crypto/rand"
|
2017-03-16 04:56:25 +03:00
|
|
|
"crypto/sha256"
|
2016-09-17 03:23:37 +03:00
|
|
|
"reflect"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/davecgh/go-spew/spew"
|
2017-08-22 08:51:19 +03:00
|
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
2016-09-17 03:23:37 +03:00
|
|
|
)
|
|
|
|
|
2017-08-22 08:51:19 +03:00
|
|
|
func randInvoice(value lnwire.MilliSatoshi) (*Invoice, error) {
|
2016-09-19 21:51:37 +03:00
|
|
|
var pre [32]byte
|
|
|
|
if _, err := rand.Read(pre[:]); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
i := &Invoice{
|
2017-08-29 06:04:06 +03:00
|
|
|
// Use single second precision to avoid false positive test
|
|
|
|
// failures due to the monotonic time component.
|
|
|
|
CreationDate: time.Unix(time.Now().Unix(), 0),
|
2016-09-19 21:51:37 +03:00
|
|
|
Terms: ContractTerm{
|
|
|
|
PaymentPreimage: pre,
|
|
|
|
Value: value,
|
|
|
|
},
|
|
|
|
}
|
2016-09-24 01:15:22 +03:00
|
|
|
i.Memo = []byte("memo")
|
2018-02-07 06:11:11 +03:00
|
|
|
i.Receipt = []byte("receipt")
|
2016-09-19 21:51:37 +03:00
|
|
|
|
2017-09-05 18:59:52 +03:00
|
|
|
// Create a random byte slice of MaxPaymentRequestSize bytes to be used
|
|
|
|
// as a dummy paymentrequest, and determine if it should be set based
|
|
|
|
// on one of the random bytes.
|
|
|
|
var r [MaxPaymentRequestSize]byte
|
|
|
|
if _, err := rand.Read(r[:]); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if r[0]&1 == 0 {
|
|
|
|
i.PaymentRequest = r[:]
|
|
|
|
} else {
|
|
|
|
i.PaymentRequest = []byte("")
|
|
|
|
}
|
|
|
|
|
2016-09-19 21:51:37 +03:00
|
|
|
return i, nil
|
|
|
|
}
|
|
|
|
|
2016-09-17 03:23:37 +03:00
|
|
|
func TestInvoiceWorkflow(t *testing.T) {
|
2017-06-17 01:59:20 +03:00
|
|
|
t.Parallel()
|
|
|
|
|
2016-09-17 03:23:37 +03:00
|
|
|
db, cleanUp, err := makeTestDB()
|
2016-12-22 23:04:41 +03:00
|
|
|
defer cleanUp()
|
2016-09-17 03:23:37 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a fake invoice which we'll use several times in the tests
|
|
|
|
// below.
|
|
|
|
fakeInvoice := &Invoice{
|
2017-08-29 06:04:06 +03:00
|
|
|
// Use single second precision to avoid false positive test
|
|
|
|
// failures due to the monotonic time component.
|
|
|
|
CreationDate: time.Unix(time.Now().Unix(), 0),
|
2016-09-17 03:23:37 +03:00
|
|
|
}
|
2016-09-24 01:15:22 +03:00
|
|
|
fakeInvoice.Memo = []byte("memo")
|
2018-02-07 06:11:11 +03:00
|
|
|
fakeInvoice.Receipt = []byte("receipt")
|
2017-09-05 18:59:52 +03:00
|
|
|
fakeInvoice.PaymentRequest = []byte("")
|
2016-09-17 03:23:37 +03:00
|
|
|
copy(fakeInvoice.Terms.PaymentPreimage[:], rev[:])
|
2017-08-22 08:51:19 +03:00
|
|
|
fakeInvoice.Terms.Value = lnwire.NewMSatFromSatoshis(10000)
|
2016-09-17 03:23:37 +03:00
|
|
|
|
2018-02-07 06:11:11 +03:00
|
|
|
// Add the invoice to the database, this should succeed as there aren't
|
2016-09-17 03:23:37 +03:00
|
|
|
// any existing invoices within the database with the same payment
|
|
|
|
// hash.
|
2018-06-30 04:05:51 +03:00
|
|
|
if _, err := db.AddInvoice(fakeInvoice); err != nil {
|
2016-09-17 03:23:37 +03:00
|
|
|
t.Fatalf("unable to find invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Attempt to retrieve the invoice which was just added to the
|
|
|
|
// database. It should be found, and the invoice returned should be
|
|
|
|
// identical to the one created above.
|
2017-03-16 04:56:25 +03:00
|
|
|
paymentHash := sha256.Sum256(fakeInvoice.Terms.PaymentPreimage[:])
|
2016-09-17 03:23:37 +03:00
|
|
|
dbInvoice, err := db.LookupInvoice(paymentHash)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to find invoice: %v", err)
|
|
|
|
}
|
2018-04-25 07:01:22 +03:00
|
|
|
if !reflect.DeepEqual(*fakeInvoice, dbInvoice) {
|
2016-09-17 03:23:37 +03:00
|
|
|
t.Fatalf("invoice fetched from db doesn't match original %v vs %v",
|
|
|
|
spew.Sdump(fakeInvoice), spew.Sdump(dbInvoice))
|
|
|
|
}
|
|
|
|
|
2018-04-25 07:01:22 +03:00
|
|
|
// The add index of the invoice retrieved from the database should now
|
|
|
|
// be fully populated. As this is the first index written to the DB,
|
|
|
|
// the addIndex should be 1.
|
|
|
|
if dbInvoice.AddIndex != 1 {
|
|
|
|
t.Fatalf("wrong add index: expected %v, got %v", 1,
|
|
|
|
dbInvoice.AddIndex)
|
|
|
|
}
|
|
|
|
|
2017-12-05 09:07:21 +03:00
|
|
|
// Settle the invoice, the version retrieved from the database should
|
|
|
|
// now have the settled bit toggle to true and a non-default
|
|
|
|
// SettledDate
|
2018-04-25 07:01:22 +03:00
|
|
|
payAmt := fakeInvoice.Terms.Value * 2
|
2018-06-29 07:43:18 +03:00
|
|
|
if _, err := db.SettleInvoice(paymentHash, payAmt); err != nil {
|
2016-09-17 03:23:37 +03:00
|
|
|
t.Fatalf("unable to settle invoice: %v", err)
|
|
|
|
}
|
|
|
|
dbInvoice2, err := db.LookupInvoice(paymentHash)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch invoice: %v", err)
|
|
|
|
}
|
2018-12-19 17:56:26 +03:00
|
|
|
if dbInvoice2.Terms.State != ContractSettled {
|
2016-09-17 03:23:37 +03:00
|
|
|
t.Fatalf("invoice should now be settled but isn't")
|
|
|
|
}
|
2017-12-05 09:07:21 +03:00
|
|
|
if dbInvoice2.SettleDate.IsZero() {
|
|
|
|
t.Fatalf("invoice should have non-zero SettledDate but isn't")
|
|
|
|
}
|
|
|
|
|
2018-04-25 07:01:22 +03:00
|
|
|
// Our 2x payment should be reflected, and also the settle index of 1
|
|
|
|
// should also have been committed for this index.
|
|
|
|
if dbInvoice2.AmtPaid != payAmt {
|
|
|
|
t.Fatalf("wrong amt paid: expected %v, got %v", payAmt,
|
|
|
|
dbInvoice2.AmtPaid)
|
|
|
|
}
|
|
|
|
if dbInvoice2.SettleIndex != 1 {
|
|
|
|
t.Fatalf("wrong settle index: expected %v, got %v", 1,
|
|
|
|
dbInvoice2.SettleIndex)
|
|
|
|
}
|
|
|
|
|
2016-09-17 03:23:37 +03:00
|
|
|
// Attempt to insert generated above again, this should fail as
|
|
|
|
// duplicates are rejected by the processing logic.
|
2018-06-30 04:05:51 +03:00
|
|
|
if _, err := db.AddInvoice(fakeInvoice); err != ErrDuplicateInvoice {
|
2016-09-17 03:23:37 +03:00
|
|
|
t.Fatalf("invoice insertion should fail due to duplication, "+
|
|
|
|
"instead %v", err)
|
|
|
|
}
|
|
|
|
|
2017-09-25 21:25:58 +03:00
|
|
|
// Attempt to look up a non-existent invoice, this should also fail but
|
2016-09-17 03:23:37 +03:00
|
|
|
// with a "not found" error.
|
|
|
|
var fakeHash [32]byte
|
|
|
|
if _, err := db.LookupInvoice(fakeHash); err != ErrInvoiceNotFound {
|
|
|
|
t.Fatalf("lookup should have failed, instead %v", err)
|
|
|
|
}
|
2016-09-19 21:51:37 +03:00
|
|
|
|
2018-04-25 07:01:22 +03:00
|
|
|
// Add 10 random invoices.
|
2016-09-19 21:51:37 +03:00
|
|
|
const numInvoices = 10
|
2017-08-22 08:51:19 +03:00
|
|
|
amt := lnwire.NewMSatFromSatoshis(1000)
|
2016-09-19 21:51:37 +03:00
|
|
|
invoices := make([]*Invoice, numInvoices+1)
|
2018-04-25 07:01:22 +03:00
|
|
|
invoices[0] = &dbInvoice2
|
2016-09-19 21:51:37 +03:00
|
|
|
for i := 1; i < len(invoices)-1; i++ {
|
|
|
|
invoice, err := randInvoice(amt)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-06-30 04:05:51 +03:00
|
|
|
if _, err := db.AddInvoice(invoice); err != nil {
|
2016-09-19 21:51:37 +03:00
|
|
|
t.Fatalf("unable to add invoice %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
invoices[i] = invoice
|
|
|
|
}
|
|
|
|
|
|
|
|
// Perform a scan to collect all the active invoices.
|
|
|
|
dbInvoices, err := db.FetchAllInvoices(false)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch all invoices: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// The retrieve list of invoices should be identical as since we're
|
2018-02-07 06:11:11 +03:00
|
|
|
// using big endian, the invoices should be retrieved in ascending
|
2016-09-19 21:51:37 +03:00
|
|
|
// order (and the primary key should be incremented with each
|
|
|
|
// insertion).
|
|
|
|
for i := 0; i < len(invoices)-1; i++ {
|
2018-04-25 07:01:22 +03:00
|
|
|
if !reflect.DeepEqual(*invoices[i], dbInvoices[i]) {
|
2018-02-07 06:11:11 +03:00
|
|
|
t.Fatalf("retrieved invoices don't match %v vs %v",
|
2016-09-19 21:51:37 +03:00
|
|
|
spew.Sdump(invoices[i]),
|
|
|
|
spew.Sdump(dbInvoices[i]))
|
|
|
|
}
|
|
|
|
}
|
2016-09-17 03:23:37 +03:00
|
|
|
}
|
2018-04-25 07:01:43 +03:00
|
|
|
|
|
|
|
// TestInvoiceTimeSeries tests that newly added invoices invoices, as well as
|
|
|
|
// settled invoices are added to the database are properly placed in the add
|
|
|
|
// add or settle index which serves as an event time series.
|
|
|
|
func TestInvoiceAddTimeSeries(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
db, cleanUp, err := makeTestDB()
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll start off by creating 20 random invoices, and inserting them
|
|
|
|
// into the database.
|
|
|
|
const numInvoices = 20
|
|
|
|
amt := lnwire.NewMSatFromSatoshis(1000)
|
|
|
|
invoices := make([]Invoice, numInvoices)
|
|
|
|
for i := 0; i < len(invoices); i++ {
|
|
|
|
invoice, err := randInvoice(amt)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-06-30 04:05:51 +03:00
|
|
|
if _, err := db.AddInvoice(invoice); err != nil {
|
2018-04-25 07:01:43 +03:00
|
|
|
t.Fatalf("unable to add invoice %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
invoices[i] = *invoice
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the invoices constructed, we'll now create a series of queries
|
|
|
|
// that we'll use to assert expected return values of
|
|
|
|
// InvoicesAddedSince.
|
|
|
|
addQueries := []struct {
|
|
|
|
sinceAddIndex uint64
|
|
|
|
|
|
|
|
resp []Invoice
|
|
|
|
}{
|
|
|
|
// If we specify a value of zero, we shouldn't get any invoices
|
|
|
|
// back.
|
|
|
|
{
|
|
|
|
sinceAddIndex: 0,
|
|
|
|
},
|
|
|
|
|
|
|
|
// If we specify a value well beyond the number of inserted
|
|
|
|
// invoices, we shouldn't get any invoices back.
|
|
|
|
{
|
|
|
|
sinceAddIndex: 99999999,
|
|
|
|
},
|
|
|
|
|
|
|
|
// Using an index of 1 should result in all values, but the
|
|
|
|
// first one being returned.
|
|
|
|
{
|
|
|
|
sinceAddIndex: 1,
|
|
|
|
resp: invoices[1:],
|
|
|
|
},
|
|
|
|
|
|
|
|
// If we use an index of 10, then we should retrieve the
|
|
|
|
// reaming 10 invoices.
|
|
|
|
{
|
|
|
|
sinceAddIndex: 10,
|
|
|
|
resp: invoices[10:],
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, query := range addQueries {
|
|
|
|
resp, err := db.InvoicesAddedSince(query.sinceAddIndex)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if !reflect.DeepEqual(query.resp, resp) {
|
|
|
|
t.Fatalf("test #%v: expected %v, got %v", i,
|
|
|
|
spew.Sdump(query.resp), spew.Sdump(resp))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll now only settle the latter half of each of those invoices.
|
|
|
|
for i := 10; i < len(invoices); i++ {
|
|
|
|
invoice := &invoices[i]
|
|
|
|
|
|
|
|
paymentHash := sha256.Sum256(
|
|
|
|
invoice.Terms.PaymentPreimage[:],
|
|
|
|
)
|
|
|
|
|
2018-06-29 07:43:18 +03:00
|
|
|
_, err := db.SettleInvoice(paymentHash, 0)
|
2018-04-25 07:01:43 +03:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to settle invoice: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
invoices, err = db.FetchAllInvoices(false)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to fetch invoices: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll slice off the first 10 invoices, as we only settled the last
|
|
|
|
// 10.
|
|
|
|
invoices = invoices[10:]
|
|
|
|
|
|
|
|
// We'll now prepare an additional set of queries to ensure the settle
|
|
|
|
// time series has properly been maintained in the database.
|
|
|
|
settleQueries := []struct {
|
|
|
|
sinceSettleIndex uint64
|
|
|
|
|
|
|
|
resp []Invoice
|
|
|
|
}{
|
|
|
|
// If we specify a value of zero, we shouldn't get any settled
|
|
|
|
// invoices back.
|
|
|
|
{
|
|
|
|
sinceSettleIndex: 0,
|
|
|
|
},
|
|
|
|
|
|
|
|
// If we specify a value well beyond the number of settled
|
|
|
|
// invoices, we shouldn't get any invoices back.
|
|
|
|
{
|
|
|
|
sinceSettleIndex: 99999999,
|
|
|
|
},
|
|
|
|
|
|
|
|
// Using an index of 1 should result in the final 10 invoices
|
|
|
|
// being returned, as we only settled those.
|
|
|
|
{
|
|
|
|
sinceSettleIndex: 1,
|
|
|
|
resp: invoices[1:],
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, query := range settleQueries {
|
|
|
|
resp, err := db.InvoicesSettledSince(query.sinceSettleIndex)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if !reflect.DeepEqual(query.resp, resp) {
|
|
|
|
t.Fatalf("test #%v: expected %v, got %v", i,
|
|
|
|
spew.Sdump(query.resp), spew.Sdump(resp))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2018-07-18 02:27:04 +03:00
|
|
|
|
|
|
|
// TestDuplicateSettleInvoice tests that if we add a new invoice and settle it
|
|
|
|
// twice, then the second time we also receive the invoice that we settled as a
|
|
|
|
// return argument.
|
|
|
|
func TestDuplicateSettleInvoice(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
db, cleanUp, err := makeTestDB()
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll start out by creating an invoice and writing it to the DB.
|
|
|
|
amt := lnwire.NewMSatFromSatoshis(1000)
|
|
|
|
invoice, err := randInvoice(amt)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if _, err := db.AddInvoice(invoice); err != nil {
|
|
|
|
t.Fatalf("unable to add invoice %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the invoice in the DB, we'll now attempt to settle the invoice.
|
|
|
|
payHash := sha256.Sum256(invoice.Terms.PaymentPreimage[:])
|
|
|
|
dbInvoice, err := db.SettleInvoice(payHash, amt)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to settle invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll update what we expect the settle invoice to be so that our
|
|
|
|
// comparison below has the correct assumption.
|
|
|
|
invoice.SettleIndex = 1
|
2018-12-19 17:56:26 +03:00
|
|
|
invoice.Terms.State = ContractSettled
|
2018-07-18 02:27:04 +03:00
|
|
|
invoice.AmtPaid = amt
|
|
|
|
invoice.SettleDate = dbInvoice.SettleDate
|
|
|
|
|
|
|
|
// We should get back the exact same invoice that we just inserted.
|
|
|
|
if !reflect.DeepEqual(dbInvoice, invoice) {
|
|
|
|
t.Fatalf("wrong invoice after settle, expected %v got %v",
|
|
|
|
spew.Sdump(invoice), spew.Sdump(dbInvoice))
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we try to settle the invoice again, then we should get the very
|
2019-01-14 14:03:26 +03:00
|
|
|
// same invoice back, but with an error this time.
|
2018-07-18 02:27:04 +03:00
|
|
|
dbInvoice, err = db.SettleInvoice(payHash, amt)
|
2019-01-14 14:03:26 +03:00
|
|
|
if err != ErrInvoiceAlreadySettled {
|
|
|
|
t.Fatalf("expected ErrInvoiceAlreadySettled")
|
2018-07-18 02:27:04 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
if dbInvoice == nil {
|
|
|
|
t.Fatalf("invoice from db is nil after settle!")
|
|
|
|
}
|
|
|
|
|
|
|
|
invoice.SettleDate = dbInvoice.SettleDate
|
|
|
|
if !reflect.DeepEqual(dbInvoice, invoice) {
|
|
|
|
t.Fatalf("wrong invoice after second settle, expected %v got %v",
|
|
|
|
spew.Sdump(invoice), spew.Sdump(dbInvoice))
|
|
|
|
}
|
|
|
|
}
|
2018-08-11 06:24:04 +03:00
|
|
|
|
|
|
|
// TestQueryInvoices ensures that we can properly query the invoice database for
|
2018-09-11 04:20:38 +03:00
|
|
|
// invoices using different types of queries.
|
2018-08-11 06:24:04 +03:00
|
|
|
func TestQueryInvoices(t *testing.T) {
|
|
|
|
t.Parallel()
|
|
|
|
|
|
|
|
db, cleanUp, err := makeTestDB()
|
|
|
|
defer cleanUp()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to make test db: %v", err)
|
|
|
|
}
|
|
|
|
|
2018-09-11 04:20:38 +03:00
|
|
|
// To begin the test, we'll add 50 invoices to the database. We'll
|
2018-08-11 06:24:04 +03:00
|
|
|
// assume that the index of the invoice within the database is the same
|
|
|
|
// as the amount of the invoice itself.
|
2018-09-11 04:20:38 +03:00
|
|
|
const numInvoices = 50
|
|
|
|
for i := lnwire.MilliSatoshi(1); i <= numInvoices; i++ {
|
2018-08-11 06:24:04 +03:00
|
|
|
invoice, err := randInvoice(i)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to create invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
if _, err := db.AddInvoice(invoice); err != nil {
|
|
|
|
t.Fatalf("unable to add invoice: %v", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll only settle half of all invoices created.
|
|
|
|
if i%2 == 0 {
|
|
|
|
paymentHash := sha256.Sum256(invoice.Terms.PaymentPreimage[:])
|
|
|
|
if _, err := db.SettleInvoice(paymentHash, i); err != nil {
|
|
|
|
t.Fatalf("unable to settle invoice: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-09-11 04:20:38 +03:00
|
|
|
// We'll then retrieve the set of all invoices and pending invoices.
|
|
|
|
// This will serve useful when comparing the expected responses of the
|
|
|
|
// query with the actual ones.
|
|
|
|
invoices, err := db.FetchAllInvoices(false)
|
2018-08-11 06:24:04 +03:00
|
|
|
if err != nil {
|
2018-09-11 04:20:38 +03:00
|
|
|
t.Fatalf("unable to retrieve invoices: %v", err)
|
2018-08-11 06:24:04 +03:00
|
|
|
}
|
2018-09-11 04:20:38 +03:00
|
|
|
pendingInvoices, err := db.FetchAllInvoices(true)
|
2018-08-11 06:24:04 +03:00
|
|
|
if err != nil {
|
2018-09-11 04:20:38 +03:00
|
|
|
t.Fatalf("unable to retrieve pending invoices: %v", err)
|
2018-08-11 06:24:04 +03:00
|
|
|
}
|
|
|
|
|
2018-09-11 04:20:38 +03:00
|
|
|
// The test will consist of several queries along with their respective
|
|
|
|
// expected response. Each query response should match its expected one.
|
|
|
|
testCases := []struct {
|
|
|
|
query InvoiceQuery
|
|
|
|
expected []Invoice
|
|
|
|
}{
|
|
|
|
// Fetch all invoices with a single query.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
expected: invoices,
|
|
|
|
},
|
2018-09-28 13:58:06 +03:00
|
|
|
// Fetch all invoices with a single query, reversed.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
expected: invoices,
|
|
|
|
},
|
2018-09-11 04:20:38 +03:00
|
|
|
// Fetch the first 25 invoices.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
NumMaxInvoices: numInvoices / 2,
|
|
|
|
},
|
|
|
|
expected: invoices[:numInvoices/2],
|
|
|
|
},
|
|
|
|
// Fetch the first 10 invoices, but this time iterating
|
|
|
|
// backwards.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 11,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
expected: invoices[:10],
|
|
|
|
},
|
|
|
|
// Fetch the last 40 invoices.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 10,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
expected: invoices[10:],
|
|
|
|
},
|
2018-09-28 13:58:06 +03:00
|
|
|
// Fetch all but the first invoice.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 1,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
expected: invoices[1:],
|
|
|
|
},
|
|
|
|
// Fetch one invoice, reversed, with index offset 3. This
|
|
|
|
// should give us the second invoice in the array.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 3,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[1:2],
|
|
|
|
},
|
|
|
|
// Same as above, at index 2.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 2,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[0:1],
|
|
|
|
},
|
2018-09-28 13:29:03 +03:00
|
|
|
// Fetch one invoice, at index 1, reversed. Since invoice#1 is
|
|
|
|
// the very first, there won't be any left in a reverse search,
|
|
|
|
// so we expect no invoices to be returned.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 1,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: nil,
|
|
|
|
},
|
|
|
|
// Same as above, but don't restrict the number of invoices to
|
|
|
|
// 1.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 1,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
expected: nil,
|
|
|
|
},
|
2018-09-28 13:58:06 +03:00
|
|
|
// Fetch one invoice, reversed, with no offset set. We expect
|
|
|
|
// the last invoice in the response.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[numInvoices-1:],
|
|
|
|
},
|
|
|
|
// Fetch one invoice, reversed, the offset set at numInvoices+1.
|
|
|
|
// We expect this to return the last invoice.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: numInvoices + 1,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[numInvoices-1:],
|
|
|
|
},
|
|
|
|
// Same as above, at offset numInvoices.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: numInvoices,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[numInvoices-2 : numInvoices-1],
|
|
|
|
},
|
|
|
|
// Fetch one invoice, at no offset (same as offset 0). We
|
|
|
|
// expect the first invoice only in the response.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[:1],
|
|
|
|
},
|
|
|
|
// Same as above, at offset 1.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 1,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[1:2],
|
|
|
|
},
|
|
|
|
// Same as above, at offset 2.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 2,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[2:3],
|
|
|
|
},
|
|
|
|
// Same as above, at offset numInvoices-1. Expect the last
|
|
|
|
// invoice to be returned.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: numInvoices - 1,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: invoices[numInvoices-1:],
|
|
|
|
},
|
|
|
|
// Same as above, at offset numInvoices. No invoices should be
|
|
|
|
// returned, as there are no invoices after this offset.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: numInvoices,
|
|
|
|
NumMaxInvoices: 1,
|
|
|
|
},
|
|
|
|
expected: nil,
|
|
|
|
},
|
2018-09-11 04:20:38 +03:00
|
|
|
// Fetch all pending invoices with a single query.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
PendingOnly: true,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
expected: pendingInvoices,
|
|
|
|
},
|
|
|
|
// Fetch the first 12 pending invoices.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
PendingOnly: true,
|
|
|
|
NumMaxInvoices: numInvoices / 4,
|
|
|
|
},
|
|
|
|
expected: pendingInvoices[:len(pendingInvoices)/2],
|
|
|
|
},
|
|
|
|
// Fetch the first 5 pending invoices, but this time iterating
|
|
|
|
// backwards.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 10,
|
|
|
|
PendingOnly: true,
|
|
|
|
Reversed: true,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
// Since we seek to the invoice with index 10 and
|
|
|
|
// iterate backwards, there should only be 5 pending
|
|
|
|
// invoices before it as every other invoice within the
|
|
|
|
// index is settled.
|
|
|
|
expected: pendingInvoices[:5],
|
|
|
|
},
|
|
|
|
// Fetch the last 15 invoices.
|
|
|
|
{
|
|
|
|
query: InvoiceQuery{
|
|
|
|
IndexOffset: 20,
|
|
|
|
PendingOnly: true,
|
|
|
|
NumMaxInvoices: numInvoices,
|
|
|
|
},
|
|
|
|
// Since we seek to the invoice with index 20, there are
|
|
|
|
// 30 invoices left. From these 30, only 15 of them are
|
|
|
|
// still pending.
|
|
|
|
expected: pendingInvoices[len(pendingInvoices)-15:],
|
|
|
|
},
|
2018-08-11 06:24:04 +03:00
|
|
|
}
|
|
|
|
|
2018-09-11 04:20:38 +03:00
|
|
|
for i, testCase := range testCases {
|
|
|
|
response, err := db.QueryInvoices(testCase.query)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("unable to query invoice database: %v", err)
|
2018-08-11 06:24:04 +03:00
|
|
|
}
|
|
|
|
|
2018-09-11 04:20:38 +03:00
|
|
|
if !reflect.DeepEqual(response.Invoices, testCase.expected) {
|
|
|
|
t.Fatalf("test #%d: query returned incorrect set of "+
|
|
|
|
"invoices: expcted %v, got %v", i,
|
|
|
|
spew.Sdump(response.Invoices),
|
|
|
|
spew.Sdump(testCase.expected))
|
2018-08-11 06:24:04 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|