582b164c46
* mod: bump btcwallet version to accept db timeout * btcwallet: add DBTimeOut in config * kvdb: add database timeout option for bbolt This commit adds a DBTimeout option in bbolt config. The relevant functions walletdb.Open/Create are updated to use this config. In addition, the bolt compacter also applies the new timeout option. * channeldb: add DBTimeout in db options This commit adds the DBTimeout option for channeldb. A new unit test file is created to test the default options. In addition, the params used in kvdb.Create inside channeldb_test is updated with a DefaultDBTimeout value. * contractcourt+routing: use DBTimeout in kvdb This commit touches multiple test files in contractcourt and routing. The call of function kvdb.Create and kvdb.Open are now updated with the new param DBTimeout, using the default value kvdb.DefaultDBTimeout. * lncfg: add DBTimeout option in db config The DBTimeout option is added to db config. A new unit test is added to check the default DB config is created as expected. * migration: add DBTimeout param in kvdb.Create/kvdb.Open * keychain: update tests to use DBTimeout param * htlcswitch+chainreg: add DBTimeout option * macaroons: support DBTimeout config in creation This commit adds the DBTimeout during the creation of macaroons.db. The usage of kvdb.Create and kvdb.Open in its tests are updated with a timeout value using kvdb.DefaultDBTimeout. * walletunlocker: add dbTimeout option in UnlockerService This commit adds a new param, dbTimeout, during the creation of UnlockerService. This param is then passed to wallet.NewLoader inside various service calls, specifying a timeout value to be used when opening the bbolt. In addition, the macaroonService is also called with this dbTimeout param. * watchtower/wtdb: add dbTimeout param during creation This commit adds the dbTimeout param for the creation of both watchtower.db and wtclient.db. * multi: add db timeout param for walletdb.Create This commit adds the db timeout param for the function call walletdb.Create. It touches only the test files found in chainntnfs, lnwallet, and routing. * lnd: pass DBTimeout config to relevant services This commit enables lnd to pass the DBTimeout config to the following services/config/functions, - chainControlConfig - walletunlocker - wallet.NewLoader - macaroons - watchtower In addition, the usage of wallet.Create is updated too. * sample-config: add dbtimeout option
218 lines
5.3 KiB
Go
218 lines
5.3 KiB
Go
package routing
|
|
|
|
import (
|
|
"io/ioutil"
|
|
"os"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/lightningnetwork/lnd/channeldb/kvdb"
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
|
"github.com/lightningnetwork/lnd/routing/route"
|
|
)
|
|
|
|
var (
|
|
mcTestRoute = &route.Route{
|
|
SourcePubKey: mcTestSelf,
|
|
Hops: []*route.Hop{
|
|
{
|
|
ChannelID: 1,
|
|
PubKeyBytes: route.Vertex{11},
|
|
AmtToForward: 1000,
|
|
LegacyPayload: true,
|
|
},
|
|
{
|
|
ChannelID: 2,
|
|
PubKeyBytes: route.Vertex{12},
|
|
LegacyPayload: true,
|
|
},
|
|
},
|
|
}
|
|
|
|
mcTestTime = time.Date(2018, time.January, 9, 14, 00, 00, 0, time.UTC)
|
|
mcTestSelf = route.Vertex{10}
|
|
mcTestNode1 = mcTestRoute.Hops[0].PubKeyBytes
|
|
mcTestNode2 = mcTestRoute.Hops[1].PubKeyBytes
|
|
|
|
testPenaltyHalfLife = 30 * time.Minute
|
|
testAprioriHopProbability = 0.9
|
|
testAprioriWeight = 0.5
|
|
)
|
|
|
|
type mcTestContext struct {
|
|
t *testing.T
|
|
mc *MissionControl
|
|
now time.Time
|
|
|
|
db kvdb.Backend
|
|
dbPath string
|
|
|
|
pid uint64
|
|
}
|
|
|
|
func createMcTestContext(t *testing.T) *mcTestContext {
|
|
ctx := &mcTestContext{
|
|
t: t,
|
|
now: mcTestTime,
|
|
}
|
|
|
|
file, err := ioutil.TempFile("", "*.db")
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
ctx.dbPath = file.Name()
|
|
|
|
ctx.db, err = kvdb.Open(
|
|
kvdb.BoltBackendName, ctx.dbPath, true, kvdb.DefaultDBTimeout,
|
|
)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
ctx.restartMc()
|
|
|
|
return ctx
|
|
}
|
|
|
|
// restartMc creates a new instances of mission control on the same database.
|
|
func (ctx *mcTestContext) restartMc() {
|
|
mc, err := NewMissionControl(
|
|
ctx.db,
|
|
&MissionControlConfig{
|
|
PenaltyHalfLife: testPenaltyHalfLife,
|
|
AprioriHopProbability: testAprioriHopProbability,
|
|
AprioriWeight: testAprioriWeight,
|
|
SelfNode: mcTestSelf,
|
|
},
|
|
)
|
|
if err != nil {
|
|
ctx.t.Fatal(err)
|
|
}
|
|
|
|
mc.now = func() time.Time { return ctx.now }
|
|
ctx.mc = mc
|
|
}
|
|
|
|
// cleanup closes the database and removes the temp file.
|
|
func (ctx *mcTestContext) cleanup() {
|
|
ctx.db.Close()
|
|
os.Remove(ctx.dbPath)
|
|
}
|
|
|
|
// Assert that mission control returns a probability for an edge.
|
|
func (ctx *mcTestContext) expectP(amt lnwire.MilliSatoshi, expected float64) {
|
|
ctx.t.Helper()
|
|
|
|
p := ctx.mc.GetProbability(mcTestNode1, mcTestNode2, amt)
|
|
if p != expected {
|
|
ctx.t.Fatalf("expected probability %v but got %v", expected, p)
|
|
}
|
|
}
|
|
|
|
// reportFailure reports a failure by using a test route.
|
|
func (ctx *mcTestContext) reportFailure(amt lnwire.MilliSatoshi,
|
|
failure lnwire.FailureMessage) {
|
|
|
|
mcTestRoute.Hops[0].AmtToForward = amt
|
|
|
|
errorSourceIdx := 1
|
|
ctx.mc.ReportPaymentFail(
|
|
ctx.pid, mcTestRoute, &errorSourceIdx, failure,
|
|
)
|
|
}
|
|
|
|
// reportSuccess reports a success by using a test route.
|
|
func (ctx *mcTestContext) reportSuccess() {
|
|
err := ctx.mc.ReportPaymentSuccess(ctx.pid, mcTestRoute)
|
|
if err != nil {
|
|
ctx.t.Fatal(err)
|
|
}
|
|
|
|
ctx.pid++
|
|
}
|
|
|
|
// TestMissionControl tests mission control probability estimation.
|
|
func TestMissionControl(t *testing.T) {
|
|
ctx := createMcTestContext(t)
|
|
defer ctx.cleanup()
|
|
|
|
ctx.now = testTime
|
|
|
|
testTime := time.Date(2018, time.January, 9, 14, 00, 00, 0, time.UTC)
|
|
|
|
// For local channels, we expect a higher probability than our a prior
|
|
// test probability.
|
|
selfP := ctx.mc.GetProbability(mcTestSelf, mcTestNode1, 100)
|
|
if selfP != prevSuccessProbability {
|
|
t.Fatalf("expected prev success prob for untried local chans")
|
|
}
|
|
|
|
// Initial probability is expected to be the a priori.
|
|
ctx.expectP(1000, testAprioriHopProbability)
|
|
|
|
// Expect probability to be zero after reporting the edge as failed.
|
|
ctx.reportFailure(1000, lnwire.NewTemporaryChannelFailure(nil))
|
|
ctx.expectP(1000, 0)
|
|
|
|
// As we reported with a min penalization amt, a lower amt than reported
|
|
// should return the node probability, which is the a priori
|
|
// probability.
|
|
ctx.expectP(500, testAprioriHopProbability)
|
|
|
|
// Edge decay started. The node probability weighted average should now
|
|
// have shifted from 1:1 to 1:0.5 -> 60%. The connection probability is
|
|
// half way through the recovery, so we expect 30% here.
|
|
ctx.now = testTime.Add(30 * time.Minute)
|
|
ctx.expectP(1000, 0.3)
|
|
|
|
// Edge fails again, this time without a min penalization amt. The edge
|
|
// should be penalized regardless of amount.
|
|
ctx.reportFailure(0, lnwire.NewTemporaryChannelFailure(nil))
|
|
ctx.expectP(1000, 0)
|
|
ctx.expectP(500, 0)
|
|
|
|
// Edge decay started.
|
|
ctx.now = testTime.Add(60 * time.Minute)
|
|
ctx.expectP(1000, 0.3)
|
|
|
|
// Restart mission control to test persistence.
|
|
ctx.restartMc()
|
|
ctx.expectP(1000, 0.3)
|
|
|
|
// A node level failure should bring probability of all known channels
|
|
// back to zero.
|
|
ctx.reportFailure(0, lnwire.NewExpiryTooSoon(lnwire.ChannelUpdate{}))
|
|
ctx.expectP(1000, 0)
|
|
|
|
// Check whether history snapshot looks sane.
|
|
history := ctx.mc.GetHistorySnapshot()
|
|
|
|
if len(history.Pairs) != 4 {
|
|
t.Fatalf("expected 4 pairs, but got %v", len(history.Pairs))
|
|
}
|
|
|
|
// Test reporting a success.
|
|
ctx.reportSuccess()
|
|
}
|
|
|
|
// TestMissionControlChannelUpdate tests that the first channel update is not
|
|
// penalizing the channel yet.
|
|
func TestMissionControlChannelUpdate(t *testing.T) {
|
|
ctx := createMcTestContext(t)
|
|
|
|
// Report a policy related failure. Because it is the first, we don't
|
|
// expect a penalty.
|
|
ctx.reportFailure(
|
|
0, lnwire.NewFeeInsufficient(0, lnwire.ChannelUpdate{}),
|
|
)
|
|
ctx.expectP(100, testAprioriHopProbability)
|
|
|
|
// Report another failure for the same channel. We expect it to be
|
|
// pruned.
|
|
ctx.reportFailure(
|
|
0, lnwire.NewFeeInsufficient(0, lnwire.ChannelUpdate{}),
|
|
)
|
|
ctx.expectP(100, 0)
|
|
}
|