2016-12-08 09:47:01 +03:00
|
|
|
package channeldb
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
2018-07-18 05:10:07 +03:00
|
|
|
"crypto/sha256"
|
2016-12-08 09:47:01 +03:00
|
|
|
"encoding/binary"
|
2018-10-18 01:35:54 +03:00
|
|
|
"errors"
|
2017-10-02 17:38:45 +03:00
|
|
|
"fmt"
|
2016-12-08 09:47:01 +03:00
|
|
|
"image/color"
|
|
|
|
"io"
|
2017-10-02 17:38:45 +03:00
|
|
|
"math"
|
2016-12-08 09:47:01 +03:00
|
|
|
"net"
|
|
|
|
"time"
|
|
|
|
|
2018-06-05 04:34:16 +03:00
|
|
|
"github.com/btcsuite/btcd/btcec"
|
|
|
|
"github.com/btcsuite/btcd/chaincfg/chainhash"
|
2018-07-18 05:10:07 +03:00
|
|
|
"github.com/btcsuite/btcd/txscript"
|
2018-06-05 04:34:16 +03:00
|
|
|
"github.com/btcsuite/btcd/wire"
|
|
|
|
"github.com/btcsuite/btcutil"
|
2018-07-22 05:49:54 +03:00
|
|
|
"github.com/coreos/bbolt"
|
|
|
|
"github.com/lightningnetwork/lnd/lnwire"
|
2016-12-08 09:47:01 +03:00
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
// nodeBucket is a bucket which houses all the vertices or nodes within
|
|
|
|
// the channel graph. This bucket has a single-sub bucket which adds an
|
|
|
|
// additional index from pubkey -> alias. Within the top-level of this
|
|
|
|
// bucket, the key space maps a node's compressed public key to the
|
|
|
|
// serialized information for that node. Additionally, there's a
|
|
|
|
// special key "source" which stores the pubkey of the source node. The
|
|
|
|
// source node is used as the starting point for all graph/queries and
|
|
|
|
// traversals. The graph is formed as a star-graph with the source node
|
|
|
|
// at the center.
|
|
|
|
//
|
2018-07-23 04:59:28 +03:00
|
|
|
// maps: pubKey -> nodeInfo
|
2016-12-08 09:47:01 +03:00
|
|
|
// maps: source -> selfPubKey
|
|
|
|
nodeBucket = []byte("graph-node")
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// nodeUpdateIndexBucket is a sub-bucket of the nodeBucket. This bucket
|
|
|
|
// will be used to quickly look up the "freshness" of a node's last
|
|
|
|
// update to the network. The bucket only contains keys, and no values,
|
|
|
|
// it's mapping:
|
|
|
|
//
|
|
|
|
// maps: updateTime || nodeID -> nil
|
|
|
|
nodeUpdateIndexBucket = []byte("graph-node-update-index")
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// sourceKey is a special key that resides within the nodeBucket. The
|
|
|
|
// sourceKey maps a key to the public key of the "self node".
|
|
|
|
sourceKey = []byte("source")
|
|
|
|
|
|
|
|
// aliasIndexBucket is a sub-bucket that's nested within the main
|
2018-02-07 06:13:07 +03:00
|
|
|
// nodeBucket. This bucket maps the public key of a node to its
|
2016-12-08 09:47:01 +03:00
|
|
|
// current alias. This bucket is provided as it can be used within a
|
|
|
|
// future UI layer to add an additional degree of confirmation.
|
|
|
|
aliasIndexBucket = []byte("alias")
|
|
|
|
|
|
|
|
// edgeBucket is a bucket which houses all of the edge or channel
|
|
|
|
// information within the channel graph. This bucket essentially acts
|
|
|
|
// as an adjacency list, which in conjunction with a range scan, can be
|
2018-06-18 13:35:22 +03:00
|
|
|
// used to iterate over all the incoming and outgoing edges for a
|
|
|
|
// particular node. Key in the bucket use a prefix scheme which leads
|
|
|
|
// with the node's public key and sends with the compact edge ID.
|
|
|
|
// For each chanID, there will be two entries within the bucket, as the
|
|
|
|
// graph is directed: nodes may have different policies w.r.t to fees
|
|
|
|
// for their respective directions.
|
2016-12-08 09:47:01 +03:00
|
|
|
//
|
2018-06-18 13:35:22 +03:00
|
|
|
// maps: pubKey || chanID -> channel edge policy for node
|
2016-12-08 09:47:01 +03:00
|
|
|
edgeBucket = []byte("graph-edge")
|
|
|
|
|
2018-06-18 13:35:22 +03:00
|
|
|
// unknownPolicy is represented as an empty slice. It is
|
|
|
|
// used as the value in edgeBucket for unknown channel edge policies.
|
|
|
|
// Unknown policies are still stored in the database to enable efficient
|
|
|
|
// lookup of incoming channel edges.
|
|
|
|
unknownPolicy = []byte{}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// chanStart is an array of all zero bytes which is used to perform
|
|
|
|
// range scans within the edgeBucket to obtain all of the outgoing
|
|
|
|
// edges for a particular node.
|
|
|
|
chanStart [8]byte
|
|
|
|
|
|
|
|
// edgeIndexBucket is an index which can be used to iterate all edges
|
2017-03-06 04:18:00 +03:00
|
|
|
// in the bucket, grouping them according to their in/out nodes.
|
|
|
|
// Additionally, the items in this bucket also contain the complete
|
|
|
|
// edge information for a channel. The edge information includes the
|
2018-04-18 05:02:04 +03:00
|
|
|
// capacity of the channel, the nodes that made the channel, etc. This
|
|
|
|
// bucket resides within the edgeBucket above. Creation of an edge
|
2016-12-08 09:47:01 +03:00
|
|
|
// proceeds in two phases: first the edge is added to the edge index,
|
|
|
|
// afterwards the edgeBucket can be updated with the latest details of
|
|
|
|
// the edge as they are announced on the network.
|
|
|
|
//
|
2017-03-06 04:18:00 +03:00
|
|
|
// maps: chanID -> pubKey1 || pubKey2 || restofEdgeInfo
|
2016-12-08 09:47:01 +03:00
|
|
|
edgeIndexBucket = []byte("edge-index")
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// edgeUpdateIndexBucket is a sub-bucket of the main edgeBucket. This
|
|
|
|
// bucket contains an index which allows us to gauge the "freshness" of
|
|
|
|
// a channel's last updates.
|
|
|
|
//
|
|
|
|
// maps: updateTime || chanID -> nil
|
|
|
|
edgeUpdateIndexBucket = []byte("edge-update-index")
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// channelPointBucket maps a channel's full outpoint (txid:index) to
|
|
|
|
// its short 8-byte channel ID. This bucket resides within the
|
|
|
|
// edgeBucket above, and can be used to quickly remove an edge due to
|
|
|
|
// the outpoint being spent, or to query for existence of a channel.
|
|
|
|
//
|
|
|
|
// maps: outPoint -> chanID
|
|
|
|
channelPointBucket = []byte("chan-index")
|
|
|
|
|
2019-03-27 23:06:12 +03:00
|
|
|
// zombieBucket is a sub-bucket of the main edgeBucket bucket
|
|
|
|
// responsible for maintaining an index of zombie channels. Each entry
|
|
|
|
// exists within the bucket as follows:
|
|
|
|
//
|
|
|
|
// maps: chanID -> pubKey1 || pubKey2
|
|
|
|
//
|
|
|
|
// The chanID represents the channel ID of the edge that is marked as a
|
|
|
|
// zombie and is used as the key, which maps to the public keys of the
|
|
|
|
// edge's participants.
|
|
|
|
zombieBucket = []byte("zombie-index")
|
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
// graphMetaBucket is a top-level bucket which stores various meta-deta
|
2017-03-06 04:18:00 +03:00
|
|
|
// related to the on-disk channel graph. Data stored in this bucket
|
2016-12-20 03:58:27 +03:00
|
|
|
// includes the block to which the graph has been synced to, the total
|
|
|
|
// number of channels, etc.
|
|
|
|
graphMetaBucket = []byte("graph-meta")
|
|
|
|
|
2017-10-02 17:38:45 +03:00
|
|
|
// pruneLogBucket is a bucket within the graphMetaBucket that stores
|
|
|
|
// a mapping from the block height to the hash for the blocks used to
|
|
|
|
// prune the graph.
|
|
|
|
// Once a new block is discovered, any channels that have been closed
|
|
|
|
// (by spending the outpoint) can safely be removed from the graph, and
|
|
|
|
// the block is added to the prune log. We need to keep such a log for
|
|
|
|
// the case where a reorg happens, and we must "rewind" the state of the
|
|
|
|
// graph by removing channels that were previously confirmed. In such a
|
|
|
|
// case we'll remove all entries from the prune log with a block height
|
|
|
|
// that no longer exists.
|
|
|
|
pruneLogBucket = []byte("prune-log")
|
2016-12-08 09:47:01 +03:00
|
|
|
)
|
|
|
|
|
2018-09-05 03:01:20 +03:00
|
|
|
const (
|
|
|
|
// MaxAllowedExtraOpaqueBytes is the largest amount of opaque bytes that
|
|
|
|
// we'll permit to be written to disk. We limit this as otherwise, it
|
|
|
|
// would be possible for a node to create a ton of updates and slowly
|
|
|
|
// fill our disk, and also waste bandwidth due to relaying.
|
|
|
|
MaxAllowedExtraOpaqueBytes = 10000
|
|
|
|
)
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// ChannelGraph is a persistent, on-disk graph representation of the Lightning
|
|
|
|
// Network. This struct can be used to implement path finding algorithms on top
|
|
|
|
// of, and also to update a node's view based on information received from the
|
|
|
|
// p2p network. Internally, the graph is stored using a modified adjacency list
|
|
|
|
// representation with some added object interaction possible with each
|
|
|
|
// serialized edge/node. The graph is stored is directed, meaning that are two
|
|
|
|
// edges stored for each channel: an inbound/outbound edge for each node pair.
|
|
|
|
// Nodes, edges, and edge information can all be added to the graph
|
|
|
|
// independently. Edge removal results in the deletion of all edge information
|
|
|
|
// for that edge.
|
|
|
|
type ChannelGraph struct {
|
|
|
|
db *DB
|
|
|
|
|
|
|
|
// TODO(roasbeef): store and update bloom filter to reduce disk access
|
|
|
|
// due to current gossip model
|
|
|
|
// * LRU cache for edges?
|
|
|
|
}
|
|
|
|
|
2017-10-11 07:36:36 +03:00
|
|
|
// Database returns a pointer to the underlying database.
|
|
|
|
func (c *ChannelGraph) Database() *DB {
|
|
|
|
return c.db
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// ForEachChannel iterates through all the channel edges stored within the
|
|
|
|
// graph and invokes the passed callback for each edge. The callback takes two
|
|
|
|
// edges as since this is a directed graph, both the in/out edges are visited.
|
|
|
|
// If the callback returns an error, then the transaction is aborted and the
|
|
|
|
// iteration stops early.
|
2017-01-18 00:01:07 +03:00
|
|
|
//
|
2017-03-06 04:18:00 +03:00
|
|
|
// NOTE: If an edge can't be found, or wasn't advertised, then a nil pointer
|
|
|
|
// for that particular channel edge routing policy will be passed into the
|
|
|
|
// callback.
|
|
|
|
func (c *ChannelGraph) ForEachChannel(cb func(*ChannelEdgeInfo, *ChannelEdgePolicy, *ChannelEdgePolicy) error) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// TODO(roasbeef): ptr map to reduce # of allocs? no duplicates
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
return c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// First, grab the node bucket. This will be used to populate
|
|
|
|
// the Node pointers in each edge read from disk.
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next, grab the edge bucket which stores the edges, and also
|
|
|
|
// the index itself so we can group the directed edges together
|
|
|
|
// logically.
|
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
2016-12-25 03:29:49 +03:00
|
|
|
return ErrGraphNoEdgesFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
2016-12-25 03:29:49 +03:00
|
|
|
return ErrGraphNoEdgesFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// For each edge pair within the edge index, we fetch each edge
|
|
|
|
// itself and also the node information in order to fully
|
2017-03-06 04:18:00 +03:00
|
|
|
// populated the object.
|
|
|
|
return edgeIndex.ForEach(func(chanID, edgeInfoBytes []byte) error {
|
|
|
|
infoReader := bytes.NewReader(edgeInfoBytes)
|
|
|
|
edgeInfo, err := deserializeChanEdgeInfo(infoReader)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-08-10 06:45:33 +03:00
|
|
|
edgeInfo.db = c.db
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
edge1, edge2, err := fetchChanEdgePolicies(
|
|
|
|
edgeIndex, edges, nodes, chanID, c.db,
|
|
|
|
)
|
|
|
|
if err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// With both edges read, execute the call back. IF this
|
|
|
|
// function returns an error then the transaction will
|
|
|
|
// be aborted.
|
2018-02-13 03:17:08 +03:00
|
|
|
return cb(&edgeInfo, edge1, edge2)
|
2016-12-08 09:47:01 +03:00
|
|
|
})
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
// ForEachNode iterates through all the stored vertices/nodes in the graph,
|
|
|
|
// executing the passed callback with each node encountered. If the callback
|
|
|
|
// returns an error, then the transaction is aborted and the iteration stops
|
|
|
|
// early.
|
2017-04-14 23:14:02 +03:00
|
|
|
//
|
|
|
|
// If the caller wishes to re-use an existing boltdb transaction, then it
|
|
|
|
// should be passed as the first argument. Otherwise the first argument should
|
|
|
|
// be nil and a fresh transaction will be created to execute the graph
|
|
|
|
// traversal
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): add iterator interface to allow for memory efficient graph
|
|
|
|
// traversal when graph gets mega
|
2018-11-30 07:04:21 +03:00
|
|
|
func (c *ChannelGraph) ForEachNode(tx *bbolt.Tx, cb func(*bbolt.Tx, *LightningNode) error) error {
|
|
|
|
traversal := func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// First grab the nodes bucket which stores the mapping from
|
|
|
|
// pubKey to node information.
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
return nodes.ForEach(func(pubKey, nodeBytes []byte) error {
|
|
|
|
// If this is the source key, then we skip this
|
|
|
|
// iteration as the value for this key is a pubKey
|
|
|
|
// rather than raw node information.
|
|
|
|
if bytes.Equal(pubKey, sourceKey) || len(pubKey) != 33 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeReader := bytes.NewReader(nodeBytes)
|
|
|
|
node, err := deserializeLightningNode(nodeReader)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
node.db = c.db
|
|
|
|
|
|
|
|
// Execute the callback, the transaction will abort if
|
|
|
|
// this returns an error.
|
2018-02-13 03:17:08 +03:00
|
|
|
return cb(tx, &node)
|
2016-12-08 09:47:01 +03:00
|
|
|
})
|
2017-04-14 23:14:02 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// If no transaction was provided, then we'll create a new transaction
|
|
|
|
// to execute the transaction within.
|
|
|
|
if tx == nil {
|
|
|
|
return c.db.View(traversal)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we re-use the existing transaction to execute the graph
|
|
|
|
// traversal.
|
|
|
|
return traversal(tx)
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// SourceNode returns the source node of the graph. The source node is treated
|
2017-01-13 08:01:50 +03:00
|
|
|
// as the center node within a star-graph. This method may be used to kick off
|
2016-12-08 09:47:01 +03:00
|
|
|
// a path finding algorithm in order to explore the reachability of another
|
|
|
|
// node based off the source node.
|
2016-12-22 23:49:30 +03:00
|
|
|
func (c *ChannelGraph) SourceNode() (*LightningNode, error) {
|
2016-12-08 09:47:01 +03:00
|
|
|
var source *LightningNode
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-12-04 13:20:38 +03:00
|
|
|
// First grab the nodes bucket which stores the mapping from
|
|
|
|
// pubKey to node information.
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
node, err := c.sourceNode(nodes)
|
2016-12-08 09:47:01 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-06-15 05:47:00 +03:00
|
|
|
source = node
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return source, nil
|
|
|
|
}
|
|
|
|
|
2018-06-15 05:47:00 +03:00
|
|
|
// sourceNode uses an existing database transaction and returns the source node
|
|
|
|
// of the graph. The source node is treated as the center node within a
|
|
|
|
// star-graph. This method may be used to kick off a path finding algorithm in
|
|
|
|
// order to explore the reachability of another node based off the source node.
|
2018-12-04 13:20:38 +03:00
|
|
|
func (c *ChannelGraph) sourceNode(nodes *bbolt.Bucket) (*LightningNode, error) {
|
2018-06-15 05:47:00 +03:00
|
|
|
selfPub := nodes.Get(sourceKey)
|
|
|
|
if selfPub == nil {
|
|
|
|
return nil, ErrSourceNodeNotSet
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the pubKey of the source node retrieved, we're able to
|
|
|
|
// fetch the full node information.
|
|
|
|
node, err := fetchLightningNode(nodes, selfPub)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
node.db = c.db
|
|
|
|
|
|
|
|
return &node, nil
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// SetSourceNode sets the source node within the graph database. The source
|
|
|
|
// node is to be used as the center of a star-graph within path finding
|
|
|
|
// algorithms.
|
2016-12-22 23:49:30 +03:00
|
|
|
func (c *ChannelGraph) SetSourceNode(node *LightningNode) error {
|
2018-01-31 07:19:40 +03:00
|
|
|
nodePubBytes := node.PubKeyBytes[:]
|
|
|
|
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// First grab the nodes bucket which stores the mapping from
|
|
|
|
// pubKey to node information.
|
|
|
|
nodes, err := tx.CreateBucketIfNotExists(nodeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next we create the mapping from source to the targeted
|
|
|
|
// public key.
|
2018-01-31 07:19:40 +03:00
|
|
|
if err := nodes.Put(sourceKey, nodePubBytes); err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, we commit the information of the lightning node
|
|
|
|
// itself.
|
|
|
|
return addLightningNode(tx, node)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2017-07-14 22:25:02 +03:00
|
|
|
// AddLightningNode adds a vertex/node to the graph database. If the node is not
|
|
|
|
// in the database from before, this will add a new, unconnected one to the
|
|
|
|
// graph. If it is present from before, this will update that node's
|
|
|
|
// information. Note that this method is expected to only be called to update
|
2017-12-18 05:40:05 +03:00
|
|
|
// an already present node from a node announcement, or to insert a node found
|
2017-07-14 22:25:02 +03:00
|
|
|
// in a channel update.
|
|
|
|
//
|
2016-12-22 23:49:30 +03:00
|
|
|
// TODO(roasbeef): also need sig of announcement
|
|
|
|
func (c *ChannelGraph) AddLightningNode(node *LightningNode) error {
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
return addLightningNode(tx, node)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func addLightningNode(tx *bbolt.Tx, node *LightningNode) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
nodes, err := tx.CreateBucketIfNotExists(nodeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
aliases, err := nodes.CreateBucketIfNotExists(aliasIndexBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
updateIndex, err := nodes.CreateBucketIfNotExists(
|
|
|
|
nodeUpdateIndexBucket,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
return putLightningNode(nodes, aliases, updateIndex, node)
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// LookupAlias attempts to return the alias as advertised by the target node.
|
2016-12-20 03:58:27 +03:00
|
|
|
// TODO(roasbeef): currently assumes that aliases are unique...
|
2016-12-22 23:49:30 +03:00
|
|
|
func (c *ChannelGraph) LookupAlias(pub *btcec.PublicKey) (string, error) {
|
2016-12-08 09:47:01 +03:00
|
|
|
var alias string
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
aliases := nodes.Bucket(aliasIndexBucket)
|
|
|
|
if aliases == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
nodePub := pub.SerializeCompressed()
|
|
|
|
a := aliases.Get(nodePub)
|
|
|
|
if a == nil {
|
|
|
|
return ErrNodeAliasNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO(roasbeef): should actually be using the utf-8
|
|
|
|
// package...
|
|
|
|
alias = string(a)
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
|
|
|
|
return alias, nil
|
|
|
|
}
|
|
|
|
|
2018-06-15 05:47:00 +03:00
|
|
|
// DeleteLightningNode starts a new database transaction to remove a vertex/node
|
|
|
|
// from the database according to the node's public key.
|
2016-12-22 23:49:30 +03:00
|
|
|
func (c *ChannelGraph) DeleteLightningNode(nodePub *btcec.PublicKey) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// TODO(roasbeef): ensure dangling edges are removed...
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2018-12-04 13:25:05 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodeNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
return c.deleteLightningNode(
|
|
|
|
nodes, nodePub.SerializeCompressed(),
|
|
|
|
)
|
2018-06-15 05:47:00 +03:00
|
|
|
})
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-06-15 05:47:00 +03:00
|
|
|
// deleteLightningNode uses an existing database transaction to remove a
|
|
|
|
// vertex/node from the database according to the node's public key.
|
2018-12-04 13:25:05 +03:00
|
|
|
func (c *ChannelGraph) deleteLightningNode(nodes *bbolt.Bucket,
|
2018-06-15 05:47:00 +03:00
|
|
|
compressedPubKey []byte) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-06-15 05:47:00 +03:00
|
|
|
aliases := nodes.Bucket(aliasIndexBucket)
|
|
|
|
if aliases == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := aliases.Delete(compressedPubKey); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-08-17 04:21:39 +03:00
|
|
|
|
|
|
|
// Before we delete the node, we'll fetch its current state so we can
|
|
|
|
// determine when its last update was to clear out the node update
|
|
|
|
// index.
|
|
|
|
node, err := fetchLightningNode(nodes, compressedPubKey)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := nodes.Delete(compressedPubKey); err != nil {
|
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, we'll delete the index entry for the node within the
|
|
|
|
// nodeUpdateIndexBucket as this node is no longer active, so we don't
|
|
|
|
// need to track its last update.
|
|
|
|
nodeUpdateIndex := nodes.Bucket(nodeUpdateIndexBucket)
|
|
|
|
if nodeUpdateIndex == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// In order to delete the entry, we'll need to reconstruct the key for
|
|
|
|
// its last update.
|
|
|
|
updateUnix := uint64(node.LastUpdate.Unix())
|
|
|
|
var indexKey [8 + 33]byte
|
|
|
|
byteOrder.PutUint64(indexKey[:8], updateUnix)
|
|
|
|
copy(indexKey[8:], compressedPubKey)
|
|
|
|
|
|
|
|
return nodeUpdateIndex.Delete(indexKey[:])
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// AddChannelEdge adds a new (undirected, blank) edge to the graph database. An
|
2017-03-06 04:18:00 +03:00
|
|
|
// undirected edge from the two target nodes are created. The information
|
|
|
|
// stored denotes the static attributes of the channel, such as the channelID,
|
|
|
|
// the keys involved in creation of the channel, and the set of features that
|
|
|
|
// the channel supports. The chanPoint and chanID are used to uniquely identify
|
|
|
|
// the edge globally within the database.
|
|
|
|
func (c *ChannelGraph) AddChannelEdge(edge *ChannelEdgeInfo) error {
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2018-12-10 06:30:26 +03:00
|
|
|
return c.addChannelEdge(tx, edge)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
// addChannelEdge is the private form of AddChannelEdge that allows callers to
|
|
|
|
// utilize an existing db transaction.
|
|
|
|
func (c *ChannelGraph) addChannelEdge(tx *bbolt.Tx, edge *ChannelEdgeInfo) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// Construct the channel's primary key which is the 8-byte channel ID.
|
|
|
|
var chanKey [8]byte
|
2017-03-06 04:18:00 +03:00
|
|
|
binary.BigEndian.PutUint64(chanKey[:], edge.ChannelID)
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
nodes, err := tx.CreateBucketIfNotExists(nodeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
edges, err := tx.CreateBucketIfNotExists(edgeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
edgeIndex, err := edges.CreateBucketIfNotExists(edgeIndexBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
chanIndex, err := edges.CreateBucketIfNotExists(channelPointBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
// First, attempt to check if this edge has already been created. If
|
|
|
|
// so, then we can exit early as this method is meant to be idempotent.
|
|
|
|
if edgeInfo := edgeIndex.Get(chanKey[:]); edgeInfo != nil {
|
|
|
|
return ErrEdgeAlreadyExist
|
|
|
|
}
|
2018-07-23 07:02:46 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
// Before we insert the channel into the database, we'll ensure that
|
|
|
|
// both nodes already exist in the channel graph. If either node
|
|
|
|
// doesn't, then we'll insert a "shell" node that just includes its
|
|
|
|
// public key, so subsequent validation and queries can work properly.
|
|
|
|
_, node1Err := fetchLightningNode(nodes, edge.NodeKey1Bytes[:])
|
|
|
|
switch {
|
|
|
|
case node1Err == ErrGraphNodeNotFound:
|
|
|
|
node1Shell := LightningNode{
|
|
|
|
PubKeyBytes: edge.NodeKey1Bytes,
|
|
|
|
HaveNodeAnnouncement: false,
|
2018-07-23 07:02:46 +03:00
|
|
|
}
|
2018-12-10 06:30:26 +03:00
|
|
|
err := addLightningNode(tx, &node1Shell)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("unable to create shell node "+
|
|
|
|
"for: %x", edge.NodeKey1Bytes)
|
2018-07-23 07:02:46 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
}
|
|
|
|
case node1Err != nil:
|
|
|
|
return err
|
|
|
|
}
|
2018-07-23 07:02:46 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
_, node2Err := fetchLightningNode(nodes, edge.NodeKey2Bytes[:])
|
|
|
|
switch {
|
|
|
|
case node2Err == ErrGraphNodeNotFound:
|
|
|
|
node2Shell := LightningNode{
|
|
|
|
PubKeyBytes: edge.NodeKey2Bytes,
|
|
|
|
HaveNodeAnnouncement: false,
|
2018-07-23 07:02:46 +03:00
|
|
|
}
|
2018-12-10 06:30:26 +03:00
|
|
|
err := addLightningNode(tx, &node2Shell)
|
|
|
|
if err != nil {
|
|
|
|
return fmt.Errorf("unable to create shell node "+
|
|
|
|
"for: %x", edge.NodeKey2Bytes)
|
2018-07-23 07:02:46 +03:00
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
2018-12-10 06:30:26 +03:00
|
|
|
case node2Err != nil:
|
|
|
|
return err
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
// If the edge hasn't been created yet, then we'll first add it to the
|
|
|
|
// edge index in order to associate the edge between two nodes and also
|
|
|
|
// store the static components of the channel.
|
|
|
|
if err := putChanEdgeInfo(edgeIndex, edge, chanKey); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-06-18 13:35:22 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
// Mark edge policies for both sides as unknown. This is to enable
|
|
|
|
// efficient incoming channel lookup for a node.
|
|
|
|
for _, key := range []*[33]byte{&edge.NodeKey1Bytes,
|
|
|
|
&edge.NodeKey2Bytes} {
|
2018-06-18 13:35:22 +03:00
|
|
|
|
2018-12-10 06:30:26 +03:00
|
|
|
err := putChanEdgePolicyUnknown(edges, edge.ChannelID,
|
|
|
|
key[:])
|
|
|
|
if err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return err
|
|
|
|
}
|
2018-12-10 06:30:26 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Finally we add it to the channel index which maps channel points
|
|
|
|
// (outpoints) to the shorter channel ID's.
|
|
|
|
var b bytes.Buffer
|
|
|
|
if err := writeOutpoint(&b, &edge.ChannelPoint); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
return chanIndex.Put(b.Bytes(), chanKey[:])
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// HasChannelEdge returns true if the database knows of a channel edge with the
|
2018-07-23 04:59:28 +03:00
|
|
|
// passed channel ID, and false otherwise. If an edge with that ID is found
|
2016-12-22 23:49:30 +03:00
|
|
|
// within the graph, then two time stamps representing the last time the edge
|
2019-03-27 23:06:57 +03:00
|
|
|
// was updated for both directed edges are returned along with the boolean. If
|
|
|
|
// it is not found, then the zombie index is checked and its result is returned
|
|
|
|
// as the second boolean.
|
|
|
|
func (c *ChannelGraph) HasChannelEdge(chanID uint64,
|
|
|
|
) (time.Time, time.Time, bool, bool, error) {
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2016-12-22 23:49:30 +03:00
|
|
|
var (
|
|
|
|
node1UpdateTime time.Time
|
|
|
|
node2UpdateTime time.Time
|
|
|
|
exists bool
|
2019-03-27 23:06:57 +03:00
|
|
|
isZombie bool
|
2016-12-22 23:49:30 +03:00
|
|
|
)
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2019-03-27 23:06:57 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
|
|
|
|
var channelID [8]byte
|
|
|
|
byteOrder.PutUint64(channelID[:], chanID)
|
2019-03-27 23:06:57 +03:00
|
|
|
|
|
|
|
// If the edge doesn't exist, then we'll also check our zombie
|
|
|
|
// index.
|
2016-12-22 23:49:30 +03:00
|
|
|
if edgeIndex.Get(channelID[:]) == nil {
|
|
|
|
exists = false
|
2019-03-27 23:06:57 +03:00
|
|
|
zombieIndex := edges.Bucket(zombieBucket)
|
|
|
|
if zombieIndex != nil {
|
|
|
|
isZombie, _, _ = isZombieEdge(zombieIndex, chanID)
|
|
|
|
}
|
|
|
|
|
2016-12-22 23:49:30 +03:00
|
|
|
return nil
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2016-12-27 07:00:35 +03:00
|
|
|
exists = true
|
2019-03-27 23:06:57 +03:00
|
|
|
isZombie = false
|
2016-12-27 07:00:35 +03:00
|
|
|
|
2016-12-22 23:49:30 +03:00
|
|
|
// If the channel has been found in the graph, then retrieve
|
|
|
|
// the edges itself so we can return the last updated
|
2018-02-07 06:11:11 +03:00
|
|
|
// timestamps.
|
2016-12-22 23:49:30 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodeNotFound
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
e1, e2, err := fetchChanEdgePolicies(edgeIndex, edges, nodes,
|
2016-12-22 23:49:30 +03:00
|
|
|
channelID[:], c.db)
|
|
|
|
if err != nil {
|
2016-12-31 03:35:23 +03:00
|
|
|
return err
|
2016-12-22 23:49:30 +03:00
|
|
|
}
|
|
|
|
|
2016-12-31 03:35:23 +03:00
|
|
|
// As we may have only one of the edges populated, only set the
|
|
|
|
// update time if the edge was found in the database.
|
|
|
|
if e1 != nil {
|
|
|
|
node1UpdateTime = e1.LastUpdate
|
|
|
|
}
|
|
|
|
if e2 != nil {
|
|
|
|
node2UpdateTime = e2.LastUpdate
|
|
|
|
}
|
2016-12-22 23:49:30 +03:00
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil
|
2019-03-27 23:06:57 +03:00
|
|
|
})
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2019-03-27 23:06:57 +03:00
|
|
|
return node1UpdateTime, node2UpdateTime, exists, isZombie, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-03-27 20:00:38 +03:00
|
|
|
// UpdateChannelEdge retrieves and update edge of the graph database. Method
|
2018-02-07 06:13:07 +03:00
|
|
|
// only reserved for updating an edge info after its already been created.
|
2017-03-27 20:00:38 +03:00
|
|
|
// In order to maintain this constraints, we return an error in the scenario
|
|
|
|
// that an edge info hasn't yet been created yet, but someone attempts to update
|
|
|
|
// it.
|
|
|
|
func (c *ChannelGraph) UpdateChannelEdge(edge *ChannelEdgeInfo) error {
|
|
|
|
// Construct the channel's primary key which is the 8-byte channel ID.
|
|
|
|
var chanKey [8]byte
|
|
|
|
binary.BigEndian.PutUint64(chanKey[:], edge.ChannelID)
|
|
|
|
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2018-12-04 13:20:38 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edge == nil {
|
|
|
|
return ErrEdgeNotFound
|
2017-03-27 20:00:38 +03:00
|
|
|
}
|
2018-12-04 13:20:38 +03:00
|
|
|
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrEdgeNotFound
|
2017-03-27 20:00:38 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
if edgeInfo := edgeIndex.Get(chanKey[:]); edgeInfo == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
return putChanEdgeInfo(edgeIndex, edge, chanKey)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
const (
|
2017-10-02 17:38:45 +03:00
|
|
|
// pruneTipBytes is the total size of the value which stores a prune
|
|
|
|
// entry of the graph in the prune log. The "prune tip" is the last
|
|
|
|
// entry in the prune log, and indicates if the channel graph is in
|
|
|
|
// sync with the current UTXO state. The structure of the value
|
|
|
|
// is: blockHash, taking 32 bytes total.
|
|
|
|
pruneTipBytes = 32
|
2016-12-20 03:58:27 +03:00
|
|
|
)
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
// PruneGraph prunes newly closed channels from the channel graph in response
|
|
|
|
// to a new block being solved on the network. Any transactions which spend the
|
2016-12-22 23:49:30 +03:00
|
|
|
// funding output of any known channels within he graph will be deleted.
|
2016-12-20 03:58:27 +03:00
|
|
|
// Additionally, the "prune tip", or the last block which has been used to
|
|
|
|
// prune the graph is stored so callers can ensure the graph is fully in sync
|
2017-03-06 04:20:46 +03:00
|
|
|
// with the current UTXO state. A slice of channels that have been closed by
|
|
|
|
// the target block are returned if the function succeeds without error.
|
2016-12-22 23:49:30 +03:00
|
|
|
func (c *ChannelGraph) PruneGraph(spentOutputs []*wire.OutPoint,
|
2017-03-06 04:20:46 +03:00
|
|
|
blockHash *chainhash.Hash, blockHeight uint32) ([]*ChannelEdgeInfo, error) {
|
2016-12-20 03:58:27 +03:00
|
|
|
|
2017-03-06 04:20:46 +03:00
|
|
|
var chansClosed []*ChannelEdgeInfo
|
2016-12-20 03:58:27 +03:00
|
|
|
|
2019-03-13 23:13:36 +03:00
|
|
|
err := c.db.Batch(func(tx *bbolt.Tx) error {
|
|
|
|
chansClosed = nil
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// First grab the edges bucket which houses the information
|
|
|
|
// we'd like to delete
|
|
|
|
edges, err := tx.CreateBucketIfNotExists(edgeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Next grab the two edge indexes which will also need to be updated.
|
|
|
|
edgeIndex, err := edges.CreateBucketIfNotExists(edgeIndexBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
chanIndex, err := edges.CreateBucketIfNotExists(channelPointBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-12-04 13:20:39 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrSourceNodeNotSet
|
2018-04-28 02:21:46 +03:00
|
|
|
}
|
2019-03-27 23:06:34 +03:00
|
|
|
zombieIndex, err := edges.CreateBucketIfNotExists(zombieBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-02-07 06:11:11 +03:00
|
|
|
// For each of the outpoints that have been spent within the
|
2016-12-20 03:58:27 +03:00
|
|
|
// block, we attempt to delete them from the graph as if that
|
|
|
|
// outpoint was a channel, then it has now been closed.
|
|
|
|
for _, chanPoint := range spentOutputs {
|
|
|
|
// TODO(roasbeef): load channel bloom filter, continue
|
|
|
|
// if NOT if filter
|
|
|
|
|
2017-03-06 04:20:46 +03:00
|
|
|
var opBytes bytes.Buffer
|
2017-07-26 06:39:59 +03:00
|
|
|
if err := writeOutpoint(&opBytes, chanPoint); err != nil {
|
2017-03-06 04:20:46 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// First attempt to see if the channel exists within
|
|
|
|
// the database, if not, then we can exit early.
|
|
|
|
chanID := chanIndex.Get(opBytes.Bytes())
|
|
|
|
if chanID == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// However, if it does, then we'll read out the full
|
|
|
|
// version so we can add it to the set of deleted
|
|
|
|
// channels.
|
|
|
|
edgeInfo, err := fetchChanEdgeInfo(edgeIndex, chanID)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Attempt to delete the channel, an ErrEdgeNotFound
|
2016-12-20 03:58:27 +03:00
|
|
|
// will be returned if that outpoint isn't known to be
|
|
|
|
// a channel. If no error is returned, then a channel
|
|
|
|
// was successfully pruned.
|
2018-04-28 02:21:46 +03:00
|
|
|
err = delChannelByEdge(
|
2019-03-27 23:06:34 +03:00
|
|
|
edges, edgeIndex, chanIndex, zombieIndex, nodes,
|
|
|
|
chanPoint, false,
|
2018-04-28 02:21:46 +03:00
|
|
|
)
|
2016-12-20 03:58:27 +03:00
|
|
|
if err != nil && err != ErrEdgeNotFound {
|
|
|
|
return err
|
|
|
|
}
|
2018-06-28 06:05:45 +03:00
|
|
|
|
|
|
|
chansClosed = append(chansClosed, &edgeInfo)
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
metaBucket, err := tx.CreateBucketIfNotExists(graphMetaBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-10-02 17:38:45 +03:00
|
|
|
pruneBucket, err := metaBucket.CreateBucketIfNotExists(pruneLogBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the graph pruned, add a new entry to the prune log,
|
|
|
|
// which can be used to check if the graph is fully synced with
|
|
|
|
// the current UTXO state.
|
|
|
|
var blockHeightBytes [4]byte
|
|
|
|
byteOrder.PutUint32(blockHeightBytes[:], blockHeight)
|
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
var newTip [pruneTipBytes]byte
|
|
|
|
copy(newTip[:], blockHash[:])
|
|
|
|
|
2018-06-28 06:05:45 +03:00
|
|
|
err = pruneBucket.Put(blockHeightBytes[:], newTip[:])
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now that the graph has been pruned, we'll also attempt to
|
|
|
|
// prune any nodes that have had a channel closed within the
|
|
|
|
// latest block.
|
2018-12-04 13:25:05 +03:00
|
|
|
return c.pruneGraphNodes(nodes, edgeIndex)
|
2016-12-20 03:58:27 +03:00
|
|
|
})
|
|
|
|
if err != nil {
|
2017-03-06 04:20:46 +03:00
|
|
|
return nil, err
|
2016-12-20 03:58:27 +03:00
|
|
|
}
|
|
|
|
|
2017-03-06 04:20:46 +03:00
|
|
|
return chansClosed, nil
|
2016-12-20 03:58:27 +03:00
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-07-22 05:49:54 +03:00
|
|
|
// PruneGraphNodes is a garbage collection method which attempts to prune out
|
|
|
|
// any nodes from the channel graph that are currently unconnected. This ensure
|
|
|
|
// that we only maintain a graph of reachable nodes. In the event that a pruned
|
|
|
|
// node gains more channels, it will be re-added back to the graph.
|
|
|
|
func (c *ChannelGraph) PruneGraphNodes() error {
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2018-12-04 13:20:38 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
2018-07-22 05:49:54 +03:00
|
|
|
}
|
2018-07-23 04:59:28 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
2018-07-22 05:49:54 +03:00
|
|
|
}
|
|
|
|
|
2018-12-04 13:25:05 +03:00
|
|
|
return c.pruneGraphNodes(nodes, edgeIndex)
|
2018-07-22 05:49:54 +03:00
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2018-06-28 06:05:45 +03:00
|
|
|
// pruneGraphNodes attempts to remove any nodes from the graph who have had a
|
|
|
|
// channel closed within the current block. If the node still has existing
|
|
|
|
// channels in the graph, this will act as a no-op.
|
2018-12-04 13:25:05 +03:00
|
|
|
func (c *ChannelGraph) pruneGraphNodes(nodes *bbolt.Bucket,
|
2018-11-30 07:04:21 +03:00
|
|
|
edgeIndex *bbolt.Bucket) error {
|
2018-06-28 06:05:45 +03:00
|
|
|
|
|
|
|
log.Trace("Pruning nodes from graph with no open channels")
|
|
|
|
|
|
|
|
// We'll retrieve the graph's source node to ensure we don't remove it
|
|
|
|
// even if it no longer has any open channels.
|
2018-12-04 13:20:38 +03:00
|
|
|
sourceNode, err := c.sourceNode(nodes)
|
2018-06-28 06:05:45 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-07-23 04:59:28 +03:00
|
|
|
// We'll use this map to keep count the number of references to a node
|
|
|
|
// in the graph. A node should only be removed once it has no more
|
|
|
|
// references in the graph.
|
|
|
|
nodeRefCounts := make(map[[33]byte]int)
|
|
|
|
err = nodes.ForEach(func(pubKey, nodeBytes []byte) error {
|
|
|
|
// If this is the source key, then we skip this
|
|
|
|
// iteration as the value for this key is a pubKey
|
|
|
|
// rather than raw node information.
|
|
|
|
if bytes.Equal(pubKey, sourceKey) || len(pubKey) != 33 {
|
|
|
|
return nil
|
2018-06-28 06:05:45 +03:00
|
|
|
}
|
|
|
|
|
2018-07-23 04:59:28 +03:00
|
|
|
var nodePub [33]byte
|
|
|
|
copy(nodePub[:], pubKey)
|
|
|
|
nodeRefCounts[nodePub] = 0
|
2018-06-28 06:05:45 +03:00
|
|
|
|
2018-07-23 04:59:28 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-06-28 06:05:45 +03:00
|
|
|
|
2018-07-23 04:59:28 +03:00
|
|
|
// To ensure we never delete the source node, we'll start off by
|
|
|
|
// bumping its ref count to 1.
|
|
|
|
nodeRefCounts[sourceNode.PubKeyBytes] = 1
|
|
|
|
|
|
|
|
// Next, we'll run through the edgeIndex which maps a channel ID to the
|
|
|
|
// edge info. We'll use this scan to populate our reference count map
|
|
|
|
// above.
|
|
|
|
err = edgeIndex.ForEach(func(chanID, edgeInfoBytes []byte) error {
|
|
|
|
// The first 66 bytes of the edge info contain the pubkeys of
|
|
|
|
// the nodes that this edge attaches. We'll extract them, and
|
|
|
|
// add them to the ref count map.
|
|
|
|
var node1, node2 [33]byte
|
|
|
|
copy(node1[:], edgeInfoBytes[:33])
|
|
|
|
copy(node2[:], edgeInfoBytes[33:])
|
|
|
|
|
|
|
|
// With the nodes extracted, we'll increase the ref count of
|
|
|
|
// each of the nodes.
|
2018-07-31 11:29:12 +03:00
|
|
|
nodeRefCounts[node1]++
|
|
|
|
nodeRefCounts[node2]++
|
2018-07-22 05:49:08 +03:00
|
|
|
|
2018-07-23 04:59:28 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, we'll make a second pass over the set of nodes, and delete
|
|
|
|
// any nodes that have a ref count of zero.
|
2018-07-23 07:00:43 +03:00
|
|
|
var numNodesPruned int
|
2018-07-23 04:59:28 +03:00
|
|
|
for nodePubKey, refCount := range nodeRefCounts {
|
|
|
|
// If the ref count of the node isn't zero, then we can safely
|
|
|
|
// skip it as it still has edges to or from it within the
|
|
|
|
// graph.
|
|
|
|
if refCount != 0 {
|
2018-06-28 06:05:45 +03:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2018-07-23 04:59:28 +03:00
|
|
|
// If we reach this point, then there are no longer any edges
|
|
|
|
// that connect this node, so we can delete it.
|
2018-12-04 13:25:05 +03:00
|
|
|
if err := c.deleteLightningNode(nodes, nodePubKey[:]); err != nil {
|
2018-07-23 04:59:28 +03:00
|
|
|
log.Warnf("Unable to prune node %x from the "+
|
|
|
|
"graph: %v", nodePubKey, err)
|
2018-07-23 07:00:43 +03:00
|
|
|
continue
|
2018-06-28 06:05:45 +03:00
|
|
|
}
|
2018-07-23 07:00:43 +03:00
|
|
|
|
|
|
|
log.Infof("Pruned unconnected node %x from channel graph",
|
|
|
|
nodePubKey[:])
|
|
|
|
|
|
|
|
numNodesPruned++
|
|
|
|
}
|
|
|
|
|
|
|
|
if numNodesPruned > 0 {
|
|
|
|
log.Infof("Pruned %v unconnected nodes from the channel graph",
|
|
|
|
numNodesPruned)
|
2018-06-28 06:05:45 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-10-02 17:38:45 +03:00
|
|
|
// DisconnectBlockAtHeight is used to indicate that the block specified
|
|
|
|
// by the passed height has been disconnected from the main chain. This
|
|
|
|
// will "rewind" the graph back to the height below, deleting channels
|
|
|
|
// that are no longer confirmed from the graph. The prune log will be
|
|
|
|
// set to the last prune height valid for the remaining chain.
|
|
|
|
// Channels that were removed from the graph resulting from the
|
|
|
|
// disconnected block are returned.
|
|
|
|
func (c *ChannelGraph) DisconnectBlockAtHeight(height uint32) ([]*ChannelEdgeInfo,
|
|
|
|
error) {
|
|
|
|
|
|
|
|
// Every channel having a ShortChannelID starting at 'height'
|
|
|
|
// will no longer be confirmed.
|
|
|
|
startShortChanID := lnwire.ShortChannelID{
|
|
|
|
BlockHeight: height,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete everything after this height from the db.
|
|
|
|
endShortChanID := lnwire.ShortChannelID{
|
|
|
|
BlockHeight: math.MaxUint32 & 0x00ffffff,
|
|
|
|
TxIndex: math.MaxUint32 & 0x00ffffff,
|
|
|
|
TxPosition: math.MaxUint16,
|
|
|
|
}
|
|
|
|
// The block height will be the 3 first bytes of the channel IDs.
|
|
|
|
var chanIDStart [8]byte
|
|
|
|
byteOrder.PutUint64(chanIDStart[:], startShortChanID.ToUint64())
|
|
|
|
var chanIDEnd [8]byte
|
|
|
|
byteOrder.PutUint64(chanIDEnd[:], endShortChanID.ToUint64())
|
|
|
|
|
|
|
|
// Keep track of the channels that are removed from the graph.
|
|
|
|
var removedChans []*ChannelEdgeInfo
|
|
|
|
|
2019-03-13 23:13:36 +03:00
|
|
|
if err := c.db.Batch(func(tx *bbolt.Tx) error {
|
|
|
|
removedChans = nil
|
|
|
|
|
2017-10-02 17:38:45 +03:00
|
|
|
edges, err := tx.CreateBucketIfNotExists(edgeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
edgeIndex, err := edges.CreateBucketIfNotExists(edgeIndexBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
chanIndex, err := edges.CreateBucketIfNotExists(channelPointBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-03-27 23:06:34 +03:00
|
|
|
zombieIndex, err := edges.CreateBucketIfNotExists(zombieBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-04-28 02:21:46 +03:00
|
|
|
nodes, err := tx.CreateBucketIfNotExists(nodeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-10-02 17:38:45 +03:00
|
|
|
|
|
|
|
// Scan from chanIDStart to chanIDEnd, deleting every
|
|
|
|
// found edge.
|
|
|
|
cursor := edgeIndex.Cursor()
|
|
|
|
for k, v := cursor.Seek(chanIDStart[:]); k != nil &&
|
|
|
|
bytes.Compare(k, chanIDEnd[:]) <= 0; k, v = cursor.Next() {
|
|
|
|
|
|
|
|
edgeInfoReader := bytes.NewReader(v)
|
|
|
|
edgeInfo, err := deserializeChanEdgeInfo(edgeInfoReader)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-04-17 05:04:29 +03:00
|
|
|
err = delChannelByEdge(
|
2019-03-27 23:06:34 +03:00
|
|
|
edges, edgeIndex, chanIndex, zombieIndex, nodes,
|
|
|
|
&edgeInfo.ChannelPoint, false,
|
2018-04-17 05:04:29 +03:00
|
|
|
)
|
2017-10-02 17:38:45 +03:00
|
|
|
if err != nil && err != ErrEdgeNotFound {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
removedChans = append(removedChans, &edgeInfo)
|
2017-10-02 17:38:45 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Delete all the entries in the prune log having a height
|
|
|
|
// greater or equal to the block disconnected.
|
|
|
|
metaBucket, err := tx.CreateBucketIfNotExists(graphMetaBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
pruneBucket, err := metaBucket.CreateBucketIfNotExists(pruneLogBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
var pruneKeyStart [4]byte
|
|
|
|
byteOrder.PutUint32(pruneKeyStart[:], height)
|
|
|
|
|
|
|
|
var pruneKeyEnd [4]byte
|
|
|
|
byteOrder.PutUint32(pruneKeyEnd[:], math.MaxUint32)
|
|
|
|
|
|
|
|
pruneCursor := pruneBucket.Cursor()
|
|
|
|
for k, _ := pruneCursor.Seek(pruneKeyStart[:]); k != nil &&
|
|
|
|
bytes.Compare(k, pruneKeyEnd[:]) <= 0; k, _ = pruneCursor.Next() {
|
|
|
|
if err := pruneCursor.Delete(); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return removedChans, nil
|
|
|
|
}
|
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
// PruneTip returns the block height and hash of the latest block that has been
|
|
|
|
// used to prune channels in the graph. Knowing the "prune tip" allows callers
|
|
|
|
// to tell if the graph is currently in sync with the current best known UTXO
|
|
|
|
// state.
|
2017-01-06 00:56:27 +03:00
|
|
|
func (c *ChannelGraph) PruneTip() (*chainhash.Hash, uint32, error) {
|
2016-12-20 03:58:27 +03:00
|
|
|
var (
|
2017-10-02 17:38:45 +03:00
|
|
|
tipHash chainhash.Hash
|
|
|
|
tipHeight uint32
|
2016-12-20 03:58:27 +03:00
|
|
|
)
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-20 03:58:27 +03:00
|
|
|
graphMeta := tx.Bucket(graphMetaBucket)
|
|
|
|
if graphMeta == nil {
|
|
|
|
return ErrGraphNotFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
2017-10-02 17:38:45 +03:00
|
|
|
pruneBucket := graphMeta.Bucket(pruneLogBucket)
|
|
|
|
if pruneBucket == nil {
|
|
|
|
return ErrGraphNeverPruned
|
|
|
|
}
|
2016-12-20 03:58:27 +03:00
|
|
|
|
2017-10-02 17:38:45 +03:00
|
|
|
pruneCursor := pruneBucket.Cursor()
|
|
|
|
|
|
|
|
// The prune key with the largest block height will be our
|
|
|
|
// prune tip.
|
|
|
|
k, v := pruneCursor.Last()
|
|
|
|
if k == nil {
|
2016-12-20 03:58:27 +03:00
|
|
|
return ErrGraphNeverPruned
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
2017-10-02 17:38:45 +03:00
|
|
|
|
|
|
|
// Once we have the prune tip, the value will be the block hash,
|
|
|
|
// and the key the block height.
|
|
|
|
copy(tipHash[:], v[:])
|
|
|
|
tipHeight = byteOrder.Uint32(k[:])
|
2016-12-20 03:58:27 +03:00
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, 0, err
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
return &tipHash, tipHeight, nil
|
|
|
|
}
|
|
|
|
|
2018-02-07 06:13:07 +03:00
|
|
|
// DeleteChannelEdge removes an edge from the database as identified by its
|
2019-03-27 23:06:34 +03:00
|
|
|
// funding outpoint and also marks it as a zombie. This ensures that we're
|
|
|
|
// unable to re-add this to our database once again. If the edge does not exist
|
|
|
|
// within the database, then ErrEdgeNotFound will be returned.
|
2016-12-22 23:49:30 +03:00
|
|
|
func (c *ChannelGraph) DeleteChannelEdge(chanPoint *wire.OutPoint) error {
|
2016-12-20 03:58:27 +03:00
|
|
|
// TODO(roasbeef): possibly delete from node bucket if node has no more
|
|
|
|
// channels
|
|
|
|
// TODO(roasbeef): don't delete both edges?
|
|
|
|
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2016-12-20 03:58:27 +03:00
|
|
|
// First grab the edges bucket which houses the information
|
|
|
|
// we'd like to delete
|
2018-12-04 13:20:38 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrEdgeNotFound
|
2016-12-20 03:58:27 +03:00
|
|
|
}
|
2018-12-04 13:20:38 +03:00
|
|
|
|
|
|
|
// Next grab the two edge indexes which will also need to be
|
|
|
|
// updated.
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrEdgeNotFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
2018-12-04 13:20:38 +03:00
|
|
|
chanIndex := edges.Bucket(channelPointBucket)
|
|
|
|
if chanIndex == nil {
|
|
|
|
return ErrEdgeNotFound
|
2016-12-20 03:58:27 +03:00
|
|
|
}
|
2018-12-04 13:20:38 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodeNotFound
|
2018-04-28 02:21:46 +03:00
|
|
|
}
|
2019-03-27 23:06:34 +03:00
|
|
|
zombieIndex, err := edges.CreateBucketIfNotExists(zombieBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2016-12-20 03:58:27 +03:00
|
|
|
|
2018-12-04 13:20:38 +03:00
|
|
|
return delChannelByEdge(
|
2019-03-27 23:06:34 +03:00
|
|
|
edges, edgeIndex, chanIndex, zombieIndex, nodes,
|
|
|
|
chanPoint, true,
|
2018-12-04 13:20:38 +03:00
|
|
|
)
|
2016-12-08 09:47:01 +03:00
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2016-12-27 06:52:03 +03:00
|
|
|
// ChannelID attempt to lookup the 8-byte compact channel ID which maps to the
|
|
|
|
// passed channel point (outpoint). If the passed channel doesn't exist within
|
|
|
|
// the database, then ErrEdgeNotFound is returned.
|
|
|
|
func (c *ChannelGraph) ChannelID(chanPoint *wire.OutPoint) (uint64, error) {
|
|
|
|
var chanID uint64
|
|
|
|
|
|
|
|
var b bytes.Buffer
|
2017-07-26 06:39:59 +03:00
|
|
|
if err := writeOutpoint(&b, chanPoint); err != nil {
|
2016-12-27 06:52:03 +03:00
|
|
|
return 0, nil
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
if err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-27 06:52:03 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
chanIndex := edges.Bucket(channelPointBucket)
|
2017-09-19 06:40:29 +03:00
|
|
|
if chanIndex == nil {
|
2016-12-27 06:52:03 +03:00
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
|
|
|
|
chanIDBytes := chanIndex.Get(b.Bytes())
|
|
|
|
if chanIDBytes == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
chanID = byteOrder.Uint64(chanIDBytes)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return 0, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return chanID, nil
|
|
|
|
}
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// TODO(roasbeef): allow updates to use Batch?
|
|
|
|
|
|
|
|
// HighestChanID returns the "highest" known channel ID in the channel graph.
|
|
|
|
// This represents the "newest" channel from the PoV of the chain. This method
|
|
|
|
// can be used by peers to quickly determine if they're graphs are in sync.
|
|
|
|
func (c *ChannelGraph) HighestChanID() (uint64, error) {
|
|
|
|
var cid uint64
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-04-17 05:04:29 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// In order to find the highest chan ID, we'll fetch a cursor
|
|
|
|
// and use that to seek to the "end" of our known rage.
|
|
|
|
cidCursor := edgeIndex.Cursor()
|
|
|
|
|
|
|
|
lastChanID, _ := cidCursor.Last()
|
|
|
|
|
|
|
|
// If there's no key, then this means that we don't actually
|
|
|
|
// know of any channels, so we'll return a predicable error.
|
|
|
|
if lastChanID == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we'll de serialize the channel ID and return it
|
|
|
|
// to the caller.
|
|
|
|
cid = byteOrder.Uint64(lastChanID)
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil && err != ErrGraphNoEdgesFound {
|
|
|
|
return 0, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return cid, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// ChannelEdge represents the complete set of information for a channel edge in
|
|
|
|
// the known channel graph. This struct couples the core information of the
|
|
|
|
// edge as well as each of the known advertised edge policies.
|
|
|
|
type ChannelEdge struct {
|
|
|
|
// Info contains all the static information describing the channel.
|
|
|
|
Info *ChannelEdgeInfo
|
|
|
|
|
|
|
|
// Policy1 points to the "first" edge policy of the channel containing
|
|
|
|
// the dynamic information required to properly route through the edge.
|
|
|
|
Policy1 *ChannelEdgePolicy
|
|
|
|
|
|
|
|
// Policy2 points to the "second" edge policy of the channel containing
|
|
|
|
// the dynamic information required to properly route through the edge.
|
|
|
|
Policy2 *ChannelEdgePolicy
|
|
|
|
}
|
|
|
|
|
|
|
|
// ChanUpdatesInHorizon returns all the known channel edges which have at least
|
|
|
|
// one edge that has an update timestamp within the specified horizon.
|
|
|
|
func (c *ChannelGraph) ChanUpdatesInHorizon(startTime, endTime time.Time) ([]ChannelEdge, error) {
|
2018-09-05 04:39:23 +03:00
|
|
|
// To ensure we don't return duplicate ChannelEdges, we'll use an
|
|
|
|
// additional map to keep track of the edges already seen to prevent
|
|
|
|
// re-adding it.
|
|
|
|
edgesSeen := make(map[uint64]struct{})
|
2018-04-17 05:04:29 +03:00
|
|
|
var edgesInHorizon []ChannelEdge
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-04-17 05:04:29 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
edgeUpdateIndex := edges.Bucket(edgeUpdateIndexBucket)
|
|
|
|
if edgeUpdateIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll now obtain a cursor to perform a range query within
|
|
|
|
// the index to find all channels within the horizon.
|
|
|
|
updateCursor := edgeUpdateIndex.Cursor()
|
|
|
|
|
|
|
|
var startTimeBytes, endTimeBytes [8 + 8]byte
|
|
|
|
byteOrder.PutUint64(
|
|
|
|
startTimeBytes[:8], uint64(startTime.Unix()),
|
|
|
|
)
|
|
|
|
byteOrder.PutUint64(
|
|
|
|
endTimeBytes[:8], uint64(endTime.Unix()),
|
|
|
|
)
|
|
|
|
|
|
|
|
// With our start and end times constructed, we'll step through
|
|
|
|
// the index collecting the info and policy of each update of
|
|
|
|
// each channel that has a last update within the time range.
|
|
|
|
for indexKey, _ := updateCursor.Seek(startTimeBytes[:]); indexKey != nil &&
|
|
|
|
bytes.Compare(indexKey, endTimeBytes[:]) <= 0; indexKey, _ = updateCursor.Next() {
|
|
|
|
|
|
|
|
// We have a new eligible entry, so we'll slice of the
|
|
|
|
// chan ID so we can query it in the DB.
|
|
|
|
chanID := indexKey[8:]
|
|
|
|
|
2018-09-05 04:39:23 +03:00
|
|
|
// If we've already retrieved the info and policies for
|
|
|
|
// this edge, then we can skip it as we don't need to do
|
|
|
|
// so again.
|
|
|
|
chanIDInt := byteOrder.Uint64(chanID)
|
|
|
|
if _, ok := edgesSeen[chanIDInt]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// First, we'll fetch the static edge information.
|
|
|
|
edgeInfo, err := fetchChanEdgeInfo(edgeIndex, chanID)
|
|
|
|
if err != nil {
|
2018-09-01 00:59:32 +03:00
|
|
|
chanID := byteOrder.Uint64(chanID)
|
|
|
|
return fmt.Errorf("unable to fetch info for "+
|
|
|
|
"edge with chan_id=%v: %v", chanID, err)
|
2018-04-17 05:04:29 +03:00
|
|
|
}
|
2018-08-10 06:45:33 +03:00
|
|
|
edgeInfo.db = c.db
|
2018-04-17 05:04:29 +03:00
|
|
|
|
|
|
|
// With the static information obtained, we'll now
|
|
|
|
// fetch the dynamic policy info.
|
|
|
|
edge1, edge2, err := fetchChanEdgePolicies(
|
|
|
|
edgeIndex, edges, nodes, chanID, c.db,
|
|
|
|
)
|
|
|
|
if err != nil {
|
2018-09-01 00:59:32 +03:00
|
|
|
chanID := byteOrder.Uint64(chanID)
|
|
|
|
return fmt.Errorf("unable to fetch policies "+
|
|
|
|
"for edge with chan_id=%v: %v", chanID,
|
|
|
|
err)
|
2018-04-17 05:04:29 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, we'll collate this edge with the rest of
|
|
|
|
// edges to be returned.
|
2018-09-05 04:39:23 +03:00
|
|
|
edgesSeen[chanIDInt] = struct{}{}
|
2018-04-17 05:04:29 +03:00
|
|
|
edgesInHorizon = append(edgesInHorizon, ChannelEdge{
|
|
|
|
Info: &edgeInfo,
|
|
|
|
Policy1: edge1,
|
|
|
|
Policy2: edge2,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
switch {
|
|
|
|
case err == ErrGraphNoEdgesFound:
|
|
|
|
fallthrough
|
|
|
|
case err == ErrGraphNodesNotFound:
|
|
|
|
break
|
|
|
|
|
|
|
|
case err != nil:
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return edgesInHorizon, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// NodeUpdatesInHorizon returns all the known lightning node which have an
|
|
|
|
// update timestamp within the passed range. This method can be used by two
|
|
|
|
// nodes to quickly determine if they have the same set of up to date node
|
|
|
|
// announcements.
|
|
|
|
func (c *ChannelGraph) NodeUpdatesInHorizon(startTime, endTime time.Time) ([]LightningNode, error) {
|
|
|
|
var nodesInHorizon []LightningNode
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-04-17 05:04:29 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeUpdateIndex := nodes.Bucket(nodeUpdateIndexBucket)
|
|
|
|
if nodeUpdateIndex == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll now obtain a cursor to perform a range query within
|
|
|
|
// the index to find all node announcements within the horizon.
|
|
|
|
updateCursor := nodeUpdateIndex.Cursor()
|
|
|
|
|
|
|
|
var startTimeBytes, endTimeBytes [8 + 33]byte
|
|
|
|
byteOrder.PutUint64(
|
|
|
|
startTimeBytes[:8], uint64(startTime.Unix()),
|
|
|
|
)
|
|
|
|
byteOrder.PutUint64(
|
|
|
|
endTimeBytes[:8], uint64(endTime.Unix()),
|
|
|
|
)
|
|
|
|
|
|
|
|
// With our start and end times constructed, we'll step through
|
|
|
|
// the index collecting info for each node within the time
|
|
|
|
// range.
|
|
|
|
for indexKey, _ := updateCursor.Seek(startTimeBytes[:]); indexKey != nil &&
|
|
|
|
bytes.Compare(indexKey, endTimeBytes[:]) <= 0; indexKey, _ = updateCursor.Next() {
|
|
|
|
|
|
|
|
nodePub := indexKey[8:]
|
|
|
|
node, err := fetchLightningNode(nodes, nodePub)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
node.db = c.db
|
|
|
|
|
|
|
|
nodesInHorizon = append(nodesInHorizon, node)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
switch {
|
|
|
|
case err == ErrGraphNoEdgesFound:
|
|
|
|
fallthrough
|
|
|
|
case err == ErrGraphNodesNotFound:
|
|
|
|
break
|
|
|
|
|
|
|
|
case err != nil:
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return nodesInHorizon, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// FilterKnownChanIDs takes a set of channel IDs and return the subset of chan
|
|
|
|
// ID's that we don't know of in the passed set. In other words, we perform a
|
|
|
|
// set difference of our set of chan ID's and the ones passed in. This method
|
|
|
|
// can be used by callers to determine the set of channels ta peer knows of
|
|
|
|
// that we don't.
|
|
|
|
func (c *ChannelGraph) FilterKnownChanIDs(chanIDs []uint64) ([]uint64, error) {
|
|
|
|
var newChanIDs []uint64
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-04-17 05:04:29 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll run through the set of chanIDs and collate only the
|
|
|
|
// set of channel that are unable to be found within our db.
|
|
|
|
var cidBytes [8]byte
|
|
|
|
for _, cid := range chanIDs {
|
|
|
|
byteOrder.PutUint64(cidBytes[:], cid)
|
|
|
|
|
|
|
|
if v := edgeIndex.Get(cidBytes[:]); v == nil {
|
|
|
|
newChanIDs = append(newChanIDs, cid)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
switch {
|
|
|
|
// If we don't know of any edges yet, then we'll return the entire set
|
|
|
|
// of chan IDs specified.
|
|
|
|
case err == ErrGraphNoEdgesFound:
|
|
|
|
return chanIDs, nil
|
|
|
|
|
|
|
|
case err != nil:
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return newChanIDs, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// FilterChannelRange returns the channel ID's of all known channels which were
|
|
|
|
// mined in a block height within the passed range. This method can be used to
|
|
|
|
// quickly share with a peer the set of channels we know of within a particular
|
|
|
|
// range to catch them up after a period of time offline.
|
|
|
|
func (c *ChannelGraph) FilterChannelRange(startHeight, endHeight uint32) ([]uint64, error) {
|
|
|
|
var chanIDs []uint64
|
|
|
|
|
|
|
|
startChanID := &lnwire.ShortChannelID{
|
|
|
|
BlockHeight: startHeight,
|
|
|
|
}
|
|
|
|
|
|
|
|
endChanID := lnwire.ShortChannelID{
|
|
|
|
BlockHeight: endHeight,
|
|
|
|
TxIndex: math.MaxUint32 & 0x00ffffff,
|
|
|
|
TxPosition: math.MaxUint16,
|
|
|
|
}
|
|
|
|
|
|
|
|
// As we need to perform a range scan, we'll convert the starting and
|
|
|
|
// ending height to their corresponding values when encoded using short
|
|
|
|
// channel ID's.
|
|
|
|
var chanIDStart, chanIDEnd [8]byte
|
|
|
|
byteOrder.PutUint64(chanIDStart[:], startChanID.ToUint64())
|
|
|
|
byteOrder.PutUint64(chanIDEnd[:], endChanID.ToUint64())
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-04-17 05:04:29 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
|
|
|
|
cursor := edgeIndex.Cursor()
|
|
|
|
|
|
|
|
// We'll now iterate through the database, and find each
|
|
|
|
// channel ID that resides within the specified range.
|
|
|
|
var cid uint64
|
|
|
|
for k, _ := cursor.Seek(chanIDStart[:]); k != nil &&
|
|
|
|
bytes.Compare(k, chanIDEnd[:]) <= 0; k, _ = cursor.Next() {
|
|
|
|
|
|
|
|
// This channel ID rests within the target range, so
|
|
|
|
// we'll convert it into an integer and add it to our
|
|
|
|
// returned set.
|
|
|
|
cid = byteOrder.Uint64(k)
|
|
|
|
chanIDs = append(chanIDs, cid)
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
switch {
|
|
|
|
// If we don't know of any channels yet, then there's nothing to
|
|
|
|
// filter, so we'll return an empty slice.
|
|
|
|
case err == ErrGraphNoEdgesFound:
|
|
|
|
return chanIDs, nil
|
|
|
|
|
|
|
|
case err != nil:
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return chanIDs, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// FetchChanInfos returns the set of channel edges that correspond to the
|
|
|
|
// passed channel ID's. This can be used to respond to peer queries that are
|
|
|
|
// seeking to fill in gaps in their view of the channel graph.
|
|
|
|
func (c *ChannelGraph) FetchChanInfos(chanIDs []uint64) ([]ChannelEdge, error) {
|
|
|
|
// TODO(roasbeef): sort cids?
|
|
|
|
|
|
|
|
var (
|
|
|
|
chanEdges []ChannelEdge
|
|
|
|
cidBytes [8]byte
|
|
|
|
)
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-04-17 05:04:29 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, cid := range chanIDs {
|
|
|
|
byteOrder.PutUint64(cidBytes[:], cid)
|
|
|
|
|
|
|
|
// First, we'll fetch the static edge information.
|
|
|
|
edgeInfo, err := fetchChanEdgeInfo(
|
|
|
|
edgeIndex, cidBytes[:],
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-08-10 06:45:33 +03:00
|
|
|
edgeInfo.db = c.db
|
2018-04-17 05:04:29 +03:00
|
|
|
|
|
|
|
// With the static information obtained, we'll now
|
|
|
|
// fetch the dynamic policy info.
|
|
|
|
edge1, edge2, err := fetchChanEdgePolicies(
|
|
|
|
edgeIndex, edges, nodes, cidBytes[:], c.db,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
chanEdges = append(chanEdges, ChannelEdge{
|
|
|
|
Info: &edgeInfo,
|
|
|
|
Policy1: edge1,
|
|
|
|
Policy2: edge2,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return chanEdges, nil
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func delEdgeUpdateIndexEntry(edgesBucket *bbolt.Bucket, chanID uint64,
|
2018-04-28 02:21:46 +03:00
|
|
|
edge1, edge2 *ChannelEdgePolicy) error {
|
|
|
|
|
|
|
|
// First, we'll fetch the edge update index bucket which currently
|
|
|
|
// stores an entry for the channel we're about to delete.
|
2018-12-04 13:20:38 +03:00
|
|
|
updateIndex := edgesBucket.Bucket(edgeUpdateIndexBucket)
|
|
|
|
if updateIndex == nil {
|
|
|
|
// No edges in bucket, return early.
|
|
|
|
return nil
|
2018-04-28 02:21:46 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Now that we have the bucket, we'll attempt to construct a template
|
|
|
|
// for the index key: updateTime || chanid.
|
|
|
|
var indexKey [8 + 8]byte
|
|
|
|
byteOrder.PutUint64(indexKey[8:], chanID)
|
|
|
|
|
|
|
|
// With the template constructed, we'll attempt to delete an entry that
|
|
|
|
// would have been created by both edges: we'll alternate the update
|
|
|
|
// times, as one may had overridden the other.
|
|
|
|
if edge1 != nil {
|
|
|
|
byteOrder.PutUint64(indexKey[:8], uint64(edge1.LastUpdate.Unix()))
|
|
|
|
if err := updateIndex.Delete(indexKey[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We'll also attempt to delete the entry that may have been created by
|
|
|
|
// the second edge.
|
|
|
|
if edge2 != nil {
|
|
|
|
byteOrder.PutUint64(indexKey[:8], uint64(edge2.LastUpdate.Unix()))
|
|
|
|
if err := updateIndex.Delete(indexKey[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-03-27 23:06:34 +03:00
|
|
|
func delChannelByEdge(edges, edgeIndex, chanIndex, zombieIndex,
|
|
|
|
nodes *bbolt.Bucket, chanPoint *wire.OutPoint, isZombie bool) error {
|
|
|
|
|
2016-12-20 03:58:27 +03:00
|
|
|
var b bytes.Buffer
|
2017-07-26 06:39:59 +03:00
|
|
|
if err := writeOutpoint(&b, chanPoint); err != nil {
|
2016-12-20 03:58:27 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-28 02:21:46 +03:00
|
|
|
// If the channel's outpoint doesn't exist within the outpoint index,
|
|
|
|
// then the edge does not exist.
|
2016-12-20 03:58:27 +03:00
|
|
|
chanID := chanIndex.Get(b.Bytes())
|
|
|
|
if chanID == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
2018-04-28 02:21:46 +03:00
|
|
|
// Otherwise we obtain the two public keys from the mapping: chanID ->
|
|
|
|
// pubKey1 || pubKey2. With this, we can construct the keys which house
|
|
|
|
// both of the directed edges for this channel.
|
2016-12-20 03:58:27 +03:00
|
|
|
nodeKeys := edgeIndex.Get(chanID)
|
2017-10-02 17:38:45 +03:00
|
|
|
if nodeKeys == nil {
|
|
|
|
return fmt.Errorf("could not find nodekeys for chanID %v",
|
|
|
|
chanID)
|
|
|
|
}
|
2016-12-20 03:58:27 +03:00
|
|
|
|
2018-06-06 03:30:50 +03:00
|
|
|
// We'll also remove the entry in the edge update index bucket before
|
|
|
|
// we delete the edges themselves so we can access their last update
|
|
|
|
// times.
|
|
|
|
cid := byteOrder.Uint64(chanID)
|
|
|
|
edge1, edge2, err := fetchChanEdgePolicies(
|
|
|
|
edgeIndex, edges, nodes, chanID, nil,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
err = delEdgeUpdateIndexEntry(edges, cid, edge1, edge2)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-28 02:21:46 +03:00
|
|
|
// The edge key is of the format pubKey || chanID. First we construct
|
|
|
|
// the latter half, populating the channel ID.
|
2016-12-20 03:58:27 +03:00
|
|
|
var edgeKey [33 + 8]byte
|
|
|
|
copy(edgeKey[33:], chanID)
|
|
|
|
|
2018-04-28 02:21:46 +03:00
|
|
|
// With the latter half constructed, copy over the first public key to
|
|
|
|
// delete the edge in this direction, then the second to delete the
|
|
|
|
// edge in the opposite direction.
|
2016-12-20 03:58:27 +03:00
|
|
|
copy(edgeKey[:33], nodeKeys[:33])
|
|
|
|
if edges.Get(edgeKey[:]) != nil {
|
|
|
|
if err := edges.Delete(edgeKey[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
copy(edgeKey[:33], nodeKeys[33:])
|
|
|
|
if edges.Get(edgeKey[:]) != nil {
|
|
|
|
if err := edges.Delete(edgeKey[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-03-27 23:06:34 +03:00
|
|
|
// With the edge data deleted, we can purge the information from the two
|
|
|
|
// edge indexes.
|
2016-12-20 03:58:27 +03:00
|
|
|
if err := edgeIndex.Delete(chanID); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-03-27 23:06:34 +03:00
|
|
|
if err := chanIndex.Delete(b.Bytes()); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, we'll mark the edge as a zombie within our index if it's
|
|
|
|
// being removed due to the channel becoming a zombie. We do this to
|
|
|
|
// ensure we don't store unnecessary data for spent channels.
|
|
|
|
if !isZombie {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
var pubKey1, pubKey2 [33]byte
|
|
|
|
copy(pubKey1[:], nodeKeys[:33])
|
|
|
|
copy(pubKey2[:], nodeKeys[33:])
|
|
|
|
|
|
|
|
return markEdgeZombie(
|
|
|
|
zombieIndex, byteOrder.Uint64(chanID), pubKey1, pubKey2,
|
|
|
|
)
|
2016-12-20 03:58:27 +03:00
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// UpdateEdgePolicy updates the edge routing policy for a single directed edge
|
2016-12-08 09:47:01 +03:00
|
|
|
// within the database for the referenced channel. The `flags` attribute within
|
2017-03-06 04:18:00 +03:00
|
|
|
// the ChannelEdgePolicy determines which of the directed edges are being
|
|
|
|
// updated. If the flag is 1, then the first node's information is being
|
|
|
|
// updated, otherwise it's the second node's information. The node ordering is
|
2017-12-18 05:40:05 +03:00
|
|
|
// determined by the lexicographical ordering of the identity public keys of
|
2017-03-06 04:18:00 +03:00
|
|
|
// the nodes on either side of the channel.
|
2017-03-09 07:44:32 +03:00
|
|
|
func (c *ChannelGraph) UpdateEdgePolicy(edge *ChannelEdgePolicy) error {
|
2019-03-13 23:13:36 +03:00
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
2018-12-10 06:30:26 +03:00
|
|
|
return updateEdgePolicy(tx, edge)
|
2018-09-01 00:59:33 +03:00
|
|
|
})
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-09-01 00:59:33 +03:00
|
|
|
// updateEdgePolicy attempts to update an edge's policy within the relevant
|
|
|
|
// buckets using an existing database transaction.
|
2018-12-10 06:30:26 +03:00
|
|
|
func updateEdgePolicy(tx *bbolt.Tx, edge *ChannelEdgePolicy) error {
|
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
|
|
|
|
}
|
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
nodes, err := tx.CreateBucketIfNotExists(nodeBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-09-01 00:59:33 +03:00
|
|
|
// Create the channelID key be converting the channel ID
|
|
|
|
// integer into a byte slice.
|
|
|
|
var chanID [8]byte
|
|
|
|
byteOrder.PutUint64(chanID[:], edge.ChannelID)
|
|
|
|
|
|
|
|
// With the channel ID, we then fetch the value storing the two
|
|
|
|
// nodes which connect this channel edge.
|
|
|
|
nodeInfo := edgeIndex.Get(chanID[:])
|
|
|
|
if nodeInfo == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// Depending on the flags value passed above, either the first
|
|
|
|
// or second edge policy is being updated.
|
|
|
|
var fromNode, toNode []byte
|
2019-01-12 20:59:43 +03:00
|
|
|
if edge.ChannelFlags&lnwire.ChanUpdateDirection == 0 {
|
2018-09-01 00:59:33 +03:00
|
|
|
fromNode = nodeInfo[:33]
|
2018-09-01 00:59:34 +03:00
|
|
|
toNode = nodeInfo[33:66]
|
2018-09-01 00:59:33 +03:00
|
|
|
} else {
|
2018-09-01 00:59:34 +03:00
|
|
|
fromNode = nodeInfo[33:66]
|
2018-09-01 00:59:33 +03:00
|
|
|
toNode = nodeInfo[:33]
|
|
|
|
}
|
|
|
|
|
|
|
|
// Finally, with the direction of the edge being updated
|
|
|
|
// identified, we update the on-disk edge representation.
|
2018-09-01 00:59:35 +03:00
|
|
|
return putChanEdgePolicy(edges, nodes, edge, fromNode, toNode)
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// LightningNode represents an individual vertex/node within the channel graph.
|
|
|
|
// A node is connected to other nodes by one or more channel edges emanating
|
|
|
|
// from it. As the graph is directed, a node will also have an incoming edge
|
|
|
|
// attached to it for each outgoing edge.
|
|
|
|
type LightningNode struct {
|
2018-01-31 07:19:40 +03:00
|
|
|
// PubKeyBytes is the raw bytes of the public key of the target node.
|
|
|
|
PubKeyBytes [33]byte
|
|
|
|
pubKey *btcec.PublicKey
|
2017-07-14 22:25:02 +03:00
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
// HaveNodeAnnouncement indicates whether we received a node
|
|
|
|
// announcement for this particular node. If true, the remaining fields
|
|
|
|
// will be set, if false only the PubKey is known for this node.
|
2017-07-14 22:25:02 +03:00
|
|
|
HaveNodeAnnouncement bool
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// LastUpdate is the last time the vertex information for this node has
|
|
|
|
// been updated.
|
|
|
|
LastUpdate time.Time
|
|
|
|
|
|
|
|
// Address is the TCP address this node is reachable over.
|
2017-02-17 12:29:23 +03:00
|
|
|
Addresses []net.Addr
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
// Color is the selected color for the node.
|
|
|
|
Color color.RGBA
|
|
|
|
|
|
|
|
// Alias is a nick-name for the node. The alias can be used to confirm
|
|
|
|
// a node's identity or to serve as a short ID for an address book.
|
|
|
|
Alias string
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
// AuthSigBytes is the raw signature under the advertised public key
|
|
|
|
// which serves to authenticate the attributes announced by this node.
|
|
|
|
AuthSigBytes []byte
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2017-03-20 12:24:55 +03:00
|
|
|
// Features is the list of protocol features supported by this node.
|
|
|
|
Features *lnwire.FeatureVector
|
|
|
|
|
2018-09-01 05:36:16 +03:00
|
|
|
// ExtraOpaqueData is the set of data that was appended to this
|
|
|
|
// message, some of which we may not actually know how to iterate or
|
|
|
|
// parse. By holding onto this data, we ensure that we're able to
|
|
|
|
// properly validate the set of signatures that cover these new fields,
|
|
|
|
// and ensure we're able to make upgrades to the network in a forwards
|
|
|
|
// compatible manner.
|
|
|
|
ExtraOpaqueData []byte
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
db *DB
|
|
|
|
|
|
|
|
// TODO(roasbeef): discovery will need storage to keep it's last IP
|
|
|
|
// address and re-announce if interface changes?
|
|
|
|
|
|
|
|
// TODO(roasbeef): add update method and fetch?
|
|
|
|
}
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
// PubKey is the node's long-term identity public key. This key will be used to
|
|
|
|
// authenticated any advertisements/updates sent by the node.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the pubkey if absolutely necessary.
|
2018-01-31 07:40:41 +03:00
|
|
|
func (l *LightningNode) PubKey() (*btcec.PublicKey, error) {
|
|
|
|
if l.pubKey != nil {
|
|
|
|
return l.pubKey, nil
|
2018-01-31 07:19:40 +03:00
|
|
|
}
|
|
|
|
|
2018-01-31 07:40:41 +03:00
|
|
|
key, err := btcec.ParsePubKey(l.PubKeyBytes[:], btcec.S256())
|
2018-01-31 07:19:40 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2018-01-31 07:40:41 +03:00
|
|
|
l.pubKey = key
|
|
|
|
l.pubKey.Curve = nil
|
2018-01-31 07:19:40 +03:00
|
|
|
|
|
|
|
return key, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// AuthSig is a signature under the advertised public key which serves to
|
|
|
|
// authenticate the attributes announced by this node.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the signature if absolutely necessary.
|
2018-01-31 07:40:41 +03:00
|
|
|
func (l *LightningNode) AuthSig() (*btcec.Signature, error) {
|
|
|
|
return btcec.ParseSignature(l.AuthSigBytes, btcec.S256())
|
2018-01-31 07:19:40 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// AddPubKey is a setter-link method that can be used to swap out the public
|
|
|
|
// key for a node.
|
2018-01-31 07:40:41 +03:00
|
|
|
func (l *LightningNode) AddPubKey(key *btcec.PublicKey) {
|
|
|
|
l.pubKey = key
|
|
|
|
copy(l.PubKeyBytes[:], key.SerializeCompressed())
|
2018-01-31 07:19:40 +03:00
|
|
|
}
|
|
|
|
|
2018-04-23 22:55:25 +03:00
|
|
|
// NodeAnnouncement retrieves the latest node announcement of the node.
|
|
|
|
func (l *LightningNode) NodeAnnouncement(signed bool) (*lnwire.NodeAnnouncement,
|
|
|
|
error) {
|
|
|
|
|
|
|
|
if !l.HaveNodeAnnouncement {
|
|
|
|
return nil, fmt.Errorf("node does not have node announcement")
|
|
|
|
}
|
|
|
|
|
|
|
|
alias, err := lnwire.NewNodeAlias(l.Alias)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeAnn := &lnwire.NodeAnnouncement{
|
|
|
|
Features: l.Features.RawFeatureVector,
|
|
|
|
NodeID: l.PubKeyBytes,
|
|
|
|
RGBColor: l.Color,
|
|
|
|
Alias: alias,
|
|
|
|
Addresses: l.Addresses,
|
|
|
|
Timestamp: uint32(l.LastUpdate.Unix()),
|
|
|
|
ExtraOpaqueData: l.ExtraOpaqueData,
|
|
|
|
}
|
|
|
|
|
|
|
|
if !signed {
|
|
|
|
return nodeAnn, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
sig, err := lnwire.NewSigFromRawSignature(l.AuthSigBytes)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeAnn.Signature = sig
|
|
|
|
|
|
|
|
return nodeAnn, nil
|
|
|
|
}
|
|
|
|
|
2018-10-18 01:35:54 +03:00
|
|
|
// isPublic determines whether the node is seen as public within the graph from
|
|
|
|
// the source node's point of view. An existing database transaction can also be
|
|
|
|
// specified.
|
2018-11-30 07:04:21 +03:00
|
|
|
func (l *LightningNode) isPublic(tx *bbolt.Tx, sourcePubKey []byte) (bool, error) {
|
2018-10-18 01:35:54 +03:00
|
|
|
// In order to determine whether this node is publicly advertised within
|
|
|
|
// the graph, we'll need to look at all of its edges and check whether
|
|
|
|
// they extend to any other node than the source node. errDone will be
|
|
|
|
// used to terminate the check early.
|
|
|
|
nodeIsPublic := false
|
|
|
|
errDone := errors.New("done")
|
2018-11-30 07:04:21 +03:00
|
|
|
err := l.ForEachChannel(tx, func(_ *bbolt.Tx, info *ChannelEdgeInfo,
|
2018-10-18 01:35:54 +03:00
|
|
|
_, _ *ChannelEdgePolicy) error {
|
|
|
|
|
|
|
|
// If this edge doesn't extend to the source node, we'll
|
|
|
|
// terminate our search as we can now conclude that the node is
|
|
|
|
// publicly advertised within the graph due to the local node
|
|
|
|
// knowing of the current edge.
|
|
|
|
if !bytes.Equal(info.NodeKey1Bytes[:], sourcePubKey) &&
|
|
|
|
!bytes.Equal(info.NodeKey2Bytes[:], sourcePubKey) {
|
|
|
|
|
|
|
|
nodeIsPublic = true
|
|
|
|
return errDone
|
|
|
|
}
|
|
|
|
|
|
|
|
// Since the edge _does_ extend to the source node, we'll also
|
|
|
|
// need to ensure that this is a public edge.
|
|
|
|
if info.AuthProof != nil {
|
|
|
|
nodeIsPublic = true
|
|
|
|
return errDone
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we'll continue our search.
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil && err != errDone {
|
|
|
|
return false, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return nodeIsPublic, nil
|
|
|
|
}
|
|
|
|
|
2016-12-22 23:49:30 +03:00
|
|
|
// FetchLightningNode attempts to look up a target node by its identity public
|
2017-03-06 04:18:00 +03:00
|
|
|
// key. If the node isn't found in the database, then ErrGraphNodeNotFound is
|
2016-12-22 23:49:30 +03:00
|
|
|
// returned.
|
2016-12-08 09:47:01 +03:00
|
|
|
func (c *ChannelGraph) FetchLightningNode(pub *btcec.PublicKey) (*LightningNode, error) {
|
2016-12-22 23:49:30 +03:00
|
|
|
var node *LightningNode
|
2016-12-08 09:47:01 +03:00
|
|
|
nodePub := pub.SerializeCompressed()
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-22 23:49:30 +03:00
|
|
|
// First grab the nodes bucket which stores the mapping from
|
2016-12-08 09:47:01 +03:00
|
|
|
// pubKey to node information.
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// If a key for this serialized public key isn't found, then
|
|
|
|
// the target node doesn't exist within the database.
|
|
|
|
nodeBytes := nodes.Get(nodePub)
|
|
|
|
if nodeBytes == nil {
|
|
|
|
return ErrGraphNodeNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the node is found, then we can de deserialize the node
|
|
|
|
// information to return to the user.
|
|
|
|
nodeReader := bytes.NewReader(nodeBytes)
|
|
|
|
n, err := deserializeLightningNode(nodeReader)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
n.db = c.db
|
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
node = &n
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return node, nil
|
|
|
|
}
|
|
|
|
|
2016-12-22 23:49:30 +03:00
|
|
|
// HasLightningNode determines if the graph has a vertex identified by the
|
|
|
|
// target node identity public key. If the node exists in the database, a
|
|
|
|
// timestamp of when the data for the node was lasted updated is returned along
|
|
|
|
// with a true boolean. Otherwise, an empty time.Time is returned with a false
|
|
|
|
// boolean.
|
2018-01-31 07:19:40 +03:00
|
|
|
func (c *ChannelGraph) HasLightningNode(nodePub [33]byte) (time.Time, bool, error) {
|
2016-12-22 23:49:30 +03:00
|
|
|
var (
|
|
|
|
updateTime time.Time
|
|
|
|
exists bool
|
|
|
|
)
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-22 23:49:30 +03:00
|
|
|
// First grab the nodes bucket which stores the mapping from
|
|
|
|
// pubKey to node information.
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// If a key for this serialized public key isn't found, we can
|
|
|
|
// exit early.
|
2018-01-31 07:19:40 +03:00
|
|
|
nodeBytes := nodes.Get(nodePub[:])
|
2016-12-22 23:49:30 +03:00
|
|
|
if nodeBytes == nil {
|
|
|
|
exists = false
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise we continue on to obtain the time stamp
|
|
|
|
// representing the last time the data for this node was
|
|
|
|
// updated.
|
|
|
|
nodeReader := bytes.NewReader(nodeBytes)
|
|
|
|
node, err := deserializeLightningNode(nodeReader)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
exists = true
|
|
|
|
updateTime = node.LastUpdate
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return time.Time{}, exists, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
return updateTime, exists, nil
|
|
|
|
}
|
|
|
|
|
2018-06-18 13:35:22 +03:00
|
|
|
// ForEachChannel iterates through all channels of this node, executing the
|
|
|
|
// passed callback with an edge info structure and the policies of each end
|
|
|
|
// of the channel. The first edge policy is the outgoing edge *to* the
|
|
|
|
// the connecting node, while the second is the incoming edge *from* the
|
|
|
|
// connecting node. If the callback returns an error, then the iteration is
|
|
|
|
// halted with the error propagated back up to the caller.
|
|
|
|
//
|
|
|
|
// Unknown policies are passed into the callback as nil values.
|
2017-04-14 23:14:02 +03:00
|
|
|
//
|
|
|
|
// If the caller wishes to re-use an existing boltdb transaction, then it
|
|
|
|
// should be passed as the first argument. Otherwise the first argument should
|
|
|
|
// be nil and a fresh transaction will be created to execute the graph
|
|
|
|
// traversal.
|
2018-11-30 07:04:21 +03:00
|
|
|
func (l *LightningNode) ForEachChannel(tx *bbolt.Tx,
|
|
|
|
cb func(*bbolt.Tx, *ChannelEdgeInfo, *ChannelEdgePolicy, *ChannelEdgePolicy) error) error {
|
2017-04-14 23:14:02 +03:00
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
nodePub := l.PubKeyBytes[:]
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
traversal := func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
2017-03-06 04:18:00 +03:00
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
// In order to reach all the edges for this node, we take
|
|
|
|
// advantage of the construction of the key-space within the
|
|
|
|
// edge bucket. The keys are stored in the form: pubKey ||
|
|
|
|
// chanID. Therefore, starting from a chanID of zero, we can
|
2016-12-22 23:49:30 +03:00
|
|
|
// scan forward in the bucket, grabbing all the edges for the
|
2016-12-08 09:47:01 +03:00
|
|
|
// node. Once the prefix no longer matches, then we know we're
|
|
|
|
// done.
|
|
|
|
var nodeStart [33 + 8]byte
|
|
|
|
copy(nodeStart[:], nodePub)
|
|
|
|
copy(nodeStart[33:], chanStart[:])
|
|
|
|
|
|
|
|
// Starting from the key pubKey || 0, we seek forward in the
|
|
|
|
// bucket until the retrieved key no longer has the public key
|
|
|
|
// as its prefix. This indicates that we've stepped over into
|
|
|
|
// another node's edges, so we can terminate our scan.
|
|
|
|
edgeCursor := edges.Cursor()
|
2018-06-18 13:35:22 +03:00
|
|
|
for nodeEdge, _ := edgeCursor.Seek(nodeStart[:]); bytes.HasPrefix(nodeEdge, nodePub); nodeEdge, _ = edgeCursor.Next() {
|
|
|
|
// If the prefix still matches, the channel id is
|
|
|
|
// returned in nodeEdge. Channel id is used to lookup
|
|
|
|
// the node at the other end of the channel and both
|
|
|
|
// edge policies.
|
2017-03-06 04:18:00 +03:00
|
|
|
chanID := nodeEdge[33:]
|
|
|
|
edgeInfo, err := fetchChanEdgeInfo(edgeIndex, chanID)
|
2016-12-08 09:47:01 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-08-10 06:45:33 +03:00
|
|
|
edgeInfo.db = l.db
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-06-18 13:35:22 +03:00
|
|
|
outgoingPolicy, err := fetchChanEdgePolicy(
|
|
|
|
edges, chanID, nodePub, nodes,
|
2017-08-22 09:03:51 +03:00
|
|
|
)
|
2018-08-10 06:45:59 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-06-18 13:35:22 +03:00
|
|
|
|
|
|
|
otherNode, err := edgeInfo.OtherNodeKeyBytes(nodePub)
|
|
|
|
if err != nil {
|
2017-08-22 09:03:51 +03:00
|
|
|
return err
|
|
|
|
}
|
2018-06-18 13:35:22 +03:00
|
|
|
|
|
|
|
incomingPolicy, err := fetchChanEdgePolicy(
|
|
|
|
edges, chanID, otherNode, nodes,
|
|
|
|
)
|
2018-08-10 06:45:59 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-08-22 09:03:51 +03:00
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// Finally, we execute the callback.
|
2018-06-18 13:35:22 +03:00
|
|
|
err = cb(tx, &edgeInfo, outgoingPolicy, incomingPolicy)
|
2017-08-22 09:03:51 +03:00
|
|
|
if err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
2016-12-15 05:57:46 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// If no transaction was provided, then we'll create a new transaction
|
|
|
|
// to execute the transaction within.
|
|
|
|
if tx == nil {
|
|
|
|
return l.db.View(traversal)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we re-use the existing transaction to execute the graph
|
|
|
|
// traversal.
|
|
|
|
return traversal(tx)
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// ChannelEdgeInfo represents a fully authenticated channel along with all its
|
|
|
|
// unique attributes. Once an authenticated channel announcement has been
|
2018-04-18 05:02:04 +03:00
|
|
|
// processed on the network, then an instance of ChannelEdgeInfo encapsulating
|
2017-03-06 04:18:00 +03:00
|
|
|
// the channels attributes is stored. The other portions relevant to routing
|
|
|
|
// policy of a channel are stored within a ChannelEdgePolicy for each direction
|
|
|
|
// of the channel.
|
|
|
|
type ChannelEdgeInfo struct {
|
2016-12-08 09:47:01 +03:00
|
|
|
// ChannelID is the unique channel ID for the channel. The first 3
|
|
|
|
// bytes are the block height, the next 3 the index within the block,
|
|
|
|
// and the last 2 bytes are the output index for the channel.
|
|
|
|
ChannelID uint64
|
|
|
|
|
2017-08-22 08:57:16 +03:00
|
|
|
// ChainHash is the hash that uniquely identifies the chain that this
|
|
|
|
// channel was opened within.
|
|
|
|
//
|
|
|
|
// TODO(roasbeef): need to modify db keying for multi-chain
|
|
|
|
// * must add chain hash to prefix as well
|
|
|
|
ChainHash chainhash.Hash
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
// NodeKey1Bytes is the raw public key of the first node.
|
|
|
|
NodeKey1Bytes [33]byte
|
|
|
|
nodeKey1 *btcec.PublicKey
|
|
|
|
|
|
|
|
// NodeKey2Bytes is the raw public key of the first node.
|
|
|
|
NodeKey2Bytes [33]byte
|
|
|
|
nodeKey2 *btcec.PublicKey
|
|
|
|
|
|
|
|
// BitcoinKey1Bytes is the raw public key of the first node.
|
|
|
|
BitcoinKey1Bytes [33]byte
|
|
|
|
bitcoinKey1 *btcec.PublicKey
|
|
|
|
|
|
|
|
// BitcoinKey2Bytes is the raw public key of the first node.
|
|
|
|
BitcoinKey2Bytes [33]byte
|
|
|
|
bitcoinKey2 *btcec.PublicKey
|
2017-03-06 04:18:00 +03:00
|
|
|
|
|
|
|
// Features is an opaque byte slice that encodes the set of channel
|
|
|
|
// specific features that this channel edge supports.
|
|
|
|
Features []byte
|
|
|
|
|
|
|
|
// AuthProof is the authentication proof for this channel. This proof
|
|
|
|
// contains a set of signatures binding four identities, which attests
|
|
|
|
// to the legitimacy of the advertised channel.
|
|
|
|
AuthProof *ChannelAuthProof
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// ChannelPoint is the funding outpoint of the channel. This can be
|
|
|
|
// used to uniquely identify the channel within the channel graph.
|
|
|
|
ChannelPoint wire.OutPoint
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// Capacity is the total capacity of the channel, this is determined by
|
|
|
|
// the value output in the outpoint that created this channel.
|
|
|
|
Capacity btcutil.Amount
|
2018-08-10 06:45:33 +03:00
|
|
|
|
2018-09-01 05:36:16 +03:00
|
|
|
// ExtraOpaqueData is the set of data that was appended to this
|
|
|
|
// message, some of which we may not actually know how to iterate or
|
|
|
|
// parse. By holding onto this data, we ensure that we're able to
|
|
|
|
// properly validate the set of signatures that cover these new fields,
|
|
|
|
// and ensure we're able to make upgrades to the network in a forwards
|
|
|
|
// compatible manner.
|
|
|
|
ExtraOpaqueData []byte
|
|
|
|
|
2018-08-10 06:45:33 +03:00
|
|
|
db *DB
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
// AddNodeKeys is a setter-like method that can be used to replace the set of
|
|
|
|
// keys for the target ChannelEdgeInfo.
|
|
|
|
func (c *ChannelEdgeInfo) AddNodeKeys(nodeKey1, nodeKey2, bitcoinKey1,
|
|
|
|
bitcoinKey2 *btcec.PublicKey) {
|
|
|
|
|
|
|
|
c.nodeKey1 = nodeKey1
|
|
|
|
copy(c.NodeKey1Bytes[:], c.nodeKey1.SerializeCompressed())
|
|
|
|
|
|
|
|
c.nodeKey2 = nodeKey2
|
|
|
|
copy(c.NodeKey2Bytes[:], nodeKey2.SerializeCompressed())
|
|
|
|
|
|
|
|
c.bitcoinKey1 = bitcoinKey1
|
|
|
|
copy(c.BitcoinKey1Bytes[:], c.bitcoinKey1.SerializeCompressed())
|
|
|
|
|
|
|
|
c.bitcoinKey2 = bitcoinKey2
|
|
|
|
copy(c.BitcoinKey2Bytes[:], bitcoinKey2.SerializeCompressed())
|
|
|
|
}
|
|
|
|
|
|
|
|
// NodeKey1 is the identity public key of the "first" node that was involved in
|
|
|
|
// the creation of this channel. A node is considered "first" if the
|
|
|
|
// lexicographical ordering the its serialized public key is "smaller" than
|
|
|
|
// that of the other node involved in channel creation.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the pubkey if absolutely necessary.
|
|
|
|
func (c *ChannelEdgeInfo) NodeKey1() (*btcec.PublicKey, error) {
|
|
|
|
if c.nodeKey1 != nil {
|
|
|
|
return c.nodeKey1, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
key, err := btcec.ParsePubKey(c.NodeKey1Bytes[:], btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
c.nodeKey1 = key
|
|
|
|
|
|
|
|
return key, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// NodeKey2 is the identity public key of the "second" node that was
|
|
|
|
// involved in the creation of this channel. A node is considered
|
|
|
|
// "second" if the lexicographical ordering the its serialized public
|
|
|
|
// key is "larger" than that of the other node involved in channel
|
|
|
|
// creation.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the pubkey if absolutely necessary.
|
|
|
|
func (c *ChannelEdgeInfo) NodeKey2() (*btcec.PublicKey, error) {
|
|
|
|
if c.nodeKey2 != nil {
|
|
|
|
return c.nodeKey2, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
key, err := btcec.ParsePubKey(c.NodeKey2Bytes[:], btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
c.nodeKey2 = key
|
|
|
|
|
|
|
|
return key, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// BitcoinKey1 is the Bitcoin multi-sig key belonging to the first
|
|
|
|
// node, that was involved in the funding transaction that originally
|
|
|
|
// created the channel that this struct represents.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the pubkey if absolutely necessary.
|
|
|
|
func (c *ChannelEdgeInfo) BitcoinKey1() (*btcec.PublicKey, error) {
|
|
|
|
if c.bitcoinKey1 != nil {
|
|
|
|
return c.bitcoinKey1, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
key, err := btcec.ParsePubKey(c.BitcoinKey1Bytes[:], btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
c.bitcoinKey1 = key
|
|
|
|
|
|
|
|
return key, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// BitcoinKey2 is the Bitcoin multi-sig key belonging to the second
|
|
|
|
// node, that was involved in the funding transaction that originally
|
|
|
|
// created the channel that this struct represents.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the pubkey if absolutely necessary.
|
|
|
|
func (c *ChannelEdgeInfo) BitcoinKey2() (*btcec.PublicKey, error) {
|
|
|
|
if c.bitcoinKey2 != nil {
|
|
|
|
return c.bitcoinKey2, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
key, err := btcec.ParsePubKey(c.BitcoinKey2Bytes[:], btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
c.bitcoinKey2 = key
|
|
|
|
|
|
|
|
return key, nil
|
|
|
|
}
|
|
|
|
|
2018-06-18 13:35:22 +03:00
|
|
|
// OtherNodeKeyBytes returns the node key bytes of the other end of
|
|
|
|
// the channel.
|
|
|
|
func (c *ChannelEdgeInfo) OtherNodeKeyBytes(thisNodeKey []byte) (
|
|
|
|
[]byte, error) {
|
|
|
|
|
|
|
|
switch {
|
|
|
|
case bytes.Equal(c.NodeKey1Bytes[:], thisNodeKey):
|
|
|
|
return c.NodeKey2Bytes[:], nil
|
|
|
|
case bytes.Equal(c.NodeKey2Bytes[:], thisNodeKey):
|
|
|
|
return c.NodeKey1Bytes[:], nil
|
|
|
|
default:
|
2018-08-10 06:45:33 +03:00
|
|
|
return nil, fmt.Errorf("node not participating in this channel")
|
2018-06-18 13:35:22 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-08-10 06:46:56 +03:00
|
|
|
// FetchOtherNode attempts to fetch the full LightningNode that's opposite of
|
|
|
|
// the target node in the channel. This is useful when one knows the pubkey of
|
|
|
|
// one of the nodes, and wishes to obtain the full LightningNode for the other
|
|
|
|
// end of the channel.
|
2018-11-30 07:04:21 +03:00
|
|
|
func (c *ChannelEdgeInfo) FetchOtherNode(tx *bbolt.Tx, thisNodeKey []byte) (*LightningNode, error) {
|
2018-08-10 06:46:56 +03:00
|
|
|
|
|
|
|
// Ensure that the node passed in is actually a member of the channel.
|
|
|
|
var targetNodeBytes [33]byte
|
|
|
|
switch {
|
|
|
|
case bytes.Equal(c.NodeKey1Bytes[:], thisNodeKey):
|
|
|
|
targetNodeBytes = c.NodeKey2Bytes
|
|
|
|
case bytes.Equal(c.NodeKey2Bytes[:], thisNodeKey):
|
|
|
|
targetNodeBytes = c.NodeKey1Bytes
|
|
|
|
default:
|
|
|
|
return nil, fmt.Errorf("node not participating in this channel")
|
|
|
|
}
|
|
|
|
|
|
|
|
var targetNode *LightningNode
|
2018-11-30 07:04:21 +03:00
|
|
|
fetchNodeFunc := func(tx *bbolt.Tx) error {
|
2018-08-10 06:46:56 +03:00
|
|
|
// First grab the nodes bucket which stores the mapping from
|
|
|
|
// pubKey to node information.
|
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
node, err := fetchLightningNode(nodes, targetNodeBytes[:])
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
node.db = c.db
|
|
|
|
|
|
|
|
targetNode = &node
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the transaction is nil, then we'll need to create a new one,
|
|
|
|
// otherwise we can use the existing db transaction.
|
|
|
|
var err error
|
|
|
|
if tx == nil {
|
|
|
|
err = c.db.View(fetchNodeFunc)
|
|
|
|
} else {
|
|
|
|
err = fetchNodeFunc(tx)
|
|
|
|
}
|
|
|
|
|
|
|
|
return targetNode, err
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// ChannelAuthProof is the authentication proof (the signature portion) for a
|
|
|
|
// channel. Using the four signatures contained in the struct, and some
|
2018-02-07 11:30:09 +03:00
|
|
|
// auxiliary knowledge (the funding script, node identities, and outpoint) nodes
|
2017-03-06 04:18:00 +03:00
|
|
|
// on the network are able to validate the authenticity and existence of a
|
|
|
|
// channel. Each of these signatures signs the following digest: chanID ||
|
|
|
|
// nodeID1 || nodeID2 || bitcoinKey1|| bitcoinKey2 || 2-byte-feature-len ||
|
|
|
|
// features.
|
|
|
|
type ChannelAuthProof struct {
|
2018-01-31 07:19:40 +03:00
|
|
|
// nodeSig1 is a cached instance of the first node signature.
|
|
|
|
nodeSig1 *btcec.Signature
|
|
|
|
|
|
|
|
// NodeSig1Bytes are the raw bytes of the first node signature encoded
|
|
|
|
// in DER format.
|
|
|
|
NodeSig1Bytes []byte
|
|
|
|
|
|
|
|
// nodeSig2 is a cached instance of the second node signature.
|
|
|
|
nodeSig2 *btcec.Signature
|
|
|
|
|
|
|
|
// NodeSig2Bytes are the raw bytes of the second node signature
|
|
|
|
// encoded in DER format.
|
|
|
|
NodeSig2Bytes []byte
|
|
|
|
|
|
|
|
// bitcoinSig1 is a cached instance of the first bitcoin signature.
|
|
|
|
bitcoinSig1 *btcec.Signature
|
|
|
|
|
|
|
|
// BitcoinSig1Bytes are the raw bytes of the first bitcoin signature
|
|
|
|
// encoded in DER format.
|
|
|
|
BitcoinSig1Bytes []byte
|
|
|
|
|
|
|
|
// bitcoinSig2 is a cached instance of the second bitcoin signature.
|
|
|
|
bitcoinSig2 *btcec.Signature
|
|
|
|
|
|
|
|
// BitcoinSig2Bytes are the raw bytes of the second bitcoin signature
|
|
|
|
// encoded in DER format.
|
|
|
|
BitcoinSig2Bytes []byte
|
|
|
|
}
|
|
|
|
|
2018-01-31 07:40:41 +03:00
|
|
|
// Node1Sig is the signature using the identity key of the node that is first
|
2018-01-31 07:19:40 +03:00
|
|
|
// in a lexicographical ordering of the serialized public keys of the two nodes
|
|
|
|
// that created the channel.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the signature if absolutely necessary.
|
|
|
|
func (c *ChannelAuthProof) Node1Sig() (*btcec.Signature, error) {
|
|
|
|
if c.nodeSig1 != nil {
|
|
|
|
return c.nodeSig1, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
sig, err := btcec.ParseSignature(c.NodeSig1Bytes, btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
c.nodeSig1 = sig
|
|
|
|
|
|
|
|
return sig, nil
|
|
|
|
}
|
|
|
|
|
2018-01-31 07:40:41 +03:00
|
|
|
// Node2Sig is the signature using the identity key of the node that is second
|
2018-01-31 07:19:40 +03:00
|
|
|
// in a lexicographical ordering of the serialized public keys of the two nodes
|
|
|
|
// that created the channel.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the signature if absolutely necessary.
|
|
|
|
func (c *ChannelAuthProof) Node2Sig() (*btcec.Signature, error) {
|
|
|
|
if c.nodeSig2 != nil {
|
|
|
|
return c.nodeSig2, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
sig, err := btcec.ParseSignature(c.NodeSig2Bytes, btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
c.nodeSig2 = sig
|
|
|
|
|
|
|
|
return sig, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// BitcoinSig1 is the signature using the public key of the first node that was
|
|
|
|
// used in the channel's multi-sig output.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the signature if absolutely necessary.
|
|
|
|
func (c *ChannelAuthProof) BitcoinSig1() (*btcec.Signature, error) {
|
|
|
|
if c.bitcoinSig1 != nil {
|
|
|
|
return c.bitcoinSig1, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
sig, err := btcec.ParseSignature(c.BitcoinSig1Bytes, btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
c.bitcoinSig1 = sig
|
|
|
|
|
|
|
|
return sig, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// BitcoinSig2 is the signature using the public key of the second node that
|
|
|
|
// was used in the channel's multi-sig output.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the signature if absolutely necessary.
|
|
|
|
func (c *ChannelAuthProof) BitcoinSig2() (*btcec.Signature, error) {
|
|
|
|
if c.bitcoinSig2 != nil {
|
|
|
|
return c.bitcoinSig2, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
sig, err := btcec.ParseSignature(c.BitcoinSig2Bytes, btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
c.bitcoinSig2 = sig
|
|
|
|
|
|
|
|
return sig, nil
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
2017-03-19 21:37:15 +03:00
|
|
|
// IsEmpty check is the authentication proof is empty Proof is empty if at
|
|
|
|
// least one of the signatures are equal to nil.
|
2018-01-31 07:40:41 +03:00
|
|
|
func (c *ChannelAuthProof) IsEmpty() bool {
|
|
|
|
return len(c.NodeSig1Bytes) == 0 ||
|
|
|
|
len(c.NodeSig2Bytes) == 0 ||
|
|
|
|
len(c.BitcoinSig1Bytes) == 0 ||
|
|
|
|
len(c.BitcoinSig2Bytes) == 0
|
2017-03-19 21:37:15 +03:00
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// ChannelEdgePolicy represents a *directed* edge within the channel graph. For
|
|
|
|
// each channel in the database, there are two distinct edges: one for each
|
|
|
|
// possible direction of travel along the channel. The edges themselves hold
|
|
|
|
// information concerning fees, and minimum time-lock information which is
|
|
|
|
// utilized during path finding.
|
|
|
|
type ChannelEdgePolicy struct {
|
2018-01-31 07:19:40 +03:00
|
|
|
// SigBytes is the raw bytes of the signature of the channel edge
|
|
|
|
// policy. We'll only parse these if the caller needs to access the
|
|
|
|
// signature for validation purposes.
|
|
|
|
SigBytes []byte
|
|
|
|
|
|
|
|
// sig is a cached fully parsed signature.
|
|
|
|
sig *btcec.Signature
|
2017-03-27 18:01:12 +03:00
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// ChannelID is the unique channel ID for the channel. The first 3
|
|
|
|
// bytes are the block height, the next 3 the index within the block,
|
|
|
|
// and the last 2 bytes are the output index for the channel.
|
|
|
|
ChannelID uint64
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// LastUpdate is the last time an authenticated edge for this channel
|
|
|
|
// was received.
|
|
|
|
LastUpdate time.Time
|
|
|
|
|
2019-01-12 20:59:43 +03:00
|
|
|
// MessageFlags is a bitfield which indicates the presence of optional
|
|
|
|
// fields (like max_htlc) in the policy.
|
|
|
|
MessageFlags lnwire.ChanUpdateMsgFlags
|
|
|
|
|
|
|
|
// ChannelFlags is a bitfield which signals the capabilities of the
|
|
|
|
// channel as well as the directed edge this update applies to.
|
|
|
|
ChannelFlags lnwire.ChanUpdateChanFlags
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// TimeLockDelta is the number of blocks this node will subtract from
|
|
|
|
// the expiry of an incoming HTLC. This value expresses the time buffer
|
|
|
|
// the node would like to HTLC exchanges.
|
|
|
|
TimeLockDelta uint16
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
// MinHTLC is the smallest value HTLC this node will accept, expressed
|
|
|
|
// in millisatoshi.
|
2017-08-22 08:54:42 +03:00
|
|
|
MinHTLC lnwire.MilliSatoshi
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2019-01-12 20:59:44 +03:00
|
|
|
// MaxHTLC is the largest value HTLC this node will accept, expressed
|
|
|
|
// in millisatoshi.
|
|
|
|
MaxHTLC lnwire.MilliSatoshi
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// FeeBaseMSat is the base HTLC fee that will be charged for forwarding
|
|
|
|
// ANY HTLC, expressed in mSAT's.
|
2017-08-22 08:54:42 +03:00
|
|
|
FeeBaseMSat lnwire.MilliSatoshi
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
// FeeProportionalMillionths is the rate that the node will charge for
|
2017-01-13 08:01:50 +03:00
|
|
|
// HTLCs for each millionth of a satoshi forwarded.
|
2017-08-22 08:54:42 +03:00
|
|
|
FeeProportionalMillionths lnwire.MilliSatoshi
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
// Node is the LightningNode that this directed edge leads to. Using
|
|
|
|
// this pointer the channel graph can further be traversed.
|
|
|
|
Node *LightningNode
|
|
|
|
|
2018-09-01 05:36:16 +03:00
|
|
|
// ExtraOpaqueData is the set of data that was appended to this
|
|
|
|
// message, some of which we may not actually know how to iterate or
|
|
|
|
// parse. By holding onto this data, we ensure that we're able to
|
|
|
|
// properly validate the set of signatures that cover these new fields,
|
|
|
|
// and ensure we're able to make upgrades to the network in a forwards
|
|
|
|
// compatible manner.
|
|
|
|
ExtraOpaqueData []byte
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
db *DB
|
|
|
|
}
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
// Signature is a channel announcement signature, which is needed for proper
|
|
|
|
// edge policy announcement.
|
|
|
|
//
|
|
|
|
// NOTE: By having this method to access an attribute, we ensure we only need
|
|
|
|
// to fully deserialize the signature if absolutely necessary.
|
|
|
|
func (c *ChannelEdgePolicy) Signature() (*btcec.Signature, error) {
|
|
|
|
if c.sig != nil {
|
|
|
|
return c.sig, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
sig, err := btcec.ParseSignature(c.SigBytes, btcec.S256())
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
c.sig = sig
|
|
|
|
|
|
|
|
return sig, nil
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
// FetchChannelEdgesByOutpoint attempts to lookup the two directed edges for
|
|
|
|
// the channel identified by the funding outpoint. If the channel can't be
|
2017-03-06 04:18:00 +03:00
|
|
|
// found, then ErrEdgeNotFound is returned. A struct which houses the general
|
|
|
|
// information for the channel itself is returned as well as two structs that
|
|
|
|
// contain the routing policies for the channel in either direction.
|
2019-03-27 23:06:57 +03:00
|
|
|
func (c *ChannelGraph) FetchChannelEdgesByOutpoint(op *wire.OutPoint,
|
|
|
|
) (*ChannelEdgeInfo, *ChannelEdgePolicy, *ChannelEdgePolicy, error) {
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
var (
|
2017-03-06 04:18:00 +03:00
|
|
|
edgeInfo *ChannelEdgeInfo
|
|
|
|
policy1 *ChannelEdgePolicy
|
|
|
|
policy2 *ChannelEdgePolicy
|
2016-12-08 09:47:01 +03:00
|
|
|
)
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// First, grab the node bucket. This will be used to populate
|
|
|
|
// the Node pointers in each edge read from disk.
|
2018-01-29 01:48:28 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Next, grab the edge bucket which stores the edges, and also
|
|
|
|
// the index itself so we can group the directed edges together
|
|
|
|
// logically.
|
2018-01-29 01:48:28 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
2018-01-29 01:48:28 +03:00
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// If the channel's outpoint doesn't exist within the outpoint
|
|
|
|
// index, then the edge does not exist.
|
2018-01-29 01:48:28 +03:00
|
|
|
chanIndex := edges.Bucket(channelPointBucket)
|
|
|
|
if chanIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
var b bytes.Buffer
|
2017-07-26 06:39:59 +03:00
|
|
|
if err := writeOutpoint(&b, op); err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
chanID := chanIndex.Get(b.Bytes())
|
|
|
|
if chanID == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// If the channel is found to exists, then we'll first retrieve
|
|
|
|
// the general information for the channel.
|
|
|
|
edge, err := fetchChanEdgeInfo(edgeIndex, chanID)
|
2016-12-08 09:47:01 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2018-02-13 03:17:08 +03:00
|
|
|
edgeInfo = &edge
|
2018-08-10 06:45:33 +03:00
|
|
|
edgeInfo.db = c.db
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// Once we have the information about the channels' parameters,
|
|
|
|
// we'll fetch the routing policies for each for the directed
|
|
|
|
// edges.
|
2018-07-23 07:02:46 +03:00
|
|
|
e1, e2, err := fetchChanEdgePolicies(
|
|
|
|
edgeIndex, edges, nodes, chanID, c.db,
|
|
|
|
)
|
2017-03-06 04:18:00 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
policy1 = e1
|
|
|
|
policy2 = e2
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return nil, nil, nil, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
return edgeInfo, policy1, policy2, nil
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// FetchChannelEdgesByID attempts to lookup the two directed edges for the
|
|
|
|
// channel identified by the channel ID. If the channel can't be found, then
|
2017-03-06 04:18:00 +03:00
|
|
|
// ErrEdgeNotFound is returned. A struct which houses the general information
|
|
|
|
// for the channel itself is returned as well as two structs that contain the
|
|
|
|
// routing policies for the channel in either direction.
|
2019-03-27 23:06:57 +03:00
|
|
|
//
|
|
|
|
// ErrZombieEdge an be returned if the edge is currently marked as a zombie
|
|
|
|
// within the database. In this case, the ChannelEdgePolicy's will be nil, and
|
|
|
|
// the ChannelEdgeInfo will only include the public keys of each node.
|
|
|
|
func (c *ChannelGraph) FetchChannelEdgesByID(chanID uint64,
|
|
|
|
) (*ChannelEdgeInfo, *ChannelEdgePolicy, *ChannelEdgePolicy, error) {
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
var (
|
2017-03-06 04:18:00 +03:00
|
|
|
edgeInfo *ChannelEdgeInfo
|
|
|
|
policy1 *ChannelEdgePolicy
|
|
|
|
policy2 *ChannelEdgePolicy
|
2016-12-08 09:47:01 +03:00
|
|
|
channelID [8]byte
|
|
|
|
)
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2016-12-08 09:47:01 +03:00
|
|
|
// First, grab the node bucket. This will be used to populate
|
|
|
|
// the Node pointers in each edge read from disk.
|
2016-12-27 07:02:40 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNotFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Next, grab the edge bucket which stores the edges, and also
|
|
|
|
// the index itself so we can group the directed edges together
|
|
|
|
// logically.
|
2016-12-27 07:02:40 +03:00
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
2016-12-27 07:02:40 +03:00
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
byteOrder.PutUint64(channelID[:], chanID)
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2019-03-27 23:06:57 +03:00
|
|
|
// Now, attempt to fetch edge.
|
2017-03-06 04:18:00 +03:00
|
|
|
edge, err := fetchChanEdgeInfo(edgeIndex, channelID[:])
|
2019-03-27 23:06:57 +03:00
|
|
|
|
|
|
|
// If it doesn't exist, we'll quickly check our zombie index to
|
|
|
|
// see if we've previously marked it as so.
|
|
|
|
if err == ErrEdgeNotFound {
|
|
|
|
// If the zombie index doesn't exist, or the edge is not
|
|
|
|
// marked as a zombie within it, then we'll return the
|
|
|
|
// original ErrEdgeNotFound error.
|
|
|
|
zombieIndex := edges.Bucket(zombieBucket)
|
|
|
|
if zombieIndex == nil {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
isZombie, pubKey1, pubKey2 := isZombieEdge(
|
|
|
|
zombieIndex, chanID,
|
|
|
|
)
|
|
|
|
if !isZombie {
|
|
|
|
return ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, the edge is marked as a zombie, so we'll
|
|
|
|
// populate the edge info with the public keys of each
|
|
|
|
// party as this is the only information we have about
|
|
|
|
// it and return an error signaling so.
|
|
|
|
edgeInfo = &ChannelEdgeInfo{
|
|
|
|
NodeKey1Bytes: pubKey1,
|
|
|
|
NodeKey2Bytes: pubKey2,
|
|
|
|
}
|
|
|
|
return ErrZombieEdge
|
|
|
|
}
|
|
|
|
|
|
|
|
// Otherwise, we'll just return the error if any.
|
2017-03-06 04:18:00 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-03-27 23:06:57 +03:00
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
edgeInfo = &edge
|
2018-08-10 06:45:33 +03:00
|
|
|
edgeInfo.db = c.db
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2019-03-27 23:06:57 +03:00
|
|
|
// Then we'll attempt to fetch the accompanying policies of this
|
|
|
|
// edge.
|
2018-07-23 07:02:46 +03:00
|
|
|
e1, e2, err := fetchChanEdgePolicies(
|
|
|
|
edgeIndex, edges, nodes, channelID[:], c.db,
|
|
|
|
)
|
2016-12-08 09:47:01 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
policy1 = e1
|
|
|
|
policy2 = e2
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil
|
|
|
|
})
|
2019-03-27 23:06:57 +03:00
|
|
|
if err == ErrZombieEdge {
|
|
|
|
return edgeInfo, nil, nil, err
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
if err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return nil, nil, nil, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
return edgeInfo, policy1, policy2, nil
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2018-10-18 01:35:54 +03:00
|
|
|
// IsPublicNode is a helper method that determines whether the node with the
|
|
|
|
// given public key is seen as a public node in the graph from the graph's
|
|
|
|
// source node's point of view.
|
|
|
|
func (c *ChannelGraph) IsPublicNode(pubKey [33]byte) (bool, error) {
|
|
|
|
var nodeIsPublic bool
|
2018-11-30 07:04:21 +03:00
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
2018-10-18 01:35:54 +03:00
|
|
|
nodes := tx.Bucket(nodeBucket)
|
|
|
|
if nodes == nil {
|
|
|
|
return ErrGraphNodesNotFound
|
|
|
|
}
|
|
|
|
ourPubKey := nodes.Get(sourceKey)
|
|
|
|
if ourPubKey == nil {
|
|
|
|
return ErrSourceNodeNotSet
|
|
|
|
}
|
|
|
|
node, err := fetchLightningNode(nodes, pubKey[:])
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
nodeIsPublic, err = node.isPublic(tx, ourPubKey)
|
|
|
|
return err
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return false, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return nodeIsPublic, nil
|
|
|
|
}
|
|
|
|
|
2018-07-18 05:10:07 +03:00
|
|
|
// genMultiSigP2WSH generates the p2wsh'd multisig script for 2 of 2 pubkeys.
|
|
|
|
func genMultiSigP2WSH(aPub, bPub []byte) ([]byte, error) {
|
|
|
|
if len(aPub) != 33 || len(bPub) != 33 {
|
|
|
|
return nil, fmt.Errorf("Pubkey size error. Compressed " +
|
|
|
|
"pubkeys only")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Swap to sort pubkeys if needed. Keys are sorted in lexicographical
|
|
|
|
// order. The signatures within the scriptSig must also adhere to the
|
|
|
|
// order, ensuring that the signatures for each public key appears in
|
|
|
|
// the proper order on the stack.
|
|
|
|
if bytes.Compare(aPub, bPub) == 1 {
|
|
|
|
aPub, bPub = bPub, aPub
|
|
|
|
}
|
|
|
|
|
|
|
|
// First, we'll generate the witness script for the multi-sig.
|
|
|
|
bldr := txscript.NewScriptBuilder()
|
|
|
|
bldr.AddOp(txscript.OP_2)
|
|
|
|
bldr.AddData(aPub) // Add both pubkeys (sorted).
|
|
|
|
bldr.AddData(bPub)
|
|
|
|
bldr.AddOp(txscript.OP_2)
|
|
|
|
bldr.AddOp(txscript.OP_CHECKMULTISIG)
|
|
|
|
witnessScript, err := bldr.Script()
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// With the witness script generated, we'll now turn it into a p2sh
|
|
|
|
// script:
|
|
|
|
// * OP_0 <sha256(script)>
|
|
|
|
bldr = txscript.NewScriptBuilder()
|
|
|
|
bldr.AddOp(txscript.OP_0)
|
|
|
|
scriptHash := sha256.Sum256(witnessScript)
|
|
|
|
bldr.AddData(scriptHash[:])
|
|
|
|
|
|
|
|
return bldr.Script()
|
|
|
|
}
|
|
|
|
|
|
|
|
// EdgePoint couples the outpoint of a channel with the funding script that it
|
|
|
|
// creates. The FilteredChainView will use this to watch for spends of this
|
|
|
|
// edge point on chain. We require both of these values as depending on the
|
|
|
|
// concrete implementation, either the pkScript, or the out point will be used.
|
|
|
|
type EdgePoint struct {
|
|
|
|
// FundingPkScript is the p2wsh multi-sig script of the target channel.
|
|
|
|
FundingPkScript []byte
|
|
|
|
|
|
|
|
// OutPoint is the outpoint of the target channel.
|
|
|
|
OutPoint wire.OutPoint
|
|
|
|
}
|
|
|
|
|
2018-08-16 07:43:33 +03:00
|
|
|
// String returns a human readable version of the target EdgePoint. We return
|
|
|
|
// the outpoint directly as it is enough to uniquely identify the edge point.
|
|
|
|
func (e *EdgePoint) String() string {
|
|
|
|
return e.OutPoint.String()
|
|
|
|
}
|
|
|
|
|
2017-05-11 02:19:42 +03:00
|
|
|
// ChannelView returns the verifiable edge information for each active channel
|
2018-07-18 05:10:07 +03:00
|
|
|
// within the known channel graph. The set of UTXO's (along with their scripts)
|
|
|
|
// returned are the ones that need to be watched on chain to detect channel
|
|
|
|
// closes on the resident blockchain.
|
|
|
|
func (c *ChannelGraph) ChannelView() ([]EdgePoint, error) {
|
|
|
|
var edgePoints []EdgePoint
|
2018-11-30 07:04:21 +03:00
|
|
|
if err := c.db.View(func(tx *bbolt.Tx) error {
|
2017-05-11 02:19:42 +03:00
|
|
|
// We're going to iterate over the entire channel index, so
|
|
|
|
// we'll need to fetch the edgeBucket to get to the index as
|
|
|
|
// it's a sub-bucket.
|
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
chanIndex := edges.Bucket(channelPointBucket)
|
|
|
|
if chanIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
2018-07-18 05:10:07 +03:00
|
|
|
edgeIndex := edges.Bucket(edgeIndexBucket)
|
|
|
|
if edgeIndex == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
2017-05-11 02:19:42 +03:00
|
|
|
|
|
|
|
// Once we have the proper bucket, we'll range over each key
|
|
|
|
// (which is the channel point for the channel) and decode it,
|
|
|
|
// accumulating each entry.
|
2018-07-18 05:10:07 +03:00
|
|
|
return chanIndex.ForEach(func(chanPointBytes, chanID []byte) error {
|
2017-05-11 02:19:42 +03:00
|
|
|
chanPointReader := bytes.NewReader(chanPointBytes)
|
|
|
|
|
|
|
|
var chanPoint wire.OutPoint
|
|
|
|
err := readOutpoint(chanPointReader, &chanPoint)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-07-18 05:10:07 +03:00
|
|
|
edgeInfo, err := fetchChanEdgeInfo(
|
|
|
|
edgeIndex, chanID,
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
pkScript, err := genMultiSigP2WSH(
|
|
|
|
edgeInfo.BitcoinKey1Bytes[:],
|
|
|
|
edgeInfo.BitcoinKey2Bytes[:],
|
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
edgePoints = append(edgePoints, EdgePoint{
|
|
|
|
FundingPkScript: pkScript,
|
|
|
|
OutPoint: chanPoint,
|
|
|
|
})
|
|
|
|
|
2017-05-11 02:19:42 +03:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
}); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
2018-07-18 05:10:07 +03:00
|
|
|
return edgePoints, nil
|
2017-05-11 02:19:42 +03:00
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
// NewChannelEdgePolicy returns a new blank ChannelEdgePolicy.
|
|
|
|
func (c *ChannelGraph) NewChannelEdgePolicy() *ChannelEdgePolicy {
|
|
|
|
return &ChannelEdgePolicy{db: c.db}
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2019-03-27 23:06:12 +03:00
|
|
|
// MarkEdgeZombie marks an edge as a zombie within the graph's zombie index.
|
|
|
|
// The public keys should represent the node public keys of the two parties
|
|
|
|
// involved in the edge.
|
|
|
|
func (c *ChannelGraph) MarkEdgeZombie(chanID uint64, pubKey1,
|
|
|
|
pubKey2 [33]byte) error {
|
|
|
|
|
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
zombieIndex, err := edges.CreateBucketIfNotExists(zombieBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
return markEdgeZombie(zombieIndex, chanID, pubKey1, pubKey2)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
// markEdgeZombie marks an edge as a zombie within our zombie index. The public
|
|
|
|
// keys should represent the node public keys of the two parties involved in the
|
|
|
|
// edge.
|
|
|
|
func markEdgeZombie(zombieIndex *bbolt.Bucket, chanID uint64, pubKey1,
|
|
|
|
pubKey2 [33]byte) error {
|
|
|
|
|
|
|
|
var k [8]byte
|
|
|
|
byteOrder.PutUint64(k[:], chanID)
|
|
|
|
|
|
|
|
var v [66]byte
|
|
|
|
copy(v[:33], pubKey1[:])
|
|
|
|
copy(v[33:], pubKey2[:])
|
|
|
|
|
|
|
|
return zombieIndex.Put(k[:], v[:])
|
|
|
|
}
|
|
|
|
|
|
|
|
// MarkEdgeLive clears an edge from our zombie index, deeming it as live.
|
|
|
|
func (c *ChannelGraph) MarkEdgeLive(chanID uint64) error {
|
|
|
|
return c.db.Batch(func(tx *bbolt.Tx) error {
|
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
zombieIndex := edges.Bucket(zombieBucket)
|
|
|
|
if zombieIndex == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
var k [8]byte
|
|
|
|
byteOrder.PutUint64(k[:], chanID)
|
|
|
|
return zombieIndex.Delete(k[:])
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
// IsZombieEdge returns whether the edge is considered zombie. If it is a
|
|
|
|
// zombie, then the two node public keys corresponding to this edge are also
|
|
|
|
// returned.
|
|
|
|
func (c *ChannelGraph) IsZombieEdge(chanID uint64) (bool, [33]byte, [33]byte) {
|
|
|
|
var (
|
|
|
|
isZombie bool
|
|
|
|
pubKey1, pubKey2 [33]byte
|
|
|
|
)
|
|
|
|
|
|
|
|
err := c.db.View(func(tx *bbolt.Tx) error {
|
|
|
|
edges := tx.Bucket(edgeBucket)
|
|
|
|
if edges == nil {
|
|
|
|
return ErrGraphNoEdgesFound
|
|
|
|
}
|
|
|
|
zombieIndex := edges.Bucket(zombieBucket)
|
|
|
|
if zombieIndex == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
isZombie, pubKey1, pubKey2 = isZombieEdge(zombieIndex, chanID)
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return false, [33]byte{}, [33]byte{}
|
|
|
|
}
|
|
|
|
|
|
|
|
return isZombie, pubKey1, pubKey2
|
|
|
|
}
|
|
|
|
|
|
|
|
// isZombieEdge returns whether an entry exists for the given channel in the
|
|
|
|
// zombie index. If an entry exists, then the two node public keys corresponding
|
|
|
|
// to this edge are also returned.
|
|
|
|
func isZombieEdge(zombieIndex *bbolt.Bucket,
|
|
|
|
chanID uint64) (bool, [33]byte, [33]byte) {
|
|
|
|
|
|
|
|
var k [8]byte
|
|
|
|
byteOrder.PutUint64(k[:], chanID)
|
|
|
|
|
|
|
|
v := zombieIndex.Get(k[:])
|
|
|
|
if v == nil {
|
|
|
|
return false, [33]byte{}, [33]byte{}
|
|
|
|
}
|
|
|
|
|
|
|
|
var pubKey1, pubKey2 [33]byte
|
|
|
|
copy(pubKey1[:], v[:33])
|
|
|
|
copy(pubKey2[:], v[33:])
|
|
|
|
|
|
|
|
return true, pubKey1, pubKey2
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func putLightningNode(nodeBucket *bbolt.Bucket, aliasBucket *bbolt.Bucket,
|
|
|
|
updateIndex *bbolt.Bucket, node *LightningNode) error {
|
2018-04-17 05:04:29 +03:00
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
var (
|
2017-02-17 12:29:23 +03:00
|
|
|
scratch [16]byte
|
2016-12-08 09:47:01 +03:00
|
|
|
b bytes.Buffer
|
|
|
|
)
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
pub, err := node.PubKey()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
nodePub := pub.SerializeCompressed()
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2017-07-14 22:25:02 +03:00
|
|
|
// If the node has the update time set, write it, else write 0.
|
|
|
|
updateUnix := uint64(0)
|
|
|
|
if node.LastUpdate.Unix() > 0 {
|
|
|
|
updateUnix = uint64(node.LastUpdate.Unix())
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-02-17 12:29:23 +03:00
|
|
|
byteOrder.PutUint64(scratch[:8], updateUnix)
|
|
|
|
if _, err := b.Write(scratch[:8]); err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if _, err := b.Write(nodePub); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// If we got a node announcement for this node, we will have the rest
|
|
|
|
// of the data available. If not we don't have more data to write.
|
2017-07-14 22:25:02 +03:00
|
|
|
if !node.HaveNodeAnnouncement {
|
|
|
|
// Write HaveNodeAnnouncement=0.
|
|
|
|
byteOrder.PutUint16(scratch[:2], 0)
|
|
|
|
if _, err := b.Write(scratch[:2]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
return nodeBucket.Put(nodePub, b.Bytes())
|
|
|
|
}
|
|
|
|
|
|
|
|
// Write HaveNodeAnnouncement=1.
|
|
|
|
byteOrder.PutUint16(scratch[:2], 1)
|
|
|
|
if _, err := b.Write(scratch[:2]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
if err := binary.Write(&b, byteOrder, node.Color.R); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(&b, byteOrder, node.Color.G); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(&b, byteOrder, node.Color.B); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := wire.WriteVarString(&b, 0, node.Alias); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-03-20 12:24:55 +03:00
|
|
|
if err := node.Features.Encode(&b); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-02-17 12:29:23 +03:00
|
|
|
numAddresses := uint16(len(node.Addresses))
|
|
|
|
byteOrder.PutUint16(scratch[:2], numAddresses)
|
|
|
|
if _, err := b.Write(scratch[:2]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, address := range node.Addresses {
|
2018-01-30 03:00:23 +03:00
|
|
|
if err := serializeAddr(&b, address); err != nil {
|
|
|
|
return err
|
2017-02-17 12:29:23 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-10-10 16:05:59 +03:00
|
|
|
sigLen := len(node.AuthSigBytes)
|
|
|
|
if sigLen > 80 {
|
|
|
|
return fmt.Errorf("max sig len allowed is 80, had %v",
|
|
|
|
sigLen)
|
|
|
|
}
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
err = wire.WriteVarBytes(&b, 0, node.AuthSigBytes)
|
2017-03-27 18:01:12 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-09-05 03:01:20 +03:00
|
|
|
if len(node.ExtraOpaqueData) > MaxAllowedExtraOpaqueBytes {
|
|
|
|
return ErrTooManyExtraOpaqueBytes(len(node.ExtraOpaqueData))
|
|
|
|
}
|
2018-09-01 05:36:16 +03:00
|
|
|
err = wire.WriteVarBytes(&b, 0, node.ExtraOpaqueData)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-07-14 22:25:02 +03:00
|
|
|
if err := aliasBucket.Put(nodePub, []byte(node.Alias)); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// With the alias bucket updated, we'll now update the index that
|
|
|
|
// tracks the time series of node updates.
|
|
|
|
var indexKey [8 + 33]byte
|
|
|
|
byteOrder.PutUint64(indexKey[:8], updateUnix)
|
|
|
|
copy(indexKey[8:], nodePub)
|
|
|
|
|
|
|
|
// If there was already an old index entry for this node, then we'll
|
|
|
|
// delete the old one before we write the new entry.
|
|
|
|
if nodeBytes := nodeBucket.Get(nodePub); nodeBytes != nil {
|
|
|
|
// Extract out the old update time to we can reconstruct the
|
|
|
|
// prior index key to delete it from the index.
|
|
|
|
oldUpdateTime := nodeBytes[:8]
|
|
|
|
|
|
|
|
var oldIndexKey [8 + 33]byte
|
|
|
|
copy(oldIndexKey[:8], oldUpdateTime)
|
|
|
|
copy(oldIndexKey[8:], nodePub)
|
|
|
|
|
|
|
|
if err := updateIndex.Delete(oldIndexKey[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := updateIndex.Put(indexKey[:], nil); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-07-14 22:25:02 +03:00
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
return nodeBucket.Put(nodePub, b.Bytes())
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func fetchLightningNode(nodeBucket *bbolt.Bucket,
|
2018-02-13 03:17:08 +03:00
|
|
|
nodePub []byte) (LightningNode, error) {
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
nodeBytes := nodeBucket.Get(nodePub)
|
|
|
|
if nodeBytes == nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, ErrGraphNodeNotFound
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
nodeReader := bytes.NewReader(nodeBytes)
|
|
|
|
return deserializeLightningNode(nodeReader)
|
|
|
|
}
|
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
func deserializeLightningNode(r io.Reader) (LightningNode, error) {
|
2018-01-31 07:19:40 +03:00
|
|
|
var (
|
2018-02-13 03:17:08 +03:00
|
|
|
node LightningNode
|
2018-01-31 07:19:40 +03:00
|
|
|
scratch [8]byte
|
|
|
|
err error
|
|
|
|
)
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
if _, err := r.Read(scratch[:]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
unix := int64(byteOrder.Uint64(scratch[:]))
|
|
|
|
node.LastUpdate = time.Unix(unix, 0)
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := io.ReadFull(r, node.PubKeyBytes[:]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-07-14 22:25:02 +03:00
|
|
|
if _, err := r.Read(scratch[:2]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2017-07-14 22:25:02 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
hasNodeAnn := byteOrder.Uint16(scratch[:2])
|
|
|
|
if hasNodeAnn == 1 {
|
|
|
|
node.HaveNodeAnnouncement = true
|
|
|
|
} else {
|
|
|
|
node.HaveNodeAnnouncement = false
|
|
|
|
}
|
|
|
|
|
|
|
|
// The rest of the data is optional, and will only be there if we got a node
|
|
|
|
// announcement for this node.
|
|
|
|
if !node.HaveNodeAnnouncement {
|
|
|
|
return node, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// We did get a node announcement for this node, so we'll have the rest
|
|
|
|
// of the data available.
|
2016-12-08 09:47:01 +03:00
|
|
|
if err := binary.Read(r, byteOrder, &node.Color.R); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
if err := binary.Read(r, byteOrder, &node.Color.G); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
if err := binary.Read(r, byteOrder, &node.Color.B); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
node.Alias, err = wire.ReadVarString(r, 0)
|
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
|
|
|
|
2017-10-11 21:37:54 +03:00
|
|
|
fv := lnwire.NewFeatureVector(nil, lnwire.GlobalFeatures)
|
|
|
|
err = fv.Decode(r)
|
2017-03-20 12:24:55 +03:00
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2017-03-20 12:24:55 +03:00
|
|
|
}
|
2017-10-11 21:37:54 +03:00
|
|
|
node.Features = fv
|
2017-03-20 12:24:55 +03:00
|
|
|
|
2017-02-17 12:29:23 +03:00
|
|
|
if _, err := r.Read(scratch[:2]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2017-02-17 12:29:23 +03:00
|
|
|
}
|
|
|
|
numAddresses := int(byteOrder.Uint16(scratch[:2]))
|
|
|
|
|
|
|
|
var addresses []net.Addr
|
|
|
|
for i := 0; i < numAddresses; i++ {
|
2018-01-30 03:00:23 +03:00
|
|
|
address, err := deserializeAddr(r)
|
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2017-02-17 12:29:23 +03:00
|
|
|
}
|
|
|
|
addresses = append(addresses, address)
|
|
|
|
}
|
|
|
|
node.Addresses = addresses
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
node.AuthSigBytes, err = wire.ReadVarBytes(r, 0, 80, "sig")
|
2017-03-27 18:01:12 +03:00
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return LightningNode{}, err
|
2017-03-27 18:01:12 +03:00
|
|
|
}
|
|
|
|
|
2018-09-01 05:36:16 +03:00
|
|
|
// We'll try and see if there are any opaque bytes left, if not, then
|
|
|
|
// we'll ignore the EOF error and return the node as is.
|
2018-09-05 03:01:20 +03:00
|
|
|
node.ExtraOpaqueData, err = wire.ReadVarBytes(
|
|
|
|
r, 0, MaxAllowedExtraOpaqueBytes, "blob",
|
|
|
|
)
|
2018-09-06 02:42:16 +03:00
|
|
|
switch {
|
|
|
|
case err == io.ErrUnexpectedEOF:
|
|
|
|
case err == io.EOF:
|
|
|
|
case err != nil:
|
2018-09-01 05:36:16 +03:00
|
|
|
return LightningNode{}, err
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
return node, nil
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func putChanEdgeInfo(edgeIndex *bbolt.Bucket, edgeInfo *ChannelEdgeInfo, chanID [8]byte) error {
|
2017-03-06 04:18:00 +03:00
|
|
|
var b bytes.Buffer
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := b.Write(edgeInfo.NodeKey1Bytes[:]); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := b.Write(edgeInfo.NodeKey2Bytes[:]); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := b.Write(edgeInfo.BitcoinKey1Bytes[:]); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := b.Write(edgeInfo.BitcoinKey2Bytes[:]); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := wire.WriteVarBytes(&b, 0, edgeInfo.Features); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
authProof := edgeInfo.AuthProof
|
2017-03-19 21:37:15 +03:00
|
|
|
var nodeSig1, nodeSig2, bitcoinSig1, bitcoinSig2 []byte
|
|
|
|
if authProof != nil {
|
2018-01-31 07:19:40 +03:00
|
|
|
nodeSig1 = authProof.NodeSig1Bytes
|
|
|
|
nodeSig2 = authProof.NodeSig2Bytes
|
|
|
|
bitcoinSig1 = authProof.BitcoinSig1Bytes
|
|
|
|
bitcoinSig2 = authProof.BitcoinSig2Bytes
|
2017-03-19 21:37:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
if err := wire.WriteVarBytes(&b, 0, nodeSig1); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
2017-03-19 21:37:15 +03:00
|
|
|
if err := wire.WriteVarBytes(&b, 0, nodeSig2); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
2017-03-19 21:37:15 +03:00
|
|
|
if err := wire.WriteVarBytes(&b, 0, bitcoinSig1); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
2017-03-19 21:37:15 +03:00
|
|
|
if err := wire.WriteVarBytes(&b, 0, bitcoinSig2); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-07-26 06:39:59 +03:00
|
|
|
if err := writeOutpoint(&b, &edgeInfo.ChannelPoint); err != nil {
|
2017-03-06 04:18:00 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(&b, byteOrder, uint64(edgeInfo.Capacity)); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if _, err := b.Write(chanID[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-08-22 08:57:16 +03:00
|
|
|
if _, err := b.Write(edgeInfo.ChainHash[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2018-09-05 03:01:20 +03:00
|
|
|
if len(edgeInfo.ExtraOpaqueData) > MaxAllowedExtraOpaqueBytes {
|
|
|
|
return ErrTooManyExtraOpaqueBytes(len(edgeInfo.ExtraOpaqueData))
|
|
|
|
}
|
2018-09-01 05:36:16 +03:00
|
|
|
err := wire.WriteVarBytes(&b, 0, edgeInfo.ExtraOpaqueData)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
return edgeIndex.Put(chanID[:], b.Bytes())
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func fetchChanEdgeInfo(edgeIndex *bbolt.Bucket,
|
2018-02-13 03:17:08 +03:00
|
|
|
chanID []byte) (ChannelEdgeInfo, error) {
|
2017-03-06 04:18:00 +03:00
|
|
|
|
|
|
|
edgeInfoBytes := edgeIndex.Get(chanID)
|
|
|
|
if edgeInfoBytes == nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, ErrEdgeNotFound
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
edgeInfoReader := bytes.NewReader(edgeInfoBytes)
|
|
|
|
return deserializeChanEdgeInfo(edgeInfoReader)
|
|
|
|
}
|
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
func deserializeChanEdgeInfo(r io.Reader) (ChannelEdgeInfo, error) {
|
2017-03-06 04:18:00 +03:00
|
|
|
var (
|
2018-01-31 07:19:40 +03:00
|
|
|
err error
|
2018-02-13 03:17:08 +03:00
|
|
|
edgeInfo ChannelEdgeInfo
|
2017-03-06 04:18:00 +03:00
|
|
|
)
|
|
|
|
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := io.ReadFull(r, edgeInfo.NodeKey1Bytes[:]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := io.ReadFull(r, edgeInfo.NodeKey2Bytes[:]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := io.ReadFull(r, edgeInfo.BitcoinKey1Bytes[:]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
2018-01-31 07:19:40 +03:00
|
|
|
if _, err := io.ReadFull(r, edgeInfo.BitcoinKey2Bytes[:]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
edgeInfo.Features, err = wire.ReadVarBytes(r, 0, 900, "features")
|
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
2017-03-19 21:37:15 +03:00
|
|
|
proof := &ChannelAuthProof{}
|
2017-03-06 04:18:00 +03:00
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
proof.NodeSig1Bytes, err = wire.ReadVarBytes(r, 0, 80, "sigs")
|
2017-03-06 04:18:00 +03:00
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
2018-02-13 03:17:08 +03:00
|
|
|
proof.NodeSig2Bytes, err = wire.ReadVarBytes(r, 0, 80, "sigs")
|
2017-03-06 04:18:00 +03:00
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
2018-02-13 03:17:08 +03:00
|
|
|
proof.BitcoinSig1Bytes, err = wire.ReadVarBytes(r, 0, 80, "sigs")
|
2017-03-06 04:18:00 +03:00
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
2018-02-13 03:17:08 +03:00
|
|
|
proof.BitcoinSig2Bytes, err = wire.ReadVarBytes(r, 0, 80, "sigs")
|
2017-03-06 04:18:00 +03:00
|
|
|
if err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
2017-03-19 21:37:15 +03:00
|
|
|
if !proof.IsEmpty() {
|
|
|
|
edgeInfo.AuthProof = proof
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
edgeInfo.ChannelPoint = wire.OutPoint{}
|
2017-07-26 06:39:59 +03:00
|
|
|
if err := readOutpoint(r, &edgeInfo.ChannelPoint); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
if err := binary.Read(r, byteOrder, &edgeInfo.Capacity); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
if err := binary.Read(r, byteOrder, &edgeInfo.ChannelID); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-08-22 08:57:16 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
if _, err := io.ReadFull(r, edgeInfo.ChainHash[:]); err != nil {
|
2018-02-13 03:17:08 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
2018-09-01 05:36:16 +03:00
|
|
|
// We'll try and see if there are any opaque bytes left, if not, then
|
|
|
|
// we'll ignore the EOF error and return the edge as is.
|
2018-09-05 03:01:20 +03:00
|
|
|
edgeInfo.ExtraOpaqueData, err = wire.ReadVarBytes(
|
|
|
|
r, 0, MaxAllowedExtraOpaqueBytes, "blob",
|
|
|
|
)
|
2018-09-06 02:42:16 +03:00
|
|
|
switch {
|
|
|
|
case err == io.ErrUnexpectedEOF:
|
|
|
|
case err == io.EOF:
|
|
|
|
case err != nil:
|
2018-09-01 05:36:16 +03:00
|
|
|
return ChannelEdgeInfo{}, err
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
return edgeInfo, nil
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func putChanEdgePolicy(edges, nodes *bbolt.Bucket, edge *ChannelEdgePolicy,
|
2018-09-01 00:59:35 +03:00
|
|
|
from, to []byte) error {
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
var edgeKey [33 + 8]byte
|
|
|
|
copy(edgeKey[:], from)
|
|
|
|
byteOrder.PutUint64(edgeKey[33:], edge.ChannelID)
|
|
|
|
|
|
|
|
var b bytes.Buffer
|
2019-01-12 20:59:44 +03:00
|
|
|
if err := serializeChanEdgePolicy(&b, edge, to); err != nil {
|
2018-09-01 05:36:16 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// Before we write out the new edge, we'll create a new entry in the
|
|
|
|
// update index in order to keep it fresh.
|
2019-01-12 20:59:44 +03:00
|
|
|
updateUnix := uint64(edge.LastUpdate.Unix())
|
2018-04-17 05:04:29 +03:00
|
|
|
var indexKey [8 + 8]byte
|
2019-01-12 20:59:44 +03:00
|
|
|
byteOrder.PutUint64(indexKey[:8], updateUnix)
|
2018-04-17 05:04:29 +03:00
|
|
|
byteOrder.PutUint64(indexKey[8:], edge.ChannelID)
|
|
|
|
|
|
|
|
updateIndex, err := edges.CreateBucketIfNotExists(edgeUpdateIndexBucket)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// If there was already an entry for this edge, then we'll need to
|
|
|
|
// delete the old one to ensure we don't leave around any after-images.
|
2018-06-18 13:35:22 +03:00
|
|
|
// An unknown policy value does not have a update time recorded, so
|
|
|
|
// it also does not need to be removed.
|
|
|
|
if edgeBytes := edges.Get(edgeKey[:]); edgeBytes != nil &&
|
|
|
|
!bytes.Equal(edgeBytes[:], unknownPolicy) {
|
|
|
|
|
2018-04-17 05:04:29 +03:00
|
|
|
// In order to delete the old entry, we'll need to obtain the
|
|
|
|
// *prior* update time in order to delete it. To do this, we'll
|
2018-09-01 00:59:35 +03:00
|
|
|
// need to deserialize the existing policy within the database
|
|
|
|
// (now outdated by the new one), and delete its corresponding
|
2019-01-12 20:59:44 +03:00
|
|
|
// entry within the update index. We'll ignore any
|
|
|
|
// ErrEdgePolicyOptionalFieldNotFound error, as we only need
|
|
|
|
// the channel ID and update time to delete the entry.
|
|
|
|
// TODO(halseth): get rid of these invalid policies in a
|
|
|
|
// migration.
|
2018-09-01 00:59:35 +03:00
|
|
|
oldEdgePolicy, err := deserializeChanEdgePolicy(
|
|
|
|
bytes.NewReader(edgeBytes), nodes,
|
|
|
|
)
|
2019-01-12 20:59:44 +03:00
|
|
|
if err != nil && err != ErrEdgePolicyOptionalFieldNotFound {
|
2018-09-01 00:59:35 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
oldUpdateTime := uint64(oldEdgePolicy.LastUpdate.Unix())
|
2018-04-17 05:04:29 +03:00
|
|
|
|
|
|
|
var oldIndexKey [8 + 8]byte
|
2018-09-01 00:59:35 +03:00
|
|
|
byteOrder.PutUint64(oldIndexKey[:8], oldUpdateTime)
|
2018-04-17 05:04:29 +03:00
|
|
|
byteOrder.PutUint64(oldIndexKey[8:], edge.ChannelID)
|
|
|
|
|
|
|
|
if err := updateIndex.Delete(oldIndexKey[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := updateIndex.Put(indexKey[:], nil); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
return edges.Put(edgeKey[:], b.Bytes()[:])
|
|
|
|
}
|
|
|
|
|
2018-06-18 13:35:22 +03:00
|
|
|
// putChanEdgePolicyUnknown marks the edge policy as unknown
|
|
|
|
// in the edges bucket.
|
2018-11-30 07:04:21 +03:00
|
|
|
func putChanEdgePolicyUnknown(edges *bbolt.Bucket, channelID uint64,
|
2018-06-18 13:35:22 +03:00
|
|
|
from []byte) error {
|
|
|
|
|
|
|
|
var edgeKey [33 + 8]byte
|
|
|
|
copy(edgeKey[:], from)
|
|
|
|
byteOrder.PutUint64(edgeKey[33:], channelID)
|
|
|
|
|
|
|
|
if edges.Get(edgeKey[:]) != nil {
|
|
|
|
return fmt.Errorf("Cannot write unknown policy for channel %v "+
|
|
|
|
" when there is already a policy present", channelID)
|
|
|
|
}
|
|
|
|
|
|
|
|
return edges.Put(edgeKey[:], unknownPolicy)
|
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func fetchChanEdgePolicy(edges *bbolt.Bucket, chanID []byte,
|
|
|
|
nodePub []byte, nodes *bbolt.Bucket) (*ChannelEdgePolicy, error) {
|
2017-03-06 04:18:00 +03:00
|
|
|
|
|
|
|
var edgeKey [33 + 8]byte
|
|
|
|
copy(edgeKey[:], nodePub)
|
|
|
|
copy(edgeKey[33:], chanID[:])
|
|
|
|
|
|
|
|
edgeBytes := edges.Get(edgeKey[:])
|
|
|
|
if edgeBytes == nil {
|
|
|
|
return nil, ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
2018-06-18 13:35:22 +03:00
|
|
|
// No need to deserialize unknown policy.
|
|
|
|
if bytes.Equal(edgeBytes[:], unknownPolicy) {
|
|
|
|
return nil, nil
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
edgeReader := bytes.NewReader(edgeBytes)
|
|
|
|
|
2019-01-12 20:59:44 +03:00
|
|
|
ep, err := deserializeChanEdgePolicy(edgeReader, nodes)
|
|
|
|
switch {
|
|
|
|
// If the db policy was missing an expected optional field, we return
|
|
|
|
// nil as if the policy was unknown.
|
|
|
|
case err == ErrEdgePolicyOptionalFieldNotFound:
|
|
|
|
return nil, nil
|
|
|
|
|
|
|
|
case err != nil:
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return ep, nil
|
2017-03-06 04:18:00 +03:00
|
|
|
}
|
|
|
|
|
2018-11-30 07:04:21 +03:00
|
|
|
func fetchChanEdgePolicies(edgeIndex *bbolt.Bucket, edges *bbolt.Bucket,
|
|
|
|
nodes *bbolt.Bucket, chanID []byte,
|
2017-03-06 04:18:00 +03:00
|
|
|
db *DB) (*ChannelEdgePolicy, *ChannelEdgePolicy, error) {
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
edgeInfo := edgeIndex.Get(chanID)
|
2017-01-23 01:28:28 +03:00
|
|
|
if edgeInfo == nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil, nil, ErrEdgeNotFound
|
|
|
|
}
|
|
|
|
|
2017-01-23 01:28:28 +03:00
|
|
|
// The first node is contained within the first half of the edge
|
|
|
|
// information. We only propagate the error here and below if it's
|
|
|
|
// something other than edge non-existence.
|
2016-12-08 09:47:01 +03:00
|
|
|
node1Pub := edgeInfo[:33]
|
2017-03-06 04:18:00 +03:00
|
|
|
edge1, err := fetchChanEdgePolicy(edges, chanID, node1Pub, nodes)
|
2018-06-18 13:35:22 +03:00
|
|
|
if err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil, nil, err
|
|
|
|
}
|
2016-12-31 03:35:23 +03:00
|
|
|
|
2017-01-23 01:28:28 +03:00
|
|
|
// As we may have a single direction of the edge but not the other,
|
|
|
|
// only fill in the database pointers if the edge is found.
|
2016-12-31 03:35:23 +03:00
|
|
|
if edge1 != nil {
|
|
|
|
edge1.db = db
|
|
|
|
edge1.Node.db = db
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
// Similarly, the second node is contained within the latter
|
|
|
|
// half of the edge information.
|
2018-09-01 00:59:34 +03:00
|
|
|
node2Pub := edgeInfo[33:66]
|
2017-03-06 04:18:00 +03:00
|
|
|
edge2, err := fetchChanEdgePolicy(edges, chanID, node2Pub, nodes)
|
2018-06-18 13:35:22 +03:00
|
|
|
if err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil, nil, err
|
|
|
|
}
|
2016-12-31 03:35:23 +03:00
|
|
|
|
|
|
|
if edge2 != nil {
|
|
|
|
edge2.db = db
|
|
|
|
edge2.Node.db = db
|
|
|
|
}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
return edge1, edge2, nil
|
|
|
|
}
|
|
|
|
|
2019-01-12 20:59:44 +03:00
|
|
|
func serializeChanEdgePolicy(w io.Writer, edge *ChannelEdgePolicy,
|
|
|
|
to []byte) error {
|
|
|
|
|
|
|
|
err := wire.WriteVarBytes(w, 0, edge.SigBytes)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := binary.Write(w, byteOrder, edge.ChannelID); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
var scratch [8]byte
|
|
|
|
updateUnix := uint64(edge.LastUpdate.Unix())
|
|
|
|
byteOrder.PutUint64(scratch[:], updateUnix)
|
|
|
|
if _, err := w.Write(scratch[:]); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := binary.Write(w, byteOrder, edge.MessageFlags); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(w, byteOrder, edge.ChannelFlags); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(w, byteOrder, edge.TimeLockDelta); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(w, byteOrder, uint64(edge.MinHTLC)); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(w, byteOrder, uint64(edge.FeeBaseMSat)); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err := binary.Write(w, byteOrder, uint64(edge.FeeProportionalMillionths)); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
if _, err := w.Write(to); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2019-01-12 20:59:44 +03:00
|
|
|
// If the max_htlc field is present, we write it. To be compatible with
|
|
|
|
// older versions that wasn't aware of this field, we write it as part
|
|
|
|
// of the opaque data.
|
|
|
|
// TODO(halseth): clean up when moving to TLV.
|
|
|
|
var opaqueBuf bytes.Buffer
|
2019-01-16 14:43:46 +03:00
|
|
|
if edge.MessageFlags.HasMaxHtlc() {
|
2019-01-12 20:59:44 +03:00
|
|
|
err := binary.Write(&opaqueBuf, byteOrder, uint64(edge.MaxHTLC))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-12 20:59:44 +03:00
|
|
|
if len(edge.ExtraOpaqueData) > MaxAllowedExtraOpaqueBytes {
|
|
|
|
return ErrTooManyExtraOpaqueBytes(len(edge.ExtraOpaqueData))
|
|
|
|
}
|
2019-01-12 20:59:44 +03:00
|
|
|
if _, err := opaqueBuf.Write(edge.ExtraOpaqueData); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-01-12 20:59:44 +03:00
|
|
|
|
2019-01-12 20:59:44 +03:00
|
|
|
if err := wire.WriteVarBytes(w, 0, opaqueBuf.Bytes()); err != nil {
|
2019-01-12 20:59:44 +03:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
func deserializeChanEdgePolicy(r io.Reader,
|
2018-11-30 07:04:21 +03:00
|
|
|
nodes *bbolt.Bucket) (*ChannelEdgePolicy, error) {
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2017-03-06 04:18:00 +03:00
|
|
|
edge := &ChannelEdgePolicy{}
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-02-13 03:17:08 +03:00
|
|
|
var err error
|
|
|
|
edge.SigBytes, err = wire.ReadVarBytes(r, 0, 80, "sig")
|
2017-03-27 18:01:12 +03:00
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
if err := binary.Read(r, byteOrder, &edge.ChannelID); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
var scratch [8]byte
|
|
|
|
if _, err := r.Read(scratch[:]); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
unix := int64(byteOrder.Uint64(scratch[:]))
|
|
|
|
edge.LastUpdate = time.Unix(unix, 0)
|
|
|
|
|
2019-01-12 20:59:43 +03:00
|
|
|
if err := binary.Read(r, byteOrder, &edge.MessageFlags); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
if err := binary.Read(r, byteOrder, &edge.ChannelFlags); err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil, err
|
|
|
|
}
|
2017-03-06 04:18:00 +03:00
|
|
|
if err := binary.Read(r, byteOrder, &edge.TimeLockDelta); err != nil {
|
2016-12-08 09:47:01 +03:00
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
var n uint64
|
|
|
|
if err := binary.Read(r, byteOrder, &n); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2017-08-22 08:54:42 +03:00
|
|
|
edge.MinHTLC = lnwire.MilliSatoshi(n)
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
if err := binary.Read(r, byteOrder, &n); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2017-08-22 08:54:42 +03:00
|
|
|
edge.FeeBaseMSat = lnwire.MilliSatoshi(n)
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
if err := binary.Read(r, byteOrder, &n); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2017-08-22 08:54:42 +03:00
|
|
|
edge.FeeProportionalMillionths = lnwire.MilliSatoshi(n)
|
2016-12-08 09:47:01 +03:00
|
|
|
|
|
|
|
var pub [33]byte
|
|
|
|
if _, err := r.Read(pub[:]); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
node, err := fetchLightningNode(nodes, pub[:])
|
|
|
|
if err != nil {
|
2018-07-23 07:02:46 +03:00
|
|
|
return nil, fmt.Errorf("unable to fetch node: %x, %v",
|
|
|
|
pub[:], err)
|
2016-12-08 09:47:01 +03:00
|
|
|
}
|
2019-01-12 20:59:44 +03:00
|
|
|
edge.Node = &node
|
2016-12-08 09:47:01 +03:00
|
|
|
|
2018-09-01 05:36:16 +03:00
|
|
|
// We'll try and see if there are any opaque bytes left, if not, then
|
|
|
|
// we'll ignore the EOF error and return the edge as is.
|
2018-09-05 03:01:20 +03:00
|
|
|
edge.ExtraOpaqueData, err = wire.ReadVarBytes(
|
|
|
|
r, 0, MaxAllowedExtraOpaqueBytes, "blob",
|
|
|
|
)
|
2018-09-06 02:42:16 +03:00
|
|
|
switch {
|
|
|
|
case err == io.ErrUnexpectedEOF:
|
|
|
|
case err == io.EOF:
|
|
|
|
case err != nil:
|
2018-09-01 05:36:16 +03:00
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
2019-01-12 20:59:44 +03:00
|
|
|
// See if optional fields are present.
|
2019-01-16 14:43:46 +03:00
|
|
|
if edge.MessageFlags.HasMaxHtlc() {
|
2019-01-12 20:59:44 +03:00
|
|
|
// The max_htlc field should be at the beginning of the opaque
|
|
|
|
// bytes.
|
|
|
|
opq := edge.ExtraOpaqueData
|
|
|
|
|
|
|
|
// If the max_htlc field is not present, it might be old data
|
|
|
|
// stored before this field was validated. We'll return the
|
|
|
|
// edge along with an error.
|
|
|
|
if len(opq) < 8 {
|
|
|
|
return edge, ErrEdgePolicyOptionalFieldNotFound
|
|
|
|
}
|
|
|
|
|
|
|
|
maxHtlc := byteOrder.Uint64(opq[:8])
|
|
|
|
edge.MaxHTLC = lnwire.MilliSatoshi(maxHtlc)
|
|
|
|
|
|
|
|
// Exclude the parsed field from the rest of the opaque data.
|
|
|
|
edge.ExtraOpaqueData = opq[8:]
|
|
|
|
}
|
|
|
|
|
2016-12-08 09:47:01 +03:00
|
|
|
return edge, nil
|
|
|
|
}
|