mirror of
https://github.com/XRPLF/rippled.git
synced 2026-04-25 21:48:03 +00:00
Compare commits
77 Commits
develop
...
bthomee/no
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
5b5ad09c53 | ||
|
|
2405a3353c | ||
|
|
705622011b | ||
|
|
ae5f5cb92b | ||
|
|
cde8f17b5d | ||
|
|
c8fb69ee1a | ||
|
|
430f770f2a | ||
|
|
44590a7008 | ||
|
|
1934c316b2 | ||
|
|
fd2a8b5825 | ||
|
|
012144c16c | ||
|
|
8e2d949680 | ||
|
|
b781018fee | ||
|
|
d867c9b26d | ||
|
|
7749ed8488 | ||
|
|
4d01cac564 | ||
|
|
98e1ad2dec | ||
|
|
342171bf20 | ||
|
|
fe74f48e7a | ||
|
|
d6e05cf513 | ||
|
|
e31419aa06 | ||
|
|
05b4c67b96 | ||
|
|
2524476124 | ||
|
|
8007788d77 | ||
|
|
809395a59c | ||
|
|
3ef64e019b | ||
|
|
228ad1e98f | ||
|
|
c4b342a027 | ||
|
|
556d80e724 | ||
|
|
044f1e67b9 | ||
|
|
0ad72fae8f | ||
|
|
cc383c5fb8 | ||
|
|
88715f1e5a | ||
|
|
e1477cef0c | ||
|
|
942874d7b0 | ||
|
|
79326fc6b5 | ||
|
|
48535d5226 | ||
|
|
d1a6558080 | ||
|
|
84f86b354f | ||
|
|
40a3985b02 | ||
|
|
208bd35d45 | ||
|
|
e90fbbf7b2 | ||
|
|
277450e648 | ||
|
|
e6993524ea | ||
|
|
b117ecc6a2 | ||
|
|
6c3b00c342 | ||
|
|
8c296a935a | ||
|
|
573ba82181 | ||
|
|
1542ab7e27 | ||
|
|
6374f4886d | ||
|
|
ebf336f472 | ||
|
|
ddc15ad612 | ||
|
|
82db6ac498 | ||
|
|
f749c41306 | ||
|
|
f25e47a58d | ||
|
|
2396799bd8 | ||
|
|
4855b9f96a | ||
|
|
b2f65cb7eb | ||
|
|
c523673885 | ||
|
|
caac4d63d3 | ||
|
|
29b0076fa8 | ||
|
|
c9aa1094a7 | ||
|
|
b86f69cb82 | ||
|
|
5d0bf78512 | ||
|
|
554df631c6 | ||
|
|
5e704bfdfb | ||
|
|
fe8cc02bfa | ||
|
|
061c033f52 | ||
|
|
832a7e7e4a | ||
|
|
b2371c4c02 | ||
|
|
b94a7c4b44 | ||
|
|
9b9027112d | ||
|
|
8e7889c66e | ||
|
|
d836c3788d | ||
|
|
1cb7c0293f | ||
|
|
52dabc1f79 | ||
|
|
2d78d41f7b |
@@ -93,7 +93,6 @@ test.core > xrpl.basics
|
|||||||
test.core > xrpl.core
|
test.core > xrpl.core
|
||||||
test.core > xrpld.core
|
test.core > xrpld.core
|
||||||
test.core > xrpl.json
|
test.core > xrpl.json
|
||||||
test.core > xrpl.protocol
|
|
||||||
test.core > xrpl.rdb
|
test.core > xrpl.rdb
|
||||||
test.core > xrpl.server
|
test.core > xrpl.server
|
||||||
test.csf > xrpl.basics
|
test.csf > xrpl.basics
|
||||||
|
|||||||
@@ -11,6 +11,7 @@
|
|||||||
#include <limits>
|
#include <limits>
|
||||||
#include <stdexcept>
|
#include <stdexcept>
|
||||||
#include <string>
|
#include <string>
|
||||||
|
#include <string_view>
|
||||||
#include <type_traits>
|
#include <type_traits>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@@ -231,4 +232,11 @@ makeSlice(std::basic_string<char, Traits, Alloc> const& s)
|
|||||||
return Slice(s.data(), s.size());
|
return Slice(s.data(), s.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
template <class Traits>
|
||||||
|
Slice
|
||||||
|
makeSlice(std::basic_string_view<char, Traits> s)
|
||||||
|
{
|
||||||
|
return Slice(s.data(), s.size());
|
||||||
|
}
|
||||||
|
|
||||||
} // namespace xrpl
|
} // namespace xrpl
|
||||||
|
|||||||
@@ -246,7 +246,15 @@ message TMGetObjectByHash {
|
|||||||
|
|
||||||
message TMLedgerNode {
|
message TMLedgerNode {
|
||||||
required bytes nodedata = 1;
|
required bytes nodedata = 1;
|
||||||
optional bytes nodeid = 2; // missing for ledger base data
|
|
||||||
|
// Used when protocol version <2.3. Not set for ledger base data.
|
||||||
|
optional bytes nodeid = 2;
|
||||||
|
|
||||||
|
// Used when protocol version >=2.3. Neither value is set for ledger base data.
|
||||||
|
oneof reference {
|
||||||
|
bytes id = 3; // Set for inner nodes.
|
||||||
|
uint32 depth = 4; // Set for leaf nodes.
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
enum TMLedgerInfoType {
|
enum TMLedgerInfoType {
|
||||||
|
|||||||
@@ -16,6 +16,7 @@
|
|||||||
|
|
||||||
#include <set>
|
#include <set>
|
||||||
#include <stack>
|
#include <stack>
|
||||||
|
#include <tuple>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
namespace xrpl {
|
namespace xrpl {
|
||||||
@@ -73,6 +74,22 @@ enum class SHAMapState {
|
|||||||
|
|
||||||
See https://en.wikipedia.org/wiki/Merkle_tree
|
See https://en.wikipedia.org/wiki/Merkle_tree
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
/** Holds a SHAMap node's identity, serialized data, and leaf status.
|
||||||
|
Used by getNodeFat to return node data for peer synchronization.
|
||||||
|
*/
|
||||||
|
struct SHAMapNodeData
|
||||||
|
{
|
||||||
|
SHAMapNodeID nodeID;
|
||||||
|
Blob data;
|
||||||
|
bool isLeaf;
|
||||||
|
|
||||||
|
SHAMapNodeData(SHAMapNodeID const& id, Blob d, bool leaf)
|
||||||
|
: nodeID(id), data(std::move(d)), isLeaf(leaf)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
class SHAMap
|
class SHAMap
|
||||||
{
|
{
|
||||||
private:
|
private:
|
||||||
@@ -85,7 +102,7 @@ private:
|
|||||||
/** The sequence of the ledger that this map references, if any. */
|
/** The sequence of the ledger that this map references, if any. */
|
||||||
std::uint32_t ledgerSeq_ = 0;
|
std::uint32_t ledgerSeq_ = 0;
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> root_;
|
SHAMapTreeNodePtr root_;
|
||||||
mutable SHAMapState state_;
|
mutable SHAMapState state_;
|
||||||
SHAMapType const type_;
|
SHAMapType const type_;
|
||||||
bool backed_ = true; // Map is backed by the database
|
bool backed_ = true; // Map is backed by the database
|
||||||
@@ -250,10 +267,10 @@ public:
|
|||||||
std::vector<std::pair<SHAMapNodeID, uint256>>
|
std::vector<std::pair<SHAMapNodeID, uint256>>
|
||||||
getMissingNodes(int maxNodes, SHAMapSyncFilter* filter);
|
getMissingNodes(int maxNodes, SHAMapSyncFilter* filter);
|
||||||
|
|
||||||
bool
|
[[nodiscard]] bool
|
||||||
getNodeFat(
|
getNodeFat(
|
||||||
SHAMapNodeID const& wanted,
|
SHAMapNodeID const& wanted,
|
||||||
std::vector<std::pair<SHAMapNodeID, Blob>>& data,
|
std::vector<SHAMapNodeData>& data,
|
||||||
bool fatLeaves,
|
bool fatLeaves,
|
||||||
std::uint32_t depth) const;
|
std::uint32_t depth) const;
|
||||||
|
|
||||||
@@ -280,10 +297,42 @@ public:
|
|||||||
void
|
void
|
||||||
serializeRoot(Serializer& s) const;
|
serializeRoot(Serializer& s) const;
|
||||||
|
|
||||||
|
/** Add a root node to the SHAMap during synchronization.
|
||||||
|
*
|
||||||
|
* This function is used when receiving the root node of a SHAMap from a peer during ledger
|
||||||
|
* synchronization. The node must already have been deserialized.
|
||||||
|
*
|
||||||
|
* @param hash The expected hash of the root node.
|
||||||
|
* @param rootNode A deserialized root node to add.
|
||||||
|
* @param filter Optional sync filter to track received nodes.
|
||||||
|
* @return Status indicating whether the node was useful, duplicate, or invalid.
|
||||||
|
*
|
||||||
|
* @note This function expects the rootNode to be a valid, deserialized SHAMapTreeNode. The
|
||||||
|
* caller is responsible for deserialization and basic validation before calling this
|
||||||
|
* function.
|
||||||
|
*/
|
||||||
SHAMapAddNode
|
SHAMapAddNode
|
||||||
addRootNode(SHAMapHash const& hash, Slice const& rootNode, SHAMapSyncFilter* filter);
|
addRootNode(SHAMapHash const& hash, SHAMapTreeNodePtr rootNode, SHAMapSyncFilter const* filter);
|
||||||
|
|
||||||
|
/** Add a known node at a specific position in the SHAMap during synchronization.
|
||||||
|
*
|
||||||
|
* This function is used when receiving nodes from peers during ledger synchronization. The node
|
||||||
|
* is inserted at the position specified by nodeID. The node must already have been
|
||||||
|
* deserialized.
|
||||||
|
*
|
||||||
|
* @param nodeID The position in the tree where this node belongs.
|
||||||
|
* @param treeNode A deserialized tree node to add.
|
||||||
|
* @param filter Optional sync filter to track received nodes.
|
||||||
|
* @return Status indicating whether the node was useful, duplicate, or invalid.
|
||||||
|
*
|
||||||
|
* @note This function expects that the caller has already validated that the nodeID is
|
||||||
|
* consistent with the node's content.
|
||||||
|
*/
|
||||||
SHAMapAddNode
|
SHAMapAddNode
|
||||||
addKnownNode(SHAMapNodeID const& nodeID, Slice const& rawNode, SHAMapSyncFilter* filter);
|
addKnownNode(
|
||||||
|
SHAMapNodeID const& nodeID,
|
||||||
|
SHAMapTreeNodePtr treeNode,
|
||||||
|
SHAMapSyncFilter const* filter);
|
||||||
|
|
||||||
// status functions
|
// status functions
|
||||||
void
|
void
|
||||||
@@ -326,36 +375,32 @@ public:
|
|||||||
invariants() const;
|
invariants() const;
|
||||||
|
|
||||||
private:
|
private:
|
||||||
using SharedPtrNodeStack =
|
using SharedPtrNodeStack = std::stack<std::pair<SHAMapTreeNodePtr, SHAMapNodeID>>;
|
||||||
std::stack<std::pair<intr_ptr::SharedPtr<SHAMapTreeNode>, SHAMapNodeID>>;
|
|
||||||
using DeltaRef =
|
using DeltaRef =
|
||||||
std::pair<boost::intrusive_ptr<SHAMapItem const>, boost::intrusive_ptr<SHAMapItem const>>;
|
std::pair<boost::intrusive_ptr<SHAMapItem const>, boost::intrusive_ptr<SHAMapItem const>>;
|
||||||
|
|
||||||
// tree node cache operations
|
// tree node cache operations
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
cacheLookup(SHAMapHash const& hash) const;
|
cacheLookup(SHAMapHash const& hash) const;
|
||||||
|
|
||||||
void
|
void
|
||||||
canonicalize(SHAMapHash const& hash, intr_ptr::SharedPtr<SHAMapTreeNode>&) const;
|
canonicalize(SHAMapHash const& hash, SHAMapTreeNodePtr&) const;
|
||||||
|
|
||||||
// database operations
|
// database operations
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
fetchNodeFromDB(SHAMapHash const& hash) const;
|
fetchNodeFromDB(SHAMapHash const& hash) const;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
fetchNodeNT(SHAMapHash const& hash) const;
|
fetchNodeNT(SHAMapHash const& hash) const;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
fetchNodeNT(SHAMapHash const& hash, SHAMapSyncFilter* filter) const;
|
fetchNodeNT(SHAMapHash const& hash, SHAMapSyncFilter const* filter) const;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
fetchNode(SHAMapHash const& hash) const;
|
fetchNode(SHAMapHash const& hash) const;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
checkFilter(SHAMapHash const& hash, SHAMapSyncFilter* filter) const;
|
checkFilter(SHAMapHash const& hash, SHAMapSyncFilter const* filter) const;
|
||||||
|
|
||||||
/** Update hashes up to the root */
|
/** Update hashes up to the root */
|
||||||
void
|
void
|
||||||
dirtyUp(
|
dirtyUp(SharedPtrNodeStack& stack, uint256 const& target, SHAMapTreeNodePtr terminal);
|
||||||
SharedPtrNodeStack& stack,
|
|
||||||
uint256 const& target,
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> terminal);
|
|
||||||
|
|
||||||
/** Walk towards the specified id, returning the node. Caller must check
|
/** Walk towards the specified id, returning the node. Caller must check
|
||||||
if the return is nullptr, and if not, if the node->peekItem()->key() ==
|
if the return is nullptr, and if not, if the node->peekItem()->key() ==
|
||||||
@@ -377,25 +422,21 @@ private:
|
|||||||
preFlushNode(intr_ptr::SharedPtr<Node> node) const;
|
preFlushNode(intr_ptr::SharedPtr<Node> node) const;
|
||||||
|
|
||||||
/** write and canonicalize modified node */
|
/** write and canonicalize modified node */
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
writeNode(NodeObjectType t, intr_ptr::SharedPtr<SHAMapTreeNode> node) const;
|
writeNode(NodeObjectType t, SHAMapTreeNodePtr node) const;
|
||||||
|
|
||||||
// returns the first item at or below this node
|
// returns the first item at or below this node
|
||||||
SHAMapLeafNode*
|
SHAMapLeafNode*
|
||||||
firstBelow(intr_ptr::SharedPtr<SHAMapTreeNode>, SharedPtrNodeStack& stack, int branch = 0)
|
firstBelow(SHAMapTreeNodePtr node, SharedPtrNodeStack& stack, int branch = 0) const;
|
||||||
const;
|
|
||||||
|
|
||||||
// returns the last item at or below this node
|
// returns the last item at or below this node
|
||||||
SHAMapLeafNode*
|
SHAMapLeafNode*
|
||||||
lastBelow(
|
lastBelow(SHAMapTreeNodePtr node, SharedPtrNodeStack& stack, int branch = branchFactor) const;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> node,
|
|
||||||
SharedPtrNodeStack& stack,
|
|
||||||
int branch = branchFactor) const;
|
|
||||||
|
|
||||||
// helper function for firstBelow and lastBelow
|
// helper function for firstBelow and lastBelow
|
||||||
SHAMapLeafNode*
|
SHAMapLeafNode*
|
||||||
belowHelper(
|
belowHelper(
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> node,
|
SHAMapTreeNodePtr node,
|
||||||
SharedPtrNodeStack& stack,
|
SharedPtrNodeStack& stack,
|
||||||
int branch,
|
int branch,
|
||||||
std::tuple<int, std::function<bool(int)>, std::function<void(int&)>> const& loopParams)
|
std::tuple<int, std::function<bool(int)>, std::function<void(int&)>> const& loopParams)
|
||||||
@@ -407,20 +448,19 @@ private:
|
|||||||
descend(SHAMapInnerNode*, int branch) const;
|
descend(SHAMapInnerNode*, int branch) const;
|
||||||
SHAMapTreeNode*
|
SHAMapTreeNode*
|
||||||
descendThrow(SHAMapInnerNode*, int branch) const;
|
descendThrow(SHAMapInnerNode*, int branch) const;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
descend(SHAMapInnerNode&, int branch) const;
|
descend(SHAMapInnerNode&, int branch) const;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
descendThrow(SHAMapInnerNode&, int branch) const;
|
descendThrow(SHAMapInnerNode&, int branch) const;
|
||||||
|
|
||||||
// Descend with filter
|
// Descend with filter
|
||||||
// If pending, callback is called as if it called fetchNodeNT
|
// If pending, callback is called as if it called fetchNodeNT
|
||||||
using descendCallback =
|
using descendCallback = std::function<void(SHAMapTreeNodePtr, SHAMapHash const&)>;
|
||||||
std::function<void(intr_ptr::SharedPtr<SHAMapTreeNode>, SHAMapHash const&)>;
|
|
||||||
SHAMapTreeNode*
|
SHAMapTreeNode*
|
||||||
descendAsync(
|
descendAsync(
|
||||||
SHAMapInnerNode* parent,
|
SHAMapInnerNode* parent,
|
||||||
int branch,
|
int branch,
|
||||||
SHAMapSyncFilter* filter,
|
SHAMapSyncFilter const* filter,
|
||||||
bool& pending,
|
bool& pending,
|
||||||
descendCallback&&) const;
|
descendCallback&&) const;
|
||||||
|
|
||||||
@@ -429,11 +469,11 @@ private:
|
|||||||
SHAMapInnerNode* parent,
|
SHAMapInnerNode* parent,
|
||||||
SHAMapNodeID const& parentID,
|
SHAMapNodeID const& parentID,
|
||||||
int branch,
|
int branch,
|
||||||
SHAMapSyncFilter* filter) const;
|
SHAMapSyncFilter const* filter) const;
|
||||||
|
|
||||||
// Non-storing
|
// Non-storing
|
||||||
// Does not hook the returned node to its parent
|
// Does not hook the returned node to its parent
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
descendNoStore(SHAMapInnerNode&, int branch) const;
|
descendNoStore(SHAMapInnerNode&, int branch) const;
|
||||||
|
|
||||||
/** If there is only one leaf below this node, get its contents */
|
/** If there is only one leaf below this node, get its contents */
|
||||||
@@ -495,10 +535,10 @@ private:
|
|||||||
|
|
||||||
// nodes we may have acquired from deferred reads
|
// nodes we may have acquired from deferred reads
|
||||||
using DeferredNode = std::tuple<
|
using DeferredNode = std::tuple<
|
||||||
SHAMapInnerNode*, // parent node
|
SHAMapInnerNode*, // parent node
|
||||||
SHAMapNodeID, // parent node ID
|
SHAMapNodeID, // parent node ID
|
||||||
int, // branch
|
int, // branch
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>>; // node
|
SHAMapTreeNodePtr>; // node
|
||||||
|
|
||||||
int deferred_;
|
int deferred_;
|
||||||
std::mutex deferLock_;
|
std::mutex deferLock_;
|
||||||
@@ -524,7 +564,7 @@ private:
|
|||||||
gmn_ProcessDeferredReads(MissingNodes&);
|
gmn_ProcessDeferredReads(MissingNodes&);
|
||||||
|
|
||||||
// fetch from DB helper function
|
// fetch from DB helper function
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
finishFetch(SHAMapHash const& hash, std::shared_ptr<NodeObject> const& object) const;
|
finishFetch(SHAMapHash const& hash, std::shared_ptr<NodeObject> const& object) const;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
@@ -27,7 +27,7 @@ public:
|
|||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
clone(std::uint32_t cowid) const final
|
clone(std::uint32_t cowid) const final
|
||||||
{
|
{
|
||||||
return intr_ptr::make_shared<SHAMapAccountStateLeafNode>(item_, cowid, hash_);
|
return intr_ptr::make_shared<SHAMapAccountStateLeafNode>(item_, cowid, hash_);
|
||||||
|
|||||||
@@ -87,7 +87,7 @@ public:
|
|||||||
void
|
void
|
||||||
partialDestructor() override;
|
partialDestructor() override;
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
clone(std::uint32_t cowid) const override;
|
clone(std::uint32_t cowid) const override;
|
||||||
|
|
||||||
SHAMapNodeType
|
SHAMapNodeType
|
||||||
@@ -121,19 +121,19 @@ public:
|
|||||||
getChildHash(int m) const;
|
getChildHash(int m) const;
|
||||||
|
|
||||||
void
|
void
|
||||||
setChild(int m, intr_ptr::SharedPtr<SHAMapTreeNode> child);
|
setChild(int m, SHAMapTreeNodePtr child);
|
||||||
|
|
||||||
void
|
void
|
||||||
shareChild(int m, intr_ptr::SharedPtr<SHAMapTreeNode> const& child);
|
shareChild(int m, SHAMapTreeNodePtr const& child);
|
||||||
|
|
||||||
SHAMapTreeNode*
|
SHAMapTreeNode*
|
||||||
getChildPointer(int branch);
|
getChildPointer(int branch);
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
getChild(int branch);
|
getChild(int branch);
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
canonicalizeChild(int branch, intr_ptr::SharedPtr<SHAMapTreeNode> node);
|
canonicalizeChild(int branch, SHAMapTreeNodePtr node);
|
||||||
|
|
||||||
// sync functions
|
// sync functions
|
||||||
bool
|
bool
|
||||||
@@ -161,10 +161,10 @@ public:
|
|||||||
void
|
void
|
||||||
invariants(bool is_root = false) const override;
|
invariants(bool is_root = false) const override;
|
||||||
|
|
||||||
static intr_ptr::SharedPtr<SHAMapTreeNode>
|
static SHAMapTreeNodePtr
|
||||||
makeFullInner(Slice data, SHAMapHash const& hash, bool hashValid);
|
makeFullInner(Slice data, SHAMapHash const& hash, bool hashValid);
|
||||||
|
|
||||||
static intr_ptr::SharedPtr<SHAMapTreeNode>
|
static SHAMapTreeNodePtr
|
||||||
makeCompressedInner(Slice data);
|
makeCompressedInner(Slice data);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
@@ -166,4 +166,6 @@ private:
|
|||||||
makeTransactionWithMeta(Slice data, SHAMapHash const& hash, bool hashValid);
|
makeTransactionWithMeta(Slice data, SHAMapHash const& hash, bool hashValid);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
using SHAMapTreeNodePtr = intr_ptr::SharedPtr<SHAMapTreeNode>;
|
||||||
|
|
||||||
} // namespace xrpl
|
} // namespace xrpl
|
||||||
|
|||||||
@@ -26,7 +26,7 @@ public:
|
|||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
clone(std::uint32_t cowid) const final
|
clone(std::uint32_t cowid) const final
|
||||||
{
|
{
|
||||||
return intr_ptr::make_shared<SHAMapTxLeafNode>(item_, cowid, hash_);
|
return intr_ptr::make_shared<SHAMapTxLeafNode>(item_, cowid, hash_);
|
||||||
|
|||||||
@@ -27,7 +27,7 @@ public:
|
|||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
clone(std::uint32_t cowid) const override
|
clone(std::uint32_t cowid) const override
|
||||||
{
|
{
|
||||||
return intr_ptr::make_shared<SHAMapTxPlusMetaLeafNode>(item_, cowid, hash_);
|
return intr_ptr::make_shared<SHAMapTxPlusMetaLeafNode>(item_, cowid, hash_);
|
||||||
|
|||||||
@@ -11,5 +11,5 @@ using TreeNodeCache = TaggedCache<
|
|||||||
SHAMapTreeNode,
|
SHAMapTreeNode,
|
||||||
/*IsKeyCache*/ false,
|
/*IsKeyCache*/ false,
|
||||||
intr_ptr::SharedWeakUnionPtr<SHAMapTreeNode>,
|
intr_ptr::SharedWeakUnionPtr<SHAMapTreeNode>,
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>>;
|
SHAMapTreeNodePtr>;
|
||||||
} // namespace xrpl
|
} // namespace xrpl
|
||||||
|
|||||||
@@ -148,7 +148,7 @@ public:
|
|||||||
/** Get the number of elements in each array and a pointer to the start
|
/** Get the number of elements in each array and a pointer to the start
|
||||||
of each array.
|
of each array.
|
||||||
*/
|
*/
|
||||||
[[nodiscard]] std::tuple<std::uint8_t, SHAMapHash*, intr_ptr::SharedPtr<SHAMapTreeNode>*>
|
[[nodiscard]] std::tuple<std::uint8_t, SHAMapHash*, SHAMapTreeNodePtr*>
|
||||||
getHashesAndChildren() const;
|
getHashesAndChildren() const;
|
||||||
|
|
||||||
/** Get the `hashes` array */
|
/** Get the `hashes` array */
|
||||||
@@ -156,7 +156,7 @@ public:
|
|||||||
getHashes() const;
|
getHashes() const;
|
||||||
|
|
||||||
/** Get the `children` array */
|
/** Get the `children` array */
|
||||||
[[nodiscard]] intr_ptr::SharedPtr<SHAMapTreeNode>*
|
[[nodiscard]] SHAMapTreeNodePtr*
|
||||||
getChildren() const;
|
getChildren() const;
|
||||||
|
|
||||||
/** Call the `f` callback for all 16 (branchFactor) branches - even if
|
/** Call the `f` callback for all 16 (branchFactor) branches - even if
|
||||||
|
|||||||
@@ -25,8 +25,7 @@ static_assert(
|
|||||||
// Terminology: A chunk is the memory being allocated from a block. A block
|
// Terminology: A chunk is the memory being allocated from a block. A block
|
||||||
// contains multiple chunks. This is the terminology the boost documentation
|
// contains multiple chunks. This is the terminology the boost documentation
|
||||||
// uses. Pools use "Simple Segregated Storage" as their storage format.
|
// uses. Pools use "Simple Segregated Storage" as their storage format.
|
||||||
constexpr size_t elementSizeBytes =
|
constexpr size_t elementSizeBytes = (sizeof(SHAMapHash) + sizeof(SHAMapTreeNodePtr));
|
||||||
(sizeof(SHAMapHash) + sizeof(intr_ptr::SharedPtr<SHAMapTreeNode>));
|
|
||||||
|
|
||||||
constexpr size_t blockSizeBytes = kilobytes(512);
|
constexpr size_t blockSizeBytes = kilobytes(512);
|
||||||
|
|
||||||
@@ -363,8 +362,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
// keep
|
// keep
|
||||||
new (&dstHashes[dstIndex]) SHAMapHash{srcHashes[srcIndex]};
|
new (&dstHashes[dstIndex]) SHAMapHash{srcHashes[srcIndex]};
|
||||||
|
|
||||||
new (&dstChildren[dstIndex])
|
new (&dstChildren[dstIndex]) SHAMapTreeNodePtr{std::move(srcChildren[srcIndex])};
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>{std::move(srcChildren[srcIndex])};
|
|
||||||
++dstIndex;
|
++dstIndex;
|
||||||
++srcIndex;
|
++srcIndex;
|
||||||
}
|
}
|
||||||
@@ -375,7 +373,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
if (dstIsDense)
|
if (dstIsDense)
|
||||||
{
|
{
|
||||||
new (&dstHashes[dstIndex]) SHAMapHash{};
|
new (&dstHashes[dstIndex]) SHAMapHash{};
|
||||||
new (&dstChildren[dstIndex]) intr_ptr::SharedPtr<SHAMapTreeNode>{};
|
new (&dstChildren[dstIndex]) SHAMapTreeNodePtr{};
|
||||||
++dstIndex;
|
++dstIndex;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -383,7 +381,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
{
|
{
|
||||||
// add
|
// add
|
||||||
new (&dstHashes[dstIndex]) SHAMapHash{};
|
new (&dstHashes[dstIndex]) SHAMapHash{};
|
||||||
new (&dstChildren[dstIndex]) intr_ptr::SharedPtr<SHAMapTreeNode>{};
|
new (&dstChildren[dstIndex]) SHAMapTreeNodePtr{};
|
||||||
++dstIndex;
|
++dstIndex;
|
||||||
if (srcIsDense)
|
if (srcIsDense)
|
||||||
{
|
{
|
||||||
@@ -396,7 +394,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
if (dstIsDense)
|
if (dstIsDense)
|
||||||
{
|
{
|
||||||
new (&dstHashes[dstIndex]) SHAMapHash{};
|
new (&dstHashes[dstIndex]) SHAMapHash{};
|
||||||
new (&dstChildren[dstIndex]) intr_ptr::SharedPtr<SHAMapTreeNode>{};
|
new (&dstChildren[dstIndex]) SHAMapTreeNodePtr{};
|
||||||
++dstIndex;
|
++dstIndex;
|
||||||
}
|
}
|
||||||
if (srcIsDense)
|
if (srcIsDense)
|
||||||
@@ -413,7 +411,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
for (int i = dstIndex; i < dstNumAllocated; ++i)
|
for (int i = dstIndex; i < dstNumAllocated; ++i)
|
||||||
{
|
{
|
||||||
new (&dstHashes[i]) SHAMapHash{};
|
new (&dstHashes[i]) SHAMapHash{};
|
||||||
new (&dstChildren[i]) intr_ptr::SharedPtr<SHAMapTreeNode>{};
|
new (&dstChildren[i]) SHAMapTreeNodePtr{};
|
||||||
}
|
}
|
||||||
*this = std::move(dst);
|
*this = std::move(dst);
|
||||||
}
|
}
|
||||||
@@ -433,7 +431,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
// allocate hashes and children, but do not run constructors
|
// allocate hashes and children, but do not run constructors
|
||||||
TaggedPointer newHashesAndChildren{RawAllocateTag{}, toAllocate};
|
TaggedPointer newHashesAndChildren{RawAllocateTag{}, toAllocate};
|
||||||
SHAMapHash *newHashes, *oldHashes;
|
SHAMapHash *newHashes, *oldHashes;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>*newChildren, *oldChildren;
|
SHAMapTreeNodePtr *newChildren, *oldChildren;
|
||||||
std::uint8_t newNumAllocated;
|
std::uint8_t newNumAllocated;
|
||||||
// structured bindings can't be captured in c++ 17; use tie instead
|
// structured bindings can't be captured in c++ 17; use tie instead
|
||||||
std::tie(newNumAllocated, newHashes, newChildren) = newHashesAndChildren.getHashesAndChildren();
|
std::tie(newNumAllocated, newHashes, newChildren) = newHashesAndChildren.getHashesAndChildren();
|
||||||
@@ -444,8 +442,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
// new arrays are dense, old arrays are sparse
|
// new arrays are dense, old arrays are sparse
|
||||||
iterNonEmptyChildIndexes(isBranch, [&](auto branchNum, auto indexNum) {
|
iterNonEmptyChildIndexes(isBranch, [&](auto branchNum, auto indexNum) {
|
||||||
new (&newHashes[branchNum]) SHAMapHash{oldHashes[indexNum]};
|
new (&newHashes[branchNum]) SHAMapHash{oldHashes[indexNum]};
|
||||||
new (&newChildren[branchNum])
|
new (&newChildren[branchNum]) SHAMapTreeNodePtr{std::move(oldChildren[indexNum])};
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>{std::move(oldChildren[indexNum])};
|
|
||||||
});
|
});
|
||||||
// Run the constructors for the remaining elements
|
// Run the constructors for the remaining elements
|
||||||
for (int i = 0; i < SHAMapInnerNode::branchFactor; ++i)
|
for (int i = 0; i < SHAMapInnerNode::branchFactor; ++i)
|
||||||
@@ -453,7 +450,7 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
if ((1 << i) & isBranch)
|
if ((1 << i) & isBranch)
|
||||||
continue;
|
continue;
|
||||||
new (&newHashes[i]) SHAMapHash{};
|
new (&newHashes[i]) SHAMapHash{};
|
||||||
new (&newChildren[i]) intr_ptr::SharedPtr<SHAMapTreeNode>{};
|
new (&newChildren[i]) SHAMapTreeNodePtr{};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else
|
else
|
||||||
@@ -463,14 +460,14 @@ inline TaggedPointer::TaggedPointer(
|
|||||||
iterNonEmptyChildIndexes(isBranch, [&](auto branchNum, auto indexNum) {
|
iterNonEmptyChildIndexes(isBranch, [&](auto branchNum, auto indexNum) {
|
||||||
new (&newHashes[curCompressedIndex]) SHAMapHash{oldHashes[indexNum]};
|
new (&newHashes[curCompressedIndex]) SHAMapHash{oldHashes[indexNum]};
|
||||||
new (&newChildren[curCompressedIndex])
|
new (&newChildren[curCompressedIndex])
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>{std::move(oldChildren[indexNum])};
|
SHAMapTreeNodePtr{std::move(oldChildren[indexNum])};
|
||||||
++curCompressedIndex;
|
++curCompressedIndex;
|
||||||
});
|
});
|
||||||
// Run the constructors for the remaining elements
|
// Run the constructors for the remaining elements
|
||||||
for (int i = curCompressedIndex; i < newNumAllocated; ++i)
|
for (int i = curCompressedIndex; i < newNumAllocated; ++i)
|
||||||
{
|
{
|
||||||
new (&newHashes[i]) SHAMapHash{};
|
new (&newHashes[i]) SHAMapHash{};
|
||||||
new (&newChildren[i]) intr_ptr::SharedPtr<SHAMapTreeNode>{};
|
new (&newChildren[i]) SHAMapTreeNodePtr{};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -484,7 +481,7 @@ inline TaggedPointer::TaggedPointer(std::uint8_t numChildren)
|
|||||||
for (std::size_t i = 0; i < numAllocated; ++i)
|
for (std::size_t i = 0; i < numAllocated; ++i)
|
||||||
{
|
{
|
||||||
new (&hashes[i]) SHAMapHash{};
|
new (&hashes[i]) SHAMapHash{};
|
||||||
new (&children[i]) intr_ptr::SharedPtr<SHAMapTreeNode>{};
|
new (&children[i]) SHAMapTreeNodePtr{};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -522,14 +519,13 @@ TaggedPointer::isDense() const
|
|||||||
return (tp_ & tagMask) == boundaries.size() - 1;
|
return (tp_ & tagMask) == boundaries.size() - 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
[[nodiscard]] inline std::tuple<std::uint8_t, SHAMapHash*, intr_ptr::SharedPtr<SHAMapTreeNode>*>
|
[[nodiscard]] inline std::tuple<std::uint8_t, SHAMapHash*, SHAMapTreeNodePtr*>
|
||||||
TaggedPointer::getHashesAndChildren() const
|
TaggedPointer::getHashesAndChildren() const
|
||||||
{
|
{
|
||||||
auto const [tag, ptr] = decode();
|
auto const [tag, ptr] = decode();
|
||||||
auto const hashes = reinterpret_cast<SHAMapHash*>(ptr);
|
auto const hashes = reinterpret_cast<SHAMapHash*>(ptr);
|
||||||
std::uint8_t numAllocated = boundaries[tag];
|
std::uint8_t numAllocated = boundaries[tag];
|
||||||
auto const children =
|
auto const children = reinterpret_cast<SHAMapTreeNodePtr*>(hashes + numAllocated);
|
||||||
reinterpret_cast<intr_ptr::SharedPtr<SHAMapTreeNode>*>(hashes + numAllocated);
|
|
||||||
return {numAllocated, hashes, children};
|
return {numAllocated, hashes, children};
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -539,7 +535,7 @@ TaggedPointer::getHashes() const
|
|||||||
return reinterpret_cast<SHAMapHash*>(tp_ & ptrMask);
|
return reinterpret_cast<SHAMapHash*>(tp_ & ptrMask);
|
||||||
};
|
};
|
||||||
|
|
||||||
[[nodiscard]] inline intr_ptr::SharedPtr<SHAMapTreeNode>*
|
[[nodiscard]] inline SHAMapTreeNodePtr*
|
||||||
TaggedPointer::getChildren() const
|
TaggedPointer::getChildren() const
|
||||||
{
|
{
|
||||||
auto [unused1, unused2, result] = getHashesAndChildren();
|
auto [unused1, unused2, result] = getHashesAndChildren();
|
||||||
|
|||||||
@@ -97,10 +97,7 @@ SHAMap::snapShot(bool isMutable) const
|
|||||||
}
|
}
|
||||||
|
|
||||||
void
|
void
|
||||||
SHAMap::dirtyUp(
|
SHAMap::dirtyUp(SharedPtrNodeStack& stack, uint256 const& target, SHAMapTreeNodePtr child)
|
||||||
SharedPtrNodeStack& stack,
|
|
||||||
uint256 const& target,
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> child)
|
|
||||||
{
|
{
|
||||||
// walk the tree up from through the inner nodes to the root_
|
// walk the tree up from through the inner nodes to the root_
|
||||||
// update hashes and links
|
// update hashes and links
|
||||||
@@ -165,7 +162,7 @@ SHAMap::findKey(uint256 const& id) const
|
|||||||
return leaf;
|
return leaf;
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::fetchNodeFromDB(SHAMapHash const& hash) const
|
SHAMap::fetchNodeFromDB(SHAMapHash const& hash) const
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(backed_, "xrpl::SHAMap::fetchNodeFromDB : is backed");
|
XRPL_ASSERT(backed_, "xrpl::SHAMap::fetchNodeFromDB : is backed");
|
||||||
@@ -173,7 +170,7 @@ SHAMap::fetchNodeFromDB(SHAMapHash const& hash) const
|
|||||||
return finishFetch(hash, obj);
|
return finishFetch(hash, obj);
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::finishFetch(SHAMapHash const& hash, std::shared_ptr<NodeObject> const& object) const
|
SHAMap::finishFetch(SHAMapHash const& hash, std::shared_ptr<NodeObject> const& object) const
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(backed_, "xrpl::SHAMap::finishFetch : is backed");
|
XRPL_ASSERT(backed_, "xrpl::SHAMap::finishFetch : is backed");
|
||||||
@@ -208,8 +205,8 @@ SHAMap::finishFetch(SHAMapHash const& hash, std::shared_ptr<NodeObject> const& o
|
|||||||
}
|
}
|
||||||
|
|
||||||
// See if a sync filter has a node
|
// See if a sync filter has a node
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::checkFilter(SHAMapHash const& hash, SHAMapSyncFilter* filter) const
|
SHAMap::checkFilter(SHAMapHash const& hash, SHAMapSyncFilter const* filter) const
|
||||||
{
|
{
|
||||||
if (auto nodeData = filter->getNode(hash))
|
if (auto nodeData = filter->getNode(hash))
|
||||||
{
|
{
|
||||||
@@ -234,8 +231,8 @@ SHAMap::checkFilter(SHAMapHash const& hash, SHAMapSyncFilter* filter) const
|
|||||||
|
|
||||||
// Get a node without throwing
|
// Get a node without throwing
|
||||||
// Used on maps where missing nodes are expected
|
// Used on maps where missing nodes are expected
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::fetchNodeNT(SHAMapHash const& hash, SHAMapSyncFilter* filter) const
|
SHAMap::fetchNodeNT(SHAMapHash const& hash, SHAMapSyncFilter const* filter) const
|
||||||
{
|
{
|
||||||
auto node = cacheLookup(hash);
|
auto node = cacheLookup(hash);
|
||||||
if (node)
|
if (node)
|
||||||
@@ -257,7 +254,7 @@ SHAMap::fetchNodeNT(SHAMapHash const& hash, SHAMapSyncFilter* filter) const
|
|||||||
return node;
|
return node;
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::fetchNodeNT(SHAMapHash const& hash) const
|
SHAMap::fetchNodeNT(SHAMapHash const& hash) const
|
||||||
{
|
{
|
||||||
auto node = cacheLookup(hash);
|
auto node = cacheLookup(hash);
|
||||||
@@ -269,7 +266,7 @@ SHAMap::fetchNodeNT(SHAMapHash const& hash) const
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Throw if the node is missing
|
// Throw if the node is missing
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::fetchNode(SHAMapHash const& hash) const
|
SHAMap::fetchNode(SHAMapHash const& hash) const
|
||||||
{
|
{
|
||||||
auto node = fetchNodeNT(hash);
|
auto node = fetchNodeNT(hash);
|
||||||
@@ -291,10 +288,10 @@ SHAMap::descendThrow(SHAMapInnerNode* parent, int branch) const
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::descendThrow(SHAMapInnerNode& parent, int branch) const
|
SHAMap::descendThrow(SHAMapInnerNode& parent, int branch) const
|
||||||
{
|
{
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> ret = descend(parent, branch);
|
SHAMapTreeNodePtr ret = descend(parent, branch);
|
||||||
|
|
||||||
if (!ret && !parent.isEmptyBranch(branch))
|
if (!ret && !parent.isEmptyBranch(branch))
|
||||||
Throw<SHAMapMissingNode>(type_, parent.getChildHash(branch));
|
Throw<SHAMapMissingNode>(type_, parent.getChildHash(branch));
|
||||||
@@ -309,7 +306,7 @@ SHAMap::descend(SHAMapInnerNode* parent, int branch) const
|
|||||||
if ((ret != nullptr) || !backed_)
|
if ((ret != nullptr) || !backed_)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> node = fetchNodeNT(parent->getChildHash(branch));
|
SHAMapTreeNodePtr node = fetchNodeNT(parent->getChildHash(branch));
|
||||||
if (!node)
|
if (!node)
|
||||||
return nullptr;
|
return nullptr;
|
||||||
|
|
||||||
@@ -317,10 +314,10 @@ SHAMap::descend(SHAMapInnerNode* parent, int branch) const
|
|||||||
return node.get();
|
return node.get();
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::descend(SHAMapInnerNode& parent, int branch) const
|
SHAMap::descend(SHAMapInnerNode& parent, int branch) const
|
||||||
{
|
{
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> node = parent.getChild(branch);
|
SHAMapTreeNodePtr node = parent.getChild(branch);
|
||||||
if (node || !backed_)
|
if (node || !backed_)
|
||||||
return node;
|
return node;
|
||||||
|
|
||||||
@@ -334,10 +331,10 @@ SHAMap::descend(SHAMapInnerNode& parent, int branch) const
|
|||||||
|
|
||||||
// Gets the node that would be hooked to this branch,
|
// Gets the node that would be hooked to this branch,
|
||||||
// but doesn't hook it up.
|
// but doesn't hook it up.
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::descendNoStore(SHAMapInnerNode& parent, int branch) const
|
SHAMap::descendNoStore(SHAMapInnerNode& parent, int branch) const
|
||||||
{
|
{
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> ret = parent.getChild(branch);
|
SHAMapTreeNodePtr ret = parent.getChild(branch);
|
||||||
if (!ret && backed_)
|
if (!ret && backed_)
|
||||||
ret = fetchNode(parent.getChildHash(branch));
|
ret = fetchNode(parent.getChildHash(branch));
|
||||||
return ret;
|
return ret;
|
||||||
@@ -348,7 +345,7 @@ SHAMap::descend(
|
|||||||
SHAMapInnerNode* parent,
|
SHAMapInnerNode* parent,
|
||||||
SHAMapNodeID const& parentID,
|
SHAMapNodeID const& parentID,
|
||||||
int branch,
|
int branch,
|
||||||
SHAMapSyncFilter* filter) const
|
SHAMapSyncFilter const* filter) const
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(parent->isInner(), "xrpl::SHAMap::descend : valid parent input");
|
XRPL_ASSERT(parent->isInner(), "xrpl::SHAMap::descend : valid parent input");
|
||||||
XRPL_ASSERT(
|
XRPL_ASSERT(
|
||||||
@@ -361,7 +358,7 @@ SHAMap::descend(
|
|||||||
if (child == nullptr)
|
if (child == nullptr)
|
||||||
{
|
{
|
||||||
auto const& childHash = parent->getChildHash(branch);
|
auto const& childHash = parent->getChildHash(branch);
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> childNode = fetchNodeNT(childHash, filter);
|
SHAMapTreeNodePtr childNode = fetchNodeNT(childHash, filter);
|
||||||
|
|
||||||
if (childNode)
|
if (childNode)
|
||||||
{
|
{
|
||||||
@@ -377,7 +374,7 @@ SHAMapTreeNode*
|
|||||||
SHAMap::descendAsync(
|
SHAMap::descendAsync(
|
||||||
SHAMapInnerNode* parent,
|
SHAMapInnerNode* parent,
|
||||||
int branch,
|
int branch,
|
||||||
SHAMapSyncFilter* filter,
|
SHAMapSyncFilter const* filter,
|
||||||
bool& pending,
|
bool& pending,
|
||||||
descendCallback&& callback) const
|
descendCallback&& callback) const
|
||||||
{
|
{
|
||||||
@@ -434,7 +431,7 @@ SHAMap::unshareNode(intr_ptr::SharedPtr<Node> node, SHAMapNodeID const& nodeID)
|
|||||||
|
|
||||||
SHAMapLeafNode*
|
SHAMapLeafNode*
|
||||||
SHAMap::belowHelper(
|
SHAMap::belowHelper(
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> node,
|
SHAMapTreeNodePtr node,
|
||||||
SharedPtrNodeStack& stack,
|
SharedPtrNodeStack& stack,
|
||||||
int branch,
|
int branch,
|
||||||
std::tuple<int, std::function<bool(int)>, std::function<void(int&)>> const& loopParams) const
|
std::tuple<int, std::function<bool(int)>, std::function<void(int&)>> const& loopParams) const
|
||||||
@@ -479,8 +476,7 @@ SHAMap::belowHelper(
|
|||||||
return nullptr;
|
return nullptr;
|
||||||
}
|
}
|
||||||
SHAMapLeafNode*
|
SHAMapLeafNode*
|
||||||
SHAMap::lastBelow(intr_ptr::SharedPtr<SHAMapTreeNode> node, SharedPtrNodeStack& stack, int branch)
|
SHAMap::lastBelow(SHAMapTreeNodePtr node, SharedPtrNodeStack& stack, int branch) const
|
||||||
const
|
|
||||||
{
|
{
|
||||||
auto init = branchFactor - 1;
|
auto init = branchFactor - 1;
|
||||||
auto cmp = [](int i) { return i >= 0; };
|
auto cmp = [](int i) { return i >= 0; };
|
||||||
@@ -489,8 +485,7 @@ SHAMap::lastBelow(intr_ptr::SharedPtr<SHAMapTreeNode> node, SharedPtrNodeStack&
|
|||||||
return belowHelper(node, stack, branch, {init, cmp, incr});
|
return belowHelper(node, stack, branch, {init, cmp, incr});
|
||||||
}
|
}
|
||||||
SHAMapLeafNode*
|
SHAMapLeafNode*
|
||||||
SHAMap::firstBelow(intr_ptr::SharedPtr<SHAMapTreeNode> node, SharedPtrNodeStack& stack, int branch)
|
SHAMap::firstBelow(SHAMapTreeNodePtr node, SharedPtrNodeStack& stack, int branch) const
|
||||||
const
|
|
||||||
{
|
{
|
||||||
auto init = 0;
|
auto init = 0;
|
||||||
auto cmp = [](int i) { return i <= branchFactor; };
|
auto cmp = [](int i) { return i <= branchFactor; };
|
||||||
@@ -699,10 +694,8 @@ SHAMap::delItem(uint256 const& id)
|
|||||||
|
|
||||||
SHAMapNodeType const type = leaf->getType();
|
SHAMapNodeType const type = leaf->getType();
|
||||||
|
|
||||||
using TreeNodeType = intr_ptr::SharedPtr<SHAMapTreeNode>;
|
|
||||||
|
|
||||||
// What gets attached to the end of the chain (For now, nothing, since we deleted the leaf)
|
// What gets attached to the end of the chain (For now, nothing, since we deleted the leaf)
|
||||||
TreeNodeType prevNode;
|
SHAMapTreeNodePtr prevNode;
|
||||||
|
|
||||||
while (!stack.empty())
|
while (!stack.empty())
|
||||||
{
|
{
|
||||||
@@ -728,7 +721,7 @@ SHAMap::delItem(uint256 const& id)
|
|||||||
// no children below this branch
|
// no children below this branch
|
||||||
//
|
//
|
||||||
// Note: This is unnecessary due to the std::move above but left here for safety
|
// Note: This is unnecessary due to the std::move above but left here for safety
|
||||||
prevNode = TreeNodeType{};
|
prevNode = SHAMapTreeNodePtr{};
|
||||||
}
|
}
|
||||||
else if (bc == 1)
|
else if (bc == 1)
|
||||||
{
|
{
|
||||||
@@ -741,7 +734,7 @@ SHAMap::delItem(uint256 const& id)
|
|||||||
{
|
{
|
||||||
if (!node->isEmptyBranch(i))
|
if (!node->isEmptyBranch(i))
|
||||||
{
|
{
|
||||||
node->setChild(i, TreeNodeType{});
|
node->setChild(i, SHAMapTreeNodePtr{});
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -937,8 +930,8 @@ SHAMap::fetchRoot(SHAMapHash const& hash, SHAMapSyncFilter* filter)
|
|||||||
@note The node must have already been unshared by having the caller
|
@note The node must have already been unshared by having the caller
|
||||||
first call SHAMapTreeNode::unshare().
|
first call SHAMapTreeNode::unshare().
|
||||||
*/
|
*/
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::writeNode(NodeObjectType t, intr_ptr::SharedPtr<SHAMapTreeNode> node) const
|
SHAMap::writeNode(NodeObjectType t, SHAMapTreeNodePtr node) const
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(node->cowid() == 0, "xrpl::SHAMap::writeNode : valid input node");
|
XRPL_ASSERT(node->cowid() == 0, "xrpl::SHAMap::writeNode : valid input node");
|
||||||
XRPL_ASSERT(backed_, "xrpl::SHAMap::writeNode : is backed");
|
XRPL_ASSERT(backed_, "xrpl::SHAMap::writeNode : is backed");
|
||||||
@@ -1155,7 +1148,7 @@ SHAMap::dump(bool hash) const
|
|||||||
JLOG(journal_.info()) << leafCount << " resident leaves";
|
JLOG(journal_.info()) << leafCount << " resident leaves";
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMap::cacheLookup(SHAMapHash const& hash) const
|
SHAMap::cacheLookup(SHAMapHash const& hash) const
|
||||||
{
|
{
|
||||||
auto ret = f_.getTreeNodeCache()->fetch(hash.as_uint256());
|
auto ret = f_.getTreeNodeCache()->fetch(hash.as_uint256());
|
||||||
@@ -1164,7 +1157,7 @@ SHAMap::cacheLookup(SHAMapHash const& hash) const
|
|||||||
}
|
}
|
||||||
|
|
||||||
void
|
void
|
||||||
SHAMap::canonicalize(SHAMapHash const& hash, intr_ptr::SharedPtr<SHAMapTreeNode>& node) const
|
SHAMap::canonicalize(SHAMapHash const& hash, SHAMapTreeNodePtr& node) const
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(backed_, "xrpl::SHAMap::canonicalize : is backed");
|
XRPL_ASSERT(backed_, "xrpl::SHAMap::canonicalize : is backed");
|
||||||
XRPL_ASSERT(node->cowid() == 0, "xrpl::SHAMap::canonicalize : valid node input");
|
XRPL_ASSERT(node->cowid() == 0, "xrpl::SHAMap::canonicalize : valid node input");
|
||||||
|
|||||||
@@ -261,7 +261,7 @@ SHAMap::walkMap(std::vector<SHAMapMissingNode>& missingNodes, int maxMissing) co
|
|||||||
{
|
{
|
||||||
if (!node->isEmptyBranch(i))
|
if (!node->isEmptyBranch(i))
|
||||||
{
|
{
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> const nextNode = descendNoStore(*node, i);
|
SHAMapTreeNodePtr const nextNode = descendNoStore(*node, i);
|
||||||
|
|
||||||
if (nextNode)
|
if (nextNode)
|
||||||
{
|
{
|
||||||
@@ -286,7 +286,7 @@ SHAMap::walkMapParallel(std::vector<SHAMapMissingNode>& missingNodes, int maxMis
|
|||||||
return false;
|
return false;
|
||||||
|
|
||||||
using StackEntry = intr_ptr::SharedPtr<SHAMapInnerNode>;
|
using StackEntry = intr_ptr::SharedPtr<SHAMapInnerNode>;
|
||||||
std::array<intr_ptr::SharedPtr<SHAMapTreeNode>, 16> topChildren;
|
std::array<SHAMapTreeNodePtr, 16> topChildren;
|
||||||
{
|
{
|
||||||
auto const& innerRoot = intr_ptr::static_pointer_cast<SHAMapInnerNode>(root_);
|
auto const& innerRoot = intr_ptr::static_pointer_cast<SHAMapInnerNode>(root_);
|
||||||
for (int i = 0; i < 16; ++i)
|
for (int i = 0; i < 16; ++i)
|
||||||
@@ -331,8 +331,7 @@ SHAMap::walkMapParallel(std::vector<SHAMapMissingNode>& missingNodes, int maxMis
|
|||||||
{
|
{
|
||||||
if (node->isEmptyBranch(i))
|
if (node->isEmptyBranch(i))
|
||||||
continue;
|
continue;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> const nextNode =
|
SHAMapTreeNodePtr const nextNode = descendNoStore(*node, i);
|
||||||
descendNoStore(*node, i);
|
|
||||||
|
|
||||||
if (nextNode)
|
if (nextNode)
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -37,7 +37,7 @@ SHAMapInnerNode::~SHAMapInnerNode() = default;
|
|||||||
void
|
void
|
||||||
SHAMapInnerNode::partialDestructor()
|
SHAMapInnerNode::partialDestructor()
|
||||||
{
|
{
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>* children = nullptr;
|
SHAMapTreeNodePtr* children = nullptr;
|
||||||
// structured bindings can't be captured in c++ 17; use tie instead
|
// structured bindings can't be captured in c++ 17; use tie instead
|
||||||
std::tie(std::ignore, std::ignore, children) = hashesAndChildren_.getHashesAndChildren();
|
std::tie(std::ignore, std::ignore, children) = hashesAndChildren_.getHashesAndChildren();
|
||||||
iterNonEmptyChildIndexes([&](auto branchNum, auto indexNum) { children[indexNum].reset(); });
|
iterNonEmptyChildIndexes([&](auto branchNum, auto indexNum) { children[indexNum].reset(); });
|
||||||
@@ -69,7 +69,7 @@ SHAMapInnerNode::getChildIndex(int i) const
|
|||||||
return hashesAndChildren_.getChildIndex(isBranch_, i);
|
return hashesAndChildren_.getChildIndex(isBranch_, i);
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapInnerNode::clone(std::uint32_t cowid) const
|
SHAMapInnerNode::clone(std::uint32_t cowid) const
|
||||||
{
|
{
|
||||||
auto const branchCount = getBranchCount();
|
auto const branchCount = getBranchCount();
|
||||||
@@ -79,7 +79,7 @@ SHAMapInnerNode::clone(std::uint32_t cowid) const
|
|||||||
p->isBranch_ = isBranch_;
|
p->isBranch_ = isBranch_;
|
||||||
p->fullBelowGen_ = fullBelowGen_;
|
p->fullBelowGen_ = fullBelowGen_;
|
||||||
SHAMapHash *cloneHashes = nullptr, *thisHashes = nullptr;
|
SHAMapHash *cloneHashes = nullptr, *thisHashes = nullptr;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>*cloneChildren = nullptr, *thisChildren = nullptr;
|
SHAMapTreeNodePtr *cloneChildren = nullptr, *thisChildren = nullptr;
|
||||||
// structured bindings can't be captured in c++ 17; use tie instead
|
// structured bindings can't be captured in c++ 17; use tie instead
|
||||||
std::tie(std::ignore, cloneHashes, cloneChildren) =
|
std::tie(std::ignore, cloneHashes, cloneChildren) =
|
||||||
p->hashesAndChildren_.getHashesAndChildren();
|
p->hashesAndChildren_.getHashesAndChildren();
|
||||||
@@ -118,7 +118,7 @@ SHAMapInnerNode::clone(std::uint32_t cowid) const
|
|||||||
return p;
|
return p;
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapInnerNode::makeFullInner(Slice data, SHAMapHash const& hash, bool hashValid)
|
SHAMapInnerNode::makeFullInner(Slice data, SHAMapHash const& hash, bool hashValid)
|
||||||
{
|
{
|
||||||
// A full inner node is serialized as 16 256-bit hashes, back to back:
|
// A full inner node is serialized as 16 256-bit hashes, back to back:
|
||||||
@@ -153,7 +153,7 @@ SHAMapInnerNode::makeFullInner(Slice data, SHAMapHash const& hash, bool hashVali
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapInnerNode::makeCompressedInner(Slice data)
|
SHAMapInnerNode::makeCompressedInner(Slice data)
|
||||||
{
|
{
|
||||||
// A compressed inner node is serialized as a series of 33 byte chunks,
|
// A compressed inner node is serialized as a series of 33 byte chunks,
|
||||||
@@ -207,7 +207,7 @@ void
|
|||||||
SHAMapInnerNode::updateHashDeep()
|
SHAMapInnerNode::updateHashDeep()
|
||||||
{
|
{
|
||||||
SHAMapHash* hashes = nullptr;
|
SHAMapHash* hashes = nullptr;
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>* children = nullptr;
|
SHAMapTreeNodePtr* children = nullptr;
|
||||||
// structured bindings can't be captured in c++ 17; use tie instead
|
// structured bindings can't be captured in c++ 17; use tie instead
|
||||||
std::tie(std::ignore, hashes, children) = hashesAndChildren_.getHashesAndChildren();
|
std::tie(std::ignore, hashes, children) = hashesAndChildren_.getHashesAndChildren();
|
||||||
iterNonEmptyChildIndexes([&](auto branchNum, auto indexNum) {
|
iterNonEmptyChildIndexes([&](auto branchNum, auto indexNum) {
|
||||||
@@ -265,7 +265,7 @@ SHAMapInnerNode::getString(SHAMapNodeID const& id) const
|
|||||||
|
|
||||||
// We are modifying an inner node
|
// We are modifying an inner node
|
||||||
void
|
void
|
||||||
SHAMapInnerNode::setChild(int m, intr_ptr::SharedPtr<SHAMapTreeNode> child)
|
SHAMapInnerNode::setChild(int m, SHAMapTreeNodePtr child)
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(
|
XRPL_ASSERT(
|
||||||
(m >= 0) && (m < branchFactor), "xrpl::SHAMapInnerNode::setChild : valid branch input");
|
(m >= 0) && (m < branchFactor), "xrpl::SHAMapInnerNode::setChild : valid branch input");
|
||||||
@@ -307,7 +307,7 @@ SHAMapInnerNode::setChild(int m, intr_ptr::SharedPtr<SHAMapTreeNode> child)
|
|||||||
|
|
||||||
// finished modifying, now make shareable
|
// finished modifying, now make shareable
|
||||||
void
|
void
|
||||||
SHAMapInnerNode::shareChild(int m, intr_ptr::SharedPtr<SHAMapTreeNode> const& child)
|
SHAMapInnerNode::shareChild(int m, SHAMapTreeNodePtr const& child)
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(
|
XRPL_ASSERT(
|
||||||
(m >= 0) && (m < branchFactor), "xrpl::SHAMapInnerNode::shareChild : valid branch input");
|
(m >= 0) && (m < branchFactor), "xrpl::SHAMapInnerNode::shareChild : valid branch input");
|
||||||
@@ -337,7 +337,7 @@ SHAMapInnerNode::getChildPointer(int branch)
|
|||||||
return hashesAndChildren_.getChildren()[index].get();
|
return hashesAndChildren_.getChildren()[index].get();
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapInnerNode::getChild(int branch)
|
SHAMapInnerNode::getChild(int branch)
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(
|
XRPL_ASSERT(
|
||||||
@@ -364,8 +364,8 @@ SHAMapInnerNode::getChildHash(int m) const
|
|||||||
return zeroSHAMapHash;
|
return zeroSHAMapHash;
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapInnerNode::canonicalizeChild(int branch, intr_ptr::SharedPtr<SHAMapTreeNode> node)
|
SHAMapInnerNode::canonicalizeChild(int branch, SHAMapTreeNodePtr node)
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(
|
XRPL_ASSERT(
|
||||||
branch >= 0 && branch < branchFactor,
|
branch >= 0 && branch < branchFactor,
|
||||||
|
|||||||
@@ -129,7 +129,9 @@ selectBranch(SHAMapNodeID const& id, uint256 const& hash)
|
|||||||
SHAMapNodeID
|
SHAMapNodeID
|
||||||
SHAMapNodeID::createID(int depth, uint256 const& key)
|
SHAMapNodeID::createID(int depth, uint256 const& key)
|
||||||
{
|
{
|
||||||
XRPL_ASSERT((depth >= 0) && (depth < 65), "xrpl::SHAMapNodeID::createID : valid branch input");
|
XRPL_ASSERT(
|
||||||
|
depth >= 0 && depth <= SHAMap::leafDepth,
|
||||||
|
"xrpl::SHAMapNodeID::createID : valid branch input");
|
||||||
return SHAMapNodeID(depth, key & depthMask(depth));
|
return SHAMapNodeID(depth, key & depthMask(depth));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -66,7 +66,7 @@ SHAMap::visitNodes(std::function<bool(SHAMapTreeNode&)> const& function) const
|
|||||||
{
|
{
|
||||||
if (!node->isEmptyBranch(pos))
|
if (!node->isEmptyBranch(pos))
|
||||||
{
|
{
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> const child = descendNoStore(*node, pos);
|
SHAMapTreeNodePtr const child = descendNoStore(*node, pos);
|
||||||
if (!function(*child))
|
if (!function(*child))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
@@ -204,8 +204,7 @@ SHAMap::gmn_ProcessNodes(MissingNodes& mn, MissingNodes::StackEntry& se)
|
|||||||
branch,
|
branch,
|
||||||
mn.filter_,
|
mn.filter_,
|
||||||
pending,
|
pending,
|
||||||
[node, nodeID, branch, &mn](
|
[node, nodeID, branch, &mn](SHAMapTreeNodePtr found, SHAMapHash const&) {
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode> found, SHAMapHash const&) {
|
|
||||||
// a read completed asynchronously
|
// a read completed asynchronously
|
||||||
std::unique_lock<std::mutex> const lock{mn.deferLock_};
|
std::unique_lock<std::mutex> const lock{mn.deferLock_};
|
||||||
mn.finishedReads_.emplace_back(node, nodeID, branch, std::move(found));
|
mn.finishedReads_.emplace_back(node, nodeID, branch, std::move(found));
|
||||||
@@ -268,8 +267,7 @@ SHAMap::gmn_ProcessDeferredReads(MissingNodes& mn)
|
|||||||
int complete = 0;
|
int complete = 0;
|
||||||
while (complete != mn.deferred_)
|
while (complete != mn.deferred_)
|
||||||
{
|
{
|
||||||
std::tuple<SHAMapInnerNode*, SHAMapNodeID, int, intr_ptr::SharedPtr<SHAMapTreeNode>>
|
std::tuple<SHAMapInnerNode*, SHAMapNodeID, int, SHAMapTreeNodePtr> deferredNode;
|
||||||
deferredNode;
|
|
||||||
{
|
{
|
||||||
std::unique_lock<std::mutex> lock{mn.deferLock_};
|
std::unique_lock<std::mutex> lock{mn.deferLock_};
|
||||||
|
|
||||||
@@ -417,7 +415,7 @@ SHAMap::getMissingNodes(int max, SHAMapSyncFilter* filter)
|
|||||||
bool
|
bool
|
||||||
SHAMap::getNodeFat(
|
SHAMap::getNodeFat(
|
||||||
SHAMapNodeID const& wanted,
|
SHAMapNodeID const& wanted,
|
||||||
std::vector<std::pair<SHAMapNodeID, Blob>>& data,
|
std::vector<SHAMapNodeData>& data,
|
||||||
bool fatLeaves,
|
bool fatLeaves,
|
||||||
std::uint32_t depth) const
|
std::uint32_t depth) const
|
||||||
{
|
{
|
||||||
@@ -463,7 +461,7 @@ SHAMap::getNodeFat(
|
|||||||
// Add this node to the reply
|
// Add this node to the reply
|
||||||
s.erase();
|
s.erase();
|
||||||
node->serializeForWire(s);
|
node->serializeForWire(s);
|
||||||
data.emplace_back(nodeID, s.getData());
|
data.emplace_back(nodeID, s.getData(), node->isLeaf());
|
||||||
|
|
||||||
if (node->isInner())
|
if (node->isInner())
|
||||||
{
|
{
|
||||||
@@ -493,7 +491,7 @@ SHAMap::getNodeFat(
|
|||||||
// Just include this node
|
// Just include this node
|
||||||
s.erase();
|
s.erase();
|
||||||
childNode->serializeForWire(s);
|
childNode->serializeForWire(s);
|
||||||
data.emplace_back(childID, s.getData());
|
data.emplace_back(childID, s.getData(), childNode->isLeaf());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -511,8 +509,18 @@ SHAMap::serializeRoot(Serializer& s) const
|
|||||||
}
|
}
|
||||||
|
|
||||||
SHAMapAddNode
|
SHAMapAddNode
|
||||||
SHAMap::addRootNode(SHAMapHash const& hash, Slice const& rootNode, SHAMapSyncFilter* filter)
|
SHAMap::addRootNode(
|
||||||
|
SHAMapHash const& hash,
|
||||||
|
SHAMapTreeNodePtr rootNode,
|
||||||
|
SHAMapSyncFilter const* filter)
|
||||||
{
|
{
|
||||||
|
XRPL_ASSERT(rootNode, "xrpl::SHAMap::addRootNode : non-null root node");
|
||||||
|
if (!rootNode)
|
||||||
|
{
|
||||||
|
JLOG(journal_.error()) << "Null node received";
|
||||||
|
return SHAMapAddNode::invalid();
|
||||||
|
}
|
||||||
|
|
||||||
// we already have a root_ node
|
// we already have a root_ node
|
||||||
if (root_->getHash().isNonZero())
|
if (root_->getHash().isNonZero())
|
||||||
{
|
{
|
||||||
@@ -522,14 +530,16 @@ SHAMap::addRootNode(SHAMapHash const& hash, Slice const& rootNode, SHAMapSyncFil
|
|||||||
}
|
}
|
||||||
|
|
||||||
XRPL_ASSERT(cowid_ >= 1, "xrpl::SHAMap::addRootNode : valid cowid");
|
XRPL_ASSERT(cowid_ >= 1, "xrpl::SHAMap::addRootNode : valid cowid");
|
||||||
auto node = SHAMapTreeNode::makeFromWire(rootNode);
|
if (rootNode->getHash() != hash)
|
||||||
if (!node || node->getHash() != hash)
|
{
|
||||||
|
JLOG(journal_.warn()) << "Corrupt node received";
|
||||||
return SHAMapAddNode::invalid();
|
return SHAMapAddNode::invalid();
|
||||||
|
}
|
||||||
|
|
||||||
if (backed_)
|
if (backed_)
|
||||||
canonicalize(hash, node);
|
canonicalize(hash, rootNode);
|
||||||
|
|
||||||
root_ = node;
|
root_ = std::move(rootNode);
|
||||||
|
|
||||||
if (root_->isLeaf())
|
if (root_->isLeaf())
|
||||||
clearSynching();
|
clearSynching();
|
||||||
@@ -546,9 +556,23 @@ SHAMap::addRootNode(SHAMapHash const& hash, Slice const& rootNode, SHAMapSyncFil
|
|||||||
}
|
}
|
||||||
|
|
||||||
SHAMapAddNode
|
SHAMapAddNode
|
||||||
SHAMap::addKnownNode(SHAMapNodeID const& node, Slice const& rawNode, SHAMapSyncFilter* filter)
|
SHAMap::addKnownNode(
|
||||||
|
SHAMapNodeID const& nodeID,
|
||||||
|
SHAMapTreeNodePtr treeNode,
|
||||||
|
SHAMapSyncFilter const* filter)
|
||||||
{
|
{
|
||||||
XRPL_ASSERT(!node.isRoot(), "xrpl::SHAMap::addKnownNode : valid node input");
|
XRPL_ASSERT(!nodeID.isRoot(), "xrpl::SHAMap::addKnownNode : valid node input");
|
||||||
|
if (nodeID.isRoot())
|
||||||
|
{
|
||||||
|
JLOG(journal_.error()) << "Root node received";
|
||||||
|
return SHAMapAddNode::invalid();
|
||||||
|
}
|
||||||
|
XRPL_ASSERT(treeNode, "xrpl::SHAMap::addKnownNode : non-null tree node");
|
||||||
|
if (!treeNode)
|
||||||
|
{
|
||||||
|
JLOG(journal_.error()) << "Null node received";
|
||||||
|
return SHAMapAddNode::invalid();
|
||||||
|
}
|
||||||
|
|
||||||
if (!isSynching())
|
if (!isSynching())
|
||||||
{
|
{
|
||||||
@@ -562,14 +586,14 @@ SHAMap::addKnownNode(SHAMapNodeID const& node, Slice const& rawNode, SHAMapSyncF
|
|||||||
|
|
||||||
while (currNode->isInner() &&
|
while (currNode->isInner() &&
|
||||||
!safe_downcast<SHAMapInnerNode*>(currNode)->isFullBelow(generation) &&
|
!safe_downcast<SHAMapInnerNode*>(currNode)->isFullBelow(generation) &&
|
||||||
(currNodeID.getDepth() < node.getDepth()))
|
(currNodeID.getDepth() < nodeID.getDepth()))
|
||||||
{
|
{
|
||||||
int const branch = selectBranch(currNodeID, node.getNodeID());
|
int const branch = selectBranch(currNodeID, nodeID.getNodeID());
|
||||||
XRPL_ASSERT(branch >= 0, "xrpl::SHAMap::addKnownNode : valid branch");
|
XRPL_ASSERT(branch >= 0, "xrpl::SHAMap::addKnownNode : valid branch");
|
||||||
auto inner = safe_downcast<SHAMapInnerNode*>(currNode);
|
auto inner = safe_downcast<SHAMapInnerNode*>(currNode);
|
||||||
if (inner->isEmptyBranch(branch))
|
if (inner->isEmptyBranch(branch))
|
||||||
{
|
{
|
||||||
JLOG(journal_.warn()) << "Add known node for empty branch" << node;
|
JLOG(journal_.warn()) << "Add known node for empty branch" << nodeID;
|
||||||
return SHAMapAddNode::invalid();
|
return SHAMapAddNode::invalid();
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -585,67 +609,44 @@ SHAMap::addKnownNode(SHAMapNodeID const& node, Slice const& rawNode, SHAMapSyncF
|
|||||||
if (currNode != nullptr)
|
if (currNode != nullptr)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
auto newNode = SHAMapTreeNode::makeFromWire(rawNode);
|
if (childHash != treeNode->getHash())
|
||||||
|
|
||||||
if (!newNode || childHash != newNode->getHash())
|
|
||||||
{
|
{
|
||||||
JLOG(journal_.warn()) << "Corrupt node received";
|
JLOG(journal_.warn()) << "Corrupt node received";
|
||||||
return SHAMapAddNode::invalid();
|
return SHAMapAddNode::invalid();
|
||||||
}
|
}
|
||||||
|
|
||||||
// In rare cases, a node can still be corrupt even after hash
|
|
||||||
// validation. For leaf nodes, we perform an additional check to
|
|
||||||
// ensure the node's position in the tree is consistent with its
|
|
||||||
// content to prevent inconsistencies that could
|
|
||||||
// propagate further down the line.
|
|
||||||
if (newNode->isLeaf())
|
|
||||||
{
|
|
||||||
auto const& actualKey =
|
|
||||||
safe_downcast<SHAMapLeafNode const*>(newNode.get())->peekItem()->key();
|
|
||||||
|
|
||||||
// Validate that this leaf belongs at the target position
|
|
||||||
auto const expectedNodeID = SHAMapNodeID::createID(node.getDepth(), actualKey);
|
|
||||||
if (expectedNodeID.getNodeID() != node.getNodeID())
|
|
||||||
{
|
|
||||||
JLOG(journal_.debug())
|
|
||||||
<< "Leaf node position mismatch: "
|
|
||||||
<< "expected=" << expectedNodeID.getNodeID() << ", actual=" << node.getNodeID();
|
|
||||||
return SHAMapAddNode::invalid();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Inner nodes must be at a level strictly less than 64
|
// Inner nodes must be at a level strictly less than 64
|
||||||
// but leaf nodes (while notionally at level 64) can be
|
// but leaf nodes (while notionally at level 64) can be
|
||||||
// at any depth up to and including 64:
|
// at any depth up to and including 64:
|
||||||
if ((currNodeID.getDepth() > leafDepth) ||
|
if ((currNodeID.getDepth() > leafDepth) ||
|
||||||
(newNode->isInner() && currNodeID.getDepth() == leafDepth))
|
(treeNode->isInner() && currNodeID.getDepth() == leafDepth))
|
||||||
{
|
{
|
||||||
// Map is provably invalid
|
// Map is provably invalid
|
||||||
state_ = SHAMapState::Invalid;
|
state_ = SHAMapState::Invalid;
|
||||||
return SHAMapAddNode::useful();
|
return SHAMapAddNode::useful();
|
||||||
}
|
}
|
||||||
|
|
||||||
if (currNodeID != node)
|
if (currNodeID != nodeID)
|
||||||
{
|
{
|
||||||
// Either this node is broken or we didn't request it (yet)
|
// Either this node is broken or we didn't request it (yet)
|
||||||
JLOG(journal_.warn()) << "unable to hook node " << node;
|
JLOG(journal_.warn()) << "unable to hook node " << nodeID;
|
||||||
JLOG(journal_.info()) << " stuck at " << currNodeID;
|
JLOG(journal_.info()) << " stuck at " << currNodeID;
|
||||||
JLOG(journal_.info()) << "got depth=" << node.getDepth()
|
JLOG(journal_.info()) << "got depth=" << nodeID.getDepth()
|
||||||
<< ", walked to= " << currNodeID.getDepth();
|
<< ", walked to= " << currNodeID.getDepth();
|
||||||
return SHAMapAddNode::useful();
|
return SHAMapAddNode::useful();
|
||||||
}
|
}
|
||||||
|
|
||||||
if (backed_)
|
if (backed_)
|
||||||
canonicalize(childHash, newNode);
|
canonicalize(childHash, treeNode);
|
||||||
|
|
||||||
newNode = prevNode->canonicalizeChild(branch, std::move(newNode));
|
treeNode = prevNode->canonicalizeChild(branch, std::move(treeNode));
|
||||||
|
|
||||||
if (filter != nullptr)
|
if (filter != nullptr)
|
||||||
{
|
{
|
||||||
Serializer s;
|
Serializer s;
|
||||||
newNode->serializeWithPrefix(s);
|
treeNode->serializeWithPrefix(s);
|
||||||
filter->gotNode(
|
filter->gotNode(
|
||||||
false, childHash, ledgerSeq_, std::move(s.modData()), newNode->getType());
|
false, childHash, ledgerSeq_, std::move(s.modData()), treeNode->getType());
|
||||||
}
|
}
|
||||||
|
|
||||||
return SHAMapAddNode::useful();
|
return SHAMapAddNode::useful();
|
||||||
|
|||||||
@@ -25,7 +25,7 @@
|
|||||||
|
|
||||||
namespace xrpl {
|
namespace xrpl {
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapTreeNode::makeTransaction(Slice data, SHAMapHash const& hash, bool hashValid)
|
SHAMapTreeNode::makeTransaction(Slice data, SHAMapHash const& hash, bool hashValid)
|
||||||
{
|
{
|
||||||
auto item = make_shamapitem(sha512Half(HashPrefix::transactionID, data), data);
|
auto item = make_shamapitem(sha512Half(HashPrefix::transactionID, data), data);
|
||||||
@@ -36,7 +36,7 @@ SHAMapTreeNode::makeTransaction(Slice data, SHAMapHash const& hash, bool hashVal
|
|||||||
return intr_ptr::make_shared<SHAMapTxLeafNode>(std::move(item), 0);
|
return intr_ptr::make_shared<SHAMapTxLeafNode>(std::move(item), 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapTreeNode::makeTransactionWithMeta(Slice data, SHAMapHash const& hash, bool hashValid)
|
SHAMapTreeNode::makeTransactionWithMeta(Slice data, SHAMapHash const& hash, bool hashValid)
|
||||||
{
|
{
|
||||||
Serializer s(data.data(), data.size());
|
Serializer s(data.data(), data.size());
|
||||||
@@ -60,7 +60,7 @@ SHAMapTreeNode::makeTransactionWithMeta(Slice data, SHAMapHash const& hash, bool
|
|||||||
return intr_ptr::make_shared<SHAMapTxPlusMetaLeafNode>(std::move(item), 0);
|
return intr_ptr::make_shared<SHAMapTxPlusMetaLeafNode>(std::move(item), 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapTreeNode::makeAccountState(Slice data, SHAMapHash const& hash, bool hashValid)
|
SHAMapTreeNode::makeAccountState(Slice data, SHAMapHash const& hash, bool hashValid)
|
||||||
{
|
{
|
||||||
Serializer s(data.data(), data.size());
|
Serializer s(data.data(), data.size());
|
||||||
@@ -87,7 +87,7 @@ SHAMapTreeNode::makeAccountState(Slice data, SHAMapHash const& hash, bool hashVa
|
|||||||
return intr_ptr::make_shared<SHAMapAccountStateLeafNode>(std::move(item), 0);
|
return intr_ptr::make_shared<SHAMapAccountStateLeafNode>(std::move(item), 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapTreeNode::makeFromWire(Slice rawNode)
|
SHAMapTreeNode::makeFromWire(Slice rawNode)
|
||||||
{
|
{
|
||||||
if (rawNode.empty())
|
if (rawNode.empty())
|
||||||
@@ -118,7 +118,7 @@ SHAMapTreeNode::makeFromWire(Slice rawNode)
|
|||||||
Throw<std::runtime_error>("wire: Unknown type (" + std::to_string(type) + ")");
|
Throw<std::runtime_error>("wire: Unknown type (" + std::to_string(type) + ")");
|
||||||
}
|
}
|
||||||
|
|
||||||
intr_ptr::SharedPtr<SHAMapTreeNode>
|
SHAMapTreeNodePtr
|
||||||
SHAMapTreeNode::makeFromPrefix(Slice rawNode, SHAMapHash const& hash)
|
SHAMapTreeNode::makeFromPrefix(Slice rawNode, SHAMapHash const& hash)
|
||||||
{
|
{
|
||||||
if (rawNode.size() < 4)
|
if (rawNode.size() < 4)
|
||||||
|
|||||||
350
src/test/app/LedgerNodeHelpers_test.cpp
Normal file
350
src/test/app/LedgerNodeHelpers_test.cpp
Normal file
@@ -0,0 +1,350 @@
|
|||||||
|
#include <xrpld/app/ledger/detail/LedgerNodeHelpers.h>
|
||||||
|
|
||||||
|
#include <xrpl/basics/IntrusivePointer.h>
|
||||||
|
#include <xrpl/basics/base_uint.h>
|
||||||
|
#include <xrpl/beast/unit_test/suite.h>
|
||||||
|
#include <xrpl/protocol/Serializer.h>
|
||||||
|
#include <xrpl/shamap/SHAMap.h>
|
||||||
|
#include <xrpl/shamap/SHAMapAccountStateLeafNode.h>
|
||||||
|
#include <xrpl/shamap/SHAMapInnerNode.h>
|
||||||
|
#include <xrpl/shamap/SHAMapItem.h>
|
||||||
|
#include <xrpl/shamap/SHAMapTreeNode.h>
|
||||||
|
|
||||||
|
#include <boost/smart_ptr/intrusive_ptr.hpp>
|
||||||
|
|
||||||
|
#include <xrpl.pb.h>
|
||||||
|
|
||||||
|
#include <bit>
|
||||||
|
#include <cstdint>
|
||||||
|
#include <string>
|
||||||
|
|
||||||
|
namespace xrpl::tests {
|
||||||
|
|
||||||
|
class LedgerNodeHelpers_test : public beast::unit_test::suite
|
||||||
|
{
|
||||||
|
static boost::intrusive_ptr<SHAMapItem>
|
||||||
|
makeTestItem(std::uint32_t seed)
|
||||||
|
{
|
||||||
|
Serializer s;
|
||||||
|
s.add32(seed);
|
||||||
|
s.add32(seed + 1);
|
||||||
|
s.add32(seed + 2);
|
||||||
|
return make_shamapitem(s.getSHA512Half(), s.slice());
|
||||||
|
}
|
||||||
|
|
||||||
|
static std::string
|
||||||
|
serializeNode(SHAMapTreeNodePtr const& node)
|
||||||
|
{
|
||||||
|
Serializer s;
|
||||||
|
node->serializeForWire(s);
|
||||||
|
auto const slice = s.slice();
|
||||||
|
return std::string(std::bit_cast<char const*>(slice.data()), slice.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
void
|
||||||
|
testValidateLedgerNode()
|
||||||
|
{
|
||||||
|
// In the tests below the validity of the content of the node data and ID fields is not
|
||||||
|
// checked - only that the fields have values when expected. The content of the fields is
|
||||||
|
// verified in the other tests in this file.
|
||||||
|
testcase("validateLedgerNode");
|
||||||
|
|
||||||
|
// Invalid: missing all fields.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode const node;
|
||||||
|
BEAST_EXPECT(!validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: missing `nodedata` field.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodeid("test_nodeid");
|
||||||
|
BEAST_EXPECT(!validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: missing `nodedata` field.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_id("test_nodeid");
|
||||||
|
BEAST_EXPECT(!validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: missing `nodedata` field.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_depth(1);
|
||||||
|
BEAST_EXPECT(!validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: legacy `nodeid` field.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_nodeid("test_nodeid");
|
||||||
|
BEAST_EXPECT(validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: has both legacy `nodeid` and new `id` fields.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_nodeid("test_nodeid");
|
||||||
|
node.set_id("test_nodeid");
|
||||||
|
BEAST_EXPECT(!validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: has both legacy `nodeid` and new `depth` fields.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_nodeid("test_nodeid");
|
||||||
|
node.set_depth(5);
|
||||||
|
BEAST_EXPECT(!validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: new `id` field.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_id("test_id");
|
||||||
|
BEAST_EXPECT(validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: new `depth` field.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_depth(5);
|
||||||
|
BEAST_EXPECT(validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: `depth` at minimum depth.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_depth(0);
|
||||||
|
BEAST_EXPECT(validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: `depth` at arbitrary depth between minimum and maximum.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_depth(10);
|
||||||
|
BEAST_EXPECT(validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: `depth` at maximum depth.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_depth(SHAMap::leafDepth);
|
||||||
|
BEAST_EXPECT(validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: `depth` is greater than maximum depth.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata("test_data");
|
||||||
|
node.set_depth(SHAMap::leafDepth + 1);
|
||||||
|
BEAST_EXPECT(!validateLedgerNode(node));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void
|
||||||
|
testGetTreeNode()
|
||||||
|
{
|
||||||
|
testcase("getTreeNode");
|
||||||
|
|
||||||
|
// Valid: inner node. It must have at least one child for `serializeNode` to work.
|
||||||
|
{
|
||||||
|
auto const innerNode = intr_ptr::make_shared<SHAMapInnerNode>(1);
|
||||||
|
auto const childNode = intr_ptr::make_shared<SHAMapInnerNode>(1);
|
||||||
|
innerNode->setChild(0, childNode);
|
||||||
|
auto const innerData = serializeNode(innerNode);
|
||||||
|
auto const result = getTreeNode(innerData);
|
||||||
|
BEAST_EXPECT(result->isInner());
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: leaf node.
|
||||||
|
{
|
||||||
|
auto const leafItem = makeTestItem(12345);
|
||||||
|
auto const leafNode = intr_ptr::make_shared<SHAMapAccountStateLeafNode>(leafItem, 1);
|
||||||
|
auto const leafData = serializeNode(leafNode);
|
||||||
|
auto result = getTreeNode(leafData);
|
||||||
|
BEAST_EXPECT(result->isLeaf());
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: empty data.
|
||||||
|
{
|
||||||
|
auto const result = getTreeNode("");
|
||||||
|
BEAST_EXPECT(!result);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: garbage data.
|
||||||
|
{
|
||||||
|
auto const result = getTreeNode("invalid");
|
||||||
|
BEAST_EXPECT(!result);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: truncated data.
|
||||||
|
{
|
||||||
|
auto const leafItem = makeTestItem(54321);
|
||||||
|
auto const leafNode = intr_ptr::make_shared<SHAMapAccountStateLeafNode>(leafItem, 1);
|
||||||
|
// Truncate the data to trigger an exception in SHAMapTreeNode::makeAccountState when
|
||||||
|
// the data is used to deserialize the node.
|
||||||
|
uint256 const tag;
|
||||||
|
auto const leafData = serializeNode(leafNode).substr(0, tag.bytes - 1);
|
||||||
|
auto const result = getTreeNode(leafData);
|
||||||
|
BEAST_EXPECT(!result);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void
|
||||||
|
testGetSHAMapNodeID()
|
||||||
|
{
|
||||||
|
testcase("getSHAMapNodeID");
|
||||||
|
|
||||||
|
{
|
||||||
|
// Tests using inner nodes at various depths.
|
||||||
|
auto const innerNode = intr_ptr::make_shared<SHAMapInnerNode>(1);
|
||||||
|
auto const childNode = intr_ptr::make_shared<SHAMapInnerNode>(1);
|
||||||
|
innerNode->setChild(0, childNode);
|
||||||
|
auto const innerData = serializeNode(innerNode);
|
||||||
|
|
||||||
|
// Valid: legacy `nodeid` field at arbitrary depth.
|
||||||
|
{
|
||||||
|
auto const innerDepth = 3;
|
||||||
|
auto const innerID = SHAMapNodeID::createID(innerDepth, uint256{});
|
||||||
|
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata(innerData);
|
||||||
|
node.set_nodeid(innerID.getRawString());
|
||||||
|
auto const result = getSHAMapNodeID(node, innerNode);
|
||||||
|
BEAST_EXPECT(result == innerID);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: new `id` field at minimum depth.
|
||||||
|
{
|
||||||
|
auto const innerDepth = 0;
|
||||||
|
auto const innerID = SHAMapNodeID::createID(innerDepth, uint256{});
|
||||||
|
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata(innerData);
|
||||||
|
node.set_id(innerID.getRawString());
|
||||||
|
auto const result = getSHAMapNodeID(node, innerNode);
|
||||||
|
BEAST_EXPECT(result == innerID);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: new `depth` field should not be used for inner nodes.
|
||||||
|
{
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata(innerData);
|
||||||
|
node.set_depth(10);
|
||||||
|
auto const result = getSHAMapNodeID(node, innerNode);
|
||||||
|
BEAST_EXPECT(!result);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
// Tests using leaf nodes at various depths.
|
||||||
|
auto const leafItem = makeTestItem(12345);
|
||||||
|
auto const leafNode = intr_ptr::make_shared<SHAMapAccountStateLeafNode>(leafItem, 1);
|
||||||
|
auto const leafData = serializeNode(leafNode);
|
||||||
|
auto const leafKey = leafItem->key();
|
||||||
|
|
||||||
|
// Valid: legacy `nodeid` field at arbitrary depth.
|
||||||
|
{
|
||||||
|
auto const leafDepth = 5;
|
||||||
|
auto const leafID = SHAMapNodeID::createID(leafDepth, leafKey);
|
||||||
|
|
||||||
|
protocol::TMLedgerNode ledgerNode;
|
||||||
|
ledgerNode.set_nodedata(leafData);
|
||||||
|
ledgerNode.set_nodeid(leafID.getRawString());
|
||||||
|
auto const result = getSHAMapNodeID(ledgerNode, leafNode);
|
||||||
|
BEAST_EXPECT(result == leafID);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: new `id` field should not be used for leaf nodes.
|
||||||
|
{
|
||||||
|
auto const leafDepth = 5;
|
||||||
|
auto const leafID = SHAMapNodeID::createID(leafDepth, leafKey);
|
||||||
|
|
||||||
|
protocol::TMLedgerNode ledgerNode;
|
||||||
|
ledgerNode.set_nodedata(leafData);
|
||||||
|
ledgerNode.set_id(leafID.getRawString());
|
||||||
|
auto const result = getSHAMapNodeID(ledgerNode, leafNode);
|
||||||
|
BEAST_EXPECT(!result);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: new `depth` field at minimum depth.
|
||||||
|
{
|
||||||
|
auto const leafDepth = 0;
|
||||||
|
auto const leafID = SHAMapNodeID::createID(leafDepth, leafKey);
|
||||||
|
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata(leafData);
|
||||||
|
node.set_depth(leafDepth);
|
||||||
|
auto const result = getSHAMapNodeID(node, leafNode);
|
||||||
|
BEAST_EXPECT(result == leafID);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: new `depth` field at arbitrary depth between minimum and maximum.
|
||||||
|
{
|
||||||
|
auto const leafDepth = 10;
|
||||||
|
auto const leafID = SHAMapNodeID::createID(leafDepth, leafKey);
|
||||||
|
|
||||||
|
protocol::TMLedgerNode ledgerNode;
|
||||||
|
ledgerNode.set_nodedata(leafData);
|
||||||
|
ledgerNode.set_depth(leafDepth);
|
||||||
|
auto const result = getSHAMapNodeID(ledgerNode, leafNode);
|
||||||
|
BEAST_EXPECT(result == leafID);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Valid: new `depth` field at maximum depth.
|
||||||
|
// Note that we do not test a depth greater than the maximum depth, because the proto
|
||||||
|
// message is assumed to have been validated by the time the getSHAMapNodeID function is
|
||||||
|
// called.
|
||||||
|
{
|
||||||
|
auto const leafDepth = SHAMap::leafDepth;
|
||||||
|
auto const leafID = SHAMapNodeID::createID(leafDepth, leafKey);
|
||||||
|
|
||||||
|
protocol::TMLedgerNode node;
|
||||||
|
node.set_nodedata(leafData);
|
||||||
|
node.set_depth(leafDepth);
|
||||||
|
auto const result = getSHAMapNodeID(node, leafNode);
|
||||||
|
BEAST_EXPECT(result == leafID);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Invalid: legacy `nodeid` field where the node ID is inconsistent with the key.
|
||||||
|
{
|
||||||
|
auto const otherItem = makeTestItem(54321);
|
||||||
|
auto const otherNode =
|
||||||
|
intr_ptr::make_shared<SHAMapAccountStateLeafNode>(otherItem, 1);
|
||||||
|
auto const otherData = serializeNode(otherNode);
|
||||||
|
auto const otherKey = otherItem->key();
|
||||||
|
auto const otherDepth = 1;
|
||||||
|
auto const otherID = SHAMapNodeID::createID(otherDepth, otherKey);
|
||||||
|
|
||||||
|
protocol::TMLedgerNode ledgerNode;
|
||||||
|
ledgerNode.set_nodedata(otherData);
|
||||||
|
ledgerNode.set_nodeid(otherID.getRawString());
|
||||||
|
auto const result = getSHAMapNodeID(ledgerNode, leafNode);
|
||||||
|
BEAST_EXPECT(!result);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public:
|
||||||
|
void
|
||||||
|
run() override
|
||||||
|
{
|
||||||
|
testValidateLedgerNode();
|
||||||
|
testGetTreeNode();
|
||||||
|
testGetSHAMapNodeID();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
BEAST_DEFINE_TESTSUITE(LedgerNodeHelpers, app, xrpl);
|
||||||
|
|
||||||
|
} // namespace xrpl::tests
|
||||||
@@ -7,7 +7,6 @@
|
|||||||
#include <xrpl/basics/BasicConfig.h>
|
#include <xrpl/basics/BasicConfig.h>
|
||||||
#include <xrpl/beast/unit_test/suite.h>
|
#include <xrpl/beast/unit_test/suite.h>
|
||||||
#include <xrpl/beast/utility/temp_dir.h>
|
#include <xrpl/beast/utility/temp_dir.h>
|
||||||
#include <xrpl/protocol/SystemParameters.h> // IWYU pragma: keep
|
|
||||||
#include <xrpl/server/Port.h>
|
#include <xrpl/server/Port.h>
|
||||||
|
|
||||||
#include <boost/filesystem/operations.hpp>
|
#include <boost/filesystem/operations.hpp>
|
||||||
|
|||||||
@@ -63,8 +63,8 @@ public:
|
|||||||
negotiateProtocolVersion("RTXP/1.2, XRPL/2.0, XRPL/2.1") == make_protocol(2, 1));
|
negotiateProtocolVersion("RTXP/1.2, XRPL/2.0, XRPL/2.1") == make_protocol(2, 1));
|
||||||
BEAST_EXPECT(negotiateProtocolVersion("XRPL/2.2") == make_protocol(2, 2));
|
BEAST_EXPECT(negotiateProtocolVersion("XRPL/2.2") == make_protocol(2, 2));
|
||||||
BEAST_EXPECT(
|
BEAST_EXPECT(
|
||||||
negotiateProtocolVersion("RTXP/1.2, XRPL/2.2, XRPL/2.3, XRPL/999.999") ==
|
negotiateProtocolVersion("RTXP/1.2, XRPL/2.3, XRPL/2.4, XRPL/999.999") ==
|
||||||
make_protocol(2, 2));
|
make_protocol(2, 3));
|
||||||
BEAST_EXPECT(negotiateProtocolVersion("XRPL/999.999, WebSocket/1.0") == std::nullopt);
|
BEAST_EXPECT(negotiateProtocolVersion("XRPL/999.999, WebSocket/1.0") == std::nullopt);
|
||||||
BEAST_EXPECT(negotiateProtocolVersion("") == std::nullopt);
|
BEAST_EXPECT(negotiateProtocolVersion("") == std::nullopt);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +1,6 @@
|
|||||||
#include <test/shamap/common.h>
|
#include <test/shamap/common.h>
|
||||||
#include <test/unit_test/SuiteJournal.h>
|
#include <test/unit_test/SuiteJournal.h>
|
||||||
|
|
||||||
#include <xrpl/basics/Blob.h>
|
|
||||||
#include <xrpl/basics/SHAMapHash.h>
|
#include <xrpl/basics/SHAMapHash.h>
|
||||||
#include <xrpl/basics/Slice.h>
|
#include <xrpl/basics/Slice.h>
|
||||||
#include <xrpl/basics/base_uint.h>
|
#include <xrpl/basics/base_uint.h>
|
||||||
@@ -115,14 +114,17 @@ public:
|
|||||||
destination.setSynching();
|
destination.setSynching();
|
||||||
|
|
||||||
{
|
{
|
||||||
std::vector<std::pair<SHAMapNodeID, Blob>> a;
|
std::vector<SHAMapNodeData> a;
|
||||||
|
|
||||||
BEAST_EXPECT(source.getNodeFat(SHAMapNodeID(), a, rand_bool(eng_), rand_int(eng_, 2)));
|
BEAST_EXPECT(source.getNodeFat(SHAMapNodeID(), a, rand_bool(eng_), rand_int(eng_, 2)));
|
||||||
|
|
||||||
unexpected(a.empty(), "NodeSize");
|
unexpected(a.empty(), "NodeSize");
|
||||||
|
|
||||||
BEAST_EXPECT(destination.addRootNode(source.getHash(), makeSlice(a[0].second), nullptr)
|
auto node = SHAMapTreeNode::makeFromWire(makeSlice(a[0].data));
|
||||||
.isGood());
|
if (!node)
|
||||||
|
fail("", __FILE__, __LINE__);
|
||||||
|
BEAST_EXPECT(
|
||||||
|
destination.addRootNode(source.getHash(), std::move(node), nullptr).isGood());
|
||||||
}
|
}
|
||||||
|
|
||||||
do
|
do
|
||||||
@@ -136,7 +138,7 @@ public:
|
|||||||
break;
|
break;
|
||||||
|
|
||||||
// get as many nodes as possible based on this information
|
// get as many nodes as possible based on this information
|
||||||
std::vector<std::pair<SHAMapNodeID, Blob>> b;
|
std::vector<SHAMapNodeData> b;
|
||||||
|
|
||||||
for (auto& it : nodesMissing)
|
for (auto& it : nodesMissing)
|
||||||
{
|
{
|
||||||
@@ -158,8 +160,10 @@ public:
|
|||||||
// Don't use BEAST_EXPECT here b/c it will be called a
|
// Don't use BEAST_EXPECT here b/c it will be called a
|
||||||
// non-deterministic number of times and the number of tests run
|
// non-deterministic number of times and the number of tests run
|
||||||
// should be deterministic
|
// should be deterministic
|
||||||
if (!destination.addKnownNode(b[i].first, makeSlice(b[i].second), nullptr)
|
auto node = SHAMapTreeNode::makeFromWire(makeSlice(b[i].data));
|
||||||
.isUseful())
|
if (!node)
|
||||||
|
fail("", __FILE__, __LINE__);
|
||||||
|
if (!destination.addKnownNode(b[i].nodeID, std::move(node), nullptr).isUseful())
|
||||||
fail("", __FILE__, __LINE__);
|
fail("", __FILE__, __LINE__);
|
||||||
}
|
}
|
||||||
} while (true);
|
} while (true);
|
||||||
|
|||||||
@@ -9,6 +9,7 @@
|
|||||||
|
|
||||||
#include <mutex>
|
#include <mutex>
|
||||||
#include <set>
|
#include <set>
|
||||||
|
#include <string_view>
|
||||||
#include <utility>
|
#include <utility>
|
||||||
|
|
||||||
namespace xrpl {
|
namespace xrpl {
|
||||||
@@ -131,16 +132,16 @@ private:
|
|||||||
processData(std::shared_ptr<Peer> peer, protocol::TMLedgerData& data);
|
processData(std::shared_ptr<Peer> peer, protocol::TMLedgerData& data);
|
||||||
|
|
||||||
bool
|
bool
|
||||||
takeHeader(std::string const& data);
|
takeHeader(std::string_view data);
|
||||||
|
|
||||||
void
|
void
|
||||||
receiveNode(protocol::TMLedgerData& packet, SHAMapAddNode&);
|
receiveNode(protocol::TMLedgerData& packet, SHAMapAddNode& san);
|
||||||
|
|
||||||
bool
|
bool
|
||||||
takeTxRootNode(Slice const& data, SHAMapAddNode&);
|
takeTxRootNode(std::string_view data, SHAMapAddNode& san);
|
||||||
|
|
||||||
bool
|
bool
|
||||||
takeAsRootNode(Slice const& data, SHAMapAddNode&);
|
takeAsRootNode(std::string_view data, SHAMapAddNode& san);
|
||||||
|
|
||||||
std::vector<uint256>
|
std::vector<uint256>
|
||||||
neededTxHashes(int max, SHAMapSyncFilter* filter) const;
|
neededTxHashes(int max, SHAMapSyncFilter* filter) const;
|
||||||
|
|||||||
@@ -4,6 +4,7 @@
|
|||||||
#include <xrpld/app/ledger/InboundLedgers.h>
|
#include <xrpld/app/ledger/InboundLedgers.h>
|
||||||
#include <xrpld/app/ledger/LedgerMaster.h>
|
#include <xrpld/app/ledger/LedgerMaster.h>
|
||||||
#include <xrpld/app/ledger/TransactionStateSF.h>
|
#include <xrpld/app/ledger/TransactionStateSF.h>
|
||||||
|
#include <xrpld/app/ledger/detail/LedgerNodeHelpers.h>
|
||||||
#include <xrpld/app/ledger/detail/TimeoutCounter.h>
|
#include <xrpld/app/ledger/detail/TimeoutCounter.h>
|
||||||
#include <xrpld/app/main/Application.h>
|
#include <xrpld/app/main/Application.h>
|
||||||
#include <xrpld/overlay/Message.h>
|
#include <xrpld/overlay/Message.h>
|
||||||
@@ -42,8 +43,8 @@
|
|||||||
#include <mutex>
|
#include <mutex>
|
||||||
#include <random>
|
#include <random>
|
||||||
#include <sstream>
|
#include <sstream>
|
||||||
#include <stdexcept>
|
|
||||||
#include <string>
|
#include <string>
|
||||||
|
#include <string_view>
|
||||||
#include <tuple>
|
#include <tuple>
|
||||||
#include <unordered_map>
|
#include <unordered_map>
|
||||||
#include <utility>
|
#include <utility>
|
||||||
@@ -793,7 +794,7 @@ InboundLedger::filterNodes(
|
|||||||
*/
|
*/
|
||||||
// data must not have hash prefix
|
// data must not have hash prefix
|
||||||
bool
|
bool
|
||||||
InboundLedger::takeHeader(std::string const& data)
|
InboundLedger::takeHeader(std::string_view data)
|
||||||
{
|
{
|
||||||
// Return value: true=normal, false=bad data
|
// Return value: true=normal, false=bad data
|
||||||
JLOG(journal_.trace()) << "got header acquiring ledger " << hash_;
|
JLOG(journal_.trace()) << "got header acquiring ledger " << hash_;
|
||||||
@@ -881,20 +882,31 @@ InboundLedger::receiveNode(protocol::TMLedgerData& packet, SHAMapAddNode& san)
|
|||||||
{
|
{
|
||||||
auto const f = filter.get();
|
auto const f = filter.get();
|
||||||
|
|
||||||
for (auto const& node : packet.nodes())
|
for (auto const& ledgerNode : packet.nodes())
|
||||||
{
|
{
|
||||||
auto const nodeID = deserializeSHAMapNodeID(node.nodeid());
|
auto treeNode = getTreeNode(ledgerNode.nodedata());
|
||||||
|
if (!treeNode)
|
||||||
|
{
|
||||||
|
JLOG(journal_.warn()) << "Got invalid node data";
|
||||||
|
san.incInvalid();
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
auto const nodeID = getSHAMapNodeID(ledgerNode, treeNode);
|
||||||
if (!nodeID)
|
if (!nodeID)
|
||||||
throw std::runtime_error("data does not properly deserialize");
|
{
|
||||||
|
JLOG(journal_.warn()) << "Got invalid node id";
|
||||||
|
san.incInvalid();
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
if (nodeID->isRoot())
|
if (nodeID->isRoot())
|
||||||
{
|
{
|
||||||
san += map.addRootNode(rootHash, makeSlice(node.nodedata()), f);
|
san += map.addRootNode(rootHash, std::move(treeNode), f);
|
||||||
}
|
}
|
||||||
else
|
else
|
||||||
{
|
{
|
||||||
san += map.addKnownNode(*nodeID, makeSlice(node.nodedata()), f);
|
san += map.addKnownNode(*nodeID, std::move(treeNode), f);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!san.isGood())
|
if (!san.isGood())
|
||||||
@@ -934,7 +946,7 @@ InboundLedger::receiveNode(protocol::TMLedgerData& packet, SHAMapAddNode& san)
|
|||||||
Call with a lock
|
Call with a lock
|
||||||
*/
|
*/
|
||||||
bool
|
bool
|
||||||
InboundLedger::takeAsRootNode(Slice const& data, SHAMapAddNode& san)
|
InboundLedger::takeAsRootNode(std::string_view data, SHAMapAddNode& san)
|
||||||
{
|
{
|
||||||
if (failed_ || mHaveState)
|
if (failed_ || mHaveState)
|
||||||
{
|
{
|
||||||
@@ -950,9 +962,17 @@ InboundLedger::takeAsRootNode(Slice const& data, SHAMapAddNode& san)
|
|||||||
// LCOV_EXCL_STOP
|
// LCOV_EXCL_STOP
|
||||||
}
|
}
|
||||||
|
|
||||||
|
auto treeNode = getTreeNode(data);
|
||||||
|
if (!treeNode)
|
||||||
|
{
|
||||||
|
JLOG(journal_.warn()) << "Got invalid node data";
|
||||||
|
san.incInvalid();
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
AccountStateSF filter(mLedger->stateMap().family().db(), app_.getLedgerMaster());
|
AccountStateSF filter(mLedger->stateMap().family().db(), app_.getLedgerMaster());
|
||||||
san +=
|
san += mLedger->stateMap().addRootNode(
|
||||||
mLedger->stateMap().addRootNode(SHAMapHash{mLedger->header().accountHash}, data, &filter);
|
SHAMapHash{mLedger->header().accountHash}, std::move(treeNode), &filter);
|
||||||
return san.isGood();
|
return san.isGood();
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -960,7 +980,7 @@ InboundLedger::takeAsRootNode(Slice const& data, SHAMapAddNode& san)
|
|||||||
Call with a lock
|
Call with a lock
|
||||||
*/
|
*/
|
||||||
bool
|
bool
|
||||||
InboundLedger::takeTxRootNode(Slice const& data, SHAMapAddNode& san)
|
InboundLedger::takeTxRootNode(std::string_view data, SHAMapAddNode& san)
|
||||||
{
|
{
|
||||||
if (failed_ || mHaveTransactions)
|
if (failed_ || mHaveTransactions)
|
||||||
{
|
{
|
||||||
@@ -976,8 +996,17 @@ InboundLedger::takeTxRootNode(Slice const& data, SHAMapAddNode& san)
|
|||||||
// LCOV_EXCL_STOP
|
// LCOV_EXCL_STOP
|
||||||
}
|
}
|
||||||
|
|
||||||
|
auto treeNode = getTreeNode(data);
|
||||||
|
if (!treeNode)
|
||||||
|
{
|
||||||
|
JLOG(journal_.warn()) << "Got invalid node data";
|
||||||
|
san.incInvalid();
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
TransactionStateSF filter(mLedger->txMap().family().db(), app_.getLedgerMaster());
|
TransactionStateSF filter(mLedger->txMap().family().db(), app_.getLedgerMaster());
|
||||||
san += mLedger->txMap().addRootNode(SHAMapHash{mLedger->header().txHash}, data, &filter);
|
san += mLedger->txMap().addRootNode(
|
||||||
|
SHAMapHash{mLedger->header().txHash}, std::move(treeNode), &filter);
|
||||||
return san.isGood();
|
return san.isGood();
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1074,13 +1103,13 @@ InboundLedger::processData(std::shared_ptr<Peer> peer, protocol::TMLedgerData& p
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (!mHaveState && (packet.nodes().size() > 1) &&
|
if (!mHaveState && (packet.nodes().size() > 1) &&
|
||||||
!takeAsRootNode(makeSlice(packet.nodes(1).nodedata()), san))
|
!takeAsRootNode(packet.nodes(1).nodedata(), san))
|
||||||
{
|
{
|
||||||
JLOG(journal_.warn()) << "Included AS root invalid";
|
JLOG(journal_.warn()) << "Included AS root invalid";
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!mHaveTransactions && (packet.nodes().size() > 2) &&
|
if (!mHaveTransactions && (packet.nodes().size() > 2) &&
|
||||||
!takeTxRootNode(makeSlice(packet.nodes(2).nodedata()), san))
|
!takeTxRootNode(packet.nodes(2).nodedata(), san))
|
||||||
{
|
{
|
||||||
JLOG(journal_.warn()) << "Included TX root invalid";
|
JLOG(journal_.warn()) << "Included TX root invalid";
|
||||||
}
|
}
|
||||||
@@ -1111,13 +1140,13 @@ InboundLedger::processData(std::shared_ptr<Peer> peer, protocol::TMLedgerData& p
|
|||||||
|
|
||||||
ScopedLockType const sl(mtx_);
|
ScopedLockType const sl(mtx_);
|
||||||
|
|
||||||
// Verify node IDs and data are complete
|
// Verify nodes are complete
|
||||||
for (auto const& node : packet.nodes())
|
for (auto const& ledgerNode : packet.nodes())
|
||||||
{
|
{
|
||||||
if (!node.has_nodeid() || !node.has_nodedata())
|
if (!validateLedgerNode(ledgerNode))
|
||||||
{
|
{
|
||||||
JLOG(journal_.warn()) << "Got bad node";
|
JLOG(journal_.warn()) << "Got malformed ledger node";
|
||||||
peer->charge(Resource::feeMalformedRequest, "ledger_data bad node");
|
peer->charge(Resource::feeMalformedRequest, "ledgerNode");
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,13 +2,13 @@
|
|||||||
|
|
||||||
#include <xrpld/app/ledger/InboundLedger.h>
|
#include <xrpld/app/ledger/InboundLedger.h>
|
||||||
#include <xrpld/app/ledger/LedgerMaster.h>
|
#include <xrpld/app/ledger/LedgerMaster.h>
|
||||||
|
#include <xrpld/app/ledger/detail/LedgerNodeHelpers.h>
|
||||||
#include <xrpld/app/main/Application.h>
|
#include <xrpld/app/main/Application.h>
|
||||||
#include <xrpld/overlay/PeerSet.h>
|
#include <xrpld/overlay/PeerSet.h>
|
||||||
|
|
||||||
#include <xrpl/basics/Blob.h>
|
#include <xrpl/basics/Blob.h>
|
||||||
#include <xrpl/basics/DecayingSample.h>
|
#include <xrpl/basics/DecayingSample.h>
|
||||||
#include <xrpl/basics/Log.h>
|
#include <xrpl/basics/Log.h>
|
||||||
#include <xrpl/basics/Slice.h>
|
|
||||||
#include <xrpl/basics/UnorderedContainers.h>
|
#include <xrpl/basics/UnorderedContainers.h>
|
||||||
#include <xrpl/basics/base_uint.h>
|
#include <xrpl/basics/base_uint.h>
|
||||||
#include <xrpl/basics/scope.h>
|
#include <xrpl/basics/scope.h>
|
||||||
@@ -248,23 +248,20 @@ public:
|
|||||||
Serializer s;
|
Serializer s;
|
||||||
try
|
try
|
||||||
{
|
{
|
||||||
for (int i = 0; i < packet_ptr->nodes().size(); ++i)
|
for (auto const& ledgerNode : packet_ptr->nodes())
|
||||||
{
|
{
|
||||||
auto const& node = packet_ptr->nodes(i);
|
if (!validateLedgerNode(ledgerNode))
|
||||||
|
|
||||||
if (!node.has_nodeid() || !node.has_nodedata())
|
|
||||||
return;
|
return;
|
||||||
|
|
||||||
auto newNode = SHAMapTreeNode::makeFromWire(makeSlice(node.nodedata()));
|
auto const treeNode = getTreeNode(ledgerNode.nodedata());
|
||||||
|
if (!treeNode)
|
||||||
if (!newNode)
|
|
||||||
return;
|
return;
|
||||||
|
|
||||||
s.erase();
|
s.erase();
|
||||||
newNode->serializeWithPrefix(s);
|
treeNode->serializeWithPrefix(s);
|
||||||
|
|
||||||
app_.getLedgerMaster().addFetchPack(
|
app_.getLedgerMaster().addFetchPack(
|
||||||
newNode->getHash().as_uint256(), std::make_shared<Blob>(s.begin(), s.end()));
|
treeNode->getHash().as_uint256(), std::make_shared<Blob>(s.begin(), s.end()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
catch (std::exception const&) // NOLINT(bugprone-empty-catch)
|
catch (std::exception const&) // NOLINT(bugprone-empty-catch)
|
||||||
|
|||||||
@@ -1,11 +1,11 @@
|
|||||||
#include <xrpld/app/ledger/InboundTransactions.h>
|
#include <xrpld/app/ledger/InboundTransactions.h>
|
||||||
|
|
||||||
|
#include <xrpld/app/ledger/detail/LedgerNodeHelpers.h>
|
||||||
#include <xrpld/app/ledger/detail/TransactionAcquire.h>
|
#include <xrpld/app/ledger/detail/TransactionAcquire.h>
|
||||||
#include <xrpld/app/main/Application.h>
|
#include <xrpld/app/main/Application.h>
|
||||||
#include <xrpld/overlay/PeerSet.h>
|
#include <xrpld/overlay/PeerSet.h>
|
||||||
|
|
||||||
#include <xrpl/basics/Log.h>
|
#include <xrpl/basics/Log.h>
|
||||||
#include <xrpl/basics/Slice.h>
|
|
||||||
#include <xrpl/basics/UnorderedContainers.h>
|
#include <xrpl/basics/UnorderedContainers.h>
|
||||||
#include <xrpl/beast/insight/Collector.h>
|
#include <xrpl/beast/insight/Collector.h>
|
||||||
#include <xrpl/protocol/RippleLedgerHash.h>
|
#include <xrpl/protocol/RippleLedgerHash.h>
|
||||||
@@ -14,6 +14,7 @@
|
|||||||
#include <xrpl/shamap/SHAMap.h>
|
#include <xrpl/shamap/SHAMap.h>
|
||||||
#include <xrpl/shamap/SHAMapMissingNode.h>
|
#include <xrpl/shamap/SHAMapMissingNode.h>
|
||||||
#include <xrpl/shamap/SHAMapNodeID.h>
|
#include <xrpl/shamap/SHAMapNodeID.h>
|
||||||
|
#include <xrpl/shamap/SHAMapTreeNode.h>
|
||||||
|
|
||||||
#include <xrpl.pb.h>
|
#include <xrpl.pb.h>
|
||||||
|
|
||||||
@@ -144,29 +145,38 @@ public:
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::pair<SHAMapNodeID, Slice>> data;
|
std::vector<std::pair<SHAMapNodeID, SHAMapTreeNodePtr>> data;
|
||||||
data.reserve(packet.nodes().size());
|
data.reserve(packet.nodes().size());
|
||||||
|
|
||||||
for (auto const& node : packet.nodes())
|
for (auto const& ledgerNode : packet.nodes())
|
||||||
{
|
{
|
||||||
if (!node.has_nodeid() || !node.has_nodedata())
|
if (!validateLedgerNode(ledgerNode))
|
||||||
{
|
{
|
||||||
peer->charge(Resource::feeMalformedRequest, "ledger_data");
|
JLOG(j_.warn()) << "Got malformed ledger node";
|
||||||
|
peer->charge(Resource::feeMalformedRequest, "ledgerNode");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
auto const id = deserializeSHAMapNodeID(node.nodeid());
|
auto treeNode = getTreeNode(ledgerNode.nodedata());
|
||||||
|
if (!treeNode)
|
||||||
if (!id)
|
|
||||||
{
|
{
|
||||||
peer->charge(Resource::feeInvalidData, "ledger_data");
|
JLOG(j_.warn()) << "Got invalid node data";
|
||||||
|
peer->charge(Resource::feeInvalidData, "node_data");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
data.emplace_back(*id, makeSlice(node.nodedata()));
|
auto const nodeID = getSHAMapNodeID(ledgerNode, treeNode);
|
||||||
|
if (!nodeID)
|
||||||
|
{
|
||||||
|
JLOG(j_.warn()) << "Got invalid node id";
|
||||||
|
peer->charge(Resource::feeInvalidData, "node_id");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
data.emplace_back(*nodeID, std::move(treeNode));
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!ta->takeNodes(data, peer).isUseful())
|
if (!ta->takeNodes(std::move(data), peer).isUseful())
|
||||||
peer->charge(Resource::feeUselessData, "ledger_data not useful");
|
peer->charge(Resource::feeUselessData, "ledger_data not useful");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
91
src/xrpld/app/ledger/detail/LedgerNodeHelpers.cpp
Normal file
91
src/xrpld/app/ledger/detail/LedgerNodeHelpers.cpp
Normal file
@@ -0,0 +1,91 @@
|
|||||||
|
#include <xrpld/app/ledger/detail/LedgerNodeHelpers.h>
|
||||||
|
|
||||||
|
#include <xrpl/basics/Slice.h>
|
||||||
|
#include <xrpl/basics/safe_cast.h>
|
||||||
|
#include <xrpl/beast/utility/instrumentation.h>
|
||||||
|
#include <xrpl/shamap/SHAMap.h>
|
||||||
|
#include <xrpl/shamap/SHAMapLeafNode.h>
|
||||||
|
#include <xrpl/shamap/SHAMapNodeID.h>
|
||||||
|
#include <xrpl/shamap/SHAMapTreeNode.h>
|
||||||
|
|
||||||
|
#include <xrpl.pb.h>
|
||||||
|
|
||||||
|
#include <exception>
|
||||||
|
#include <optional>
|
||||||
|
#include <string_view>
|
||||||
|
|
||||||
|
namespace xrpl {
|
||||||
|
|
||||||
|
bool
|
||||||
|
validateLedgerNode(protocol::TMLedgerNode const& ledgerNode)
|
||||||
|
{
|
||||||
|
if (!ledgerNode.has_nodedata())
|
||||||
|
return false;
|
||||||
|
|
||||||
|
if (ledgerNode.has_nodeid())
|
||||||
|
return !ledgerNode.has_id() && !ledgerNode.has_depth();
|
||||||
|
|
||||||
|
return ledgerNode.has_id() ||
|
||||||
|
(ledgerNode.has_depth() && ledgerNode.depth() <= SHAMap::leafDepth);
|
||||||
|
}
|
||||||
|
|
||||||
|
SHAMapTreeNodePtr
|
||||||
|
getTreeNode(std::string_view data)
|
||||||
|
{
|
||||||
|
auto const slice = makeSlice(data);
|
||||||
|
try
|
||||||
|
{
|
||||||
|
return SHAMapTreeNode::makeFromWire(slice);
|
||||||
|
}
|
||||||
|
catch (std::exception const&)
|
||||||
|
{
|
||||||
|
return {};
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
std::optional<SHAMapNodeID>
|
||||||
|
getSHAMapNodeID(protocol::TMLedgerNode const& ledgerNode, SHAMapTreeNodePtr const& treeNode)
|
||||||
|
{
|
||||||
|
if (ledgerNode.has_id() || ledgerNode.has_depth())
|
||||||
|
{
|
||||||
|
if (treeNode->isInner())
|
||||||
|
{
|
||||||
|
if (!ledgerNode.has_id())
|
||||||
|
return std::nullopt;
|
||||||
|
|
||||||
|
return deserializeSHAMapNodeID(ledgerNode.id());
|
||||||
|
}
|
||||||
|
|
||||||
|
if (treeNode->isLeaf())
|
||||||
|
{
|
||||||
|
if (!ledgerNode.has_depth())
|
||||||
|
return std::nullopt;
|
||||||
|
|
||||||
|
auto const key =
|
||||||
|
safe_downcast<SHAMapLeafNode const*>(treeNode.get())->peekItem()->key();
|
||||||
|
return SHAMapNodeID::createID(ledgerNode.depth(), key);
|
||||||
|
}
|
||||||
|
|
||||||
|
UNREACHABLE("xrpl::getSHAMapNodeID : tree node is neither inner nor leaf");
|
||||||
|
return std::nullopt;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!ledgerNode.has_nodeid())
|
||||||
|
return std::nullopt;
|
||||||
|
|
||||||
|
auto nodeID = deserializeSHAMapNodeID(ledgerNode.nodeid());
|
||||||
|
if (!nodeID.has_value())
|
||||||
|
return std::nullopt;
|
||||||
|
|
||||||
|
if (treeNode->isLeaf())
|
||||||
|
{
|
||||||
|
auto const key = safe_downcast<SHAMapLeafNode const*>(treeNode.get())->peekItem()->key();
|
||||||
|
auto const expected_id = SHAMapNodeID::createID(static_cast<int>(nodeID->getDepth()), key);
|
||||||
|
if (nodeID->getNodeID() != expected_id.getNodeID())
|
||||||
|
return std::nullopt;
|
||||||
|
}
|
||||||
|
|
||||||
|
return nodeID;
|
||||||
|
}
|
||||||
|
|
||||||
|
} // namespace xrpl
|
||||||
72
src/xrpld/app/ledger/detail/LedgerNodeHelpers.h
Normal file
72
src/xrpld/app/ledger/detail/LedgerNodeHelpers.h
Normal file
@@ -0,0 +1,72 @@
|
|||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <xrpl/basics/IntrusivePointer.h>
|
||||||
|
#include <xrpl/shamap/SHAMapNodeID.h>
|
||||||
|
#include <xrpl/shamap/SHAMapTreeNode.h>
|
||||||
|
|
||||||
|
#include <optional>
|
||||||
|
#include <string_view>
|
||||||
|
|
||||||
|
namespace protocol {
|
||||||
|
class TMLedgerNode;
|
||||||
|
} // namespace protocol
|
||||||
|
|
||||||
|
namespace xrpl {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Validates a ledger node proto message.
|
||||||
|
*
|
||||||
|
* This function checks whether a ledger node has the expected fields (for non-ledger base data):
|
||||||
|
* - The node must have `nodedata`.
|
||||||
|
* - If the legacy `nodeid` field is present then the new `id` and `depth` fields must not be
|
||||||
|
* present.
|
||||||
|
* - If the new `id` or `depth` fields are present (it is a oneof field, so only one of the two can
|
||||||
|
* be set) then the legacy `nodeid` must not be present.
|
||||||
|
* - If the `depth` field is present then it must be between 0 and SHAMap::leafDepth (inclusive).
|
||||||
|
*
|
||||||
|
* @param ledgerNode The ledger node to validate.
|
||||||
|
* @return true if the ledger node has the expected fields, false otherwise.
|
||||||
|
*/
|
||||||
|
[[nodiscard]] bool
|
||||||
|
validateLedgerNode(protocol::TMLedgerNode const& ledgerNode);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Deserializes a SHAMapTreeNode from wire format data.
|
||||||
|
*
|
||||||
|
* This function attempts to create a SHAMapTreeNode from the provided data string. If the data is
|
||||||
|
* malformed or deserialization fails, the function returns a nullptr instead of throwing an
|
||||||
|
* exception.
|
||||||
|
*
|
||||||
|
* @param data The serialized node data in wire format.
|
||||||
|
* @return The deserialized tree node if successful, or a nullptr if deserialization fails.
|
||||||
|
*/
|
||||||
|
[[nodiscard]] SHAMapTreeNodePtr
|
||||||
|
getTreeNode(std::string_view data);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @brief Extracts or reconstructs the SHAMapNodeID from a ledger node proto message.
|
||||||
|
*
|
||||||
|
* This function retrieves the SHAMapNodeID for a tree node, with behavior that depends on which
|
||||||
|
* field is set and the node type (inner vs. leaf).
|
||||||
|
*
|
||||||
|
* When the legacy `nodeid` field is set in the message:
|
||||||
|
* - For all nodes: Deserializes the node ID from the field.
|
||||||
|
* - For leaf nodes: Validates that the node ID is consistent with the leaf's key.
|
||||||
|
*
|
||||||
|
* When the new `id` or `depth` field is set in the message:
|
||||||
|
* - For inner nodes: Deserializes the node ID from the `id` field.
|
||||||
|
* - For leaf nodes: Reconstructs the node ID using both the depth from the `depth` field and the
|
||||||
|
* key from the leaf node's item.
|
||||||
|
* Note that root nodes may be inner nodes or leaf nodes.
|
||||||
|
*
|
||||||
|
* @param ledgerNode The validated protocol message containing the ledger node data.
|
||||||
|
* @param treeNode The deserialized tree node (inner or leaf node).
|
||||||
|
* @return An optional containing the node ID if extraction/reconstruction succeeds, or std::nullopt
|
||||||
|
* if the required fields are missing or validation fails.
|
||||||
|
* @note This function expects that the caller has already validated the ledger node by calling the
|
||||||
|
* `validateLedgerNode` function and obtained a valid tree node by calling `getTreeNode`.
|
||||||
|
*/
|
||||||
|
[[nodiscard]] std::optional<SHAMapNodeID>
|
||||||
|
getSHAMapNodeID(protocol::TMLedgerNode const& ledgerNode, SHAMapTreeNodePtr const& treeNode);
|
||||||
|
|
||||||
|
} // namespace xrpl
|
||||||
@@ -7,13 +7,13 @@
|
|||||||
#include <xrpld/overlay/PeerSet.h>
|
#include <xrpld/overlay/PeerSet.h>
|
||||||
|
|
||||||
#include <xrpl/basics/Log.h>
|
#include <xrpl/basics/Log.h>
|
||||||
#include <xrpl/basics/Slice.h>
|
|
||||||
#include <xrpl/basics/base_uint.h>
|
#include <xrpl/basics/base_uint.h>
|
||||||
#include <xrpl/core/Job.h>
|
#include <xrpl/core/Job.h>
|
||||||
#include <xrpl/server/NetworkOPs.h>
|
#include <xrpl/server/NetworkOPs.h>
|
||||||
#include <xrpl/shamap/SHAMap.h>
|
#include <xrpl/shamap/SHAMap.h>
|
||||||
#include <xrpl/shamap/SHAMapAddNode.h>
|
#include <xrpl/shamap/SHAMapAddNode.h>
|
||||||
#include <xrpl/shamap/SHAMapMissingNode.h>
|
#include <xrpl/shamap/SHAMapMissingNode.h>
|
||||||
|
#include <xrpl/shamap/SHAMapTreeNode.h>
|
||||||
|
|
||||||
#include <xrpl.pb.h>
|
#include <xrpl.pb.h>
|
||||||
|
|
||||||
@@ -173,7 +173,7 @@ TransactionAcquire::trigger(std::shared_ptr<Peer> const& peer)
|
|||||||
|
|
||||||
SHAMapAddNode
|
SHAMapAddNode
|
||||||
TransactionAcquire::takeNodes(
|
TransactionAcquire::takeNodes(
|
||||||
std::vector<std::pair<SHAMapNodeID, Slice>> const& data,
|
std::vector<std::pair<SHAMapNodeID, SHAMapTreeNodePtr>> data,
|
||||||
std::shared_ptr<Peer> const& peer)
|
std::shared_ptr<Peer> const& peer)
|
||||||
{
|
{
|
||||||
ScopedLockType const sl(mtx_);
|
ScopedLockType const sl(mtx_);
|
||||||
|
|||||||
@@ -21,8 +21,8 @@ public:
|
|||||||
|
|
||||||
SHAMapAddNode
|
SHAMapAddNode
|
||||||
takeNodes(
|
takeNodes(
|
||||||
std::vector<std::pair<SHAMapNodeID, Slice>> const& data,
|
std::vector<std::pair<SHAMapNodeID, SHAMapTreeNodePtr>> data,
|
||||||
std::shared_ptr<Peer> const&);
|
std::shared_ptr<Peer> const& peer);
|
||||||
|
|
||||||
void
|
void
|
||||||
init(int startPeers);
|
init(int startPeers);
|
||||||
|
|||||||
@@ -17,6 +17,7 @@ enum class ProtocolFeature {
|
|||||||
ValidatorListPropagation,
|
ValidatorListPropagation,
|
||||||
ValidatorList2Propagation,
|
ValidatorList2Propagation,
|
||||||
LedgerReplay,
|
LedgerReplay,
|
||||||
|
LedgerNodeDepth,
|
||||||
};
|
};
|
||||||
|
|
||||||
/** Represents a peer connection in the overlay. */
|
/** Represents a peer connection in the overlay. */
|
||||||
|
|||||||
@@ -61,6 +61,7 @@
|
|||||||
#include <xrpl/server/Handoff.h>
|
#include <xrpl/server/Handoff.h>
|
||||||
#include <xrpl/server/LoadFeeTrack.h>
|
#include <xrpl/server/LoadFeeTrack.h>
|
||||||
#include <xrpl/server/NetworkOPs.h>
|
#include <xrpl/server/NetworkOPs.h>
|
||||||
|
#include <xrpl/shamap/SHAMap.h>
|
||||||
#include <xrpl/shamap/SHAMapNodeID.h>
|
#include <xrpl/shamap/SHAMapNodeID.h>
|
||||||
#include <xrpl/tx/apply.h>
|
#include <xrpl/tx/apply.h>
|
||||||
|
|
||||||
@@ -565,6 +566,8 @@ PeerImp::supportsFeature(ProtocolFeature f) const
|
|||||||
return protocol_ >= make_protocol(2, 1);
|
return protocol_ >= make_protocol(2, 1);
|
||||||
case ProtocolFeature::ValidatorList2Propagation:
|
case ProtocolFeature::ValidatorList2Propagation:
|
||||||
return protocol_ >= make_protocol(2, 2);
|
return protocol_ >= make_protocol(2, 2);
|
||||||
|
case ProtocolFeature::LedgerNodeDepth:
|
||||||
|
return protocol_ >= make_protocol(2, 3);
|
||||||
case ProtocolFeature::LedgerReplay:
|
case ProtocolFeature::LedgerReplay:
|
||||||
return ledgerReplayEnabled_;
|
return ledgerReplayEnabled_;
|
||||||
}
|
}
|
||||||
@@ -1611,7 +1614,8 @@ PeerImp::onMessage(std::shared_ptr<protocol::TMGetLedger> const& m)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify ledger node IDs
|
// Verify and parse ledger node IDs
|
||||||
|
std::vector<SHAMapNodeID> nodeIDs;
|
||||||
if (itype != protocol::liBASE)
|
if (itype != protocol::liBASE)
|
||||||
{
|
{
|
||||||
if (m->nodeids_size() <= 0)
|
if (m->nodeids_size() <= 0)
|
||||||
@@ -1620,13 +1624,16 @@ PeerImp::onMessage(std::shared_ptr<protocol::TMGetLedger> const& m)
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
nodeIDs.reserve(m->nodeids_size());
|
||||||
for (auto const& nodeId : m->nodeids())
|
for (auto const& nodeId : m->nodeids())
|
||||||
{
|
{
|
||||||
if (deserializeSHAMapNodeID(nodeId) == std::nullopt)
|
auto parsed = deserializeSHAMapNodeID(nodeId);
|
||||||
|
if (!parsed)
|
||||||
{
|
{
|
||||||
badData("Invalid SHAMap node ID");
|
badData("Invalid SHAMap node ID");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
nodeIDs.push_back(std::move(*parsed));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1649,10 +1656,11 @@ PeerImp::onMessage(std::shared_ptr<protocol::TMGetLedger> const& m)
|
|||||||
|
|
||||||
// Queue a job to process the request
|
// Queue a job to process the request
|
||||||
std::weak_ptr<PeerImp> const weak = shared_from_this();
|
std::weak_ptr<PeerImp> const weak = shared_from_this();
|
||||||
app_.getJobQueue().addJob(jtLEDGER_REQ, "RcvGetLedger", [weak, m]() {
|
app_.getJobQueue().addJob(
|
||||||
if (auto peer = weak.lock())
|
jtLEDGER_REQ, "RcvGetLedger", [weak, m, nodeIDs = std::move(nodeIDs)]() mutable {
|
||||||
peer->processLedgerRequest(m);
|
if (auto peer = weak.lock())
|
||||||
});
|
peer->processLedgerRequest(m, std::move(nodeIDs));
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
void
|
void
|
||||||
@@ -3361,7 +3369,9 @@ PeerImp::getTxSet(std::shared_ptr<protocol::TMGetLedger> const& m) const
|
|||||||
}
|
}
|
||||||
|
|
||||||
void
|
void
|
||||||
PeerImp::processLedgerRequest(std::shared_ptr<protocol::TMGetLedger> const& m)
|
PeerImp::processLedgerRequest(
|
||||||
|
std::shared_ptr<protocol::TMGetLedger> const& m,
|
||||||
|
std::vector<SHAMapNodeID> nodeIDs)
|
||||||
{
|
{
|
||||||
// Do not resource charge a peer responding to a relay
|
// Do not resource charge a peer responding to a relay
|
||||||
if (!m->has_requestcookie())
|
if (!m->has_requestcookie())
|
||||||
@@ -3446,26 +3456,25 @@ PeerImp::processLedgerRequest(std::shared_ptr<protocol::TMGetLedger> const& m)
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Add requested node data to reply
|
// Add requested node data to reply
|
||||||
if (m->nodeids_size() > 0)
|
if (!nodeIDs.empty())
|
||||||
{
|
{
|
||||||
std::uint32_t const defaultDepth = isHighLatency() ? 2 : 1;
|
std::uint32_t const defaultDepth = isHighLatency() ? 2 : 1;
|
||||||
auto const queryDepth{m->has_querydepth() ? m->querydepth() : defaultDepth};
|
auto const queryDepth{m->has_querydepth() ? m->querydepth() : defaultDepth};
|
||||||
|
|
||||||
std::vector<std::pair<SHAMapNodeID, Blob>> data;
|
std::vector<SHAMapNodeData> data;
|
||||||
|
auto const useLedgerNodeDepth = supportsFeature(ProtocolFeature::LedgerNodeDepth);
|
||||||
|
|
||||||
for (int i = 0;
|
for (auto const& nodeID : nodeIDs)
|
||||||
i < m->nodeids_size() && ledgerData.nodes_size() < Tuning::softMaxReplyNodes;
|
|
||||||
++i)
|
|
||||||
{
|
{
|
||||||
auto const shaMapNodeId{deserializeSHAMapNodeID(m->nodeids(i))};
|
if (ledgerData.nodes_size() >= Tuning::softMaxReplyNodes)
|
||||||
|
break;
|
||||||
|
|
||||||
data.clear();
|
data.clear();
|
||||||
data.reserve(Tuning::softMaxReplyNodes);
|
data.reserve(Tuning::softMaxReplyNodes);
|
||||||
|
|
||||||
try
|
try
|
||||||
{
|
{
|
||||||
// NOLINTNEXTLINE(bugprone-unchecked-optional-access) nodeids checked in onGetLedger
|
if (map->getNodeFat(nodeID, data, fatLeaves, queryDepth))
|
||||||
if (map->getNodeFat(*shaMapNodeId, data, fatLeaves, queryDepth))
|
|
||||||
{
|
{
|
||||||
JLOG(p_journal_.trace())
|
JLOG(p_journal_.trace())
|
||||||
<< "processLedgerRequest: getNodeFat got " << data.size() << " nodes";
|
<< "processLedgerRequest: getNodeFat got " << data.size() << " nodes";
|
||||||
@@ -3474,9 +3483,26 @@ PeerImp::processLedgerRequest(std::shared_ptr<protocol::TMGetLedger> const& m)
|
|||||||
{
|
{
|
||||||
if (ledgerData.nodes_size() >= Tuning::hardMaxReplyNodes)
|
if (ledgerData.nodes_size() >= Tuning::hardMaxReplyNodes)
|
||||||
break;
|
break;
|
||||||
|
|
||||||
protocol::TMLedgerNode* node{ledgerData.add_nodes()};
|
protocol::TMLedgerNode* node{ledgerData.add_nodes()};
|
||||||
node->set_nodeid(d.first.getRawString());
|
|
||||||
node->set_nodedata(d.second.data(), d.second.size());
|
node->set_nodedata(d.data.data(), d.data.size());
|
||||||
|
|
||||||
|
// When the LedgerNodeDepth protocol feature is not supported by the peer,
|
||||||
|
// we always set the `nodeid` field. However, when it is supported then we
|
||||||
|
// set the `id` field for inner nodes and the `depth` field for leaf nodes.
|
||||||
|
if (!useLedgerNodeDepth)
|
||||||
|
{
|
||||||
|
node->set_nodeid(d.nodeID.getRawString());
|
||||||
|
}
|
||||||
|
else if (d.isLeaf)
|
||||||
|
{
|
||||||
|
node->set_depth(d.nodeID.getDepth());
|
||||||
|
}
|
||||||
|
else
|
||||||
|
{
|
||||||
|
node->set_id(d.nodeID.getRawString());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else
|
else
|
||||||
@@ -3515,7 +3541,7 @@ PeerImp::processLedgerRequest(std::shared_ptr<protocol::TMGetLedger> const& m)
|
|||||||
info += ", no hash specified";
|
info += ", no hash specified";
|
||||||
|
|
||||||
JLOG(p_journal_.warn())
|
JLOG(p_journal_.warn())
|
||||||
<< "processLedgerRequest: getNodeFat with nodeId " << *shaMapNodeId
|
<< "processLedgerRequest: getNodeFat with nodeId " << nodeID
|
||||||
<< " and ledger info type " << info << " throws exception: " << e.what();
|
<< " and ledger info type " << info << " throws exception: " << e.what();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -14,6 +14,7 @@
|
|||||||
#include <xrpl/protocol/STTx.h>
|
#include <xrpl/protocol/STTx.h>
|
||||||
#include <xrpl/protocol/STValidation.h>
|
#include <xrpl/protocol/STValidation.h>
|
||||||
#include <xrpl/resource/Fees.h>
|
#include <xrpl/resource/Fees.h>
|
||||||
|
#include <xrpl/shamap/SHAMapNodeID.h>
|
||||||
|
|
||||||
#include <boost/circular_buffer.hpp>
|
#include <boost/circular_buffer.hpp>
|
||||||
#include <boost/endian/conversion.hpp>
|
#include <boost/endian/conversion.hpp>
|
||||||
@@ -792,7 +793,9 @@ private:
|
|||||||
getTxSet(std::shared_ptr<protocol::TMGetLedger> const& m) const;
|
getTxSet(std::shared_ptr<protocol::TMGetLedger> const& m) const;
|
||||||
|
|
||||||
void
|
void
|
||||||
processLedgerRequest(std::shared_ptr<protocol::TMGetLedger> const& m);
|
processLedgerRequest(
|
||||||
|
std::shared_ptr<protocol::TMGetLedger> const& m,
|
||||||
|
std::vector<SHAMapNodeID> nodeIDs);
|
||||||
};
|
};
|
||||||
|
|
||||||
//------------------------------------------------------------------------------
|
//------------------------------------------------------------------------------
|
||||||
|
|||||||
@@ -28,6 +28,7 @@ namespace xrpl {
|
|||||||
constexpr ProtocolVersion const supportedProtocolList[]{
|
constexpr ProtocolVersion const supportedProtocolList[]{
|
||||||
{2, 1},
|
{2, 1},
|
||||||
{2, 2},
|
{2, 2},
|
||||||
|
{2, 3},
|
||||||
};
|
};
|
||||||
|
|
||||||
// This ugly construct ensures that supportedProtocolList is sorted in strictly
|
// This ugly construct ensures that supportedProtocolList is sorted in strictly
|
||||||
|
|||||||
Reference in New Issue
Block a user