1#include <xrpld/app/consensus/RCLConsensus.h>
2#include <xrpld/app/consensus/RCLValidations.h>
3#include <xrpld/app/ledger/AcceptedLedger.h>
4#include <xrpld/app/ledger/InboundLedgers.h>
5#include <xrpld/app/ledger/LedgerMaster.h>
6#include <xrpld/app/ledger/LedgerToJson.h>
7#include <xrpld/app/ledger/LocalTxs.h>
8#include <xrpld/app/ledger/OpenLedger.h>
9#include <xrpld/app/ledger/OrderBookDB.h>
10#include <xrpld/app/ledger/TransactionMaster.h>
11#include <xrpld/app/main/LoadManager.h>
12#include <xrpld/app/main/Tuning.h>
13#include <xrpld/app/misc/AmendmentTable.h>
14#include <xrpld/app/misc/DeliverMax.h>
15#include <xrpld/app/misc/HashRouter.h>
16#include <xrpld/app/misc/LoadFeeTrack.h>
17#include <xrpld/app/misc/NetworkOPs.h>
18#include <xrpld/app/misc/Transaction.h>
19#include <xrpld/app/misc/TxQ.h>
20#include <xrpld/app/misc/ValidatorKeys.h>
21#include <xrpld/app/misc/ValidatorList.h>
22#include <xrpld/app/misc/detail/AccountTxPaging.h>
23#include <xrpld/app/rdb/backend/SQLiteDatabase.h>
24#include <xrpld/app/tx/apply.h>
25#include <xrpld/consensus/Consensus.h>
26#include <xrpld/consensus/ConsensusParms.h>
27#include <xrpld/overlay/Cluster.h>
28#include <xrpld/overlay/Overlay.h>
29#include <xrpld/overlay/predicates.h>
30#include <xrpld/perflog/PerfLog.h>
31#include <xrpld/rpc/BookChanges.h>
32#include <xrpld/rpc/CTID.h>
33#include <xrpld/rpc/DeliveredAmount.h>
34#include <xrpld/rpc/MPTokenIssuanceID.h>
35#include <xrpld/rpc/ServerHandler.h>
37#include <xrpl/basics/UptimeClock.h>
38#include <xrpl/basics/mulDiv.h>
39#include <xrpl/basics/safe_cast.h>
40#include <xrpl/basics/scope.h>
41#include <xrpl/beast/utility/rngfill.h>
42#include <xrpl/crypto/RFC1751.h>
43#include <xrpl/crypto/csprng.h>
44#include <xrpl/protocol/BuildInfo.h>
45#include <xrpl/protocol/Feature.h>
46#include <xrpl/protocol/MultiApiJson.h>
47#include <xrpl/protocol/NFTSyntheticSerializer.h>
48#include <xrpl/protocol/RPCErr.h>
49#include <xrpl/protocol/TxFlags.h>
50#include <xrpl/protocol/jss.h>
51#include <xrpl/resource/Fees.h>
52#include <xrpl/resource/ResourceManager.h>
54#include <boost/asio/ip/host_name.hpp>
55#include <boost/asio/steady_timer.hpp>
94 "ripple::NetworkOPsImp::TransactionStatus::TransactionStatus : "
137 std::chrono::steady_clock::time_point
start_ =
198 return !(*
this != b);
217 boost::asio::io_context& io_svc,
231 app_.logs().journal(
"FeeVote")),
234 app.getInboundTransactions(),
235 beast::get_abstract_clock<
std::chrono::steady_clock>(),
237 app_.logs().journal(
"LedgerConsensus"))
239 validatorKeys.keys ? validatorKeys.keys->publicKey
242 validatorKeys.keys ? validatorKeys.keys->masterPublicKey
437 getServerInfo(
bool human,
bool admin,
bool counters)
override;
464 TER result)
override;
498 bool historyOnly)
override;
504 bool historyOnly)
override;
576 catch (boost::system::system_error
const& e)
579 <<
"NetworkOPs: heartbeatTimer cancel error: " << e.what();
586 catch (boost::system::system_error
const& e)
589 <<
"NetworkOPs: clusterTimer cancel error: " << e.what();
596 catch (boost::system::system_error
const& e)
599 <<
"NetworkOPs: accountHistoryTxTimer cancel error: "
604 using namespace std::chrono_literals;
614 boost::asio::steady_timer& timer,
797 template <
class Handler>
799 Handler
const& handler,
801 :
hook(collector->make_hook(handler))
804 "Disconnected_duration"))
807 "Connected_duration"))
809 collector->make_gauge(
"State_Accounting",
"Syncing_duration"))
812 "Tracking_duration"))
814 collector->make_gauge(
"State_Accounting",
"Full_duration"))
817 "Disconnected_transitions"))
820 "Connected_transitions"))
823 "Syncing_transitions"))
826 "Tracking_transitions"))
828 collector->make_gauge(
"State_Accounting",
"Full_transitions"))
857 {
"disconnected",
"connected",
"syncing",
"tracking",
"full"}};
919 static std::string const hostname = boost::asio::ip::host_name();
926 static std::string const shroudedHostId = [
this]() {
932 return shroudedHostId;
947 boost::asio::steady_timer& timer,
954 [
this, onExpire, onError](boost::system::error_code
const& e) {
955 if ((e.value() == boost::system::errc::success) &&
956 (!m_job_queue.isStopped()))
961 if (e.value() != boost::system::errc::success &&
962 e.value() != boost::asio::error::operation_aborted)
965 JLOG(m_journal.error())
966 <<
"Timer got error '" << e.message()
967 <<
"'. Restarting timer.";
972 timer.expires_after(expiry_time);
973 timer.async_wait(std::move(*optionalCountedHandler));
978NetworkOPsImp::setHeartbeatTimer()
982 mConsensus.parms().ledgerGRANULARITY,
984 m_job_queue.addJob(jtNETOP_TIMER,
"NetOPs.heartbeat", [this]() {
985 processHeartbeatTimer();
988 [
this]() { setHeartbeatTimer(); });
992NetworkOPsImp::setClusterTimer()
994 using namespace std::chrono_literals;
1001 processClusterTimer();
1004 [
this]() { setClusterTimer(); });
1010 JLOG(m_journal.debug()) <<
"Scheduling AccountHistory job for account "
1012 using namespace std::chrono_literals;
1014 accountHistoryTxTimer_,
1016 [
this, subInfo]() { addAccountHistoryJob(subInfo); },
1017 [
this, subInfo]() { setAccountHistoryJobTimer(subInfo); });
1021NetworkOPsImp::processHeartbeatTimer()
1024 "Heartbeat Timer", mConsensus.validating(), m_journal);
1032 std::size_t const numPeers = app_.overlay().size();
1035 if (numPeers < minPeerCount_)
1037 if (mMode != OperatingMode::DISCONNECTED)
1039 setMode(OperatingMode::DISCONNECTED);
1041 ss <<
"Node count (" << numPeers <<
") has fallen "
1042 <<
"below required minimum (" << minPeerCount_ <<
").";
1043 JLOG(m_journal.warn()) << ss.
str();
1044 CLOG(clog.
ss()) <<
"set mode to DISCONNECTED: " << ss.
str();
1049 <<
"already DISCONNECTED. too few peers (" << numPeers
1050 <<
"), need at least " << minPeerCount_;
1057 setHeartbeatTimer();
1062 if (mMode == OperatingMode::DISCONNECTED)
1064 setMode(OperatingMode::CONNECTED);
1065 JLOG(m_journal.info())
1066 <<
"Node count (" << numPeers <<
") is sufficient.";
1067 CLOG(clog.
ss()) <<
"setting mode to CONNECTED based on " << numPeers
1073 auto origMode = mMode.load();
1074 CLOG(clog.
ss()) <<
"mode: " << strOperatingMode(origMode,
true);
1075 if (mMode == OperatingMode::SYNCING)
1076 setMode(OperatingMode::SYNCING);
1077 else if (mMode == OperatingMode::CONNECTED)
1078 setMode(OperatingMode::CONNECTED);
1079 auto newMode = mMode.load();
1080 if (origMode != newMode)
1083 <<
", changing to " << strOperatingMode(newMode,
true);
1085 CLOG(clog.
ss()) <<
". ";
1088 mConsensus.timerEntry(app_.timeKeeper().closeTime(), clog.
ss());
1090 CLOG(clog.
ss()) <<
"consensus phase " << to_string(mLastConsensusPhase);
1092 if (mLastConsensusPhase != currPhase)
1094 reportConsensusStateChange(currPhase);
1095 mLastConsensusPhase = currPhase;
1096 CLOG(clog.
ss()) <<
" changed to " << to_string(mLastConsensusPhase);
1098 CLOG(clog.
ss()) <<
". ";
1100 setHeartbeatTimer();
1104NetworkOPsImp::processClusterTimer()
1106 if (app_.cluster().size() == 0)
1109 using namespace std::chrono_literals;
1111 bool const update = app_.cluster().update(
1112 app_.nodeIdentity().first,
1114 (m_ledgerMaster.getValidatedLedgerAge() <= 4min)
1115 ? app_.getFeeTrack().getLocalFee()
1117 app_.timeKeeper().now());
1121 JLOG(m_journal.debug()) <<
"Too soon to send cluster update";
1126 protocol::TMCluster cluster;
1127 app_.cluster().for_each([&cluster](
ClusterNode const& node) {
1128 protocol::TMClusterNode& n = *cluster.add_clusternodes();
1133 n.set_nodename(node.
name());
1137 for (
auto& item : gossip.
items)
1139 protocol::TMLoadSource& node = *cluster.add_loadsources();
1140 node.set_name(to_string(item.address));
1141 node.set_cost(item.balance);
1143 app_.overlay().foreach(
send_if(
1155 if (mode == OperatingMode::FULL && admin)
1157 auto const consensusMode = mConsensus.mode();
1158 if (consensusMode != ConsensusMode::wrongLedger)
1160 if (consensusMode == ConsensusMode::proposing)
1163 if (mConsensus.validating())
1164 return "validating";
1174 if (isNeedNetworkLedger())
1182 m_ledgerMaster.getValidatedRules().enabled(featureBatch))
1184 JLOG(m_journal.error())
1185 <<
"Submitted transaction invalid: tfInnerBatchTxn flag present.";
1192 auto const txid = trans->getTransactionID();
1193 auto const flags = app_.getHashRouter().getFlags(txid);
1195 if ((flags & HashRouterFlags::BAD) != HashRouterFlags::UNDEFINED)
1197 JLOG(m_journal.warn()) <<
"Submitted transaction cached bad";
1204 app_.getHashRouter(),
1206 m_ledgerMaster.getValidatedRules(),
1209 if (validity != Validity::Valid)
1211 JLOG(m_journal.warn())
1212 <<
"Submitted transaction invalid: " << reason;
1218 JLOG(m_journal.warn())
1219 <<
"Exception checking transaction " << txid <<
": " << ex.
what();
1228 m_job_queue.addJob(
jtTRANSACTION,
"submitTxn", [
this, tx]() {
1230 processTransaction(t,
false,
false, FailHard::no);
1237 auto const newFlags = app_.getHashRouter().getFlags(transaction->getID());
1239 if ((newFlags & HashRouterFlags::BAD) != HashRouterFlags::UNDEFINED)
1242 JLOG(m_journal.warn()) << transaction->getID() <<
": cached bad!\n";
1243 transaction->setStatus(
INVALID);
1248 auto const view = m_ledgerMaster.getCurrentLedger();
1253 auto const sttx = *transaction->getSTransaction();
1254 if (sttx.isFlag(
tfInnerBatchTxn) && view->rules().enabled(featureBatch))
1256 transaction->setStatus(
INVALID);
1258 app_.getHashRouter().setFlags(
1259 transaction->getID(), HashRouterFlags::BAD);
1266 auto const [validity, reason] =
1267 checkValidity(app_.getHashRouter(), sttx, view->rules(), app_.config());
1269 validity == Validity::Valid,
1270 "ripple::NetworkOPsImp::processTransaction : valid validity");
1273 if (validity == Validity::SigBad)
1275 JLOG(m_journal.info()) <<
"Transaction has bad signature: " << reason;
1276 transaction->setStatus(
INVALID);
1278 app_.getHashRouter().setFlags(
1279 transaction->getID(), HashRouterFlags::BAD);
1284 app_.getMasterTransaction().canonicalize(&transaction);
1290NetworkOPsImp::processTransaction(
1296 auto ev = m_job_queue.makeLoadEvent(
jtTXN_PROC,
"ProcessTXN");
1299 if (!preProcessTransaction(transaction))
1303 doTransactionSync(transaction, bUnlimited, failType);
1305 doTransactionAsync(transaction, bUnlimited, failType);
1309NetworkOPsImp::doTransactionAsync(
1316 if (transaction->getApplying())
1319 mTransactions.push_back(
1321 transaction->setApplying();
1323 if (mDispatchState == DispatchState::none)
1325 if (m_job_queue.addJob(
1326 jtBATCH,
"transactionBatch", [
this]() { transactionBatch(); }))
1328 mDispatchState = DispatchState::scheduled;
1334NetworkOPsImp::doTransactionSync(
1341 if (!transaction->getApplying())
1343 mTransactions.push_back(
1345 transaction->setApplying();
1348 doTransactionSyncBatch(
1350 return transaction->getApplying();
1355NetworkOPsImp::doTransactionSyncBatch(
1361 if (mDispatchState == DispatchState::running)
1370 if (mTransactions.size())
1373 if (m_job_queue.addJob(
jtBATCH,
"transactionBatch", [
this]() {
1377 mDispatchState = DispatchState::scheduled;
1381 }
while (retryCallback(lock));
1387 auto ev = m_job_queue.makeLoadEvent(
jtTXN_PROC,
"ProcessTXNSet");
1390 for (
auto const& [_, tx] :
set)
1395 if (transaction->getStatus() ==
INVALID)
1397 if (!reason.
empty())
1399 JLOG(m_journal.trace())
1400 <<
"Exception checking transaction: " << reason;
1402 app_.getHashRouter().setFlags(
1403 tx->getTransactionID(), HashRouterFlags::BAD);
1408 if (!preProcessTransaction(transaction))
1419 for (
auto& transaction : candidates)
1421 if (!transaction->getApplying())
1423 transactions.
emplace_back(transaction,
false,
false, FailHard::no);
1424 transaction->setApplying();
1428 if (mTransactions.empty())
1429 mTransactions.swap(transactions);
1432 mTransactions.reserve(mTransactions.size() + transactions.
size());
1433 for (
auto& t : transactions)
1434 mTransactions.push_back(std::move(t));
1436 if (mTransactions.empty())
1438 JLOG(m_journal.debug()) <<
"No transaction to process!";
1445 "ripple::NetworkOPsImp::processTransactionSet has lock");
1447 mTransactions.begin(), mTransactions.end(), [](
auto const& t) {
1448 return t.transaction->getApplying();
1454NetworkOPsImp::transactionBatch()
1458 if (mDispatchState == DispatchState::running)
1461 while (mTransactions.size())
1472 mTransactions.
swap(transactions);
1474 !transactions.
empty(),
1475 "ripple::NetworkOPsImp::apply : non-empty transactions");
1477 mDispatchState != DispatchState::running,
1478 "ripple::NetworkOPsImp::apply : is not running");
1480 mDispatchState = DispatchState::running;
1486 bool changed =
false;
1500 if (e.failType == FailHard::yes)
1503 auto const result = app_.getTxQ().apply(
1504 app_, view, e.transaction->getSTransaction(), flags, j);
1505 e.result = result.ter;
1506 e.applied = result.applied;
1507 changed = changed || result.applied;
1516 if (
auto const l = m_ledgerMaster.getValidatedLedger())
1517 validatedLedgerIndex = l->info().seq;
1519 auto newOL = app_.openLedger().current();
1522 e.transaction->clearSubmitResult();
1526 pubProposedTransaction(
1527 newOL, e.transaction->getSTransaction(), e.result);
1528 e.transaction->setApplied();
1531 e.transaction->setResult(e.result);
1534 app_.getHashRouter().setFlags(
1535 e.transaction->getID(), HashRouterFlags::BAD);
1544 JLOG(m_journal.info())
1545 <<
"TransactionResult: " << token <<
": " << human;
1550 bool addLocal = e.local;
1554 JLOG(m_journal.debug())
1555 <<
"Transaction is now included in open ledger";
1556 e.transaction->setStatus(
INCLUDED);
1561 auto const& txCur = e.transaction->getSTransaction();
1564 for (
auto txNext = m_ledgerMaster.popAcctTransaction(txCur);
1566 txNext = m_ledgerMaster.popAcctTransaction(txCur), ++count)
1573 if (t->getApplying())
1575 submit_held.
emplace_back(t,
false,
false, FailHard::no);
1584 JLOG(m_journal.info()) <<
"Transaction is obsolete";
1585 e.transaction->setStatus(
OBSOLETE);
1589 JLOG(m_journal.debug())
1590 <<
"Transaction is likely to claim a"
1591 <<
" fee, but is queued until fee drops";
1593 e.transaction->setStatus(
HELD);
1597 m_ledgerMaster.addHeldTransaction(e.transaction);
1598 e.transaction->setQueued();
1599 e.transaction->setKept();
1605 if (e.failType != FailHard::yes)
1607 auto const lastLedgerSeq =
1608 e.transaction->getSTransaction()->at(
1609 ~sfLastLedgerSequence);
1610 auto const ledgersLeft = lastLedgerSeq
1612 m_ledgerMaster.getCurrentLedgerIndex()
1631 (ledgersLeft && ledgersLeft <= LocalTxs::holdLedgers) ||
1632 app_.getHashRouter().setFlags(
1633 e.transaction->getID(), HashRouterFlags::HELD))
1636 JLOG(m_journal.debug())
1637 <<
"Transaction should be held: " << e.result;
1638 e.transaction->setStatus(
HELD);
1639 m_ledgerMaster.addHeldTransaction(e.transaction);
1640 e.transaction->setKept();
1643 JLOG(m_journal.debug())
1644 <<
"Not holding transaction "
1645 << e.transaction->getID() <<
": "
1646 << (e.local ?
"local" :
"network") <<
", "
1647 <<
"result: " << e.result <<
" ledgers left: "
1648 << (ledgersLeft ? to_string(*ledgersLeft)
1654 JLOG(m_journal.debug())
1655 <<
"Status other than success " << e.result;
1656 e.transaction->setStatus(
INVALID);
1659 auto const enforceFailHard =
1660 e.failType == FailHard::yes && !
isTesSuccess(e.result);
1662 if (addLocal && !enforceFailHard)
1664 m_localTX->push_back(
1665 m_ledgerMaster.getCurrentLedgerIndex(),
1666 e.transaction->getSTransaction());
1667 e.transaction->setKept();
1671 ((mMode != OperatingMode::FULL) &&
1672 (e.failType != FailHard::yes) && e.local) ||
1677 app_.getHashRouter().shouldRelay(e.transaction->getID());
1678 if (
auto const sttx = *(e.transaction->getSTransaction());
1686 protocol::TMTransaction tx;
1690 tx.set_rawtransaction(s.
data(), s.
size());
1691 tx.set_status(protocol::tsCURRENT);
1692 tx.set_receivetimestamp(
1693 app_.timeKeeper().now().time_since_epoch().count());
1696 app_.overlay().relay(e.transaction->getID(), tx, *toSkip);
1697 e.transaction->setBroadcast();
1701 if (validatedLedgerIndex)
1703 auto [fee, accountSeq, availableSeq] =
1704 app_.getTxQ().getTxRequiredFeeAndSeq(
1705 *newOL, e.transaction->getSTransaction());
1706 e.transaction->setCurrentLedgerState(
1707 *validatedLedgerIndex, fee, accountSeq, availableSeq);
1715 e.transaction->clearApplying();
1717 if (!submit_held.
empty())
1719 if (mTransactions.empty())
1720 mTransactions.swap(submit_held);
1723 mTransactions.reserve(mTransactions.size() + submit_held.
size());
1724 for (
auto& e : submit_held)
1725 mTransactions.push_back(std::move(e));
1731 mDispatchState = DispatchState::none;
1739NetworkOPsImp::getOwnerInfo(
1744 auto root = keylet::ownerDir(account);
1745 auto sleNode = lpLedger->read(keylet::page(
root));
1752 for (
auto const& uDirEntry : sleNode->getFieldV256(sfIndexes))
1754 auto sleCur = lpLedger->read(keylet::child(uDirEntry));
1757 "ripple::NetworkOPsImp::getOwnerInfo : non-null child SLE");
1759 switch (sleCur->getType())
1762 if (!jvObjects.
isMember(jss::offers))
1763 jvObjects[jss::offers] =
1766 jvObjects[jss::offers].
append(
1767 sleCur->getJson(JsonOptions::none));
1770 case ltRIPPLE_STATE:
1771 if (!jvObjects.
isMember(jss::ripple_lines))
1773 jvObjects[jss::ripple_lines] =
1777 jvObjects[jss::ripple_lines].
append(
1778 sleCur->getJson(JsonOptions::none));
1781 case ltACCOUNT_ROOT:
1786 "ripple::NetworkOPsImp::getOwnerInfo : invalid "
1793 uNodeDir = sleNode->getFieldU64(sfIndexNext);
1797 sleNode = lpLedger->read(keylet::page(
root, uNodeDir));
1800 "ripple::NetworkOPsImp::getOwnerInfo : read next page");
1813NetworkOPsImp::isBlocked()
1815 return isAmendmentBlocked() || isUNLBlocked();
1819NetworkOPsImp::isAmendmentBlocked()
1821 return amendmentBlocked_;
1825NetworkOPsImp::setAmendmentBlocked()
1827 amendmentBlocked_ =
true;
1828 setMode(OperatingMode::CONNECTED);
1832NetworkOPsImp::isAmendmentWarned()
1834 return !amendmentBlocked_ && amendmentWarned_;
1838NetworkOPsImp::setAmendmentWarned()
1840 amendmentWarned_ =
true;
1844NetworkOPsImp::clearAmendmentWarned()
1846 amendmentWarned_ =
false;
1850NetworkOPsImp::isUNLBlocked()
1856NetworkOPsImp::setUNLBlocked()
1859 setMode(OperatingMode::CONNECTED);
1863NetworkOPsImp::clearUNLBlocked()
1865 unlBlocked_ =
false;
1869NetworkOPsImp::checkLastClosedLedger(
1878 JLOG(m_journal.trace()) <<
"NetworkOPsImp::checkLastClosedLedger";
1880 auto const ourClosed = m_ledgerMaster.getClosedLedger();
1885 uint256 closedLedger = ourClosed->info().hash;
1886 uint256 prevClosedLedger = ourClosed->info().parentHash;
1887 JLOG(m_journal.trace()) <<
"OurClosed: " << closedLedger;
1888 JLOG(m_journal.trace()) <<
"PrevClosed: " << prevClosedLedger;
1893 auto& validations = app_.getValidations();
1894 JLOG(m_journal.debug())
1895 <<
"ValidationTrie " <<
Json::Compact(validations.getJsonTrie());
1899 peerCounts[closedLedger] = 0;
1900 if (mMode >= OperatingMode::TRACKING)
1901 peerCounts[closedLedger]++;
1903 for (
auto& peer : peerList)
1905 uint256 peerLedger = peer->getClosedLedgerHash();
1908 ++peerCounts[peerLedger];
1911 for (
auto const& it : peerCounts)
1912 JLOG(m_journal.debug()) <<
"L: " << it.first <<
" n=" << it.second;
1914 uint256 preferredLCL = validations.getPreferredLCL(
1916 m_ledgerMaster.getValidLedgerIndex(),
1919 bool switchLedgers = preferredLCL != closedLedger;
1921 closedLedger = preferredLCL;
1923 if (switchLedgers && (closedLedger == prevClosedLedger))
1926 JLOG(m_journal.info()) <<
"We won't switch to our own previous ledger";
1927 networkClosed = ourClosed->info().hash;
1928 switchLedgers =
false;
1931 networkClosed = closedLedger;
1936 auto consensus = m_ledgerMaster.getLedgerByHash(closedLedger);
1939 consensus = app_.getInboundLedgers().acquire(
1940 closedLedger, 0, InboundLedger::Reason::CONSENSUS);
1943 (!m_ledgerMaster.canBeCurrent(consensus) ||
1944 !m_ledgerMaster.isCompatible(
1945 *consensus, m_journal.debug(),
"Not switching")))
1949 networkClosed = ourClosed->info().hash;
1953 JLOG(m_journal.warn()) <<
"We are not running on the consensus ledger";
1954 JLOG(m_journal.info()) <<
"Our LCL: " << ourClosed->info().hash
1956 JLOG(m_journal.info()) <<
"Net LCL " << closedLedger;
1958 if ((mMode == OperatingMode::TRACKING) || (mMode == OperatingMode::FULL))
1960 setMode(OperatingMode::CONNECTED);
1968 switchLastClosedLedger(consensus);
1975NetworkOPsImp::switchLastClosedLedger(
1979 JLOG(m_journal.error())
1980 <<
"JUMP last closed ledger to " << newLCL->info().hash;
1982 clearNeedNetworkLedger();
1985 app_.getTxQ().processClosedLedger(app_, *newLCL,
true);
1992 auto retries = m_localTX->getTxSet();
1993 auto const lastVal = app_.getLedgerMaster().getValidatedLedger();
1998 rules.
emplace(app_.config().features);
1999 app_.openLedger().accept(
2010 return app_.getTxQ().accept(app_, view);
2014 m_ledgerMaster.switchLCL(newLCL);
2016 protocol::TMStatusChange s;
2017 s.set_newevent(protocol::neSWITCHED_LEDGER);
2018 s.set_ledgerseq(newLCL->info().seq);
2019 s.set_networktime(app_.timeKeeper().now().time_since_epoch().count());
2020 s.set_ledgerhashprevious(
2021 newLCL->info().parentHash.begin(), newLCL->info().parentHash.size());
2022 s.set_ledgerhash(newLCL->info().hash.begin(), newLCL->info().hash.size());
2024 app_.overlay().foreach(
2029NetworkOPsImp::beginConsensus(
2035 "ripple::NetworkOPsImp::beginConsensus : nonzero input");
2037 auto closingInfo = m_ledgerMaster.getCurrentLedger()->info();
2039 JLOG(m_journal.info()) <<
"Consensus time for #" << closingInfo.seq
2040 <<
" with LCL " << closingInfo.parentHash;
2042 auto prevLedger = m_ledgerMaster.getLedgerByHash(closingInfo.parentHash);
2047 if (mMode == OperatingMode::FULL)
2049 JLOG(m_journal.warn()) <<
"Don't have LCL, going to tracking";
2050 setMode(OperatingMode::TRACKING);
2051 CLOG(clog) <<
"beginConsensus Don't have LCL, going to tracking. ";
2054 CLOG(clog) <<
"beginConsensus no previous ledger. ";
2059 prevLedger->info().hash == closingInfo.parentHash,
2060 "ripple::NetworkOPsImp::beginConsensus : prevLedger hash matches "
2063 closingInfo.parentHash == m_ledgerMaster.getClosedLedger()->info().hash,
2064 "ripple::NetworkOPsImp::beginConsensus : closedLedger parent matches "
2067 app_.validators().setNegativeUNL(prevLedger->negativeUNL());
2068 TrustChanges const changes = app_.validators().updateTrusted(
2069 app_.getValidations().getCurrentNodeIDs(),
2070 closingInfo.parentCloseTime,
2073 app_.getHashRouter());
2075 if (!changes.
added.empty() || !changes.
removed.empty())
2077 app_.getValidations().trustChanged(changes.
added, changes.
removed);
2079 app_.getAmendmentTable().trustChanged(
2080 app_.validators().getQuorumKeys().second);
2083 mConsensus.startRound(
2084 app_.timeKeeper().closeTime(),
2092 if (mLastConsensusPhase != currPhase)
2094 reportConsensusStateChange(currPhase);
2095 mLastConsensusPhase = currPhase;
2098 JLOG(m_journal.debug()) <<
"Initiating consensus engine";
2105 auto const& peerKey = peerPos.
publicKey();
2106 if (validatorPK_ == peerKey || validatorMasterPK_ == peerKey)
2117 JLOG(m_journal.error())
2118 <<
"Received a proposal signed by MY KEY from a peer. This may "
2119 "indicate a misconfiguration where another node has the same "
2120 "validator key, or may be caused by unusual message routing and "
2125 return mConsensus.peerProposal(app_.timeKeeper().closeTime(), peerPos);
2136 protocol::TMHaveTransactionSet msg;
2137 msg.set_hash(map->getHash().as_uint256().begin(), 256 / 8);
2138 msg.set_status(protocol::tsHAVE);
2139 app_.overlay().foreach(
2144 mConsensus.gotTxSet(app_.timeKeeper().closeTime(),
RCLTxSet{map});
2150 uint256 deadLedger = m_ledgerMaster.getClosedLedger()->info().parentHash;
2152 for (
auto const& it : app_.overlay().getActivePeers())
2154 if (it && (it->getClosedLedgerHash() == deadLedger))
2156 JLOG(m_journal.trace()) <<
"Killing obsolete peer status";
2163 checkLastClosedLedger(app_.overlay().getActivePeers(), networkClosed);
2165 if (networkClosed.
isZero())
2167 CLOG(clog) <<
"endConsensus last closed ledger is zero. ";
2177 if (((mMode == OperatingMode::CONNECTED) ||
2178 (mMode == OperatingMode::SYNCING)) &&
2184 if (!needNetworkLedger_)
2185 setMode(OperatingMode::TRACKING);
2188 if (((mMode == OperatingMode::CONNECTED) ||
2189 (mMode == OperatingMode::TRACKING)) &&
2195 auto current = m_ledgerMaster.getCurrentLedger();
2196 if (app_.timeKeeper().now() < (
current->info().parentCloseTime +
2197 2 *
current->info().closeTimeResolution))
2199 setMode(OperatingMode::FULL);
2203 beginConsensus(networkClosed, clog);
2207NetworkOPsImp::consensusViewChange()
2209 if ((mMode == OperatingMode::FULL) || (mMode == OperatingMode::TRACKING))
2211 setMode(OperatingMode::CONNECTED);
2221 if (!mStreamMaps[sManifests].empty())
2225 jvObj[jss::type] =
"manifestReceived";
2228 jvObj[jss::signing_key] =
2232 jvObj[jss::signature] =
strHex(*sig);
2235 jvObj[jss::domain] = mo.
domain;
2238 for (
auto i = mStreamMaps[sManifests].begin();
2239 i != mStreamMaps[sManifests].end();)
2241 if (
auto p = i->second.lock())
2243 p->send(jvObj,
true);
2248 i = mStreamMaps[sManifests].erase(i);
2254NetworkOPsImp::ServerFeeSummary::ServerFeeSummary(
2258 : loadFactorServer{loadFeeTrack.getLoadFactor()}
2259 , loadBaseServer{loadFeeTrack.getLoadBase()}
2261 , em{
std::move(escalationMetrics)}
2271 em.has_value() != b.
em.has_value())
2277 em->minProcessingFeeLevel != b.
em->minProcessingFeeLevel ||
2278 em->openLedgerFeeLevel != b.
em->openLedgerFeeLevel ||
2279 em->referenceFeeLevel != b.
em->referenceFeeLevel);
2312 jvObj[jss::type] =
"serverStatus";
2314 jvObj[jss::load_base] = f.loadBaseServer;
2315 jvObj[jss::load_factor_server] = f.loadFactorServer;
2316 jvObj[jss::base_fee] = f.baseFee.jsonClipped();
2321 safe_cast<std::uint64_t>(f.loadFactorServer),
2323 f.em->openLedgerFeeLevel,
2325 f.em->referenceFeeLevel)
2328 jvObj[jss::load_factor] =
trunc32(loadFactor);
2329 jvObj[jss::load_factor_fee_escalation] =
2330 f.em->openLedgerFeeLevel.jsonClipped();
2331 jvObj[jss::load_factor_fee_queue] =
2332 f.em->minProcessingFeeLevel.jsonClipped();
2333 jvObj[jss::load_factor_fee_reference] =
2334 f.em->referenceFeeLevel.jsonClipped();
2337 jvObj[jss::load_factor] = f.loadFactorServer;
2351 p->send(jvObj,
true);
2368 if (!streamMap.empty())
2371 jvObj[jss::type] =
"consensusPhase";
2372 jvObj[jss::consensus] =
to_string(phase);
2374 for (
auto i = streamMap.begin(); i != streamMap.end();)
2376 if (
auto p = i->second.lock())
2378 p->send(jvObj,
true);
2383 i = streamMap.erase(i);
2399 auto const signerPublic = val->getSignerPublic();
2401 jvObj[jss::type] =
"validationReceived";
2402 jvObj[jss::validation_public_key] =
2404 jvObj[jss::ledger_hash] =
to_string(val->getLedgerHash());
2405 jvObj[jss::signature] =
strHex(val->getSignature());
2406 jvObj[jss::full] = val->isFull();
2407 jvObj[jss::flags] = val->getFlags();
2408 jvObj[jss::signing_time] = *(*val)[~sfSigningTime];
2409 jvObj[jss::data] =
strHex(val->getSerializer().slice());
2412 if (
auto version = (*val)[~sfServerVersion])
2415 if (
auto cookie = (*val)[~sfCookie])
2418 if (
auto hash = (*val)[~sfValidatedHash])
2419 jvObj[jss::validated_hash] =
strHex(*hash);
2421 auto const masterKey =
2424 if (masterKey != signerPublic)
2429 if (
auto const seq = (*val)[~sfLedgerSequence])
2430 jvObj[jss::ledger_index] = *seq;
2432 if (val->isFieldPresent(sfAmendments))
2435 for (
auto const& amendment : val->getFieldV256(sfAmendments))
2436 jvObj[jss::amendments].append(
to_string(amendment));
2439 if (
auto const closeTime = (*val)[~sfCloseTime])
2440 jvObj[jss::close_time] = *closeTime;
2442 if (
auto const loadFee = (*val)[~sfLoadFee])
2443 jvObj[jss::load_fee] = *loadFee;
2445 if (
auto const baseFee = val->at(~sfBaseFee))
2446 jvObj[jss::base_fee] =
static_cast<double>(*baseFee);
2448 if (
auto const reserveBase = val->at(~sfReserveBase))
2449 jvObj[jss::reserve_base] = *reserveBase;
2451 if (
auto const reserveInc = val->at(~sfReserveIncrement))
2452 jvObj[jss::reserve_inc] = *reserveInc;
2456 if (
auto const baseFeeXRP = ~val->at(~sfBaseFeeDrops);
2457 baseFeeXRP && baseFeeXRP->native())
2458 jvObj[jss::base_fee] = baseFeeXRP->xrp().jsonClipped();
2460 if (
auto const reserveBaseXRP = ~val->at(~sfReserveBaseDrops);
2461 reserveBaseXRP && reserveBaseXRP->native())
2462 jvObj[jss::reserve_base] = reserveBaseXRP->xrp().jsonClipped();
2464 if (
auto const reserveIncXRP = ~val->at(~sfReserveIncrementDrops);
2465 reserveIncXRP && reserveIncXRP->native())
2466 jvObj[jss::reserve_inc] = reserveIncXRP->xrp().jsonClipped();
2475 if (jvTx.
isMember(jss::ledger_index))
2477 jvTx[jss::ledger_index] =
2485 if (
auto p = i->second.lock())
2489 [&](
Json::Value const& jv) { p->send(jv,
true); });
2509 jvObj[jss::type] =
"peerStatusChange";
2518 p->send(jvObj,
true);
2532 using namespace std::chrono_literals;
2564 <<
"recvValidation " << val->getLedgerHash() <<
" from " << source;
2580 <<
"Exception thrown for handling new validation "
2581 << val->getLedgerHash() <<
": " << e.
what();
2586 <<
"Unknown exception thrown for handling new validation "
2587 << val->getLedgerHash();
2599 ss <<
"VALIDATION: " << val->render() <<
" master_key: ";
2636 "This server is amendment blocked, and must be updated to be "
2637 "able to stay in sync with the network.";
2644 "This server has an expired validator list. validators.txt "
2645 "may be incorrectly configured or some [validator_list_sites] "
2646 "may be unreachable.";
2653 "One or more unsupported amendments have reached majority. "
2654 "Upgrade to the latest version before they are activated "
2655 "to avoid being amendment blocked.";
2656 if (
auto const expected =
2660 d[jss::expected_date] = expected->time_since_epoch().count();
2661 d[jss::expected_date_UTC] =
to_string(*expected);
2665 if (warnings.size())
2666 info[jss::warnings] = std::move(warnings);
2681 info[jss::time] =
to_string(std::chrono::floor<std::chrono::microseconds>(
2685 info[jss::network_ledger] =
"waiting";
2687 info[jss::validation_quorum] =
2695 info[jss::node_size] =
"tiny";
2698 info[jss::node_size] =
"small";
2701 info[jss::node_size] =
"medium";
2704 info[jss::node_size] =
"large";
2707 info[jss::node_size] =
"huge";
2716 info[jss::validator_list_expires] =
2717 safe_cast<Json::UInt>(when->time_since_epoch().count());
2719 info[jss::validator_list_expires] = 0;
2729 if (*when == TimeKeeper::time_point::max())
2731 x[jss::expiration] =
"never";
2732 x[jss::status] =
"active";
2739 x[jss::status] =
"active";
2741 x[jss::status] =
"expired";
2746 x[jss::status] =
"unknown";
2747 x[jss::expiration] =
"unknown";
2751#if defined(GIT_COMMIT_HASH) || defined(GIT_BRANCH)
2754#ifdef GIT_COMMIT_HASH
2755 x[jss::hash] = GIT_COMMIT_HASH;
2758 x[jss::branch] = GIT_BRANCH;
2763 info[jss::io_latency_ms] =
2771 info[jss::pubkey_validator] =
2776 info[jss::pubkey_validator] =
"none";
2786 info[jss::counters][jss::nodestore] = nodestore;
2790 info[jss::pubkey_node] =
2796 info[jss::amendment_blocked] =
true;
2810 lastClose[jss::converge_time_s] =
2815 lastClose[jss::converge_time] =
2819 info[jss::last_close] = lastClose;
2827 info[jss::network_id] =
static_cast<Json::UInt>(*netid);
2829 auto const escalationMetrics =
2837 auto const loadFactorFeeEscalation =
2839 escalationMetrics.openLedgerFeeLevel,
2841 escalationMetrics.referenceFeeLevel)
2845 safe_cast<std::uint64_t>(loadFactorServer), loadFactorFeeEscalation);
2849 info[jss::load_base] = loadBaseServer;
2850 info[jss::load_factor] =
trunc32(loadFactor);
2851 info[jss::load_factor_server] = loadFactorServer;
2858 info[jss::load_factor_fee_escalation] =
2859 escalationMetrics.openLedgerFeeLevel.jsonClipped();
2860 info[jss::load_factor_fee_queue] =
2861 escalationMetrics.minProcessingFeeLevel.jsonClipped();
2862 info[jss::load_factor_fee_reference] =
2863 escalationMetrics.referenceFeeLevel.jsonClipped();
2867 info[jss::load_factor] =
2868 static_cast<double>(loadFactor) / loadBaseServer;
2870 if (loadFactorServer != loadFactor)
2871 info[jss::load_factor_server] =
2872 static_cast<double>(loadFactorServer) / loadBaseServer;
2877 if (fee != loadBaseServer)
2878 info[jss::load_factor_local] =
2879 static_cast<double>(fee) / loadBaseServer;
2881 if (fee != loadBaseServer)
2882 info[jss::load_factor_net] =
2883 static_cast<double>(fee) / loadBaseServer;
2885 if (fee != loadBaseServer)
2886 info[jss::load_factor_cluster] =
2887 static_cast<double>(fee) / loadBaseServer;
2889 if (escalationMetrics.openLedgerFeeLevel !=
2890 escalationMetrics.referenceFeeLevel &&
2891 (admin || loadFactorFeeEscalation != loadFactor))
2892 info[jss::load_factor_fee_escalation] =
2893 escalationMetrics.openLedgerFeeLevel.decimalFromReference(
2894 escalationMetrics.referenceFeeLevel);
2895 if (escalationMetrics.minProcessingFeeLevel !=
2896 escalationMetrics.referenceFeeLevel)
2897 info[jss::load_factor_fee_queue] =
2898 escalationMetrics.minProcessingFeeLevel.decimalFromReference(
2899 escalationMetrics.referenceFeeLevel);
2912 XRPAmount const baseFee = lpClosed->fees().base;
2914 l[jss::seq] =
Json::UInt(lpClosed->info().seq);
2915 l[jss::hash] =
to_string(lpClosed->info().hash);
2920 l[jss::reserve_base] = lpClosed->fees().reserve.jsonClipped();
2921 l[jss::reserve_inc] = lpClosed->fees().increment.jsonClipped();
2923 lpClosed->info().closeTime.time_since_epoch().count());
2928 l[jss::reserve_base_xrp] = lpClosed->fees().reserve.decimalXRP();
2929 l[jss::reserve_inc_xrp] = lpClosed->fees().increment.decimalXRP();
2932 std::abs(closeOffset.count()) >= 60)
2933 l[jss::close_time_offset] =
2941 Json::UInt(age < highAgeThreshold ? age.count() : 0);
2945 auto lCloseTime = lpClosed->info().closeTime;
2947 if (lCloseTime <= closeTime)
2949 using namespace std::chrono_literals;
2950 auto age = closeTime - lCloseTime;
2952 Json::UInt(age < highAgeThreshold ? age.count() : 0);
2958 info[jss::validated_ledger] = l;
2960 info[jss::closed_ledger] = l;
2964 info[jss::published_ledger] =
"none";
2965 else if (lpPublished->info().seq != lpClosed->info().seq)
2966 info[jss::published_ledger] = lpPublished->info().seq;
2971 info[jss::jq_trans_overflow] =
2973 info[jss::peer_disconnects] =
2975 info[jss::peer_disconnects_resources] =
2980 "http",
"https",
"peer",
"ws",
"ws2",
"wss",
"wss2"};
2988 !(port.admin_nets_v4.empty() && port.admin_nets_v6.empty() &&
2989 port.admin_user.empty() && port.admin_password.empty()))
3003 for (
auto const& p : proto)
3004 jv[jss::protocol].append(p);
3011 auto const optPort = grpcSection.
get(
"port");
3012 if (optPort && grpcSection.get(
"ip"))
3015 jv[jss::port] = *optPort;
3017 jv[jss::protocol].
append(
"grpc");
3020 info[jss::ports] = std::move(ports);
3066 [&](
Json::Value const& jv) { p->send(jv, true); });
3091 lpAccepted->info().hash, alpAccepted);
3095 alpAccepted->getLedger().
get() == lpAccepted.
get(),
3096 "ripple::NetworkOPsImp::pubLedger : accepted input");
3100 <<
"Publishing ledger " << lpAccepted->info().seq <<
" "
3101 << lpAccepted->info().hash;
3109 jvObj[jss::type] =
"ledgerClosed";
3110 jvObj[jss::ledger_index] = lpAccepted->info().seq;
3111 jvObj[jss::ledger_hash] =
to_string(lpAccepted->info().hash);
3113 lpAccepted->info().closeTime.time_since_epoch().count());
3117 if (!lpAccepted->rules().enabled(featureXRPFees))
3119 jvObj[jss::fee_base] = lpAccepted->fees().base.jsonClipped();
3120 jvObj[jss::reserve_base] = lpAccepted->fees().reserve.jsonClipped();
3121 jvObj[jss::reserve_inc] =
3122 lpAccepted->fees().increment.jsonClipped();
3124 jvObj[jss::txn_count] =
Json::UInt(alpAccepted->size());
3128 jvObj[jss::validated_ledgers] =
3138 p->send(jvObj,
true);
3156 p->send(jvObj,
true);
3165 static bool firstTime =
true;
3172 for (
auto& inner : outer.second)
3174 auto& subInfo = inner.second;
3175 if (subInfo.index_->separationLedgerSeq_ == 0)
3178 alpAccepted->getLedger(), subInfo);
3187 for (
auto const& accTx : *alpAccepted)
3191 lpAccepted, *accTx, accTx == *(--alpAccepted->end()));
3218 "reportConsensusStateChange->pubConsensus",
3249 jvObj[jss::type] =
"transaction";
3253 jvObj[jss::transaction] =
3260 jvObj[jss::meta], *ledger, transaction, meta->
get());
3263 jvObj[jss::meta], transaction, meta->
get());
3267 if (
auto const& lookup = ledger->txRead(transaction->getTransactionID());
3268 lookup.second && lookup.second->isFieldPresent(sfTransactionIndex))
3270 uint32_t
const txnSeq = lookup.second->getFieldU32(sfTransactionIndex);
3272 if (transaction->isFieldPresent(sfNetworkID))
3273 netID = transaction->getFieldU32(sfNetworkID);
3278 jvObj[jss::ctid] = *ctid;
3280 if (!ledger->open())
3281 jvObj[jss::ledger_hash] =
to_string(ledger->info().hash);
3285 jvObj[jss::ledger_index] = ledger->info().seq;
3286 jvObj[jss::transaction][jss::date] =
3287 ledger->info().closeTime.time_since_epoch().count();
3288 jvObj[jss::validated] =
true;
3289 jvObj[jss::close_time_iso] =
to_string_iso(ledger->info().closeTime);
3295 jvObj[jss::validated] =
false;
3296 jvObj[jss::ledger_current_index] = ledger->info().seq;
3299 jvObj[jss::status] = validated ?
"closed" :
"proposed";
3300 jvObj[jss::engine_result] = sToken;
3301 jvObj[jss::engine_result_code] = result;
3302 jvObj[jss::engine_result_message] = sHuman;
3304 if (transaction->getTxnType() == ttOFFER_CREATE)
3306 auto const account = transaction->getAccountID(sfAccount);
3307 auto const amount = transaction->getFieldAmount(sfTakerGets);
3310 if (account != amount.issue().account)
3318 jvObj[jss::transaction][jss::owner_funds] = ownerFunds.getText();
3326 [&]<
unsigned Version>(
3328 RPC::insertDeliverMax(
3329 jvTx[jss::transaction], transaction->getTxnType(), Version);
3331 if constexpr (Version > 1)
3333 jvTx[jss::tx_json] = jvTx.removeMember(jss::transaction);
3334 jvTx[jss::hash] = hash;
3338 jvTx[jss::transaction][jss::hash] = hash;
3351 auto const& stTxn = transaction.
getTxn();
3355 auto const trResult = transaction.
getResult();
3370 [&](
Json::Value const& jv) { p->send(jv, true); });
3387 [&](
Json::Value const& jv) { p->send(jv, true); });
3412 auto const currLedgerSeq = ledger->seq();
3419 for (
auto const& affectedAccount : transaction.
getAffected())
3424 auto it = simiIt->second.begin();
3426 while (it != simiIt->second.end())
3437 it = simiIt->second.erase(it);
3444 auto it = simiIt->second.begin();
3445 while (it != simiIt->second.end())
3456 it = simiIt->second.erase(it);
3463 auto& subs = histoIt->second;
3464 auto it = subs.begin();
3465 while (it != subs.end())
3468 if (currLedgerSeq <= info.index_->separationLedgerSeq_)
3482 it = subs.erase(it);
3493 <<
"pubAccountTransaction: "
3494 <<
"proposed=" << iProposed <<
", accepted=" << iAccepted;
3496 if (!notify.
empty() || !accountHistoryNotify.
empty())
3498 auto const& stTxn = transaction.
getTxn();
3502 auto const trResult = transaction.
getResult();
3508 isrListener->getApiVersion(),
3509 [&](
Json::Value const& jv) { isrListener->send(jv,
true); });
3513 jvObj.
set(jss::account_history_boundary,
true);
3516 jvObj.
isMember(jss::account_history_tx_stream) ==
3518 "ripple::NetworkOPsImp::pubAccountTransaction : "
3519 "account_history_tx_stream not set");
3520 for (
auto& info : accountHistoryNotify)
3522 auto& index = info.index_;
3523 if (index->forwardTxIndex_ == 0 && !index->haveHistorical_)
3524 jvObj.
set(jss::account_history_tx_first,
true);
3526 jvObj.
set(jss::account_history_tx_index, index->forwardTxIndex_++);
3529 info.sink_->getApiVersion(),
3530 [&](
Json::Value const& jv) { info.sink_->send(jv,
true); });
3555 for (
auto const& affectedAccount : tx->getMentionedAccounts())
3560 auto it = simiIt->second.begin();
3562 while (it != simiIt->second.end())
3573 it = simiIt->second.erase(it);
3580 JLOG(
m_journal.
trace()) <<
"pubProposedAccountTransaction: " << iProposed;
3582 if (!notify.
empty() || !accountHistoryNotify.
empty())
3589 isrListener->getApiVersion(),
3590 [&](
Json::Value const& jv) { isrListener->send(jv,
true); });
3593 jvObj.
isMember(jss::account_history_tx_stream) ==
3595 "ripple::NetworkOPs::pubProposedAccountTransaction : "
3596 "account_history_tx_stream not set");
3597 for (
auto& info : accountHistoryNotify)
3599 auto& index = info.index_;
3600 if (index->forwardTxIndex_ == 0 && !index->haveHistorical_)
3601 jvObj.
set(jss::account_history_tx_first,
true);
3602 jvObj.
set(jss::account_history_tx_index, index->forwardTxIndex_++);
3604 info.sink_->getApiVersion(),
3605 [&](
Json::Value const& jv) { info.sink_->send(jv,
true); });
3622 for (
auto const& naAccountID : vnaAccountIDs)
3625 <<
"subAccount: account: " <<
toBase58(naAccountID);
3627 isrListener->insertSubAccountInfo(naAccountID, rt);
3632 for (
auto const& naAccountID : vnaAccountIDs)
3634 auto simIterator = subMap.
find(naAccountID);
3635 if (simIterator == subMap.
end())
3639 usisElement[isrListener->getSeq()] = isrListener;
3641 subMap.
insert(simIterator, make_pair(naAccountID, usisElement));
3646 simIterator->second[isrListener->getSeq()] = isrListener;
3657 for (
auto const& naAccountID : vnaAccountIDs)
3660 isrListener->deleteSubAccountInfo(naAccountID, rt);
3677 for (
auto const& naAccountID : vnaAccountIDs)
3679 auto simIterator = subMap.
find(naAccountID);
3681 if (simIterator != subMap.
end())
3684 simIterator->second.erase(uSeq);
3686 if (simIterator->second.empty())
3689 subMap.
erase(simIterator);
3698 enum DatabaseType { Sqlite,
None };
3699 static auto const databaseType = [&]() -> DatabaseType {
3704 return DatabaseType::Sqlite;
3706 return DatabaseType::None;
3709 if (databaseType == DatabaseType::None)
3713 "ripple::NetworkOPsImp::addAccountHistoryJob : no database");
3715 <<
"AccountHistory job for account "
3728 "AccountHistoryTxStream",
3729 [
this, dbType = databaseType, subInfo]() {
3730 auto const& accountId = subInfo.
index_->accountId_;
3731 auto& lastLedgerSeq = subInfo.
index_->historyLastLedgerSeq_;
3732 auto& txHistoryIndex = subInfo.
index_->historyTxIndex_;
3735 <<
"AccountHistory job for account " <<
toBase58(accountId)
3736 <<
" started. lastLedgerSeq=" << lastLedgerSeq;
3746 auto stx = tx->getSTransaction();
3747 if (stx->getAccountID(sfAccount) == accountId &&
3748 stx->getSeqValue() == 1)
3752 for (
auto& node : meta->getNodes())
3754 if (node.getFieldU16(sfLedgerEntryType) != ltACCOUNT_ROOT)
3757 if (node.isFieldPresent(sfNewFields))
3759 if (
auto inner =
dynamic_cast<STObject const*
>(
3760 node.peekAtPField(sfNewFields));
3763 if (inner->isFieldPresent(sfAccount) &&
3764 inner->getAccountID(sfAccount) == accountId)
3776 bool unsubscribe) ->
bool {
3779 sptr->send(jvObj,
true);
3789 bool unsubscribe) ->
bool {
3793 sptr->getApiVersion(),
3794 [&](
Json::Value const& jv) { sptr->send(jv,
true); });
3817 accountId, minLedger, maxLedger, marker, 0,
true};
3818 return db->newestAccountTxPage(options);
3823 "ripple::NetworkOPsImp::addAccountHistoryJob : "
3824 "getMoreTxns : invalid database type");
3834 while (lastLedgerSeq >= 2 && !subInfo.
index_->stopHistorical_)
3836 int feeChargeCount = 0;
3845 <<
"AccountHistory job for account "
3846 <<
toBase58(accountId) <<
" no InfoSub. Fee charged "
3847 << feeChargeCount <<
" times.";
3852 auto startLedgerSeq =
3853 (lastLedgerSeq > 1024 + 2 ? lastLedgerSeq - 1024 : 2);
3855 <<
"AccountHistory job for account " <<
toBase58(accountId)
3856 <<
", working on ledger range [" << startLedgerSeq <<
","
3857 << lastLedgerSeq <<
"]";
3859 auto haveRange = [&]() ->
bool {
3862 auto haveSomeValidatedLedgers =
3864 validatedMin, validatedMax);
3866 return haveSomeValidatedLedgers &&
3867 validatedMin <= startLedgerSeq &&
3868 lastLedgerSeq <= validatedMax;
3874 <<
"AccountHistory reschedule job for account "
3875 <<
toBase58(accountId) <<
", incomplete ledger range ["
3876 << startLedgerSeq <<
"," << lastLedgerSeq <<
"]";
3882 while (!subInfo.
index_->stopHistorical_)
3885 getMoreTxns(startLedgerSeq, lastLedgerSeq, marker);
3890 "ripple::NetworkOPsImp::addAccountHistoryJob : "
3891 "getMoreTxns failed");
3893 <<
"AccountHistory job for account "
3894 <<
toBase58(accountId) <<
" getMoreTxns failed.";
3900 auto const& txns = dbResult->first;
3901 marker = dbResult->second;
3902 size_t num_txns = txns.size();
3903 for (
size_t i = 0; i < num_txns; ++i)
3905 auto const& [tx, meta] = txns[i];
3910 <<
"AccountHistory job for account "
3911 <<
toBase58(accountId) <<
" empty tx or meta.";
3922 "ripple::NetworkOPsImp::addAccountHistoryJob : "
3923 "getLedgerBySeq failed");
3925 <<
"AccountHistory job for account "
3926 <<
toBase58(accountId) <<
" no ledger.";
3932 tx->getSTransaction();
3937 "NetworkOPsImp::addAccountHistoryJob : "
3938 "getSTransaction failed");
3940 <<
"AccountHistory job for account "
3942 <<
" getSTransaction failed.";
3949 auto const trR = meta->getResultTER();
3951 transJson(stTxn, trR,
true, curTxLedger, mRef);
3954 jss::account_history_tx_index, txHistoryIndex--);
3955 if (i + 1 == num_txns ||
3956 txns[i + 1].first->getLedger() != tx->getLedger())
3957 jvTx.
set(jss::account_history_boundary,
true);
3959 if (isFirstTx(tx, meta))
3961 jvTx.
set(jss::account_history_tx_first,
true);
3962 sendMultiApiJson(jvTx,
false);
3965 <<
"AccountHistory job for account "
3967 <<
" done, found last tx.";
3972 sendMultiApiJson(jvTx,
false);
3979 <<
"AccountHistory job for account "
3981 <<
" paging, marker=" << marker->ledgerSeq <<
":"
3990 if (!subInfo.
index_->stopHistorical_)
3992 lastLedgerSeq = startLedgerSeq - 1;
3993 if (lastLedgerSeq <= 1)
3996 <<
"AccountHistory job for account "
3998 <<
" done, reached genesis ledger.";
4011 subInfo.
index_->separationLedgerSeq_ = ledger->seq();
4012 auto const& accountId = subInfo.
index_->accountId_;
4014 if (!ledger->exists(accountKeylet))
4017 <<
"subAccountHistoryStart, no account " <<
toBase58(accountId)
4018 <<
", no need to add AccountHistory job.";
4023 if (
auto const sleAcct = ledger->read(accountKeylet); sleAcct)
4025 if (sleAcct->getFieldU32(sfSequence) == 1)
4028 <<
"subAccountHistoryStart, genesis account "
4030 <<
" does not have tx, no need to add AccountHistory job.";
4038 "ripple::NetworkOPsImp::subAccountHistoryStart : failed to "
4039 "access genesis account");
4044 subInfo.
index_->historyLastLedgerSeq_ = ledger->seq();
4045 subInfo.
index_->haveHistorical_ =
true;
4048 <<
"subAccountHistoryStart, add AccountHistory job: accountId="
4049 <<
toBase58(accountId) <<
", currentLedgerSeq=" << ledger->seq();
4059 if (!isrListener->insertSubAccountHistory(accountId))
4062 <<
"subAccountHistory, already subscribed to account "
4074 inner.
emplace(isrListener->getSeq(), ahi);
4080 simIterator->second.emplace(isrListener->getSeq(), ahi);
4094 <<
"subAccountHistory, no validated ledger yet, delay start";
4107 isrListener->deleteSubAccountHistory(account);
4121 auto& subInfoMap = simIterator->second;
4122 auto subInfoIter = subInfoMap.find(seq);
4123 if (subInfoIter != subInfoMap.end())
4125 subInfoIter->second.index_->stopHistorical_ =
true;
4130 simIterator->second.erase(seq);
4131 if (simIterator->second.empty())
4137 <<
"unsubAccountHistory, account " <<
toBase58(account)
4138 <<
", historyOnly = " << (historyOnly ?
"true" :
"false");
4146 listeners->addSubscriber(isrListener);
4150 UNREACHABLE(
"ripple::NetworkOPsImp::subBook : null book listeners");
4160 listeners->removeSubscriber(uSeq);
4172 m_standalone,
"ripple::NetworkOPsImp::acceptLedger : is standalone");
4175 Throw<std::runtime_error>(
4176 "Operation only possible in STANDALONE mode.");
4191 jvResult[jss::ledger_index] = lpClosed->info().seq;
4192 jvResult[jss::ledger_hash] =
to_string(lpClosed->info().hash);
4194 lpClosed->info().closeTime.time_since_epoch().count());
4195 if (!lpClosed->rules().enabled(featureXRPFees))
4197 jvResult[jss::fee_base] = lpClosed->fees().base.jsonClipped();
4198 jvResult[jss::reserve_base] = lpClosed->fees().reserve.jsonClipped();
4199 jvResult[jss::reserve_inc] = lpClosed->fees().increment.jsonClipped();
4205 jvResult[jss::validated_ledgers] =
4211 .emplace(isrListener->getSeq(), isrListener)
4221 .emplace(isrListener->getSeq(), isrListener)
4247 .emplace(isrListener->getSeq(), isrListener)
4275 jvResult[jss::random] =
to_string(uRandom);
4277 jvResult[jss::load_base] = feeTrack.getLoadBase();
4278 jvResult[jss::load_factor] = feeTrack.getLoadFactor();
4279 jvResult[jss::hostid] =
getHostId(admin);
4280 jvResult[jss::pubkey_node] =
4285 .emplace(isrListener->getSeq(), isrListener)
4303 .emplace(isrListener->getSeq(), isrListener)
4321 .emplace(isrListener->getSeq(), isrListener)
4339 .emplace(isrListener->getSeq(), isrListener)
4363 .emplace(isrListener->getSeq(), isrListener)
4381 .emplace(isrListener->getSeq(), isrListener)
4429 if (map.find(pInfo->getSeq()) != map.end())
4436#ifndef USE_NEW_BOOK_PAGE
4447 unsigned int iLimit,
4457 uint256 uTipIndex = uBookBase;
4461 stream <<
"getBookPage:" << book;
4462 stream <<
"getBookPage: uBookBase=" << uBookBase;
4463 stream <<
"getBookPage: uBookEnd=" << uBookEnd;
4464 stream <<
"getBookPage: uTipIndex=" << uTipIndex;
4473 bool bDirectAdvance =
true;
4477 unsigned int uBookEntry;
4483 while (!bDone && iLimit-- > 0)
4487 bDirectAdvance =
false;
4491 auto const ledgerIndex = view.
succ(uTipIndex, uBookEnd);
4495 sleOfferDir.
reset();
4504 uTipIndex = sleOfferDir->key();
4507 cdirFirst(view, uTipIndex, sleOfferDir, uBookEntry, offerIndex);
4510 <<
"getBookPage: uTipIndex=" << uTipIndex;
4512 <<
"getBookPage: offerIndex=" << offerIndex;
4522 auto const uOfferOwnerID = sleOffer->getAccountID(sfAccount);
4523 auto const& saTakerGets = sleOffer->getFieldAmount(sfTakerGets);
4524 auto const& saTakerPays = sleOffer->getFieldAmount(sfTakerPays);
4526 bool firstOwnerOffer(
true);
4532 saOwnerFunds = saTakerGets;
4534 else if (bGlobalFreeze)
4542 auto umBalanceEntry = umBalance.
find(uOfferOwnerID);
4543 if (umBalanceEntry != umBalance.
end())
4547 saOwnerFunds = umBalanceEntry->second;
4548 firstOwnerOffer =
false;
4562 if (saOwnerFunds < beast::zero)
4566 saOwnerFunds.
clear();
4574 STAmount saOwnerFundsLimit = saOwnerFunds;
4586 saOwnerFundsLimit =
divide(saOwnerFunds, offerRate);
4589 if (saOwnerFundsLimit >= saTakerGets)
4592 saTakerGetsFunded = saTakerGets;
4598 saTakerGetsFunded = saOwnerFundsLimit;
4600 saTakerGetsFunded.
setJson(jvOffer[jss::taker_gets_funded]);
4604 saTakerGetsFunded, saDirRate, saTakerPays.
issue()))
4605 .setJson(jvOffer[jss::taker_pays_funded]);
4611 saOwnerFunds,
multiply(saTakerGetsFunded, offerRate));
4613 umBalance[uOfferOwnerID] = saOwnerFunds - saOwnerPays;
4617 jvOf[jss::quality] = saDirRate.
getText();
4619 if (firstOwnerOffer)
4620 jvOf[jss::owner_funds] = saOwnerFunds.
getText();
4627 if (!
cdirNext(view, uTipIndex, sleOfferDir, uBookEntry, offerIndex))
4629 bDirectAdvance =
true;
4634 <<
"getBookPage: offerIndex=" << offerIndex;
4654 unsigned int iLimit,
4662 MetaView lesActive(lpLedger,
tapNONE,
true);
4663 OrderBookIterator obIterator(lesActive, book);
4667 bool const bGlobalFreeze = lesActive.isGlobalFrozen(book.
out.
account) ||
4668 lesActive.isGlobalFrozen(book.
in.
account);
4670 while (iLimit-- > 0 && obIterator.nextOffer())
4675 auto const uOfferOwnerID = sleOffer->getAccountID(sfAccount);
4676 auto const& saTakerGets = sleOffer->getFieldAmount(sfTakerGets);
4677 auto const& saTakerPays = sleOffer->getFieldAmount(sfTakerPays);
4678 STAmount saDirRate = obIterator.getCurrentRate();
4684 saOwnerFunds = saTakerGets;
4686 else if (bGlobalFreeze)
4694 auto umBalanceEntry = umBalance.
find(uOfferOwnerID);
4696 if (umBalanceEntry != umBalance.
end())
4700 saOwnerFunds = umBalanceEntry->second;
4706 saOwnerFunds = lesActive.accountHolds(
4712 if (saOwnerFunds.isNegative())
4716 saOwnerFunds.zero();
4723 STAmount saTakerGetsFunded;
4724 STAmount saOwnerFundsLimit = saOwnerFunds;
4736 saOwnerFundsLimit =
divide(saOwnerFunds, offerRate);
4739 if (saOwnerFundsLimit >= saTakerGets)
4742 saTakerGetsFunded = saTakerGets;
4747 saTakerGetsFunded = saOwnerFundsLimit;
4749 saTakerGetsFunded.setJson(jvOffer[jss::taker_gets_funded]);
4755 multiply(saTakerGetsFunded, saDirRate, saTakerPays.issue()))
4756 .setJson(jvOffer[jss::taker_pays_funded]);
4759 STAmount saOwnerPays = (
parityRate == offerRate)
4762 saOwnerFunds,
multiply(saTakerGetsFunded, offerRate));
4764 umBalance[uOfferOwnerID] = saOwnerFunds - saOwnerPays;
4766 if (!saOwnerFunds.isZero() || uOfferOwnerID == uTakerID)
4770 jvOf[jss::quality] = saDirRate.
getText();
4785 auto const current = std::chrono::duration_cast<std::chrono::microseconds>(
4825 ++counters_[
static_cast<std::size_t>(om)].transitions;
4827 counters_[
static_cast<std::size_t>(om)].transitions == 1)
4829 initialSyncUs_ = std::chrono::duration_cast<std::chrono::microseconds>(
4830 now - processStart_)
4834 std::chrono::duration_cast<std::chrono::microseconds>(now - start_);
4843 auto [counters, mode, start, initialSync] = getCounterData();
4844 auto const current = std::chrono::duration_cast<std::chrono::microseconds>(
4854 auto& state = obj[jss::state_accounting][
states_[i]];
4855 state[jss::transitions] =
std::to_string(counters[i].transitions);
4856 state[jss::duration_us] =
std::to_string(counters[i].dur.count());
4860 obj[jss::initial_sync_duration_us] =
std::to_string(initialSync);
4875 boost::asio::io_context& io_svc,
T back_inserter(T... args)
Decorator for streaming out compact json.
Lightweight wrapper to tag static string.
Value & append(Value const &value)
Append value to array at the end.
bool isMember(char const *key) const
Return true if the object has a member named key.
Value get(UInt index, Value const &defaultValue) const
If the array contains at least index+1 elements, returns the element value, otherwise returns default...
A generic endpoint for log messages.
Stream trace() const
Severity stream access functions.
A metric for measuring an integral value.
void set(value_type value) const
Set the value on the gauge.
A reference to a handler for performing polled collection.
A transaction that is in a closed ledger.
boost::container::flat_set< AccountID > const & getAffected() const
std::shared_ptr< STTx const > const & getTxn() const
TxMeta const & getMeta() const
virtual std::optional< NetClock::time_point > firstUnsupportedExpected() const =0
virtual Config & config()=0
virtual Overlay & overlay()=0
virtual LoadFeeTrack & getFeeTrack()=0
virtual OpenLedger & openLedger()=0
virtual beast::Journal journal(std::string const &name)=0
virtual NodeStore::Database & getNodeStore()=0
virtual ServerHandler & getServerHandler()=0
virtual std::chrono::milliseconds getIOLatency()=0
virtual OrderBookDB & getOrderBookDB()=0
virtual TimeKeeper & timeKeeper()=0
virtual TaggedCache< uint256, AcceptedLedger > & getAcceptedLedgerCache()=0
virtual JobQueue & getJobQueue()=0
virtual InboundLedgers & getInboundLedgers()=0
virtual ValidatorList & validators()=0
virtual std::optional< PublicKey const > getValidationPublicKey() const =0
virtual LedgerMaster & getLedgerMaster()=0
virtual RelationalDatabase & getRelationalDatabase()=0
virtual ManifestCache & validatorManifests()=0
virtual perf::PerfLog & getPerfLog()=0
virtual Cluster & cluster()=0
virtual AmendmentTable & getAmendmentTable()=0
virtual std::pair< PublicKey, SecretKey > const & nodeIdentity()=0
bool exists(std::string const &name) const
Returns true if a section with the given name exists.
Section & section(std::string const &name)
Returns the section with the given name.
Holds transactions which were deferred to the next pass of consensus.
The role of a ClosureCounter is to assist in shutdown by letting callers wait for the completion of c...
std::string const & name() const
std::uint32_t getLoadFee() const
NetClock::time_point getReportTime() const
PublicKey const & identity() const
std::size_t size() const
The number of nodes in the cluster list.
std::string SERVER_DOMAIN
static constexpr std::uint32_t FEE_UNITS_DEPRECATED
int RELAY_UNTRUSTED_VALIDATIONS
virtual void clearFailures()=0
virtual Json::Value getInfo()=0
std::shared_ptr< InfoSub > pointer
A pool of threads to perform work.
Json::Value getJson(int c=0)
bool addJob(JobType type, std::string const &name, JobHandler &&jobHandler)
Adds a job to the JobQueue.
std::shared_ptr< Ledger const > getValidatedLedger()
bool haveValidated()
Whether we have ever fully validated a ledger.
std::shared_ptr< ReadView const > getCurrentLedger()
bool getValidatedRange(std::uint32_t &minVal, std::uint32_t &maxVal)
std::shared_ptr< Ledger const > getClosedLedger()
std::string getCompleteLedgers()
std::size_t getFetchPackCacheSize() const
std::shared_ptr< ReadView const > getPublishedLedger()
std::shared_ptr< Ledger const > getLedgerBySeq(std::uint32_t index)
std::chrono::seconds getValidatedLedgerAge()
Manages the current fee schedule.
std::uint32_t getClusterFee() const
std::uint32_t getLocalFee() const
std::uint32_t getLoadBase() const
std::uint32_t getRemoteFee() const
std::uint32_t getLoadFactor() const
void heartbeat()
Reset the stall detection timer.
PublicKey getMasterKey(PublicKey const &pk) const
Returns ephemeral signing key's master public key.
State accounting records two attributes for each possible server state: 1) Amount of time spent in ea...
void mode(OperatingMode om)
Record state transition.
void json(Json::Value &obj) const
Output state counters in JSON format.
std::array< Counters, 5 > counters_
std::uint64_t initialSyncUs_
CounterData getCounterData() const
std::chrono::steady_clock::time_point start_
static std::array< Json::StaticString const, 5 > const states_
std::chrono::steady_clock::time_point const processStart_
Transaction with input flags and results to be applied in batches.
TransactionStatus(std::shared_ptr< Transaction > t, bool a, bool l, FailHard f)
std::shared_ptr< Transaction > const transaction
void processClusterTimer()
boost::asio::steady_timer accountHistoryTxTimer_
void pubProposedTransaction(std::shared_ptr< ReadView const > const &ledger, std::shared_ptr< STTx const > const &transaction, TER result) override
OperatingMode getOperatingMode() const override
std::string strOperatingMode(OperatingMode const mode, bool const admin) const override
bool preProcessTransaction(std::shared_ptr< Transaction > &transaction)
std::vector< TransactionStatus > mTransactions
bool unsubBookChanges(std::uint64_t uListener) override
std::atomic< OperatingMode > mMode
Json::Value getLedgerFetchInfo() override
bool isUNLBlocked() override
void unsubAccount(InfoSub::ref ispListener, hash_set< AccountID > const &vnaAccountIDs, bool rt) override
Json::Value getOwnerInfo(std::shared_ptr< ReadView const > lpLedger, AccountID const &account) override
void setNeedNetworkLedger() override
void setUNLBlocked() override
void pubConsensus(ConsensusPhase phase)
void transactionBatch()
Apply transactions in batches.
void apply(std::unique_lock< std::mutex > &batchLock)
Attempt to apply transactions and post-process based on the results.
void setAmendmentBlocked() override
bool checkLastClosedLedger(Overlay::PeerSequence const &, uint256 &networkClosed)
void processTransaction(std::shared_ptr< Transaction > &transaction, bool bUnlimited, bool bLocal, FailHard failType) override
Process transactions as they arrive from the network or which are submitted by clients.
void processTransactionSet(CanonicalTXSet const &set) override
Process a set of transactions synchronously, and ensuring that they are processed in one batch.
void clearUNLBlocked() override
boost::asio::steady_timer heartbeatTimer_
void updateLocalTx(ReadView const &view) override
bool unsubManifests(std::uint64_t uListener) override
DispatchState
Synchronization states for transaction batches.
std::optional< PublicKey > const validatorPK_
bool unsubTransactions(std::uint64_t uListener) override
void clearAmendmentWarned() override
std::size_t getLocalTxCount() override
std::unique_ptr< LocalTxs > m_localTX
bool subValidations(InfoSub::ref ispListener) override
bool subLedger(InfoSub::ref ispListener, Json::Value &jvResult) override
~NetworkOPsImp() override
bool isAmendmentBlocked() override
void unsubAccountHistoryInternal(std::uint64_t seq, AccountID const &account, bool historyOnly) override
SubAccountHistoryMapType mSubAccountHistory
Json::Value getServerInfo(bool human, bool admin, bool counters) override
InfoSub::pointer addRpcSub(std::string const &strUrl, InfoSub::ref) override
boost::asio::steady_timer clusterTimer_
bool isAmendmentWarned() override
static std::array< char const *, 5 > const states_
bool subServer(InfoSub::ref ispListener, Json::Value &jvResult, bool admin) override
void unsubAccountInternal(std::uint64_t seq, hash_set< AccountID > const &vnaAccountIDs, bool rt) override
std::atomic< bool > amendmentBlocked_
SubInfoMapType mSubAccount
std::optional< PublicKey > const validatorMasterPK_
void unsubAccountHistory(InfoSub::ref ispListener, AccountID const &account, bool historyOnly) override
unsubscribe an account's transactions
std::set< uint256 > pendingValidations_
bool beginConsensus(uint256 const &networkClosed, std::unique_ptr< std::stringstream > const &clog) override
void doTransactionAsync(std::shared_ptr< Transaction > transaction, bool bUnlimited, FailHard failtype)
For transactions not submitted by a locally connected client, fire and forget.
void setAccountHistoryJobTimer(SubAccountHistoryInfoWeak subInfo)
bool unsubValidations(std::uint64_t uListener) override
void endConsensus(std::unique_ptr< std::stringstream > const &clog) override
ClosureCounter< void, boost::system::error_code const & > waitHandlerCounter_
void pubLedger(std::shared_ptr< ReadView const > const &lpAccepted) override
void addAccountHistoryJob(SubAccountHistoryInfoWeak subInfo)
void doTransactionSync(std::shared_ptr< Transaction > transaction, bool bUnlimited, FailHard failType)
For transactions submitted directly by a client, apply batch of transactions and wait for this transa...
void setTimer(boost::asio::steady_timer &timer, std::chrono::milliseconds const &expiry_time, std::function< void()> onExpire, std::function< void()> onError)
std::array< SubMapType, SubTypes::sLastEntry > mStreamMaps
bool unsubPeerStatus(std::uint64_t uListener) override
void pubValidation(std::shared_ptr< STValidation > const &val) override
std::size_t const minPeerCount_
std::atomic< bool > unlBlocked_
bool subBook(InfoSub::ref ispListener, Book const &) override
std::uint32_t acceptLedger(std::optional< std::chrono::milliseconds > consensusDelay) override
Accepts the current transaction tree, return the new ledger's sequence.
void stateAccounting(Json::Value &obj) override
void submitTransaction(std::shared_ptr< STTx const > const &) override
bool unsubRTTransactions(std::uint64_t uListener) override
Json::Value getConsensusInfo() override
std::recursive_mutex mSubLock
std::atomic< bool > needNetworkLedger_
bool recvValidation(std::shared_ptr< STValidation > const &val, std::string const &source) override
void switchLastClosedLedger(std::shared_ptr< Ledger const > const &newLCL)
StateAccounting accounting_
void reportConsensusStateChange(ConsensusPhase phase)
bool subConsensus(InfoSub::ref ispListener) override
bool isNeedNetworkLedger() override
void setAmendmentWarned() override
bool processTrustedProposal(RCLCxPeerPos proposal) override
void doTransactionSyncBatch(std::unique_lock< std::mutex > &lock, std::function< bool(std::unique_lock< std::mutex > const &)> retryCallback)
bool subPeerStatus(InfoSub::ref ispListener) override
void mapComplete(std::shared_ptr< SHAMap > const &map, bool fromAcquire) override
bool tryRemoveRpcSub(std::string const &strUrl) override
void pubAccountTransaction(std::shared_ptr< ReadView const > const &ledger, AcceptedLedgerTx const &transaction, bool last)
LedgerMaster & m_ledgerMaster
void clearLedgerFetch() override
bool isBlocked() override
void consensusViewChange() override
void setStateTimer() override
Called to initially start our timers.
bool subManifests(InfoSub::ref ispListener) override
void pubValidatedTransaction(std::shared_ptr< ReadView const > const &ledger, AcceptedLedgerTx const &transaction, bool last)
void subAccount(InfoSub::ref ispListener, hash_set< AccountID > const &vnaAccountIDs, bool rt) override
bool unsubServer(std::uint64_t uListener) override
MultiApiJson transJson(std::shared_ptr< STTx const > const &transaction, TER result, bool validated, std::shared_ptr< ReadView const > const &ledger, std::optional< std::reference_wrapper< TxMeta const > > meta)
ServerFeeSummary mLastFeeSummary
void pubPeerStatus(std::function< Json::Value(void)> const &) override
void setStandAlone() override
bool subRTTransactions(InfoSub::ref ispListener) override
void pubProposedAccountTransaction(std::shared_ptr< ReadView const > const &ledger, std::shared_ptr< STTx const > const &transaction, TER result)
std::condition_variable mCond
void setMode(OperatingMode om) override
void getBookPage(std::shared_ptr< ReadView const > &lpLedger, Book const &, AccountID const &uTakerID, bool const bProof, unsigned int iLimit, Json::Value const &jvMarker, Json::Value &jvResult) override
void clearNeedNetworkLedger() override
NetworkOPsImp(Application &app, NetworkOPs::clock_type &clock, bool standalone, std::size_t minPeerCount, bool start_valid, JobQueue &job_queue, LedgerMaster &ledgerMaster, ValidatorKeys const &validatorKeys, boost::asio::io_context &io_svc, beast::Journal journal, beast::insight::Collector::ptr const &collector)
DispatchState mDispatchState
bool subBookChanges(InfoSub::ref ispListener) override
SubInfoMapType mSubRTAccount
void reportFeeChange() override
void processHeartbeatTimer()
bool unsubBook(std::uint64_t uListener, Book const &) override
void subAccountHistoryStart(std::shared_ptr< ReadView const > const &ledger, SubAccountHistoryInfoWeak &subInfo)
error_code_i subAccountHistory(InfoSub::ref ispListener, AccountID const &account) override
subscribe an account's new transactions and retrieve the account's historical transactions
std::mutex validationsMutex_
void pubManifest(Manifest const &) override
ConsensusPhase mLastConsensusPhase
bool subTransactions(InfoSub::ref ispListener) override
std::atomic< bool > amendmentWarned_
InfoSub::pointer findRpcSub(std::string const &strUrl) override
bool unsubLedger(std::uint64_t uListener) override
std::string getHostId(bool forAdmin)
bool unsubConsensus(std::uint64_t uListener) override
Provides server functionality for clients.
void getCountsJson(Json::Value &obj)
std::shared_ptr< OpenView const > current() const
Returns a view to the current open ledger.
Writable ledger view that accumulates state and tx changes.
BookListeners::pointer getBookListeners(Book const &)
void processTxn(std::shared_ptr< ReadView const > const &ledger, AcceptedLedgerTx const &alTx, MultiApiJson const &jvObj)
BookListeners::pointer makeBookListeners(Book const &)
virtual std::optional< std::uint32_t > networkID() const =0
Returns the ID of the network this server is configured for, if any.
virtual std::uint64_t getPeerDisconnect() const =0
virtual std::size_t size() const =0
Returns the number of active peers.
virtual std::uint64_t getJqTransOverflow() const =0
virtual std::uint64_t getPeerDisconnectCharges() const =0
Manages the generic consensus algorithm for use by the RCL.
std::size_t prevProposers() const
Get the number of proposing peers that participated in the previous round.
void simulate(NetClock::time_point const &now, std::optional< std::chrono::milliseconds > consensusDelay)
std::chrono::milliseconds prevRoundTime() const
Get duration of the previous round.
Json::Value getJson(bool full) const
A peer's signed, proposed position for use in RCLConsensus.
PublicKey const & publicKey() const
Public key of peer that sent the proposal.
Represents a set of transactions in RCLConsensus.
Wraps a ledger instance for use in generic Validations LedgerTrie.
static std::string getWordFromBlob(void const *blob, size_t bytes)
Chooses a single dictionary word from the data.
Collects logging information.
std::unique_ptr< std::stringstream > const & ss()
virtual std::shared_ptr< SLE const > read(Keylet const &k) const =0
Return the state item associated with a key.
virtual std::optional< key_type > succ(key_type const &key, std::optional< key_type > const &last=std::nullopt) const =0
Return the key of the next state item.
void setJson(Json::Value &) const
std::string getText() const override
Issue const & issue() const
std::optional< T > get(std::string const &name) const
std::size_t size() const noexcept
void const * data() const noexcept
void setup(Setup const &setup, beast::Journal journal)
time_point now() const override
Returns the current time, using the server's clock.
std::chrono::seconds closeOffset() const
time_point closeTime() const
Returns the predicted close time, in network time.
Metrics getMetrics(OpenView const &view) const
Returns fee metrics in reference fee level units.
Validator keys and manifest as set in configuration file.
std::size_t count() const
Return the number of configured validator list sites.
std::optional< PublicKey > getTrustedKey(PublicKey const &identity) const
Returns master public key if public key is trusted.
std::optional< PublicKey > localPublicKey() const
This function returns the local validator public key or a std::nullopt.
std::optional< TimeKeeper::time_point > expires() const
Return the time when the validator list will expire.
std::size_t quorum() const
Get quorum value for current trusted key set.
constexpr double decimalXRP() const
Json::Value jsonClipped() const
static constexpr std::size_t size()
virtual Json::Value currentJson() const =0
Render currently executing jobs and RPC calls and durations in Json.
virtual Json::Value countersJson() const =0
Render performance counters in Json.
Automatically unlocks and re-locks a unique_lock object.
T emplace_back(T... args)
@ arrayValue
array value (ordered list)
@ objectValue
object value (collection of name/value pairs).
void rngfill(void *const buffer, std::size_t const bytes, Generator &g)
std::string const & getVersionString()
Server version.
std::optional< std::string > encodeCTID(uint32_t ledgerSeq, uint32_t txnIndex, uint32_t networkID) noexcept
Encodes ledger sequence, transaction index, and network ID into a CTID string.
Json::Value computeBookChanges(std::shared_ptr< L const > const &lpAccepted)
void insertNFTSyntheticInJson(Json::Value &, std::shared_ptr< STTx const > const &, TxMeta const &)
Adds common synthetic fields to transaction-related JSON responses.
void insertMPTokenIssuanceID(Json::Value &response, std::shared_ptr< STTx const > const &transaction, TxMeta const &transactionMeta)
void insertDeliveredAmount(Json::Value &meta, ReadView const &, std::shared_ptr< STTx const > const &serializedTx, TxMeta const &)
Add a delivered_amount field to the meta input/output parameter.
Charge const feeMediumBurdenRPC
TER valid(STTx const &tx, ReadView const &view, AccountID const &src, beast::Journal j)
Keylet account(AccountID const &id) noexcept
AccountID root.
Keylet page(uint256 const &root, std::uint64_t index=0) noexcept
A page in a directory.
Keylet offer(AccountID const &id, std::uint32_t seq) noexcept
An offer from an account.
Rate rate(Env &env, Account const &account, std::uint32_t const &seq)
Use hash_* containers for keys that do not need a cryptographically secure hashing algorithm.
std::unique_ptr< NetworkOPs > make_NetworkOPs(Application &app, NetworkOPs::clock_type &clock, bool standalone, std::size_t minPeerCount, bool startvalid, JobQueue &job_queue, LedgerMaster &ledgerMaster, ValidatorKeys const &validatorKeys, boost::asio::io_context &io_svc, beast::Journal journal, beast::insight::Collector::ptr const &collector)
std::string toBase58(AccountID const &v)
Convert AccountID to base58 checked string.
STAmount divide(STAmount const &amount, Rate const &rate)
std::shared_ptr< STTx const > sterilize(STTx const &stx)
Sterilize a transaction.
STAmount accountFunds(ReadView const &view, AccountID const &id, STAmount const &saDefault, FreezeHandling freezeHandling, beast::Journal j)
std::uint64_t getQuality(uint256 const &uBase)
std::pair< PublicKey, SecretKey > generateKeyPair(KeyType type, Seed const &seed)
Generate a key pair deterministically.
auto constexpr muldiv_max
std::unique_ptr< LocalTxs > make_LocalTxs()
STAmount amountFromQuality(std::uint64_t rate)
void handleNewValidation(Application &app, std::shared_ptr< STValidation > const &val, std::string const &source, BypassAccept const bypassAccept, std::optional< beast::Journal > j)
Handle a new validation.
@ warnRPC_EXPIRED_VALIDATOR_LIST
@ warnRPC_UNSUPPORTED_MAJORITY
@ warnRPC_AMENDMENT_BLOCKED
bool set(T &target, std::string const &name, Section const §ion)
Set a value from a configuration Section If the named value is not found or doesn't parse as a T,...
std::unique_ptr< FeeVote > make_FeeVote(FeeSetup const &setup, beast::Journal journal)
Create an instance of the FeeVote logic.
OperatingMode
Specifies the mode under which the server believes it's operating.
@ TRACKING
convinced we agree with the network
@ DISCONNECTED
not ready to process requests
@ CONNECTED
convinced we are talking to the network
@ FULL
we have the ledger and can even validate
@ SYNCING
fallen slightly behind
STAmount multiply(STAmount const &amount, Rate const &rate)
AccountID calcAccountID(PublicKey const &pk)
@ current
This was a new validation and was added.
csprng_engine & crypto_prng()
The default cryptographically secure PRNG.
Json::Value rpcError(int iError)
bool isTefFailure(TER x) noexcept
ConsensusPhase
Phases of consensus for a single ledger round.
static std::array< char const *, 5 > const stateNames
std::string strHex(FwdIt begin, FwdIt end)
Rate transferRate(ReadView const &view, AccountID const &issuer)
Returns IOU issuer transfer fee as Rate.
void forAllApiVersions(Fn const &fn, Args &&... args)
bool isTerRetry(TER x) noexcept
send_if_pred< Predicate > send_if(std::shared_ptr< Message > const &m, Predicate const &f)
Helper function to aid in type deduction.
uint256 getQualityNext(uint256 const &uBase)
STAmount accountHolds(ReadView const &view, AccountID const &account, Currency const ¤cy, AccountID const &issuer, FreezeHandling zeroIfFrozen, beast::Journal j)
bool isTesSuccess(TER x) noexcept
Rules makeRulesGivenLedger(DigestAwareReadView const &ledger, Rules const ¤t)
std::string to_string_iso(date::sys_time< Duration > tp)
bool cdirFirst(ReadView const &view, uint256 const &root, std::shared_ptr< SLE const > &page, unsigned int &index, uint256 &entry)
Returns the first entry in the directory, advancing the index.
std::string to_string(base_uint< Bits, Tag > const &a)
FeeSetup setup_FeeVote(Section const §ion)
bool isTemMalformed(TER x) noexcept
Number root(Number f, unsigned d)
std::optional< std::uint64_t > mulDiv(std::uint64_t value, std::uint64_t mul, std::uint64_t div)
Return value*mul/div accurately.
Json::Value getJson(LedgerFill const &fill)
Return a new Json::Value representing the ledger with given options.
@ ledgerMaster
ledger master data for signing
@ proposal
proposal for signing
bool cdirNext(ReadView const &view, uint256 const &root, std::shared_ptr< SLE const > &page, unsigned int &index, uint256 &entry)
Returns the next entry in the directory, advancing the index.
std::pair< Validity, std::string > checkValidity(HashRouter &router, STTx const &tx, Rules const &rules, Config const &config)
Checks transaction signature and local checks.
Seed generateSeed(std::string const &passPhrase)
Generate a seed deterministically.
constexpr std::size_t maxPoppedTransactions
bool transResultInfo(TER code, std::string &token, std::string &text)
bool isTelLocal(TER x) noexcept
uint256 getBookBase(Book const &book)
constexpr std::uint32_t tfInnerBatchTxn
Rate const parityRate
A transfer rate signifying a 1:1 exchange.
bool isGlobalFrozen(ReadView const &view, AccountID const &issuer)
static std::uint32_t trunc32(std::uint64_t v)
static auto const genesisAccountId
T set_intersection(T... args)
std::string serialized
The manifest in serialized form.
std::uint32_t sequence
The sequence number of this manifest.
std::string domain
The domain, if one was specified in the manifest; empty otherwise.
std::optional< Blob > getSignature() const
Returns manifest signature.
std::optional< PublicKey > signingKey
The ephemeral key associated with this manifest.
Blob getMasterSignature() const
Returns manifest master key signature.
PublicKey masterKey
The master key associated with this manifest.
Server fees published on server subscription.
bool operator!=(ServerFeeSummary const &b) const
ServerFeeSummary()=default
std::optional< TxQ::Metrics > em
std::uint32_t loadFactorServer
bool operator==(ServerFeeSummary const &b) const
std::uint32_t loadBaseServer
decltype(initialSyncUs_) initialSyncUs
decltype(counters_) counters
std::uint64_t transitions
std::chrono::microseconds dur
beast::insight::Gauge full_transitions
Stats(Handler const &handler, beast::insight::Collector::ptr const &collector)
beast::insight::Hook hook
beast::insight::Gauge connected_duration
beast::insight::Gauge tracking_duration
beast::insight::Gauge connected_transitions
beast::insight::Gauge disconnected_transitions
beast::insight::Gauge syncing_duration
beast::insight::Gauge tracking_transitions
beast::insight::Gauge full_duration
beast::insight::Gauge disconnected_duration
beast::insight::Gauge syncing_transitions
std::uint32_t historyLastLedgerSeq_
std::uint32_t separationLedgerSeq_
AccountID const accountId_
std::uint32_t forwardTxIndex_
std::atomic< bool > stopHistorical_
std::int32_t historyTxIndex_
SubAccountHistoryIndex(AccountID const &accountId)
std::shared_ptr< SubAccountHistoryIndex > index_
std::shared_ptr< SubAccountHistoryIndex > index_
Represents a transfer rate.
Data format for exchanging consumption information across peers.
std::vector< Item > items
Changes in trusted nodes after updating validator list.
hash_set< NodeID > removed
Structure returned by TxQ::getMetrics, expressed in reference fee level units.
IsMemberResult isMember(char const *key) const
void set(char const *key, auto const &v)
Select all peers (except optional excluded) that are in our cluster.
Sends a message to all peers.
T time_since_epoch(T... args)